summaryrefslogtreecommitdiffstats
path: root/target/linux/coldfire/patches/023-mcfv4e_cache_base_update.patch
diff options
context:
space:
mode:
authorImre Kaloz <kaloz@openwrt.org>2009-06-23 21:04:37 +0000
committerImre Kaloz <kaloz@openwrt.org>2009-06-23 21:04:37 +0000
commitc49f135f72f98633653300bab5b0ac993b03c6be (patch)
tree50929b27bcea9ab4dc74b041d6e90967473301fa /target/linux/coldfire/patches/023-mcfv4e_cache_base_update.patch
parent50f2abfa16a2b509955312d1776beeece662c61b (diff)
downloadmaster-31e0f0ae-c49f135f72f98633653300bab5b0ac993b03c6be.tar.gz
master-31e0f0ae-c49f135f72f98633653300bab5b0ac993b03c6be.tar.bz2
master-31e0f0ae-c49f135f72f98633653300bab5b0ac993b03c6be.zip
use broken-out patches for the coldfire to make it easier to follow differences against the bsp
SVN-Revision: 16547
Diffstat (limited to 'target/linux/coldfire/patches/023-mcfv4e_cache_base_update.patch')
-rw-r--r--target/linux/coldfire/patches/023-mcfv4e_cache_base_update.patch1194
1 files changed, 1194 insertions, 0 deletions
diff --git a/target/linux/coldfire/patches/023-mcfv4e_cache_base_update.patch b/target/linux/coldfire/patches/023-mcfv4e_cache_base_update.patch
new file mode 100644
index 0000000000..9215085a06
--- /dev/null
+++ b/target/linux/coldfire/patches/023-mcfv4e_cache_base_update.patch
@@ -0,0 +1,1194 @@
+From 3592cd3db82e5b010df590079f1e310b5d317248 Mon Sep 17 00:00:00 2001
+From: Kurt Mahan <kmahan@freescale.com>
+Date: Mon, 3 Dec 2007 23:03:07 -0700
+Subject: [PATCH] Rewrite Coldfire cache code.
+
+LTIBName: mcfv4e-cache-base-update
+Signed-off-by: Kurt Mahan <kmahan@freescale.com>
+---
+ arch/m68k/coldfire/cache.c | 196 +-------------
+ arch/m68k/coldfire/head.S | 6 +-
+ arch/m68k/coldfire/signal.c | 4 +-
+ arch/m68k/kernel/sys_m68k.c | 16 ++
+ arch/m68k/mm/cache.c | 31 +---
+ arch/m68k/mm/memory.c | 76 +-----
+ include/asm-m68k/cf_cacheflush.h | 525 +++++++++++++++++++++++++++++---------
+ include/asm-m68k/cfcache.h | 95 ++++----
+ 8 files changed, 495 insertions(+), 454 deletions(-)
+
+--- a/arch/m68k/coldfire/cache.c
++++ b/arch/m68k/coldfire/cache.c
+@@ -1,7 +1,8 @@
+ /*
+- * linux/arch/m68k/coldifre/cache.c
++ * linux/arch/m68k/coldfire/cache.c
+ *
+ * Matt Waddel Matt.Waddel@freescale.com
++ * Kurt Mahan kmahan@freescale.com
+ * Copyright Freescale Semiconductor, Inc. 2007
+ *
+ * This program is free software; you can redistribute it and/or modify
+@@ -15,191 +16,13 @@
+ #include <asm/coldfire.h>
+ #include <asm/system.h>
+
+-#define _DCACHE_SIZE (2*16384)
+-#define _ICACHE_SIZE (2*16384)
+-
+-#define _SET_SHIFT 4
+-
+-/*
+- * Masks for cache sizes. Programming note: because the set size is a
+- * power of two, the mask is also the last address in the set.
+- */
+-
+-#define _DCACHE_SET_MASK ((_DCACHE_SIZE/64-1)<<_SET_SHIFT)
+-#define _ICACHE_SET_MASK ((_ICACHE_SIZE/64-1)<<_SET_SHIFT)
+-#define LAST_DCACHE_ADDR _DCACHE_SET_MASK
+-#define LAST_ICACHE_ADDR _ICACHE_SET_MASK
+-
+-/************************************************************
+- * Routine to cleanly flush the cache, pushing all lines and
+- * invalidating them.
+- *
+- * The is the flash-resident version, used after copying the .text
+- * segment from flash to ram.
+- *************************************************************/
+-void FLASHDcacheFlushInvalidate(void)
+- __attribute__ ((section (".text_loader")));
+-
+-void FLASHDcacheFlushInvalidate()
+-{
+- unsigned long set;
+- unsigned long start_set;
+- unsigned long end_set;
+-
+- start_set = 0;
+- end_set = (unsigned long)LAST_DCACHE_ADDR;
+-
+- for (set = start_set; set < end_set; set += (0x10 - 3))
+- asm volatile("cpushl %%dc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%dc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%dc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%dc,(%0)" : : "a" (set));
+-}
+-
+-/************************************************************
+- * Routine to cleanly flush the cache, pushing all lines and
+- * invalidating them.
+- *
+- *************************************************************/
+-void DcacheFlushInvalidate()
+-{
+- unsigned long set;
+- unsigned long start_set;
+- unsigned long end_set;
+-
+- start_set = 0;
+- end_set = (unsigned long)LAST_DCACHE_ADDR;
+-
+- for (set = start_set; set < end_set; set += (0x10 - 3))
+- asm volatile("cpushl %%dc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%dc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%dc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%dc,(%0)" : : "a" (set));
+-}
+-
+-
+-
+-/******************************************************************************
+- * Routine to cleanly flush the a block of cache, pushing all relevant lines
+- * and invalidating them.
+- *
+- ******************************************************************************/
+-void DcacheFlushInvalidateCacheBlock(void *start, unsigned long size)
+-{
+- unsigned long set;
+- unsigned long start_set;
+- unsigned long end_set;
+-
+- /* if size is bigger than the cache can store
+- * set the size to the maximum amount
+- */
+-
+- if (size > LAST_DCACHE_ADDR)
+- size = LAST_DCACHE_ADDR;
+-
+- start_set = ((unsigned long)start) & _DCACHE_SET_MASK;
+- end_set = ((unsigned long)(start+size-1)) & _DCACHE_SET_MASK;
+-
+- if (start_set > end_set) {
+- /* from the begining to the lowest address */
+- for (set = 0; set <= end_set; set += (0x10 - 3))
+- asm volatile("cpushl %%dc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%dc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%dc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%dc,(%0)" : : "a" (set));
+-
+- /* next loop will finish the cache ie pass the hole */
+- end_set = LAST_DCACHE_ADDR;
+- }
+- for (set = start_set; set <= end_set; set += (0x10 - 3))
+- asm volatile("cpushl %%dc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%dc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%dc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%dc,(%0)" : : "a" (set));
+-}
+-
+-
+-void IcacheInvalidateCacheBlock(void *start, unsigned long size)
+-{
+- unsigned long set;
+- unsigned long start_set;
+- unsigned long end_set;
+-
+- /* if size is bigger than the cache can store
+- * set the size to the maximum ammount
+- */
+-
+- if (size > LAST_ICACHE_ADDR)
+- size = LAST_ICACHE_ADDR;
+-
+- start_set = ((unsigned long)start) & _ICACHE_SET_MASK;
+- end_set = ((unsigned long)(start+size-1)) & _ICACHE_SET_MASK;
+-
+- if (start_set > end_set) {
+- /* from the begining to the lowest address */
+- for (set = 0; set <= end_set; set += (0x10 - 3))
+- asm volatile("cpushl %%ic,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%ic,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%ic,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%ic,(%0)" : : "a" (set));
+-
+- /* next loop will finish the cache ie pass the hole */
+- end_set = LAST_ICACHE_ADDR;
+- }
+- for (set = start_set; set <= end_set; set += (0x10 - 3))
+- asm volatile("cpushl %%ic,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%ic,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%ic,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%ic,(%0)" : : "a" (set));
+-}
+-
+-
+-/********************************************************************
+- * Disable the data cache completely
+- ********************************************************************/
+-void DcacheDisable(void)
+-{
+- int newValue;
+- unsigned long flags;
+-
+- local_save_flags(flags);
+- local_irq_disable();
+-
+- DcacheFlushInvalidate(); /* begin by flushing the cache */
+- newValue = CACHE_DISABLE_MODE; /* disable it */
+- cacr_set(newValue);
+- local_irq_restore(flags);
+-}
+-
+-/********************************************************************
+- * Unconditionally enable the data cache
+- ********************************************************************/
+-void DcacheEnable(void)
+-{
+- cacr_set(CACHE_INITIAL_MODE);
+-}
+-
+-
++/* Cache Control Reg shadow reg */
+ unsigned long shadow_cacr;
+
++/**
++ * cacr_set - Set the Cache Control Register
++ * @x Value to set
++ */
+ void cacr_set(unsigned long x)
+ {
+ shadow_cacr = x;
+@@ -209,6 +32,11 @@ void cacr_set(unsigned long x)
+ : "r" (shadow_cacr));
+ }
+
++/**
++ * cacr_get - Get the current value of the Cache Control Register
++ *
++ * @return CACR value
++ */
+ unsigned long cacr_get(void)
+ {
+ return shadow_cacr;
+--- a/arch/m68k/coldfire/head.S
++++ b/arch/m68k/coldfire/head.S
+@@ -244,7 +244,7 @@ ENTRY(__start)
+ /* Setup initial stack pointer */
+ movel #0x40001000,%sp
+
+-/* Clear usp */
++/* Setup usp */
+ subl %a0,%a0
+ movel %a0,%usp
+
+@@ -252,6 +252,10 @@ ENTRY(__start)
+ movec %d0, %rambar1
+ movew #0x2700,%sr
+
++/* reset cache */
++ movel #(CF_CACR_ICINVA + CF_CACR_DCINVA),%d0
++ movecl %d0,%cacr
++
+ movel #(MMU_BASE+1),%d0
+ movecl %d0,%mmubar
+ movel #MMUOR_CA,%a0 /* Clear tlb entries */
+--- a/arch/m68k/coldfire/signal.c
++++ b/arch/m68k/coldfire/signal.c
+@@ -37,6 +37,7 @@
+ #include <asm/cf_pgtable.h>
+ #include <asm/traps.h>
+ #include <asm/ucontext.h>
++#include <asm/cacheflush.h>
+
+ #define _BLOCKABLE (~(sigmask(SIGKILL) | sigmask(SIGSTOP)))
+
+@@ -605,10 +606,9 @@ static inline int rt_setup_ucontext(stru
+ return err;
+ }
+
+-extern void IcacheInvalidateCacheBlock(void *, unsigned long);
+ static inline void push_cache(unsigned long vaddr)
+ {
+- IcacheInvalidateCacheBlock((void *)vaddr, 8);
++ cf_cache_push(__pa(vaddr), 8);
+ }
+
+ static inline void __user *
+--- a/arch/m68k/kernel/sys_m68k.c
++++ b/arch/m68k/kernel/sys_m68k.c
+@@ -29,6 +29,9 @@
+ #include <asm/traps.h>
+ #include <asm/page.h>
+ #include <asm/unistd.h>
++#ifdef CONFIG_COLDFIRE
++#include <asm/cacheflush.h>
++#endif
+
+ /*
+ * sys_pipe() is the normal C calling standard for creating
+@@ -257,6 +260,7 @@ asmlinkage int sys_ipc (uint call, int f
+ return -EINVAL;
+ }
+
++#ifndef CONFIG_COLDFIRE
+ /* Convert virtual (user) address VADDR to physical address PADDR */
+ #define virt_to_phys_040(vaddr) \
+ ({ \
+@@ -580,6 +584,7 @@ cache_flush_060 (unsigned long addr, int
+ }
+ return 0;
+ }
++#endif /* CONFIG_COLDFIRE */
+
+ /* sys_cacheflush -- flush (part of) the processor cache. */
+ asmlinkage int
+@@ -612,6 +617,7 @@ sys_cacheflush (unsigned long addr, int
+ goto out;
+ }
+
++#ifndef CONFIG_COLDFIRE
+ if (CPU_IS_020_OR_030) {
+ if (scope == FLUSH_SCOPE_LINE && len < 256) {
+ unsigned long cacr;
+@@ -656,6 +662,16 @@ sys_cacheflush (unsigned long addr, int
+ ret = cache_flush_060 (addr, scope, cache, len);
+ }
+ }
++#else /* CONFIG_COLDFIRE */
++ if ((cache & FLUSH_CACHE_INSN) && (cache & FLUSH_CACHE_DATA))
++ flush_bcache();
++ else if (cache & FLUSH_CACHE_INSN)
++ flush_icache();
++ else
++ flush_dcache();
++
++ ret = 0;
++#endif /* CONFIG_COLDFIRE */
+ out:
+ unlock_kernel();
+ return ret;
+--- a/arch/m68k/mm/cache.c
++++ b/arch/m68k/mm/cache.c
+@@ -81,36 +81,7 @@ static unsigned long virt_to_phys_slow(u
+ void flush_icache_range(unsigned long address, unsigned long endaddr)
+ {
+ #ifdef CONFIG_COLDFIRE
+- unsigned long set;
+- unsigned long start_set;
+- unsigned long end_set;
+-
+- start_set = address & _ICACHE_SET_MASK;
+- end_set = endaddr & _ICACHE_SET_MASK;
+-
+- if (start_set > end_set) {
+- /* from the begining to the lowest address */
+- for (set = 0; set <= end_set; set += (0x10 - 3))
+- asm volatile ("cpushl %%ic,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%ic,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%ic,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%ic,(%0)" : : "a" (set));
+-
+- /* next loop will finish the cache ie pass the hole */
+- end_set = LAST_ICACHE_ADDR;
+- }
+- for (set = start_set; set <= end_set; set += (0x10 - 3))
+- asm volatile ("cpushl %%ic,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%ic,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%ic,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%ic,(%0)" : : "a" (set));
+-
++ cf_icache_flush_range(address, endaddr);
+ #else /* !CONFIG_COLDFIRE */
+
+ if (CPU_IS_040_OR_060) {
+--- a/arch/m68k/mm/memory.c
++++ b/arch/m68k/mm/memory.c
+@@ -127,6 +127,7 @@ int free_pointer_table (pmd_t *ptable)
+ return 0;
+ }
+
++#ifndef CONFIG_COLDFIRE
+ /* invalidate page in both caches */
+ static inline void clear040(unsigned long paddr)
+ {
+@@ -173,6 +174,7 @@ static inline void pushcl040(unsigned lo
+ clear040(paddr);
+ local_irq_restore(flags);
+ }
++#endif /* CONFIG_COLDFIRE */
+
+ /*
+ * 040: Hit every page containing an address in the range paddr..paddr+len-1.
+@@ -203,38 +205,10 @@ static inline void pushcl040(unsigned lo
+
+ void cache_clear (unsigned long paddr, int len)
+ {
+- if (CPU_IS_CFV4E) {
+- unsigned long set;
+- unsigned long start_set;
+- unsigned long end_set;
+-
+- start_set = paddr & _ICACHE_SET_MASK;
+- end_set = (paddr+len-1) & _ICACHE_SET_MASK;
+-
+- if (start_set > end_set) {
+- /* from the begining to the lowest address */
+- for (set = 0; set <= end_set; set += (0x10 - 3))
+- asm volatile("cpushl %%bc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%bc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%bc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%bc,(%0)" : : "a" (set));
+-
+- /* next loop will finish the cache ie pass the hole */
+- end_set = LAST_ICACHE_ADDR;
+- }
+- for (set = start_set; set <= end_set; set += (0x10 - 3))
+- asm volatile("cpushl %%bc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%bc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%bc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%bc,(%0)" : : "a" (set));
+-
+- } else if (CPU_IS_040_OR_060) {
++#ifdef CONFIG_COLDFIRE
++ cf_cache_clear(paddr, len);
++#else
++ if (CPU_IS_040_OR_060) {
+ int tmp;
+
+ /*
+@@ -268,6 +242,7 @@ void cache_clear (unsigned long paddr, i
+ if(mach_l2_flush)
+ mach_l2_flush(0);
+ #endif
++#endif /* CONFIG_COLDFIRE */
+ }
+ EXPORT_SYMBOL(cache_clear);
+
+@@ -281,38 +256,10 @@ EXPORT_SYMBOL(cache_clear);
+
+ void cache_push (unsigned long paddr, int len)
+ {
+- if (CPU_IS_CFV4E) {
+- unsigned long set;
+- unsigned long start_set;
+- unsigned long end_set;
+-
+- start_set = paddr & _ICACHE_SET_MASK;
+- end_set = (paddr+len-1) & _ICACHE_SET_MASK;
+-
+- if (start_set > end_set) {
+- /* from the begining to the lowest address */
+- for (set = 0; set <= end_set; set += (0x10 - 3))
+- asm volatile("cpushl %%bc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%bc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%bc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%bc,(%0)" : : "a" (set));
+-
+- /* next loop will finish the cache ie pass the hole */
+- end_set = LAST_ICACHE_ADDR;
+- }
+- for (set = start_set; set <= end_set; set += (0x10 - 3))
+- asm volatile("cpushl %%bc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%bc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%bc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%bc,(%0)" : : "a" (set));
+-
+- } else if (CPU_IS_040_OR_060) {
++#ifdef CONFIG_COLDFIRE
++ cf_cache_push(paddr, len);
++#else
++ if (CPU_IS_040_OR_060) {
+ int tmp = PAGE_SIZE;
+
+ /*
+@@ -352,6 +299,7 @@ void cache_push (unsigned long paddr, in
+ if(mach_l2_flush)
+ mach_l2_flush(1);
+ #endif
++#endif /* CONFIG_COLDFIRE */
+ }
+ EXPORT_SYMBOL(cache_push);
+
+--- a/include/asm-m68k/cf_cacheflush.h
++++ b/include/asm-m68k/cf_cacheflush.h
+@@ -1,160 +1,439 @@
++/*
++ * include/asm-m68k/cf_cacheflush.h - Coldfire Cache
++ *
++ * Based on include/asm-m68k/cacheflush.h
++ *
++ * Coldfire pieces by:
++ * Kurt Mahan kmahan@freescale.com
++ *
++ * Copyright Freescale Semiconductor, Inc. 2007
++ *
++ * This program is free software; you can redistribute it and/or modify it
++ * under the terms of the GNU General Public License as published by the
++ * Free Software Foundation; either version 2 of the License, or (at your
++ * option) any later version.
++ */
+ #ifndef M68K_CF_CACHEFLUSH_H
+ #define M68K_CF_CACHEFLUSH_H
+
+ #include <asm/cfcache.h>
+
+ /*
+- * Cache handling functions
++ * Coldfire Cache Model
++ *
++ * The Coldfire processors use a Harvard architecture cache configured
++ * as four-way set associative. The cache does not implement bus snooping
++ * so cache coherency with other masters must be maintained in software.
++ *
++ * The cache is managed via the CPUSHL instruction in conjunction with
++ * bits set in the CACR (cache control register). Currently the code
++ * uses the CPUSHL enhancement which adds the ability to
++ * invalidate/clear/push a cacheline by physical address. This feature
++ * is designated in the Hardware Configuration Register [D1-CPES].
++ *
++ * CACR Bits:
++ * DPI[28] cpushl invalidate disable for d-cache
++ * IDPI[12] cpushl invalidate disable for i-cache
++ * SPA[14] cpushl search by physical address
++ * IVO[20] cpushl invalidate only
++ *
++ * Random Terminology:
++ * * invalidate = reset the cache line's valid bit
++ * * push = generate a line-sized store of the data if its contents are marked
++ * as modifed (the modified flag is cleared after the store)
++ * * clear = push + invalidate
+ */
+
+-#define flush_icache() \
+-({ \
+- unsigned long set; \
+- unsigned long start_set; \
+- unsigned long end_set; \
+- \
+- start_set = 0; \
+- end_set = (unsigned long)LAST_DCACHE_ADDR; \
+- \
+- for (set = start_set; set <= end_set; set += (0x10 - 3)) \
+- asm volatile("cpushl %%ic,(%0)\n" \
+- "\taddq%.l #1,%0\n" \
+- "\tcpushl %%ic,(%0)\n" \
+- "\taddq%.l #1,%0\n" \
+- "\tcpushl %%ic,(%0)\n" \
+- "\taddq%.l #1,%0\n" \
+- "\tcpushl %%ic,(%0)" : : "a" (set)); \
+-})
++/**
++ * flush_icache - Flush all of the instruction cache
++ */
++static inline void flush_icache(void)
++{
++ asm volatile("nop\n"
++ "moveq%.l #0,%%d0\n"
++ "moveq%.l #0,%%d1\n"
++ "move%.l %%d0,%%a0\n"
++ "1:\n"
++ "cpushl %%ic,(%%a0)\n"
++ "add%.l #0x0010,%%a0\n"
++ "addq%.l #1,%%d1\n"
++ "cmpi%.l %0,%%d1\n"
++ "bne 1b\n"
++ "moveq%.l #0,%%d1\n"
++ "addq%.l #1,%%d0\n"
++ "move%.l %%d0,%%a0\n"
++ "cmpi%.l #4,%%d0\n"
++ "bne 1b\n"
++ : : "i" (CACHE_SETS)
++ : "a0", "d0", "d1");
++}
+
+-/*
+- * invalidate the cache for the specified memory range.
+- * It starts at the physical address specified for
+- * the given number of bytes.
++/**
++ * flush_dcache - Flush all of the data cache
+ */
+-extern void cache_clear(unsigned long paddr, int len);
+-/*
+- * push any dirty cache in the specified memory range.
+- * It starts at the physical address specified for
+- * the given number of bytes.
++static inline void flush_dcache(void)
++{
++ asm volatile("nop\n"
++ "moveq%.l #0,%%d0\n"
++ "moveq%.l #0,%%d1\n"
++ "move%.l %%d0,%%a0\n"
++ "1:\n"
++ "cpushl %%dc,(%%a0)\n"
++ "add%.l #0x0010,%%a0\n"
++ "addq%.l #1,%%d1\n"
++ "cmpi%.l %0,%%d1\n"
++ "bne 1b\n"
++ "moveq%.l #0,%%d1\n"
++ "addq%.l #1,%%d0\n"
++ "move%.l %%d0,%%a0\n"
++ "cmpi%.l #4,%%d0\n"
++ "bne 1b\n"
++ : : "i" (CACHE_SETS)
++ : "a0", "d0", "d1");
++}
++
++/**
++ * flush_bcache - Flush all of both caches
+ */
+-extern void cache_push(unsigned long paddr, int len);
++static inline void flush_bcache(void)
++{
++ asm volatile("nop\n"
++ "moveq%.l #0,%%d0\n"
++ "moveq%.l #0,%%d1\n"
++ "move%.l %%d0,%%a0\n"
++ "1:\n"
++ "cpushl %%bc,(%%a0)\n"
++ "add%.l #0x0010,%%a0\n"
++ "addq%.l #1,%%d1\n"
++ "cmpi%.l %0,%%d1\n"
++ "bne 1b\n"
++ "moveq%.l #0,%%d1\n"
++ "addq%.l #1,%%d0\n"
++ "move%.l %%d0,%%a0\n"
++ "cmpi%.l #4,%%d0\n"
++ "bne 1b\n"
++ : : "i" (CACHE_SETS)
++ : "a0", "d0", "d1");
++}
+
+-/*
+- * push and invalidate pages in the specified user virtual
+- * memory range.
++/**
++ * cf_cache_clear - invalidate cache
++ * @paddr: starting physical address
++ * @len: number of bytes
++ *
++ * Invalidate cache lines starting at paddr for len bytes.
++ * Those lines are not pushed.
++ */
++static inline void cf_cache_clear(unsigned long paddr, int len)
++{
++ /* number of lines */
++ len = (len + (CACHE_LINE_SIZE-1)) / CACHE_LINE_SIZE;
++
++ /* align on set boundary */
++ paddr &= 0xfffffff0;
++
++ asm volatile("nop\n"
++ "move%.l %2,%%d0\n"
++ "or%.l %3,%%d0\n"
++ "movec %%d0,%%cacr\n"
++ "move%.l %0,%%a0\n"
++ "move%.l %1,%%d0\n"
++ "1:\n"
++ "cpushl %%bc,(%%a0)\n"
++ "lea 0x10(%%a0),%%a0\n"
++ "subq%.l #1,%%d0\n"
++ "bne%.b 1b\n"
++ "movec %2,%%cacr\n"
++ : : "a" (paddr), "r" (len),
++ "r" (shadow_cacr),
++ "i" (CF_CACR_SPA+CF_CACR_IVO)
++ : "a0", "d0");
++}
++
++/**
++ * cf_cache_push - Push dirty cache out with no invalidate
++ * @paddr: starting physical address
++ * @len: number of bytes
++ *
++ * Push the any dirty lines starting at paddr for len bytes.
++ * Those lines are not invalidated.
++ */
++static inline void cf_cache_push(unsigned long paddr, int len)
++{
++ /* number of lines */
++ len = (len + (CACHE_LINE_SIZE-1)) / CACHE_LINE_SIZE;
++
++ /* align on set boundary */
++ paddr &= 0xfffffff0;
++
++ asm volatile("nop\n"
++ "move%.l %2,%%d0\n"
++ "or%.l %3,%%d0\n"
++ "movec %%d0,%%cacr\n"
++ "move%.l %0,%%a0\n"
++ "move%.l %1,%%d0\n"
++ "1:\n"
++ "cpushl %%bc,(%%a0)\n"
++ "lea 0x10(%%a0),%%a0\n"
++ "subq%.l #1,%%d0\n"
++ "bne.b 1b\n"
++ "movec %2,%%cacr\n"
++ : : "a" (paddr), "r" (len),
++ "r" (shadow_cacr),
++ "i" (CF_CACR_SPA+CF_CACR_DPI+CF_CACR_IDPI)
++ : "a0", "d0");
++}
++
++/**
++ * cf_cache_flush - Push dirty cache out and invalidate
++ * @paddr: starting physical address
++ * @len: number of bytes
++ *
++ * Push the any dirty lines starting at paddr for len bytes and
++ * invalidate those lines.
++ */
++static inline void cf_cache_flush(unsigned long paddr, int len)
++{
++ /* number of lines */
++ len = (len + (CACHE_LINE_SIZE-1)) / CACHE_LINE_SIZE;
++
++ /* align on set boundary */
++ paddr &= 0xfffffff0;
++
++ asm volatile("nop\n"
++ "move%.l %2,%%d0\n"
++ "or%.l %3,%%d0\n"
++ "movec %%d0,%%cacr\n"
++ "move%.l %0,%%a0\n"
++ "move%.l %1,%%d0\n"
++ "1:\n"
++ "cpushl %%bc,(%%a0)\n"
++ "lea 0x10(%%a0),%%a0\n"
++ "subq%.l #1,%%d0\n"
++ "bne.b 1b\n"
++ "movec %2,%%cacr\n"
++ : : "a" (paddr), "r" (len),
++ "r" (shadow_cacr),
++ "i" (CF_CACR_SPA)
++ : "a0", "d0");
++}
++
++/**
++ * cf_cache_flush_range - Push dirty data/inst cache in range out and invalidate
++ * @vstart - starting virtual address
++ * @vend: ending virtual address
++ *
++ * Push the any dirty data/instr lines starting at paddr for len bytes and
++ * invalidate those lines.
++ */
++static inline void cf_cache_flush_range(unsigned long vstart, unsigned long vend)
++{
++ int len;
++
++ /* align on set boundary */
++ vstart &= 0xfffffff0;
++ vend = PAGE_ALIGN((vend + (CACHE_LINE_SIZE-1))) & 0xfffffff0;
++ len = vend - vstart;
++ vstart = __pa(vstart);
++ vend = vstart + len;
++
++ asm volatile("nop\n"
++ "move%.l %2,%%d0\n"
++ "or%.l %3,%%d0\n"
++ "movec %%d0,%%cacr\n"
++ "move%.l %0,%%a0\n"
++ "move%.l %1,%%a1\n"
++ "1:\n"
++ "cpushl %%bc,(%%a0)\n"
++ "lea 0x10(%%a0),%%a0\n"
++ "cmpa%.l %%a0,%%a1\n"
++ "bne.b 1b\n"
++ "movec %2,%%cacr\n"
++ : /* no return */
++ : "a" (vstart), "a" (vend),
++ "r" (shadow_cacr),
++ "i" (CF_CACR_SPA)
++ : "a0", "a1", "d0");
++}
++
++/**
++ * cf_dcache_flush_range - Push dirty data cache in range out and invalidate
++ * @vstart - starting virtual address
++ * @vend: ending virtual address
++ *
++ * Push the any dirty data lines starting at paddr for len bytes and
++ * invalidate those lines.
++ */
++static inline void cf_dcache_flush_range(unsigned long vstart, unsigned long vend)
++{
++ /* align on set boundary */
++ vstart &= 0xfffffff0;
++ vend = (vend + (CACHE_LINE_SIZE-1)) & 0xfffffff0;
++
++ asm volatile("nop\n"
++ "move%.l %2,%%d0\n"
++ "or%.l %3,%%d0\n"
++ "movec %%d0,%%cacr\n"
++ "move%.l %0,%%a0\n"
++ "move%.l %1,%%a1\n"
++ "1:\n"
++ "cpushl %%dc,(%%a0)\n"
++ "lea 0x10(%%a0),%%a0\n"
++ "cmpa%.l %%a0,%%a1\n"
++ "bne.b 1b\n"
++ "movec %2,%%cacr\n"
++ : /* no return */
++ : "a" (__pa(vstart)), "a" (__pa(vend)),
++ "r" (shadow_cacr),
++ "i" (CF_CACR_SPA)
++ : "a0", "a1", "d0");
++}
++
++/**
++ * cf_icache_flush_range - Push dirty inst cache in range out and invalidate
++ * @vstart - starting virtual address
++ * @vend: ending virtual address
++ *
++ * Push the any dirty instr lines starting at paddr for len bytes and
++ * invalidate those lines. This should just be an invalidate since you
++ * shouldn't be able to have dirty instruction cache.
+ */
+-extern void cache_push_v(unsigned long vaddr, int len);
++static inline void cf_icache_flush_range(unsigned long vstart, unsigned long vend)
++{
++ /* align on set boundary */
++ vstart &= 0xfffffff0;
++ vend = (vend + (CACHE_LINE_SIZE-1)) & 0xfffffff0;
++
++ asm volatile("nop\n"
++ "move%.l %2,%%d0\n"
++ "or%.l %3,%%d0\n"
++ "movec %%d0,%%cacr\n"
++ "move%.l %0,%%a0\n"
++ "move%.l %1,%%a1\n"
++ "1:\n"
++ "cpushl %%ic,(%%a0)\n"
++ "lea 0x10(%%a0),%%a0\n"
++ "cmpa%.l %%a0,%%a1\n"
++ "bne.b 1b\n"
++ "movec %2,%%cacr\n"
++ : /* no return */
++ : "a" (__pa(vstart)), "a" (__pa(vend)),
++ "r" (shadow_cacr),
++ "i" (CF_CACR_SPA)
++ : "a0", "a1", "d0");
++}
+
+-/* This is needed whenever the virtual mapping of the current
+- process changes. */
++/**
++ * flush_cache_mm - Flush an mm_struct
++ * @mm: mm_struct to flush
++ */
++static inline void flush_cache_mm(struct mm_struct *mm)
++{
++ if (mm == current->mm)
++ flush_bcache();
++}
+
++#define flush_cache_dup_mm(mm) flush_cache_mm(mm)
+
+-#define flush_cache_all() do { } while (0)
+-#define flush_cache_mm(mm) do { } while (0)
+-#define flush_cache_range(mm, a, b) do { } while (0)
+-#define flush_cache_page(vma, vmaddr, pfn) do { } while (0)
+-
+-#define flush_dcache_range(paddr, len) do { } while (0)
+-
+-/* Push the page at kernel virtual address and clear the icache */
+-/* use cpush %bc instead of cpush %dc, cinv %ic */
+-#define flush_page_to_ram(page) __flush_page_to_ram((void *) page_address(page))
+-extern inline void __flush_page_to_ram(void *address)
+-{
+- unsigned long set;
+- unsigned long start_set;
+- unsigned long end_set;
+- unsigned long addr = (unsigned long) address;
+-
+- addr &= ~(PAGE_SIZE - 1); /* round down to page start address */
+-
+- start_set = addr & _ICACHE_SET_MASK;
+- end_set = (addr + PAGE_SIZE-1) & _ICACHE_SET_MASK;
+-
+- if (start_set > end_set) {
+- /* from the begining to the lowest address */
+- for (set = 0; set <= end_set; set += (0x10 - 3))
+- asm volatile("cpushl %%bc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%bc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%bc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%bc,(%0)" : : "a" (set));
+-
+- /* next loop will finish the cache ie pass the hole */
+- end_set = LAST_ICACHE_ADDR;
+- }
+- for (set = start_set; set <= end_set; set += (0x10 - 3))
+- asm volatile("cpushl %%bc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%bc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%bc,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%bc,(%0)" : : "a" (set));
+-}
+-
+-#define flush_dcache_page(page) do { } while (0)
+-#define flush_icache_page(vma, pg) do { } while (0)
+-#define flush_icache_user_range(adr, len) do { } while (0)
+-/* NL */
+-#define flush_icache_user_page(vma, page, addr, len) do { } while (0)
+-
+-/* Push n pages at kernel virtual address and clear the icache */
+-/* use cpush %bc instead of cpush %dc, cinv %ic */
+-extern inline void flush_icache_range(unsigned long address,
+- unsigned long endaddr)
+-{
+- unsigned long set;
+- unsigned long start_set;
+- unsigned long end_set;
+-
+- start_set = address & _ICACHE_SET_MASK;
+- end_set = endaddr & _ICACHE_SET_MASK;
+-
+- if (start_set > end_set) {
+- /* from the begining to the lowest address */
+- for (set = 0; set <= end_set; set += (0x10 - 3))
+- asm volatile("cpushl %%ic,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%ic,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%ic,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%ic,(%0)" : : "a" (set));
+-
+- /* next loop will finish the cache ie pass the hole */
+- end_set = LAST_ICACHE_ADDR;
+- }
+- for (set = start_set; set <= end_set; set += (0x10 - 3))
+- asm volatile("cpushl %%ic,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%ic,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%ic,(%0)\n"
+- "\taddq%.l #1,%0\n"
+- "\tcpushl %%ic,(%0)" : : "a" (set));
++/**
++ * flush_cache_range - Flush a cache range
++ * @vma: vma struct
++ * @start: Starting address
++ * @end: Ending address
++ *
++ * flush_cache_range must be a macro to avoid a dependency on
++ * linux/mm.h which includes this file.
++ */
++static inline void flush_cache_range(struct vm_area_struct *vma,
++ unsigned long start, unsigned long end)
++{
++ if (vma->vm_mm == current->mm)
++ cf_cache_flush_range(start, end);
+ }
+
++/**
++ * flush_cache_page - Flush a page of the cache
++ * @vma: vma struct
++ * @vmaddr:
++ * @pfn: page numer
++ *
++ * flush_cache_page must be a macro to avoid a dependency on
++ * linux/mm.h which includes this file.
++ */
++static inline void flush_cache_page(struct vm_area_struct *vma,
++ unsigned long vmaddr, unsigned long pfn)
++{
++ if (vma->vm_mm == current->mm)
++ cf_cache_flush_range(vmaddr, vmaddr+PAGE_SIZE);
++}
++
++/**
++ * __flush_page_to_ram - Push a page out of the cache
++ * @vaddr: Virtual address at start of page
++ *
++ * Push the page at kernel virtual address *vaddr* and clear
++ * the icache.
++ */
++static inline void __flush_page_to_ram(void *vaddr)
++{
++ asm volatile("nop\n"
++ "move%.l %2,%%d0\n"
++ "or%.l %3,%%d0\n"
++ "movec %%d0,%%cacr\n"
++ "move%.l %0,%%d0\n"
++ "and%.l #0xfffffff0,%%d0\n"
++ "move%.l %%d0,%%a0\n"
++ "move%.l %1,%%d0\n"
++ "1:\n"
++ "cpushl %%bc,(%%a0)\n"
++ "lea 0x10(%%a0),%%a0\n"
++ "subq%.l #1,%%d0\n"
++ "bne.b 1b\n"
++ "movec %2,%%cacr\n"
++ : : "a" (__pa(vaddr)), "i" (PAGE_SIZE / CACHE_LINE_SIZE),
++ "r" (shadow_cacr), "i" (CF_CACR_SPA)
++ : "a0", "d0");
++}
++
++/*
++ * Various defines for the kernel.
++ */
++
++extern void cache_clear(unsigned long paddr, int len);
++extern void cache_push(unsigned long paddr, int len);
++extern void flush_icache_range(unsigned long address, unsigned long endaddr);
++
++#define flush_cache_all() flush_bcache()
++#define flush_cache_vmap(start, end) flush_bcache()
++#define flush_cache_vunmap(start, end) flush_bcache()
++
++#define flush_dcache_range(vstart, vend) cf_dcache_flush_range(vstart, vend)
++#define flush_dcache_page(page) __flush_page_to_ram(page_address(page))
++#define flush_dcache_mmap_lock(mapping) do { } while (0)
++#define flush_dcache_mmap_unlock(mapping) do { } while (0)
++
++#define flush_icache_page(vma, page) __flush_page_to_ram(page_address(page))
++
++/**
++ * copy_to_user_page - Copy memory to user page
++ */
+ static inline void copy_to_user_page(struct vm_area_struct *vma,
+ struct page *page, unsigned long vaddr,
+ void *dst, void *src, int len)
+ {
+ memcpy(dst, src, len);
+- flush_icache_user_page(vma, page, vaddr, len);
++ cf_cache_flush(page_to_phys(page), PAGE_SIZE);
+ }
++
++/**
++ * copy_from_user_page - Copy memory from user page
++ */
+ static inline void copy_from_user_page(struct vm_area_struct *vma,
+ struct page *page, unsigned long vaddr,
+ void *dst, void *src, int len)
+ {
++ cf_cache_flush(page_to_phys(page), PAGE_SIZE);
+ memcpy(dst, src, len);
+ }
+
+-#define flush_cache_dup_mm(mm) flush_cache_mm(mm)
+-#define flush_cache_vmap(start, end) flush_cache_all()
+-#define flush_cache_vunmap(start, end) flush_cache_all()
+-#define flush_dcache_mmap_lock(mapping) do { } while (0)
+-#define flush_dcache_mmap_unlock(mapping) do { } while (0)
+-
+ #endif /* M68K_CF_CACHEFLUSH_H */
+--- a/include/asm-m68k/cfcache.h
++++ b/include/asm-m68k/cfcache.h
+@@ -1,19 +1,32 @@
+ /*
+- * include/asm-m68k/cfcache.h
++ * include/asm-m68k/cfcache.h - Coldfire Cache Controller
++ *
++ * Kurt Mahan kmahan@freescale.com
++ *
++ * Copyright Freescale Semiconductor, Inc. 2007
++ *
++ * This program is free software; you can redistribute it and/or modify it
++ * under the terms of the GNU General Public License as published by the
++ * Free Software Foundation; either version 2 of the License, or (at your
++ * option) any later version.
+ */
+ #ifndef CF_CFCACHE_H
+ #define CF_CFCACHE_H
+
++/*
++ * CACR Cache Control Register
++ */
+ #define CF_CACR_DEC (0x80000000) /* Data Cache Enable */
+ #define CF_CACR_DW (0x40000000) /* Data default Write-protect */
+ #define CF_CACR_DESB (0x20000000) /* Data Enable Store Buffer */
+-#define CF_CACR_DDPI (0x10000000) /* Data Disable CPUSHL Invalidate */
++#define CF_CACR_DPI (0x10000000) /* Data Disable CPUSHL Invalidate */
+ #define CF_CACR_DHLCK (0x08000000) /* 1/2 Data Cache Lock Mode */
+ #define CF_CACR_DDCM_00 (0x00000000) /* Cacheable writethrough imprecise */
+ #define CF_CACR_DDCM_01 (0x02000000) /* Cacheable copyback */
+ #define CF_CACR_DDCM_10 (0x04000000) /* Noncacheable precise */
+ #define CF_CACR_DDCM_11 (0x06000000) /* Noncacheable imprecise */
+ #define CF_CACR_DCINVA (0x01000000) /* Data Cache Invalidate All */
++#define CF_CACR_DDSP (0x00800000) /* Data default supervisor-protect */
+ #define CF_CACR_IVO (0x00100000) /* Invalidate only */
+ #define CF_CACR_BEC (0x00080000) /* Branch Cache Enable */
+ #define CF_CACR_BCINVA (0x00040000) /* Branch Cache Invalidate All */
+@@ -24,61 +37,43 @@
+ #define CF_CACR_IHLCK (0x00000800) /* 1/2 Instruction Cache Lock Mode */
+ #define CF_CACR_IDCM (0x00000400) /* Noncacheable Instr default mode */
+ #define CF_CACR_ICINVA (0x00000100) /* Instr Cache Invalidate All */
++#define CF_CACR_IDSP (0x00000080) /* Ins default supervisor-protect */
+ #define CF_CACR_EUSP (0x00000020) /* Switch stacks in user mode */
+
+-#define DCACHE_LINE_SIZE 0x0010 /* bytes per line */
+-#define DCACHE_WAY_SIZE 0x2000 /* words per cache block */
+-#define CACHE_DISABLE_MODE (CF_CACR_DCINVA+CF_CACR_BCINVA+CF_CACR_ICINVA)
+-#ifdef CONFIG_M5445X_DISABLE_CACHE
+-/* disable cache for testing rev0 silicon */
+-#define CACHE_INITIAL_MODE (CF_CACR_EUSP)
+-#else
+-#define CACHE_INITIAL_MODE (CF_CACR_DEC+CF_CACR_BEC+CF_CACR_IEC+CF_CACR_EUSP)
+-#endif
+-
+-#define _DCACHE_SIZE (2*16384)
+-#define _ICACHE_SIZE (2*16384)
+-
+-#define _SET_SHIFT 4
+-
++#ifdef CONFIG_M54455
+ /*
+- * Masks for cache sizes. Programming note: because the set size is a
+- * power of two, the mask is also the last address in the set.
+- * This may need to be #ifdef for other Coldfire processors.
++ * M5445x Cache Configuration
++ * - cache line size is 16 bytes
++ * - cache is 4-way set associative
++ * - each cache has 256 sets (64k / 16bytes / 4way)
++ * - I-Cache size is 16KB
++ * - D-Cache size is 16KB
+ */
++#define ICACHE_SIZE 0x4000 /* instruction - 16k */
++#define DCACHE_SIZE 0x4000 /* data - 16k */
+
+-#define _DCACHE_SET_MASK ((_DCACHE_SIZE/64-1)<<_SET_SHIFT)
+-#define _ICACHE_SET_MASK ((_ICACHE_SIZE/64-1)<<_SET_SHIFT)
+-#define LAST_DCACHE_ADDR _DCACHE_SET_MASK
+-#define LAST_ICACHE_ADDR _ICACHE_SET_MASK
+-
++#define CACHE_LINE_SIZE 0x0010 /* 16 bytes */
++#define CACHE_SETS 0x0100 /* 256 sets */
++#define CACHE_WAYS 0x0004 /* 4 way */
++
++#define CACHE_DISABLE_MODE (CF_CACR_DCINVA+ \
++ CF_CACR_BCINVA+ \
++ CF_CACR_ICINVA)
++
++#ifndef CONFIG_M5445X_DISABLE_CACHE
++#define CACHE_INITIAL_MODE (CF_CACR_DEC+ \
++ CF_CACR_BEC+ \
++ CF_CACR_IEC+ \
++ CF_CACR_EUSP)
++#else
++/* cache disabled for testing */
++#define CACHE_INITIAL_MODE (CF_CACR_EUSP)
++#endif /* CONFIG_M5445X_DISABLE_CACHE */
++#endif /* CONFIG_M54455 */
+
+ #ifndef __ASSEMBLY__
+
+-extern void DcacheFlushInvalidate(void);
+-
+-extern void DcacheDisable(void);
+-extern void DcacheEnable(void);
+-
+-/******************************************************************************/
+-/*** Unimplemented Cache functionality ***/
+-/******************************************************************************/
+-#define preDcacheInvalidateBlockMark()
+-#define postDcacheInvalidateBlockMark()
+-#define DcacheZeroBlock(p, l) fast_bzero((char *)(p), (long)(l))
+-#define loadDcacheInvalidateBlock() ASSERT(!"Not Implemented on V4e")
+-#define IcacheInvalidateBlock() ASSERT(!"Not Implemented on V4e")
+-
+-/******************************************************************************/
+-/*** Redundant Cache functionality on ColdFire ***/
+-/******************************************************************************/
+-#define DcacheInvalidateBlock(p, l) DcacheFlushInvalidateCacheBlock(p, l)
+-#define DcacheFlushCacheBlock(p, l) DcacheFlushInvalidateCacheBlock(p, l)
+-#define DcacheFlushBlock(p, l) DcacheFlushInvalidateCacheBlock(p, l)
+-
+-extern void DcacheFlushInvalidateCacheBlock(void *start, unsigned long size);
+-extern void FLASHDcacheFlushInvalidate(void);
+-
++extern unsigned long shadow_cacr;
+ extern void cacr_set(unsigned long x);
+
+ #endif /* !__ASSEMBLY__ */