]> rtime.felk.cvut.cz Git - mcf548x/linux.git/blobdiff - arch/m68k/include/asm/cacheflush_coldfire_m547x_8x.h
Cache headers
[mcf548x/linux.git] / arch / m68k / include / asm / cacheflush_coldfire_m547x_8x.h
diff --git a/arch/m68k/include/asm/cacheflush_coldfire_m547x_8x.h b/arch/m68k/include/asm/cacheflush_coldfire_m547x_8x.h
new file mode 100644 (file)
index 0000000..762f9fa
--- /dev/null
@@ -0,0 +1,258 @@
+/*
+ * arch/m68k/include/asm/cacheflush_coldfire_m547x_8x.h - Coldfire 547x/548x Cache
+ *
+ * Based on arch/m68k/include/asm/cacheflush.h
+ *
+ * Coldfire pieces by:
+ *   Kurt Mahan kmahan@freescale.com
+ *
+ * Copyright Freescale Semiconductor, Inc. 2007, 2008
+ *
+ * This program is free software; you can redistribute  it and/or modify it
+ * under  the terms of  the GNU General  Public License as published by the
+ * Free Software Foundation;  either version 2 of the  License, or (at your
+ * option) any later version.
+ */
+#ifndef _M68K_CACHEFLUSH_CODLFIRE_M547X_8X_H
+#define _M68K_CACHEFLUSH_CODLFIRE_M547X_8X_H
+
+/*
+ * Cache handling functions
+ */
+
+#define flush_icache()                                         \
+({                                                             \
+  unsigned long set;                                           \
+  unsigned long start_set;                                     \
+  unsigned long end_set;                                       \
+                                                               \
+  start_set = 0;                                               \
+  end_set = (unsigned long)LAST_DCACHE_ADDR;                   \
+                                                               \
+  for (set = start_set; set <= end_set; set += (0x10 - 3)) {   \
+    asm volatile("cpushl %%ic,(%0)\n"                          \
+                 "\taddq%.l #1,%0\n"                           \
+                 "\tcpushl %%ic,(%0)\n"                                \
+                 "\taddq%.l #1,%0\n"                           \
+                 "\tcpushl %%ic,(%0)\n"                                \
+                 "\taddq%.l #1,%0\n"                           \
+                 "\tcpushl %%ic,(%0)" : "=a" (set) : "a" (set));               \
+  }                                                            \
+})
+
+#define flush_dcache()                                         \
+({                                                             \
+  unsigned long set;                                           \
+  unsigned long start_set;                                     \
+  unsigned long end_set;                                       \
+                                                               \
+  start_set = 0;                                               \
+  end_set = (unsigned long)LAST_DCACHE_ADDR;                   \
+                                                               \
+  for (set = start_set; set <= end_set; set += (0x10 - 3)) {   \
+    asm volatile("cpushl %%dc,(%0)\n"                          \
+                 "\taddq%.l #1,%0\n"                           \
+                 "\tcpushl %%dc,(%0)\n"                                \
+                 "\taddq%.l #1,%0\n"                           \
+                 "\tcpushl %%dc,(%0)\n"                                \
+                 "\taddq%.l #1,%0\n"                           \
+                 "\tcpushl %%dc,(%0)" : "=a" (set) : "a" (set));               \
+  }                                                            \
+})
+
+#define flush_bcache()                                         \
+({                                                             \
+  unsigned long set;                                           \
+  unsigned long start_set;                                     \
+  unsigned long end_set;                                       \
+                                                               \
+  start_set = 0;                                               \
+  end_set = (unsigned long)LAST_DCACHE_ADDR;                   \
+                                                               \
+  for (set = start_set; set <= end_set; set += (0x10 - 3)) {   \
+    asm volatile("cpushl %%bc,(%0)\n"                          \
+                 "\taddq%.l #1,%0\n"                           \
+                 "\tcpushl %%bc,(%0)\n"                                \
+                 "\taddq%.l #1,%0\n"                           \
+                 "\tcpushl %%bc,(%0)\n"                                \
+                 "\taddq%.l #1,%0\n"                           \
+                 "\tcpushl %%bc,(%0)" : "=a" (set) : "a" (set));               \
+  }                                                            \
+})
+
+/*
+ * invalidate the cache for the specified memory range.
+ * It starts at the physical address specified for
+ * the given number of bytes.
+ */
+extern void cache_clear(unsigned long paddr, int len);
+/*
+ * push any dirty cache in the specified memory range.
+ * It starts at the physical address specified for
+ * the given number of bytes.
+ */
+extern void cache_push(unsigned long paddr, int len);
+
+/*
+ * push and invalidate pages in the specified user virtual
+ * memory range.
+ */
+extern void cache_push_v(unsigned long vaddr, int len);
+
+/* This is needed whenever the virtual mapping of the current
+   process changes.  */
+
+/**
+ * flush_cache_mm - Flush an mm_struct
+ * @mm: mm_struct to flush
+ */
+static inline void flush_cache_mm(struct mm_struct *mm)
+{
+       if (mm == current->mm)
+               flush_bcache();
+}
+
+#define flush_cache_dup_mm(mm) flush_cache_mm(mm)
+
+#define flush_cache_all()              flush_bcache()
+
+/**
+ * flush_cache_range - Flush a cache range
+ * @vma: vma struct
+ * @start: Starting address
+ * @end: Ending address
+ *
+ * flush_cache_range must be a macro to avoid a dependency on
+ * linux/mm.h which includes this file.
+ */
+static inline void flush_cache_range(struct vm_area_struct *vma,
+       unsigned long start, unsigned long end)
+{
+       if (vma->vm_mm == current->mm)
+               flush_bcache();
+//             cf_cache_flush_range(start, end);
+}
+
+/**
+ * flush_cache_page - Flush a page of the cache
+ * @vma: vma struct
+ * @vmaddr:
+ * @pfn: page numer
+ *
+ * flush_cache_page must be a macro to avoid a dependency on
+ * linux/mm.h which includes this file.
+ */
+static inline void flush_cache_page(struct vm_area_struct *vma,
+       unsigned long vmaddr, unsigned long pfn)
+{
+       if (vma->vm_mm == current->mm)
+               flush_bcache();
+//             cf_cache_flush_range(vmaddr, vmaddr+PAGE_SIZE);
+}
+
+/* Push the page at kernel virtual address and clear the icache */
+/* RZ: use cpush %bc instead of cpush %dc, cinv %ic */
+#define flush_page_to_ram(page) __flush_page_to_ram((void *) page_address(page))
+extern inline void __flush_page_to_ram(void *address)
+{
+  unsigned long set;
+  unsigned long start_set;
+  unsigned long end_set;
+  unsigned long addr = (unsigned long) address;
+
+  addr &= ~(PAGE_SIZE - 1); /* round down to page start address */
+
+  start_set = addr & _ICACHE_SET_MASK;
+  end_set = (addr + PAGE_SIZE-1) & _ICACHE_SET_MASK;
+
+  if (start_set > end_set) {
+    /* from the begining to the lowest address */
+    for (set = 0; set <= end_set; set += (0x10 - 3)) {
+      asm volatile("cpushl %%bc,(%0)\n"
+                   "\taddq%.l #1,%0\n"
+                   "\tcpushl %%bc,(%0)\n"
+                   "\taddq%.l #1,%0\n"
+                   "\tcpushl %%bc,(%0)\n"
+                   "\taddq%.l #1,%0\n"
+                   "\tcpushl %%bc,(%0)" : "=a" (set) : "a" (set));
+    }
+    /* next loop will finish the cache ie pass the hole */
+    end_set = LAST_ICACHE_ADDR;    
+  }
+  for (set = start_set; set <= end_set; set += (0x10 - 3)) {
+    asm volatile("cpushl %%bc,(%0)\n"
+                 "\taddq%.l #1,%0\n"
+                 "\tcpushl %%bc,(%0)\n"
+                 "\taddq%.l #1,%0\n"
+                 "\tcpushl %%bc,(%0)\n"
+                 "\taddq%.l #1,%0\n"
+                 "\tcpushl %%bc,(%0)" : "=a" (set) : "a" (set));
+  }
+}
+
+/* Use __flush_page_to_ram() for flush_dcache_page all values are same - MW */
+#define flush_dcache_page(page)                        \
+       __flush_page_to_ram((void *) page_address(page))
+#define flush_icache_page(vma,pg)              \
+       __flush_page_to_ram((void *) page_address(pg))
+#define flush_icache_user_range(adr,len)       do { } while (0)
+/* NL */
+#define flush_icache_user_page(vma,page,addr,len)      do { } while (0)
+
+/* Push n pages at kernel virtual address and clear the icache */
+/* RZ: use cpush %bc instead of cpush %dc, cinv %ic */
+extern inline void flush_icache_range (unsigned long address,
+                                      unsigned long endaddr)
+{
+  unsigned long set;
+  unsigned long start_set;
+  unsigned long end_set;
+
+  start_set = address & _ICACHE_SET_MASK;
+  end_set = endaddr & _ICACHE_SET_MASK;
+
+  if (start_set > end_set) {
+    /* from the begining to the lowest address */
+    for (set = 0; set <= end_set; set += (0x10 - 3)) {
+      asm volatile("cpushl %%ic,(%0)\n"
+                   "\taddq%.l #1,%0\n"
+                   "\tcpushl %%ic,(%0)\n"
+                   "\taddq%.l #1,%0\n"
+                   "\tcpushl %%ic,(%0)\n"
+                   "\taddq%.l #1,%0\n"
+                   "\tcpushl %%ic,(%0)" : "=a" (set) : "a" (set));
+    }
+    /* next loop will finish the cache ie pass the hole */
+    end_set = LAST_ICACHE_ADDR;    
+  }
+  for (set = start_set; set <= end_set; set += (0x10 - 3)) {
+    asm volatile("cpushl %%ic,(%0)\n"
+                 "\taddq%.l #1,%0\n"
+                 "\tcpushl %%ic,(%0)\n"
+                 "\taddq%.l #1,%0\n"
+                 "\tcpushl %%ic,(%0)\n"
+                 "\taddq%.l #1,%0\n"
+                 "\tcpushl %%ic,(%0)" : "=a" (set) : "a" (set));
+  }
+}
+
+static inline void copy_to_user_page(struct vm_area_struct *vma,
+                                    struct page *page, unsigned long vaddr,
+                                    void *dst, void *src, int len)
+{
+       memcpy(dst, src, len);
+       flush_icache_user_page(vma, page, vaddr, len);
+}
+static inline void copy_from_user_page(struct vm_area_struct *vma,
+                                      struct page *page, unsigned long vaddr,
+                                      void *dst, void *src, int len)
+{
+       memcpy(dst, src, len);
+}
+
+#define flush_cache_vmap(start, end)           flush_cache_all()
+#define flush_cache_vunmap(start, end)         flush_cache_all()
+#define flush_dcache_mmap_lock(mapping)                do { } while (0)
+#define flush_dcache_mmap_unlock(mapping)      do { } while (0)
+
+#endif /* _M68K_CACHEFLUSH_CODLFIRE_M547X_8X_H */