2 * include/asm-ppc/cache.h
4 #ifndef __ARCH_PPC_CACHE_H
5 #define __ARCH_PPC_CACHE_H
7 #include <asm/processor.h>
9 /* bytes per L1 cache line */
10 #if defined(CONFIG_MPC8xx)
11 #define L1_CACHE_SHIFT 4
12 #elif defined(CONFIG_PPC64BRIDGE)
13 #define L1_CACHE_SHIFT 7
14 #elif defined(CONFIG_E500MC)
15 #define L1_CACHE_SHIFT 6
17 #define L1_CACHE_SHIFT 5
20 #define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT)
23 * Use the L1 data cache line size value for the minimum DMA buffer alignment
26 #define ARCH_DMA_MINALIGN L1_CACHE_BYTES
28 #define L1_CACHE_ALIGN(x) (((x)+(L1_CACHE_BYTES-1))&~(L1_CACHE_BYTES-1))
29 #define L1_CACHE_PAGES 8
31 #define SMP_CACHE_BYTES L1_CACHE_BYTES
34 #define __cacheline_aligned __attribute__((__aligned__(L1_CACHE_BYTES)))
36 #define __cacheline_aligned \
37 __attribute__((__aligned__(L1_CACHE_BYTES))) \
38 __section(".data.cacheline_aligned")
41 #if defined(__KERNEL__) && !defined(__ASSEMBLY__)
42 extern void flush_dcache_range(unsigned long start, unsigned long stop);
43 extern void clean_dcache_range(unsigned long start, unsigned long stop);
44 extern void invalidate_dcache_range(unsigned long start, unsigned long stop);
45 extern void flush_dcache(void);
46 extern void invalidate_dcache(void);
47 extern void invalidate_icache(void);
48 #ifdef CONFIG_SYS_INIT_RAM_LOCK
49 extern void unlock_ram_in_cache(void);
50 #endif /* CONFIG_SYS_INIT_RAM_LOCK */
51 #endif /* __ASSEMBLY__ */
53 #if defined(__KERNEL__) && !defined(__ASSEMBLY__)
54 int l2cache_init(void);
55 void enable_cpc(void);
56 void disable_cpc_sram(void);
59 /* prep registers for L2 */
60 #define CACHECRBA 0x80000823 /* Cache configuration register address */
61 #define L2CACHE_MASK 0x03 /* Mask for 2 L2 Cache bits */
62 #define L2CACHE_512KB 0x00 /* 512KB */
63 #define L2CACHE_256KB 0x01 /* 256KB */
64 #define L2CACHE_1MB 0x02 /* 1MB */
65 #define L2CACHE_NONE 0x03 /* NONE */
66 #define L2CACHE_PARITY 0x08 /* Mask for L2 Cache Parity Protected bit */
69 /* Cache control on the MPC8xx is provided through some additional
70 * special purpose registers.
72 #define IC_CST 560 /* Instruction cache control/status */
73 #define IC_ADR 561 /* Address needed for some commands */
74 #define IC_DAT 562 /* Read-only data register */
75 #define DC_CST 568 /* Data cache control/status */
76 #define DC_ADR 569 /* Address needed for some commands */
77 #define DC_DAT 570 /* Read-only data register */
79 /* Commands. Only the first few are available to the instruction cache.
81 #define IDC_ENABLE 0x02000000 /* Cache enable */
82 #define IDC_DISABLE 0x04000000 /* Cache disable */
83 #define IDC_LDLCK 0x06000000 /* Load and lock */
84 #define IDC_UNLINE 0x08000000 /* Unlock line */
85 #define IDC_UNALL 0x0a000000 /* Unlock all */
86 #define IDC_INVALL 0x0c000000 /* Invalidate all */
88 #define DC_FLINE 0x0e000000 /* Flush data cache line */
89 #define DC_SFWT 0x01000000 /* Set forced writethrough mode */
90 #define DC_CFWT 0x03000000 /* Clear forced writethrough mode */
91 #define DC_SLES 0x05000000 /* Set little endian swap mode */
92 #define DC_CLES 0x07000000 /* Clear little endian swap mode */
96 #define IDC_ENABLED 0x80000000 /* Cache is enabled */
97 #define IDC_CERR1 0x00200000 /* Cache error 1 */
98 #define IDC_CERR2 0x00100000 /* Cache error 2 */
99 #define IDC_CERR3 0x00080000 /* Cache error 3 */
101 #define DC_DFWT 0x40000000 /* Data cache is forced write through */
102 #define DC_LES 0x20000000 /* Caches are little endian mode */
104 #if !defined(__ASSEMBLY__)
105 static inline uint rd_ic_cst(void)
107 return mfspr(IC_CST);
110 static inline void wr_ic_cst(uint val)
115 static inline void wr_ic_adr(uint val)
120 static inline uint rd_dc_cst(void)
122 return mfspr(DC_CST);
125 static inline void wr_dc_cst(uint val)
130 static inline void wr_dc_adr(uint val)
135 #endif /* CONFIG_MPC8xx */