Merge git://www.denx.de/git/u-boot into uboot
[platform/kernel/u-boot.git] / include / asm-mips / system.h
1 /*
2  * This file is subject to the terms and conditions of the GNU General Public
3  * License.  See the file "COPYING" in the main directory of this archive
4  * for more details.
5  *
6  * Copyright (C) 1994 - 1999 by Ralf Baechle
7  * Copyright (C) 1996 by Paul M. Antoine
8  * Copyright (C) 1994 - 1999 by Ralf Baechle
9  *
10  * Changed set_except_vector declaration to allow return of previous
11  * vector address value - necessary for "borrowing" vectors.
12  *
13  * Kevin D. Kissell, kevink@mips.org and Carsten Langgaard, carstenl@mips.com
14  * Copyright (C) 2000 MIPS Technologies, Inc.
15  */
16 #ifndef _ASM_SYSTEM_H
17 #define _ASM_SYSTEM_H
18
19 #include <linux/config.h>
20 #include <asm/sgidefs.h>
21 #include <asm/ptrace.h>
22 #if 0
23 #include <linux/kernel.h>
24 #endif
25
26 extern __inline__ void
27 __sti(void)
28 {
29         __asm__ __volatile__(
30                 ".set\tpush\n\t"
31                 ".set\treorder\n\t"
32                 ".set\tnoat\n\t"
33                 "mfc0\t$1,$12\n\t"
34                 "ori\t$1,0x1f\n\t"
35                 "xori\t$1,0x1e\n\t"
36                 "mtc0\t$1,$12\n\t"
37                 ".set\tpop\n\t"
38                 : /* no outputs */
39                 : /* no inputs */
40                 : "$1", "memory");
41 }
42
43 /*
44  * For cli() we have to insert nops to make shure that the new value
45  * has actually arrived in the status register before the end of this
46  * macro.
47  * R4000/R4400 need three nops, the R4600 two nops and the R10000 needs
48  * no nops at all.
49  */
50 extern __inline__ void
51 __cli(void)
52 {
53         __asm__ __volatile__(
54                 ".set\tpush\n\t"
55                 ".set\treorder\n\t"
56                 ".set\tnoat\n\t"
57                 "mfc0\t$1,$12\n\t"
58                 "ori\t$1,1\n\t"
59                 "xori\t$1,1\n\t"
60                 ".set\tnoreorder\n\t"
61                 "mtc0\t$1,$12\n\t"
62                 "nop\n\t"
63                 "nop\n\t"
64                 "nop\n\t"
65                 ".set\tpop\n\t"
66                 : /* no outputs */
67                 : /* no inputs */
68                 : "$1", "memory");
69 }
70
71 #define __save_flags(x)                                                 \
72 __asm__ __volatile__(                                                   \
73         ".set\tpush\n\t"                                                \
74         ".set\treorder\n\t"                                             \
75         "mfc0\t%0,$12\n\t"                                              \
76         ".set\tpop\n\t"                                                 \
77         : "=r" (x))
78
79 #define __save_and_cli(x)                                               \
80 __asm__ __volatile__(                                                   \
81         ".set\tpush\n\t"                                                \
82         ".set\treorder\n\t"                                             \
83         ".set\tnoat\n\t"                                                \
84         "mfc0\t%0,$12\n\t"                                              \
85         "ori\t$1,%0,1\n\t"                                              \
86         "xori\t$1,1\n\t"                                                \
87         ".set\tnoreorder\n\t"                                           \
88         "mtc0\t$1,$12\n\t"                                              \
89         "nop\n\t"                                                       \
90         "nop\n\t"                                                       \
91         "nop\n\t"                                                       \
92         ".set\tpop\n\t"                                                 \
93         : "=r" (x)                                                      \
94         : /* no inputs */                                               \
95         : "$1", "memory")
96
97 #define __restore_flags(flags)                                          \
98 do {                                                                    \
99         unsigned long __tmp1;                                           \
100                                                                         \
101         __asm__ __volatile__(                                           \
102                 ".set\tnoreorder\t\t\t# __restore_flags\n\t"            \
103                 ".set\tnoat\n\t"                                        \
104                 "mfc0\t$1, $12\n\t"                                     \
105                 "andi\t%0, 1\n\t"                                       \
106                 "ori\t$1, 1\n\t"                                        \
107                 "xori\t$1, 1\n\t"                                       \
108                 "or\t%0, $1\n\t"                                        \
109                 "mtc0\t%0, $12\n\t"                                     \
110                 "nop\n\t"                                               \
111                 "nop\n\t"                                               \
112                 "nop\n\t"                                               \
113                 ".set\tat\n\t"                                          \
114                 ".set\treorder"                                         \
115                 : "=r" (__tmp1)                                         \
116                 : "0" (flags)                                           \
117                 : "$1", "memory");                                      \
118 } while(0)
119
120 #ifdef CONFIG_SMP
121
122 extern void __global_sti(void);
123 extern void __global_cli(void);
124 extern unsigned long __global_save_flags(void);
125 extern void __global_restore_flags(unsigned long);
126 #  define sti() __global_sti()
127 #  define cli() __global_cli()
128 #  define save_flags(x) do { x = __global_save_flags(); } while (0)
129 #  define restore_flags(x) __global_restore_flags(x)
130 #  define save_and_cli(x) do { save_flags(x); cli(); } while(0)
131
132 #else /* Single processor */
133
134 #  define sti() __sti()
135 #  define cli() __cli()
136 #  define save_flags(x) __save_flags(x)
137 #  define save_and_cli(x) __save_and_cli(x)
138 #  define restore_flags(x) __restore_flags(x)
139
140 #endif /* SMP */
141
142 /* For spinlocks etc */
143 #define local_irq_save(x)       __save_and_cli(x);
144 #define local_irq_restore(x)    __restore_flags(x);
145 #define local_irq_disable()     __cli();
146 #define local_irq_enable()      __sti();
147
148 /*
149  * These are probably defined overly paranoid ...
150  */
151 #ifdef CONFIG_CPU_HAS_WB
152
153 #include <asm/wbflush.h>
154 #define rmb()   do { } while(0)
155 #define wmb()   wbflush()
156 #define mb()    wbflush()
157
158 #else /* CONFIG_CPU_HAS_WB  */
159
160 #define mb()                                            \
161 __asm__ __volatile__(                                   \
162         "# prevent instructions being moved around\n\t" \
163         ".set\tnoreorder\n\t"                           \
164         "# 8 nops to fool the R4400 pipeline\n\t"       \
165         "nop;nop;nop;nop;nop;nop;nop;nop\n\t"           \
166         ".set\treorder"                                 \
167         : /* no output */                               \
168         : /* no input */                                \
169         : "memory")
170 #define rmb() mb()
171 #define wmb() mb()
172
173 #endif /* CONFIG_CPU_HAS_WB  */
174
175 #ifdef CONFIG_SMP
176 #define smp_mb()        mb()
177 #define smp_rmb()       rmb()
178 #define smp_wmb()       wmb()
179 #else
180 #define smp_mb()        barrier()
181 #define smp_rmb()       barrier()
182 #define smp_wmb()       barrier()
183 #endif
184
185 #define set_mb(var, value) \
186 do { var = value; mb(); } while (0)
187
188 #define set_wmb(var, value) \
189 do { var = value; wmb(); } while (0)
190
191 #if !defined (_LANGUAGE_ASSEMBLY)
192 /*
193  * switch_to(n) should switch tasks to task nr n, first
194  * checking that n isn't the current task, in which case it does nothing.
195  */
196 #if 0
197 extern asmlinkage void *resume(void *last, void *next);
198 #endif
199 #endif /* !defined (_LANGUAGE_ASSEMBLY) */
200
201 #define prepare_to_switch()     do { } while(0)
202 #define switch_to(prev,next,last) \
203 do { \
204         (last) = resume(prev, next); \
205 } while(0)
206
207 /*
208  * For 32 and 64 bit operands we can take advantage of ll and sc.
209  * FIXME: This doesn't work for R3000 machines.
210  */
211 extern __inline__ unsigned long xchg_u32(volatile int * m, unsigned long val)
212 {
213 #ifdef CONFIG_CPU_HAS_LLSC
214         unsigned long dummy;
215
216         __asm__ __volatile__(
217                 ".set\tnoreorder\t\t\t# xchg_u32\n\t"
218                 ".set\tnoat\n\t"
219                 "ll\t%0, %3\n"
220                 "1:\tmove\t$1, %2\n\t"
221                 "sc\t$1, %1\n\t"
222                 "beqzl\t$1, 1b\n\t"
223                 " ll\t%0, %3\n\t"
224                 ".set\tat\n\t"
225                 ".set\treorder"
226                 : "=r" (val), "=o" (*m), "=r" (dummy)
227                 : "o" (*m), "2" (val)
228                 : "memory");
229
230         return val;
231 #else
232         unsigned long flags, retval;
233
234         save_flags(flags);
235         cli();
236         retval = *m;
237         *m = val;
238         restore_flags(flags);
239         return retval;
240 #endif /* Processor-dependent optimization */
241 }
242
243 #define xchg(ptr,x) ((__typeof__(*(ptr)))__xchg((unsigned long)(x),(ptr),sizeof(*(ptr))))
244 #define tas(ptr) (xchg((ptr),1))
245
246 static __inline__ unsigned long
247 __xchg(unsigned long x, volatile void * ptr, int size)
248 {
249         switch (size) {
250                 case 4:
251                         return xchg_u32(ptr, x);
252         }
253         return x;
254 }
255
256 extern void *set_except_vector(int n, void *addr);
257
258 extern void __die(const char *, struct pt_regs *, const char *where,
259         unsigned long line) __attribute__((noreturn));
260 extern void __die_if_kernel(const char *, struct pt_regs *, const char *where,
261         unsigned long line);
262
263 #define die(msg, regs)                                                  \
264         __die(msg, regs, __FILE__ ":"__FUNCTION__, __LINE__)
265 #define die_if_kernel(msg, regs)                                        \
266         __die_if_kernel(msg, regs, __FILE__ ":"__FUNCTION__, __LINE__)
267
268 #endif /* _ASM_SYSTEM_H */