4 * Copyright 2005-2009 Analog Devices Inc.
5 * D. Jeff Dionne <jeff@ryeham.ee.ryerson.ca>
6 * Kenneth Albanowski <kjahds@kjahds.com>
8 * Licensed under the GPL-2 or later.
11 #include <asm/blackfin.h>
13 #include <linux/linkage.h>
14 #include <asm/entry.h>
15 #include <asm/asm-offsets.h>
16 #include <asm/trace.h>
17 #include <asm/traps.h>
18 #include <asm/thread_info.h>
20 #include <asm/context.S>
22 .extern _ret_from_exception
24 #ifdef CONFIG_I_ENTRY_L1
30 .align 4 /* just in case */
32 /* Common interrupt entry code. First we do CLI, then push
33 * RETI, to keep interrupts disabled, but to allow this state to be changed
35 * R0 contains the interrupt number, while R1 may contain the value of IPEND,
36 * or garbage if IPEND won't be needed by the ISR. */
74 [--sp] = r0; /* Skip reserved */
82 [--sp] = r1; /* IPEND - R1 may or may not be set up before jumping here. */
84 /* Switch to other method of keeping interrupts disabled. */
85 #ifdef CONFIG_DEBUG_HWERR
91 #ifdef CONFIG_TRACE_IRQFLAGS
94 call _trace_hardirqs_off;
98 [--sp] = RETI; /* orig_pc */
99 /* Clear all L registers. */
105 #ifdef CONFIG_FRAME_POINTER
109 ANOMALY_283_315_WORKAROUND(p5, r7)
114 call ___ipipe_grab_irq
117 if cc jump .Lcommon_restore_context;
118 #else /* CONFIG_IPIPE */
120 #ifdef CONFIG_PREEMPT
122 r4.l = lo(ALIGN_PAGE_MASK);
123 r4.h = hi(ALIGN_PAGE_MASK);
126 r7 = [p5 + TI_PREEMPT]; /* get preempt count */
127 r7 += 1; /* increment it */
128 [p5 + TI_PREEMPT] = r7;
130 pseudo_long_call _do_irq, p2;
132 #ifdef CONFIG_PREEMPT
134 [p5 + TI_PREEMPT] = r7; /* restore preempt count */
138 #endif /* CONFIG_IPIPE */
139 pseudo_long_call _return_from_int, p2;
140 .Lcommon_restore_context:
144 /* interrupt routine for ivhw - 5 */
146 /* In case a single action kicks off multiple memory transactions, (like
147 * a cache line fetch, - this can cause multiple hardware errors, let's
148 * catch them all. First - make sure all the actions are complete, and
149 * the core sees the hardware errors.
155 #ifdef CONFIG_FRAME_POINTER
159 ANOMALY_283_315_WORKAROUND(p5, r7)
161 /* Handle all stacked hardware errors
162 * To make sure we don't hang forever, only do it 10 times
170 CC = BITTST(R1, EVT_IVHW_P);
172 /* OK a hardware error is pending - clear it */
180 # We are going to dump something out, so make sure we print IPEND properly
184 [sp + PT_IPEND] = r0;
186 /* set the EXCAUSE to HWERR for trap_c */
187 r0 = [sp + PT_SEQSTAT];
188 R1.L = LO(VEC_HWERR);
189 R1.H = HI(VEC_HWERR);
191 [sp + PT_SEQSTAT] = R0;
193 r0 = sp; /* stack frame pt_regs pointer argument ==> r0 */
195 pseudo_long_call _trap_c, p5;
199 /* make sure EBIU_ERRMST is clear */
200 p0.l = LO(EBIU_ERRMST);
201 p0.h = HI(EBIU_ERRMST);
202 r0.l = (CORE_ERROR | CORE_MERROR);
206 pseudo_long_call _ret_from_exception, p2;
208 .Lcommon_restore_all_sys:
213 /* Interrupt routine for evt2 (NMI).
214 * For inner circle type details, please see:
215 * http://docs.blackfin.uclinux.org/doku.php?id=linux-kernel:nmi
218 #ifndef CONFIG_NMI_WATCHDOG
221 /* Not take account of CPLBs, this handler will not return */
226 trace_buffer_save(p4,r5);
228 ANOMALY_283_315_WORKAROUND(p4, r5)
239 /* interrupt routine for core timer - 6 */
241 TIMER_INTERRUPT_ENTRY(EVT_IVTMR_P)
243 /* interrupt routine for evt7 - 7 */
245 INTERRUPT_ENTRY(EVT_IVG7_P)
247 INTERRUPT_ENTRY(EVT_IVG8_P)
249 INTERRUPT_ENTRY(EVT_IVG9_P)
251 INTERRUPT_ENTRY(EVT_IVG10_P)
253 INTERRUPT_ENTRY(EVT_IVG11_P)
255 INTERRUPT_ENTRY(EVT_IVG12_P)
257 INTERRUPT_ENTRY(EVT_IVG13_P)
260 /* interrupt routine for system_call - 15 */
261 ENTRY(_evt_system_call)
263 #ifdef CONFIG_FRAME_POINTER
266 pseudo_long_call _system_call, p2;
267 jump .Lcommon_restore_context;
268 ENDPROC(_evt_system_call)
272 * __ipipe_call_irqtail: lowers the current priority level to EVT15
273 * before running a user-defined routine, then raises the priority
274 * level to EVT14 to prepare the caller for a normal interrupt
275 * return through RTI.
277 * We currently use this facility in two occasions:
279 * - to branch to __ipipe_irq_tail_hook as requested by a high
280 * priority domain after the pipeline delivered an interrupt,
281 * e.g. such as Xenomai, in order to start its rescheduling
282 * procedure, since we may not switch tasks when IRQ levels are
283 * nested on the Blackfin, so we have to fake an interrupt return
284 * so that we may reschedule immediately.
286 * - to branch to sync_root_irqs, in order to play any interrupt
287 * pending for the root domain (i.e. the Linux kernel). This lowers
288 * the core priority level enough so that Linux IRQ handlers may
289 * never delay interrupts handled by high priority domains; we defer
290 * those handlers until this point instead. This is a substitute
291 * to using a threaded interrupt model for the Linux kernel.
293 * r0: address of user-defined routine
294 * context: caller must have preempted EVT15, hw interrupts must be off.
296 ENTRY(___ipipe_call_irqtail)
304 [--sp] = ( r7:4, p5:3 );
308 ( r7:4, p5:3 ) = [sp++];
311 #ifdef CONFIG_DEBUG_HWERR
312 /* enable irq14 & hwerr interrupt, until we transition to _evt_evt14 */
313 r0 = (EVT_IVG14 | EVT_IVHW | \
314 EVT_IRPTEN | EVT_EVX | EVT_NMI | EVT_RST | EVT_EMU);
316 /* Only enable irq14 interrupt, until we transition to _evt_evt14 */
318 EVT_IRPTEN | EVT_EVX | EVT_NMI | EVT_RST | EVT_EMU);
321 raise 14; /* Branches to _evt_evt14 */
323 jump 2b; /* Likely paranoid. */
324 ENDPROC(___ipipe_call_irqtail)
326 #endif /* CONFIG_IPIPE */