Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
awilliam
GitHub Repository: awilliam/linux-vfio
Path: blob/master/arch/blackfin/mach-common/interrupt.S
10817 views
1
/*
2
* Interrupt Entries
3
*
4
* Copyright 2005-2009 Analog Devices Inc.
5
* D. Jeff Dionne <[email protected]>
6
* Kenneth Albanowski <[email protected]>
7
*
8
* Licensed under the GPL-2 or later.
9
*/
10
11
#include <asm/blackfin.h>
12
#include <mach/irq.h>
13
#include <linux/linkage.h>
14
#include <asm/entry.h>
15
#include <asm/asm-offsets.h>
16
#include <asm/trace.h>
17
#include <asm/traps.h>
18
#include <asm/thread_info.h>
19
20
#include <asm/context.S>
21
22
.extern _ret_from_exception
23
24
#ifdef CONFIG_I_ENTRY_L1
25
.section .l1.text
26
#else
27
.text
28
#endif
29
30
.align 4 /* just in case */
31
32
/* Common interrupt entry code. First we do CLI, then push
33
* RETI, to keep interrupts disabled, but to allow this state to be changed
34
* by local_bh_enable.
35
* R0 contains the interrupt number, while R1 may contain the value of IPEND,
36
* or garbage if IPEND won't be needed by the ISR. */
37
__common_int_entry:
38
[--sp] = fp;
39
[--sp] = usp;
40
41
[--sp] = i0;
42
[--sp] = i1;
43
[--sp] = i2;
44
[--sp] = i3;
45
46
[--sp] = m0;
47
[--sp] = m1;
48
[--sp] = m2;
49
[--sp] = m3;
50
51
[--sp] = l0;
52
[--sp] = l1;
53
[--sp] = l2;
54
[--sp] = l3;
55
56
[--sp] = b0;
57
[--sp] = b1;
58
[--sp] = b2;
59
[--sp] = b3;
60
[--sp] = a0.x;
61
[--sp] = a0.w;
62
[--sp] = a1.x;
63
[--sp] = a1.w;
64
65
[--sp] = LC0;
66
[--sp] = LC1;
67
[--sp] = LT0;
68
[--sp] = LT1;
69
[--sp] = LB0;
70
[--sp] = LB1;
71
72
[--sp] = ASTAT;
73
74
[--sp] = r0; /* Skip reserved */
75
[--sp] = RETS;
76
r2 = RETI;
77
[--sp] = r2;
78
[--sp] = RETX;
79
[--sp] = RETN;
80
[--sp] = RETE;
81
[--sp] = SEQSTAT;
82
[--sp] = r1; /* IPEND - R1 may or may not be set up before jumping here. */
83
84
/* Switch to other method of keeping interrupts disabled. */
85
#ifdef CONFIG_DEBUG_HWERR
86
r1 = 0x3f;
87
sti r1;
88
#else
89
cli r1;
90
#endif
91
#ifdef CONFIG_TRACE_IRQFLAGS
92
[--sp] = r0;
93
sp += -12;
94
call _trace_hardirqs_off;
95
sp += 12;
96
r0 = [sp++];
97
#endif
98
[--sp] = RETI; /* orig_pc */
99
/* Clear all L registers. */
100
r1 = 0 (x);
101
l0 = r1;
102
l1 = r1;
103
l2 = r1;
104
l3 = r1;
105
#ifdef CONFIG_FRAME_POINTER
106
fp = 0;
107
#endif
108
109
ANOMALY_283_315_WORKAROUND(p5, r7)
110
111
r1 = sp;
112
SP += -12;
113
#ifdef CONFIG_IPIPE
114
call ___ipipe_grab_irq
115
SP += 12;
116
cc = r0 == 0;
117
if cc jump .Lcommon_restore_context;
118
#else /* CONFIG_IPIPE */
119
120
#ifdef CONFIG_PREEMPT
121
r7 = sp;
122
r4.l = lo(ALIGN_PAGE_MASK);
123
r4.h = hi(ALIGN_PAGE_MASK);
124
r7 = r7 & r4;
125
p5 = r7;
126
r7 = [p5 + TI_PREEMPT]; /* get preempt count */
127
r7 += 1; /* increment it */
128
[p5 + TI_PREEMPT] = r7;
129
#endif
130
pseudo_long_call _do_irq, p2;
131
132
#ifdef CONFIG_PREEMPT
133
r7 += -1;
134
[p5 + TI_PREEMPT] = r7; /* restore preempt count */
135
#endif
136
137
SP += 12;
138
#endif /* CONFIG_IPIPE */
139
pseudo_long_call _return_from_int, p2;
140
.Lcommon_restore_context:
141
RESTORE_CONTEXT
142
rti;
143
144
/* interrupt routine for ivhw - 5 */
145
ENTRY(_evt_ivhw)
146
/* In case a single action kicks off multiple memory transactions, (like
147
* a cache line fetch, - this can cause multiple hardware errors, let's
148
* catch them all. First - make sure all the actions are complete, and
149
* the core sees the hardware errors.
150
*/
151
SSYNC;
152
SSYNC;
153
154
SAVE_ALL_SYS
155
#ifdef CONFIG_FRAME_POINTER
156
fp = 0;
157
#endif
158
159
ANOMALY_283_315_WORKAROUND(p5, r7)
160
161
/* Handle all stacked hardware errors
162
* To make sure we don't hang forever, only do it 10 times
163
*/
164
R0 = 0;
165
R2 = 10;
166
1:
167
P0.L = LO(ILAT);
168
P0.H = HI(ILAT);
169
R1 = [P0];
170
CC = BITTST(R1, EVT_IVHW_P);
171
IF ! CC JUMP 2f;
172
/* OK a hardware error is pending - clear it */
173
R1 = EVT_IVHW_P;
174
[P0] = R1;
175
R0 += 1;
176
CC = R1 == R2;
177
if CC JUMP 2f;
178
JUMP 1b;
179
2:
180
# We are going to dump something out, so make sure we print IPEND properly
181
p2.l = lo(IPEND);
182
p2.h = hi(IPEND);
183
r0 = [p2];
184
[sp + PT_IPEND] = r0;
185
186
/* set the EXCAUSE to HWERR for trap_c */
187
r0 = [sp + PT_SEQSTAT];
188
R1.L = LO(VEC_HWERR);
189
R1.H = HI(VEC_HWERR);
190
R0 = R0 | R1;
191
[sp + PT_SEQSTAT] = R0;
192
193
r0 = sp; /* stack frame pt_regs pointer argument ==> r0 */
194
SP += -12;
195
pseudo_long_call _trap_c, p5;
196
SP += 12;
197
198
#ifdef EBIU_ERRMST
199
/* make sure EBIU_ERRMST is clear */
200
p0.l = LO(EBIU_ERRMST);
201
p0.h = HI(EBIU_ERRMST);
202
r0.l = (CORE_ERROR | CORE_MERROR);
203
w[p0] = r0.l;
204
#endif
205
206
pseudo_long_call _ret_from_exception, p2;
207
208
.Lcommon_restore_all_sys:
209
RESTORE_ALL_SYS
210
rti;
211
ENDPROC(_evt_ivhw)
212
213
/* Interrupt routine for evt2 (NMI).
214
* For inner circle type details, please see:
215
* http://docs.blackfin.uclinux.org/doku.php?id=linux-kernel:nmi
216
*/
217
ENTRY(_evt_nmi)
218
#ifndef CONFIG_NMI_WATCHDOG
219
.weak _evt_nmi
220
#else
221
/* Not take account of CPLBs, this handler will not return */
222
SAVE_ALL_SYS
223
r0 = sp;
224
r1 = retn;
225
[sp + PT_PC] = r1;
226
trace_buffer_save(p4,r5);
227
228
ANOMALY_283_315_WORKAROUND(p4, r5)
229
230
SP += -12;
231
call _do_nmi;
232
SP += 12;
233
1:
234
jump 1b;
235
#endif
236
rtn;
237
ENDPROC(_evt_nmi)
238
239
/* interrupt routine for core timer - 6 */
240
ENTRY(_evt_timer)
241
TIMER_INTERRUPT_ENTRY(EVT_IVTMR_P)
242
243
/* interrupt routine for evt7 - 7 */
244
ENTRY(_evt_evt7)
245
INTERRUPT_ENTRY(EVT_IVG7_P)
246
ENTRY(_evt_evt8)
247
INTERRUPT_ENTRY(EVT_IVG8_P)
248
ENTRY(_evt_evt9)
249
INTERRUPT_ENTRY(EVT_IVG9_P)
250
ENTRY(_evt_evt10)
251
INTERRUPT_ENTRY(EVT_IVG10_P)
252
ENTRY(_evt_evt11)
253
INTERRUPT_ENTRY(EVT_IVG11_P)
254
ENTRY(_evt_evt12)
255
INTERRUPT_ENTRY(EVT_IVG12_P)
256
ENTRY(_evt_evt13)
257
INTERRUPT_ENTRY(EVT_IVG13_P)
258
259
260
/* interrupt routine for system_call - 15 */
261
ENTRY(_evt_system_call)
262
SAVE_CONTEXT_SYSCALL
263
#ifdef CONFIG_FRAME_POINTER
264
fp = 0;
265
#endif
266
pseudo_long_call _system_call, p2;
267
jump .Lcommon_restore_context;
268
ENDPROC(_evt_system_call)
269
270
#ifdef CONFIG_IPIPE
271
/*
272
* __ipipe_call_irqtail: lowers the current priority level to EVT15
273
* before running a user-defined routine, then raises the priority
274
* level to EVT14 to prepare the caller for a normal interrupt
275
* return through RTI.
276
*
277
* We currently use this feature in two occasions:
278
*
279
* - before branching to __ipipe_irq_tail_hook as requested by a high
280
* priority domain after the pipeline delivered an interrupt,
281
* e.g. such as Xenomai, in order to start its rescheduling
282
* procedure, since we may not switch tasks when IRQ levels are
283
* nested on the Blackfin, so we have to fake an interrupt return
284
* so that we may reschedule immediately.
285
*
286
* - before branching to __ipipe_sync_root(), in order to play any interrupt
287
* pending for the root domain (i.e. the Linux kernel). This lowers
288
* the core priority level enough so that Linux IRQ handlers may
289
* never delay interrupts handled by high priority domains; we defer
290
* those handlers until this point instead. This is a substitute
291
* to using a threaded interrupt model for the Linux kernel.
292
*
293
* r0: address of user-defined routine
294
* context: caller must have preempted EVT15, hw interrupts must be off.
295
*/
296
ENTRY(___ipipe_call_irqtail)
297
p0 = r0;
298
r0.l = 1f;
299
r0.h = 1f;
300
reti = r0;
301
rti;
302
1:
303
[--sp] = rets;
304
[--sp] = ( r7:4, p5:3 );
305
sp += -12;
306
call (p0);
307
sp += 12;
308
( r7:4, p5:3 ) = [sp++];
309
rets = [sp++];
310
311
#ifdef CONFIG_DEBUG_HWERR
312
/* enable irq14 & hwerr interrupt, until we transition to _evt_evt14 */
313
r0 = (EVT_IVG14 | EVT_IVHW | \
314
EVT_IRPTEN | EVT_EVX | EVT_NMI | EVT_RST | EVT_EMU);
315
#else
316
/* Only enable irq14 interrupt, until we transition to _evt_evt14 */
317
r0 = (EVT_IVG14 | \
318
EVT_IRPTEN | EVT_EVX | EVT_NMI | EVT_RST | EVT_EMU);
319
#endif
320
sti r0;
321
raise 14; /* Branches to _evt_evt14 */
322
2:
323
jump 2b; /* Likely paranoid. */
324
ENDPROC(___ipipe_call_irqtail)
325
326
#endif /* CONFIG_IPIPE */
327
328