2 * ARM Generic/Distributed Interrupt Controller
4 * Copyright (c) 2006-2007 CodeSourcery.
5 * Written by Paul Brook
7 * This code is licenced under the GPL.
10 /* This file contains implementation code for the RealView EB interrupt
11 controller, MPCore distributed interrupt controller and ARMv7-M
12 Nested Vectored Interrupt Controller. */
17 #define DPRINTF(fmt, args...) \
18 do { printf("arm_gic: " fmt , ##args); } while (0)
20 #define DPRINTF(fmt, args...) do {} while(0)
24 static const uint8_t gic_id
[] =
25 { 0x00, 0xb0, 0x1b, 0x00, 0x0d, 0xe0, 0x05, 0xb1 };
26 #define GIC_DIST_OFFSET 0
27 /* The NVIC has 16 internal vectors. However these are not exposed
28 through the normal GIC interface. */
29 #define GIC_BASE_IRQ 32
31 static const uint8_t gic_id
[] =
32 { 0x90, 0x13, 0x04, 0x00, 0x0d, 0xf0, 0x05, 0xb1 };
33 #define GIC_DIST_OFFSET 0x1000
34 #define GIC_BASE_IRQ 0
37 typedef struct gic_irq_state
39 /* ??? The documentation seems to imply the enable bits are global, even
40 for per-cpu interrupts. This seems strange. */
42 unsigned pending
:NCPU
;
45 unsigned model
:1; /* 0 = N:N, 1 = 1:N */
46 unsigned trigger
:1; /* nonzero = edge triggered. */
49 #define ALL_CPU_MASK ((1 << NCPU) - 1)
51 #define GIC_SET_ENABLED(irq) s->irq_state[irq].enabled = 1
52 #define GIC_CLEAR_ENABLED(irq) s->irq_state[irq].enabled = 0
53 #define GIC_TEST_ENABLED(irq) s->irq_state[irq].enabled
54 #define GIC_SET_PENDING(irq, cm) s->irq_state[irq].pending |= (cm)
55 #define GIC_CLEAR_PENDING(irq, cm) s->irq_state[irq].pending &= ~(cm)
56 #define GIC_TEST_PENDING(irq, cm) ((s->irq_state[irq].pending & (cm)) != 0)
57 #define GIC_SET_ACTIVE(irq, cm) s->irq_state[irq].active |= (cm)
58 #define GIC_CLEAR_ACTIVE(irq, cm) s->irq_state[irq].active &= ~(cm)
59 #define GIC_TEST_ACTIVE(irq, cm) ((s->irq_state[irq].active & (cm)) != 0)
60 #define GIC_SET_MODEL(irq) s->irq_state[irq].model = 1
61 #define GIC_CLEAR_MODEL(irq) s->irq_state[irq].model = 0
62 #define GIC_TEST_MODEL(irq) s->irq_state[irq].model
63 #define GIC_SET_LEVEL(irq, cm) s->irq_state[irq].level = (cm)
64 #define GIC_CLEAR_LEVEL(irq, cm) s->irq_state[irq].level &= ~(cm)
65 #define GIC_TEST_LEVEL(irq, cm) ((s->irq_state[irq].level & (cm)) != 0)
66 #define GIC_SET_TRIGGER(irq) s->irq_state[irq].trigger = 1
67 #define GIC_CLEAR_TRIGGER(irq) s->irq_state[irq].trigger = 0
68 #define GIC_TEST_TRIGGER(irq) s->irq_state[irq].trigger
69 #define GIC_GET_PRIORITY(irq, cpu) \
70 (((irq) < 32) ? s->priority1[irq][cpu] : s->priority2[(irq) - 32])
72 #define GIC_TARGET(irq) 1
74 #define GIC_TARGET(irq) s->irq_target[irq]
77 typedef struct gic_state
80 qemu_irq parent_irq
[NCPU
];
82 int cpu_enabled
[NCPU
];
84 gic_irq_state irq_state
[GIC_NIRQ
];
86 int irq_target
[GIC_NIRQ
];
88 int priority1
[32][NCPU
];
89 int priority2
[GIC_NIRQ
- 32];
90 int last_active
[GIC_NIRQ
][NCPU
];
92 int priority_mask
[NCPU
];
93 int running_irq
[NCPU
];
94 int running_priority
[NCPU
];
95 int current_pending
[NCPU
];
103 /* TODO: Many places that call this routine could be optimized. */
104 /* Update interrupt status after enabled or pending bits have been changed. */
105 static void gic_update(gic_state
*s
)
114 for (cpu
= 0; cpu
< NCPU
; cpu
++) {
116 s
->current_pending
[cpu
] = 1023;
117 if (!s
->enabled
|| !s
->cpu_enabled
[cpu
]) {
118 qemu_irq_lower(s
->parent_irq
[cpu
]);
123 for (irq
= 0; irq
< GIC_NIRQ
; irq
++) {
124 if (GIC_TEST_ENABLED(irq
) && GIC_TEST_PENDING(irq
, cm
)) {
125 if (GIC_GET_PRIORITY(irq
, cpu
) < best_prio
) {
126 best_prio
= GIC_GET_PRIORITY(irq
, cpu
);
132 if (best_prio
<= s
->priority_mask
[cpu
]) {
133 s
->current_pending
[cpu
] = best_irq
;
134 if (best_prio
< s
->running_priority
[cpu
]) {
135 DPRINTF("Raised pending IRQ %d\n", best_irq
);
139 qemu_set_irq(s
->parent_irq
[cpu
], level
);
143 static void __attribute__((unused
))
144 gic_set_pending_private(gic_state
*s
, int cpu
, int irq
)
148 if (GIC_TEST_PENDING(irq
, cm
))
151 DPRINTF("Set %d pending cpu %d\n", irq
, cpu
);
152 GIC_SET_PENDING(irq
, cm
);
156 /* Process a change in an external IRQ input. */
157 static void gic_set_irq(void *opaque
, int irq
, int level
)
159 gic_state
*s
= (gic_state
*)opaque
;
160 /* The first external input line is internal interrupt 32. */
162 if (level
== GIC_TEST_LEVEL(irq
, ALL_CPU_MASK
))
166 GIC_SET_LEVEL(irq
, ALL_CPU_MASK
);
167 if (GIC_TEST_TRIGGER(irq
) || GIC_TEST_ENABLED(irq
)) {
168 DPRINTF("Set %d pending mask %x\n", irq
, GIC_TARGET(irq
));
169 GIC_SET_PENDING(irq
, GIC_TARGET(irq
));
172 GIC_CLEAR_LEVEL(irq
, ALL_CPU_MASK
);
177 static void gic_set_running_irq(gic_state
*s
, int cpu
, int irq
)
179 s
->running_irq
[cpu
] = irq
;
181 s
->running_priority
[cpu
] = 0x100;
183 s
->running_priority
[cpu
] = GIC_GET_PRIORITY(irq
, cpu
);
188 static uint32_t gic_acknowledge_irq(gic_state
*s
, int cpu
)
192 new_irq
= s
->current_pending
[cpu
];
194 || GIC_GET_PRIORITY(new_irq
, cpu
) >= s
->running_priority
[cpu
]) {
195 DPRINTF("ACK no pending IRQ\n");
198 s
->last_active
[new_irq
][cpu
] = s
->running_irq
[cpu
];
199 /* Clear pending flags for both level and edge triggered interrupts.
200 Level triggered IRQs will be reasserted once they become inactive. */
201 GIC_CLEAR_PENDING(new_irq
, GIC_TEST_MODEL(new_irq
) ? ALL_CPU_MASK
: cm
);
202 gic_set_running_irq(s
, cpu
, new_irq
);
203 DPRINTF("ACK %d\n", new_irq
);
207 static void gic_complete_irq(gic_state
* s
, int cpu
, int irq
)
211 DPRINTF("EOI %d\n", irq
);
212 if (s
->running_irq
[cpu
] == 1023)
213 return; /* No active IRQ. */
215 /* Mark level triggered interrupts as pending if they are still
217 if (!GIC_TEST_TRIGGER(irq
) && GIC_TEST_ENABLED(irq
)
218 && GIC_TEST_LEVEL(irq
, cm
) && (GIC_TARGET(irq
) & cm
) != 0) {
219 DPRINTF("Set %d pending mask %x\n", irq
, cm
);
220 GIC_SET_PENDING(irq
, cm
);
224 if (irq
!= s
->running_irq
[cpu
]) {
225 /* Complete an IRQ that is not currently running. */
226 int tmp
= s
->running_irq
[cpu
];
227 while (s
->last_active
[tmp
][cpu
] != 1023) {
228 if (s
->last_active
[tmp
][cpu
] == irq
) {
229 s
->last_active
[tmp
][cpu
] = s
->last_active
[irq
][cpu
];
232 tmp
= s
->last_active
[tmp
][cpu
];
238 /* Complete the current running IRQ. */
239 gic_set_running_irq(s
, cpu
, s
->last_active
[s
->running_irq
[cpu
]][cpu
]);
243 static uint32_t gic_dist_readb(void *opaque
, target_phys_addr_t offset
)
245 gic_state
*s
= (gic_state
*)opaque
;
253 cpu
= gic_get_current_cpu();
255 offset
-= s
->base
+ GIC_DIST_OFFSET
;
256 if (offset
< 0x100) {
261 return ((GIC_NIRQ
/ 32) - 1) | ((NCPU
- 1) << 5);
266 } else if (offset
< 0x200) {
267 /* Interrupt Set/Clear Enable. */
269 irq
= (offset
- 0x100) * 8;
271 irq
= (offset
- 0x180) * 8;
276 for (i
= 0; i
< 8; i
++) {
277 if (GIC_TEST_ENABLED(irq
+ i
)) {
281 } else if (offset
< 0x300) {
282 /* Interrupt Set/Clear Pending. */
284 irq
= (offset
- 0x200) * 8;
286 irq
= (offset
- 0x280) * 8;
291 mask
= (irq
< 32) ? cm
: ALL_CPU_MASK
;
292 for (i
= 0; i
< 8; i
++) {
293 if (GIC_TEST_PENDING(irq
+ i
, mask
)) {
297 } else if (offset
< 0x400) {
298 /* Interrupt Active. */
299 irq
= (offset
- 0x300) * 8 + GIC_BASE_IRQ
;
303 mask
= (irq
< 32) ? cm
: ALL_CPU_MASK
;
304 for (i
= 0; i
< 8; i
++) {
305 if (GIC_TEST_ACTIVE(irq
+ i
, mask
)) {
309 } else if (offset
< 0x800) {
310 /* Interrupt Priority. */
311 irq
= (offset
- 0x400) + GIC_BASE_IRQ
;
314 res
= GIC_GET_PRIORITY(irq
, cpu
);
316 } else if (offset
< 0xc00) {
317 /* Interrupt CPU Target. */
318 irq
= (offset
- 0x800) + GIC_BASE_IRQ
;
321 if (irq
>= 29 && irq
<= 31) {
324 res
= GIC_TARGET(irq
);
326 } else if (offset
< 0xf00) {
327 /* Interrupt Configuration. */
328 irq
= (offset
- 0xc00) * 2 + GIC_BASE_IRQ
;
332 for (i
= 0; i
< 4; i
++) {
333 if (GIC_TEST_MODEL(irq
+ i
))
334 res
|= (1 << (i
* 2));
335 if (GIC_TEST_TRIGGER(irq
+ i
))
336 res
|= (2 << (i
* 2));
339 } else if (offset
< 0xfe0) {
341 } else /* offset >= 0xfe0 */ {
345 res
= gic_id
[(offset
- 0xfe0) >> 2];
350 cpu_abort(cpu_single_env
, "gic_dist_readb: Bad offset %x\n", (int)offset
);
354 static uint32_t gic_dist_readw(void *opaque
, target_phys_addr_t offset
)
357 val
= gic_dist_readb(opaque
, offset
);
358 val
|= gic_dist_readb(opaque
, offset
+ 1) << 8;
362 static uint32_t gic_dist_readl(void *opaque
, target_phys_addr_t offset
)
366 gic_state
*s
= (gic_state
*)opaque
;
368 addr
= offset
- s
->base
;
369 if (addr
< 0x100 || addr
> 0xd00)
370 return nvic_readl(s
->nvic
, addr
);
372 val
= gic_dist_readw(opaque
, offset
);
373 val
|= gic_dist_readw(opaque
, offset
+ 2) << 16;
377 static void gic_dist_writeb(void *opaque
, target_phys_addr_t offset
,
380 gic_state
*s
= (gic_state
*)opaque
;
385 cpu
= gic_get_current_cpu();
386 offset
-= s
->base
+ GIC_DIST_OFFSET
;
387 if (offset
< 0x100) {
392 s
->enabled
= (value
& 1);
393 DPRINTF("Distribution %sabled\n", s
->enabled
? "En" : "Dis");
394 } else if (offset
< 4) {
400 } else if (offset
< 0x180) {
401 /* Interrupt Set Enable. */
402 irq
= (offset
- 0x100) * 8 + GIC_BASE_IRQ
;
407 for (i
= 0; i
< 8; i
++) {
408 if (value
& (1 << i
)) {
409 int mask
= (irq
< 32) ? (1 << cpu
) : GIC_TARGET(irq
);
410 if (!GIC_TEST_ENABLED(irq
+ i
))
411 DPRINTF("Enabled IRQ %d\n", irq
+ i
);
412 GIC_SET_ENABLED(irq
+ i
);
413 /* If a raised level triggered IRQ enabled then mark
415 if (GIC_TEST_LEVEL(irq
+ i
, mask
)
416 && !GIC_TEST_TRIGGER(irq
+ i
)) {
417 DPRINTF("Set %d pending mask %x\n", irq
+ i
, mask
);
418 GIC_SET_PENDING(irq
+ i
, mask
);
422 } else if (offset
< 0x200) {
423 /* Interrupt Clear Enable. */
424 irq
= (offset
- 0x180) * 8 + GIC_BASE_IRQ
;
429 for (i
= 0; i
< 8; i
++) {
430 if (value
& (1 << i
)) {
431 if (GIC_TEST_ENABLED(irq
+ i
))
432 DPRINTF("Disabled IRQ %d\n", irq
+ i
);
433 GIC_CLEAR_ENABLED(irq
+ i
);
436 } else if (offset
< 0x280) {
437 /* Interrupt Set Pending. */
438 irq
= (offset
- 0x200) * 8 + GIC_BASE_IRQ
;
444 for (i
= 0; i
< 8; i
++) {
445 if (value
& (1 << i
)) {
446 GIC_SET_PENDING(irq
+ i
, GIC_TARGET(irq
));
449 } else if (offset
< 0x300) {
450 /* Interrupt Clear Pending. */
451 irq
= (offset
- 0x280) * 8 + GIC_BASE_IRQ
;
454 for (i
= 0; i
< 8; i
++) {
455 /* ??? This currently clears the pending bit for all CPUs, even
456 for per-CPU interrupts. It's unclear whether this is the
458 if (value
& (1 << i
)) {
459 GIC_CLEAR_PENDING(irq
+ i
, ALL_CPU_MASK
);
462 } else if (offset
< 0x400) {
463 /* Interrupt Active. */
465 } else if (offset
< 0x800) {
466 /* Interrupt Priority. */
467 irq
= (offset
- 0x400) + GIC_BASE_IRQ
;
471 s
->priority1
[irq
][cpu
] = value
;
473 s
->priority2
[irq
- 32] = value
;
476 } else if (offset
< 0xc00) {
477 /* Interrupt CPU Target. */
478 irq
= (offset
- 0x800) + GIC_BASE_IRQ
;
484 value
= ALL_CPU_MASK
;
485 s
->irq_target
[irq
] = value
& ALL_CPU_MASK
;
486 } else if (offset
< 0xf00) {
487 /* Interrupt Configuration. */
488 irq
= (offset
- 0xc00) * 4 + GIC_BASE_IRQ
;
493 for (i
= 0; i
< 4; i
++) {
494 if (value
& (1 << (i
* 2))) {
495 GIC_SET_MODEL(irq
+ i
);
497 GIC_CLEAR_MODEL(irq
+ i
);
499 if (value
& (2 << (i
* 2))) {
500 GIC_SET_TRIGGER(irq
+ i
);
502 GIC_CLEAR_TRIGGER(irq
+ i
);
507 /* 0xf00 is only handled for 32-bit writes. */
513 cpu_abort(cpu_single_env
, "gic_dist_writeb: Bad offset %x\n", (int)offset
);
516 static void gic_dist_writew(void *opaque
, target_phys_addr_t offset
,
519 gic_dist_writeb(opaque
, offset
, value
& 0xff);
520 gic_dist_writeb(opaque
, offset
+ 1, value
>> 8);
523 static void gic_dist_writel(void *opaque
, target_phys_addr_t offset
,
526 gic_state
*s
= (gic_state
*)opaque
;
529 addr
= offset
- s
->base
;
530 if (addr
< 0x100 || (addr
> 0xd00 && addr
!= 0xf00)) {
531 nvic_writel(s
->nvic
, addr
, value
);
535 if (offset
- s
->base
== GIC_DIST_OFFSET
+ 0xf00) {
540 cpu
= gic_get_current_cpu();
542 switch ((value
>> 24) & 3) {
544 mask
= (value
>> 16) & ALL_CPU_MASK
;
550 mask
= ALL_CPU_MASK
^ (1 << cpu
);
553 DPRINTF("Bad Soft Int target filter\n");
557 GIC_SET_PENDING(irq
, mask
);
561 gic_dist_writew(opaque
, offset
, value
& 0xffff);
562 gic_dist_writew(opaque
, offset
+ 2, value
>> 16);
565 static CPUReadMemoryFunc
*gic_dist_readfn
[] = {
571 static CPUWriteMemoryFunc
*gic_dist_writefn
[] = {
578 static uint32_t gic_cpu_read(gic_state
*s
, int cpu
, int offset
)
581 case 0x00: /* Control */
582 return s
->cpu_enabled
[cpu
];
583 case 0x04: /* Priority mask */
584 return s
->priority_mask
[cpu
];
585 case 0x08: /* Binary Point */
586 /* ??? Not implemented. */
588 case 0x0c: /* Acknowledge */
589 return gic_acknowledge_irq(s
, cpu
);
590 case 0x14: /* Runing Priority */
591 return s
->running_priority
[cpu
];
592 case 0x18: /* Highest Pending Interrupt */
593 return s
->current_pending
[cpu
];
595 cpu_abort(cpu_single_env
, "gic_cpu_read: Bad offset %x\n",
601 static void gic_cpu_write(gic_state
*s
, int cpu
, int offset
, uint32_t value
)
604 case 0x00: /* Control */
605 s
->cpu_enabled
[cpu
] = (value
& 1);
606 DPRINTF("CPU %sabled\n", s
->cpu_enabled
? "En" : "Dis");
608 case 0x04: /* Priority mask */
609 s
->priority_mask
[cpu
] = (value
& 0xff);
611 case 0x08: /* Binary Point */
612 /* ??? Not implemented. */
614 case 0x10: /* End Of Interrupt */
615 return gic_complete_irq(s
, cpu
, value
& 0x3ff);
617 cpu_abort(cpu_single_env
, "gic_cpu_write: Bad offset %x\n",
625 static void gic_reset(gic_state
*s
)
628 memset(s
->irq_state
, 0, GIC_NIRQ
* sizeof(gic_irq_state
));
629 for (i
= 0 ; i
< NCPU
; i
++) {
630 s
->priority_mask
[i
] = 0xf0;
631 s
->current_pending
[i
] = 1023;
632 s
->running_irq
[i
] = 1023;
633 s
->running_priority
[i
] = 0x100;
635 /* The NVIC doesn't have per-cpu interfaces, so enable by default. */
636 s
->cpu_enabled
[i
] = 1;
638 s
->cpu_enabled
[i
] = 0;
641 for (i
= 0; i
< 16; i
++) {
646 /* The NVIC is always enabled. */
653 static void gic_save(QEMUFile
*f
, void *opaque
)
655 gic_state
*s
= (gic_state
*)opaque
;
659 qemu_put_be32(f
, s
->enabled
);
660 for (i
= 0; i
< NCPU
; i
++) {
661 qemu_put_be32(f
, s
->cpu_enabled
[i
]);
663 qemu_put_be32(f
, s
->irq_target
[i
]);
665 for (j
= 0; j
< 32; j
++)
666 qemu_put_be32(f
, s
->priority1
[j
][i
]);
667 for (j
= 0; j
< GIC_NIRQ
; j
++)
668 qemu_put_be32(f
, s
->last_active
[j
][i
]);
669 qemu_put_be32(f
, s
->priority_mask
[i
]);
670 qemu_put_be32(f
, s
->running_irq
[i
]);
671 qemu_put_be32(f
, s
->running_priority
[i
]);
672 qemu_put_be32(f
, s
->current_pending
[i
]);
674 for (i
= 0; i
< GIC_NIRQ
- 32; i
++) {
675 qemu_put_be32(f
, s
->priority2
[i
]);
677 for (i
= 0; i
< GIC_NIRQ
; i
++) {
678 qemu_put_byte(f
, s
->irq_state
[i
].enabled
);
679 qemu_put_byte(f
, s
->irq_state
[i
].pending
);
680 qemu_put_byte(f
, s
->irq_state
[i
].active
);
681 qemu_put_byte(f
, s
->irq_state
[i
].level
);
682 qemu_put_byte(f
, s
->irq_state
[i
].model
);
683 qemu_put_byte(f
, s
->irq_state
[i
].trigger
);
687 static int gic_load(QEMUFile
*f
, void *opaque
, int version_id
)
689 gic_state
*s
= (gic_state
*)opaque
;
696 s
->enabled
= qemu_get_be32(f
);
697 for (i
= 0; i
< NCPU
; i
++) {
698 s
->cpu_enabled
[i
] = qemu_get_be32(f
);
700 s
->irq_target
[i
] = qemu_get_be32(f
);
702 for (j
= 0; j
< 32; j
++)
703 s
->priority1
[j
][i
] = qemu_get_be32(f
);
704 for (j
= 0; j
< GIC_NIRQ
; j
++)
705 s
->last_active
[j
][i
] = qemu_get_be32(f
);
706 s
->priority_mask
[i
] = qemu_get_be32(f
);
707 s
->running_irq
[i
] = qemu_get_be32(f
);
708 s
->running_priority
[i
] = qemu_get_be32(f
);
709 s
->current_pending
[i
] = qemu_get_be32(f
);
711 for (i
= 0; i
< GIC_NIRQ
- 32; i
++) {
712 s
->priority2
[i
] = qemu_get_be32(f
);
714 for (i
= 0; i
< GIC_NIRQ
; i
++) {
715 s
->irq_state
[i
].enabled
= qemu_get_byte(f
);
716 s
->irq_state
[i
].pending
= qemu_get_byte(f
);
717 s
->irq_state
[i
].active
= qemu_get_byte(f
);
718 s
->irq_state
[i
].level
= qemu_get_byte(f
);
719 s
->irq_state
[i
].model
= qemu_get_byte(f
);
720 s
->irq_state
[i
].trigger
= qemu_get_byte(f
);
726 static gic_state
*gic_init(uint32_t base
, qemu_irq
*parent_irq
)
732 s
= (gic_state
*)qemu_mallocz(sizeof(gic_state
));
735 s
->in
= qemu_allocate_irqs(gic_set_irq
, s
, GIC_NIRQ
);
736 for (i
= 0; i
< NCPU
; i
++) {
737 s
->parent_irq
[i
] = parent_irq
[i
];
739 iomemtype
= cpu_register_io_memory(0, gic_dist_readfn
,
740 gic_dist_writefn
, s
);
741 cpu_register_physical_memory(base
+ GIC_DIST_OFFSET
, 0x00001000,
745 register_savevm("arm_gic", -1, 1, gic_save
, gic_load
, s
);