PATCH] cifs: fix broken mounts when a SSH tunnel is used (try #4)
[linux-2.6/linux-acpi-2.6/ibm-acpi-2.6.git] / arch / arm / common / gic.c
blob337741f734ac08d7f438e5d6240293adf3c5c366
1 /*
2 * linux/arch/arm/common/gic.c
4 * Copyright (C) 2002 ARM Limited, All Rights Reserved.
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License version 2 as
8 * published by the Free Software Foundation.
10 * Interrupt architecture for the GIC:
12 * o There is one Interrupt Distributor, which receives interrupts
13 * from system devices and sends them to the Interrupt Controllers.
15 * o There is one CPU Interface per CPU, which sends interrupts sent
16 * by the Distributor, and interrupts generated locally, to the
17 * associated CPU. The base address of the CPU interface is usually
18 * aliased so that the same address points to different chips depending
19 * on the CPU it is accessed from.
21 * Note that IRQs 0-31 are special - they are local to each CPU.
22 * As such, the enable set/clear, pending set/clear and active bit
23 * registers are banked per-cpu for these sources.
25 #include <linux/init.h>
26 #include <linux/kernel.h>
27 #include <linux/list.h>
28 #include <linux/smp.h>
29 #include <linux/cpumask.h>
30 #include <linux/io.h>
32 #include <asm/irq.h>
33 #include <asm/mach/irq.h>
34 #include <asm/hardware/gic.h>
36 static DEFINE_SPINLOCK(irq_controller_lock);
38 struct gic_chip_data {
39 unsigned int irq_offset;
40 void __iomem *dist_base;
41 void __iomem *cpu_base;
44 #ifndef MAX_GIC_NR
45 #define MAX_GIC_NR 1
46 #endif
48 static struct gic_chip_data gic_data[MAX_GIC_NR];
50 static inline void __iomem *gic_dist_base(unsigned int irq)
52 struct gic_chip_data *gic_data = get_irq_chip_data(irq);
53 return gic_data->dist_base;
56 static inline void __iomem *gic_cpu_base(unsigned int irq)
58 struct gic_chip_data *gic_data = get_irq_chip_data(irq);
59 return gic_data->cpu_base;
62 static inline unsigned int gic_irq(unsigned int irq)
64 struct gic_chip_data *gic_data = get_irq_chip_data(irq);
65 return irq - gic_data->irq_offset;
69 * Routines to acknowledge, disable and enable interrupts
71 * Linux assumes that when we're done with an interrupt we need to
72 * unmask it, in the same way we need to unmask an interrupt when
73 * we first enable it.
75 * The GIC has a separate notion of "end of interrupt" to re-enable
76 * an interrupt after handling, in order to support hardware
77 * prioritisation.
79 * We can make the GIC behave in the way that Linux expects by making
80 * our "acknowledge" routine disable the interrupt, then mark it as
81 * complete.
83 static void gic_ack_irq(unsigned int irq)
85 u32 mask = 1 << (irq % 32);
87 spin_lock(&irq_controller_lock);
88 writel(mask, gic_dist_base(irq) + GIC_DIST_ENABLE_CLEAR + (gic_irq(irq) / 32) * 4);
89 writel(gic_irq(irq), gic_cpu_base(irq) + GIC_CPU_EOI);
90 spin_unlock(&irq_controller_lock);
93 static void gic_mask_irq(unsigned int irq)
95 u32 mask = 1 << (irq % 32);
97 spin_lock(&irq_controller_lock);
98 writel(mask, gic_dist_base(irq) + GIC_DIST_ENABLE_CLEAR + (gic_irq(irq) / 32) * 4);
99 spin_unlock(&irq_controller_lock);
102 static void gic_unmask_irq(unsigned int irq)
104 u32 mask = 1 << (irq % 32);
106 spin_lock(&irq_controller_lock);
107 writel(mask, gic_dist_base(irq) + GIC_DIST_ENABLE_SET + (gic_irq(irq) / 32) * 4);
108 spin_unlock(&irq_controller_lock);
111 #ifdef CONFIG_SMP
112 static int gic_set_cpu(unsigned int irq, const struct cpumask *mask_val)
114 void __iomem *reg = gic_dist_base(irq) + GIC_DIST_TARGET + (gic_irq(irq) & ~3);
115 unsigned int shift = (irq % 4) * 8;
116 unsigned int cpu = cpumask_first(mask_val);
117 u32 val;
119 spin_lock(&irq_controller_lock);
120 irq_desc[irq].node = cpu;
121 val = readl(reg) & ~(0xff << shift);
122 val |= 1 << (cpu + shift);
123 writel(val, reg);
124 spin_unlock(&irq_controller_lock);
126 return 0;
128 #endif
130 static void gic_handle_cascade_irq(unsigned int irq, struct irq_desc *desc)
132 struct gic_chip_data *chip_data = get_irq_data(irq);
133 struct irq_chip *chip = get_irq_chip(irq);
134 unsigned int cascade_irq, gic_irq;
135 unsigned long status;
137 /* primary controller ack'ing */
138 chip->ack(irq);
140 spin_lock(&irq_controller_lock);
141 status = readl(chip_data->cpu_base + GIC_CPU_INTACK);
142 spin_unlock(&irq_controller_lock);
144 gic_irq = (status & 0x3ff);
145 if (gic_irq == 1023)
146 goto out;
148 cascade_irq = gic_irq + chip_data->irq_offset;
149 if (unlikely(gic_irq < 32 || gic_irq > 1020 || cascade_irq >= NR_IRQS))
150 do_bad_IRQ(cascade_irq, desc);
151 else
152 generic_handle_irq(cascade_irq);
154 out:
155 /* primary controller unmasking */
156 chip->unmask(irq);
159 static struct irq_chip gic_chip = {
160 .name = "GIC",
161 .ack = gic_ack_irq,
162 .mask = gic_mask_irq,
163 .unmask = gic_unmask_irq,
164 #ifdef CONFIG_SMP
165 .set_affinity = gic_set_cpu,
166 #endif
169 void __init gic_cascade_irq(unsigned int gic_nr, unsigned int irq)
171 if (gic_nr >= MAX_GIC_NR)
172 BUG();
173 if (set_irq_data(irq, &gic_data[gic_nr]) != 0)
174 BUG();
175 set_irq_chained_handler(irq, gic_handle_cascade_irq);
178 void __init gic_dist_init(unsigned int gic_nr, void __iomem *base,
179 unsigned int irq_start)
181 unsigned int max_irq, i;
182 u32 cpumask = 1 << smp_processor_id();
184 if (gic_nr >= MAX_GIC_NR)
185 BUG();
187 cpumask |= cpumask << 8;
188 cpumask |= cpumask << 16;
190 gic_data[gic_nr].dist_base = base;
191 gic_data[gic_nr].irq_offset = (irq_start - 1) & ~31;
193 writel(0, base + GIC_DIST_CTRL);
196 * Find out how many interrupts are supported.
198 max_irq = readl(base + GIC_DIST_CTR) & 0x1f;
199 max_irq = (max_irq + 1) * 32;
202 * The GIC only supports up to 1020 interrupt sources.
203 * Limit this to either the architected maximum, or the
204 * platform maximum.
206 if (max_irq > max(1020, NR_IRQS))
207 max_irq = max(1020, NR_IRQS);
210 * Set all global interrupts to be level triggered, active low.
212 for (i = 32; i < max_irq; i += 16)
213 writel(0, base + GIC_DIST_CONFIG + i * 4 / 16);
216 * Set all global interrupts to this CPU only.
218 for (i = 32; i < max_irq; i += 4)
219 writel(cpumask, base + GIC_DIST_TARGET + i * 4 / 4);
222 * Set priority on all interrupts.
224 for (i = 0; i < max_irq; i += 4)
225 writel(0xa0a0a0a0, base + GIC_DIST_PRI + i * 4 / 4);
228 * Disable all interrupts.
230 for (i = 0; i < max_irq; i += 32)
231 writel(0xffffffff, base + GIC_DIST_ENABLE_CLEAR + i * 4 / 32);
234 * Setup the Linux IRQ subsystem.
236 for (i = irq_start; i < gic_data[gic_nr].irq_offset + max_irq; i++) {
237 set_irq_chip(i, &gic_chip);
238 set_irq_chip_data(i, &gic_data[gic_nr]);
239 set_irq_handler(i, handle_level_irq);
240 set_irq_flags(i, IRQF_VALID | IRQF_PROBE);
243 writel(1, base + GIC_DIST_CTRL);
246 void __cpuinit gic_cpu_init(unsigned int gic_nr, void __iomem *base)
248 if (gic_nr >= MAX_GIC_NR)
249 BUG();
251 gic_data[gic_nr].cpu_base = base;
253 writel(0xf0, base + GIC_CPU_PRIMASK);
254 writel(1, base + GIC_CPU_CTRL);
257 #ifdef CONFIG_SMP
258 void gic_raise_softirq(const struct cpumask *mask, unsigned int irq)
260 unsigned long map = *cpus_addr(*mask);
262 /* this always happens on GIC0 */
263 writel(map << 16 | irq, gic_data[0].dist_base + GIC_DIST_SOFTINT);
265 #endif