x86: kexec: Use one page table in x86_64 machine_kexec
[linux-2.6/mini2440.git] / drivers / ide / icside.c
blob2d848010499dfed1997e0f0eed459cb0cff847ec
1 /*
2 * Copyright (c) 1996-2004 Russell King.
4 * Please note that this platform does not support 32-bit IDE IO.
5 */
7 #include <linux/string.h>
8 #include <linux/module.h>
9 #include <linux/ioport.h>
10 #include <linux/slab.h>
11 #include <linux/blkdev.h>
12 #include <linux/errno.h>
13 #include <linux/ide.h>
14 #include <linux/dma-mapping.h>
15 #include <linux/device.h>
16 #include <linux/init.h>
17 #include <linux/scatterlist.h>
18 #include <linux/io.h>
20 #include <asm/dma.h>
21 #include <asm/ecard.h>
23 #define DRV_NAME "icside"
25 #define ICS_IDENT_OFFSET 0x2280
27 #define ICS_ARCIN_V5_INTRSTAT 0x0000
28 #define ICS_ARCIN_V5_INTROFFSET 0x0004
29 #define ICS_ARCIN_V5_IDEOFFSET 0x2800
30 #define ICS_ARCIN_V5_IDEALTOFFSET 0x2b80
31 #define ICS_ARCIN_V5_IDESTEPPING 6
33 #define ICS_ARCIN_V6_IDEOFFSET_1 0x2000
34 #define ICS_ARCIN_V6_INTROFFSET_1 0x2200
35 #define ICS_ARCIN_V6_INTRSTAT_1 0x2290
36 #define ICS_ARCIN_V6_IDEALTOFFSET_1 0x2380
37 #define ICS_ARCIN_V6_IDEOFFSET_2 0x3000
38 #define ICS_ARCIN_V6_INTROFFSET_2 0x3200
39 #define ICS_ARCIN_V6_INTRSTAT_2 0x3290
40 #define ICS_ARCIN_V6_IDEALTOFFSET_2 0x3380
41 #define ICS_ARCIN_V6_IDESTEPPING 6
43 struct cardinfo {
44 unsigned int dataoffset;
45 unsigned int ctrloffset;
46 unsigned int stepping;
49 static struct cardinfo icside_cardinfo_v5 = {
50 .dataoffset = ICS_ARCIN_V5_IDEOFFSET,
51 .ctrloffset = ICS_ARCIN_V5_IDEALTOFFSET,
52 .stepping = ICS_ARCIN_V5_IDESTEPPING,
55 static struct cardinfo icside_cardinfo_v6_1 = {
56 .dataoffset = ICS_ARCIN_V6_IDEOFFSET_1,
57 .ctrloffset = ICS_ARCIN_V6_IDEALTOFFSET_1,
58 .stepping = ICS_ARCIN_V6_IDESTEPPING,
61 static struct cardinfo icside_cardinfo_v6_2 = {
62 .dataoffset = ICS_ARCIN_V6_IDEOFFSET_2,
63 .ctrloffset = ICS_ARCIN_V6_IDEALTOFFSET_2,
64 .stepping = ICS_ARCIN_V6_IDESTEPPING,
67 struct icside_state {
68 unsigned int channel;
69 unsigned int enabled;
70 void __iomem *irq_port;
71 void __iomem *ioc_base;
72 unsigned int sel;
73 unsigned int type;
74 struct ide_host *host;
77 #define ICS_TYPE_A3IN 0
78 #define ICS_TYPE_A3USER 1
79 #define ICS_TYPE_V6 3
80 #define ICS_TYPE_V5 15
81 #define ICS_TYPE_NOTYPE ((unsigned int)-1)
83 /* ---------------- Version 5 PCB Support Functions --------------------- */
84 /* Prototype: icside_irqenable_arcin_v5 (struct expansion_card *ec, int irqnr)
85 * Purpose : enable interrupts from card
87 static void icside_irqenable_arcin_v5 (struct expansion_card *ec, int irqnr)
89 struct icside_state *state = ec->irq_data;
91 writeb(0, state->irq_port + ICS_ARCIN_V5_INTROFFSET);
94 /* Prototype: icside_irqdisable_arcin_v5 (struct expansion_card *ec, int irqnr)
95 * Purpose : disable interrupts from card
97 static void icside_irqdisable_arcin_v5 (struct expansion_card *ec, int irqnr)
99 struct icside_state *state = ec->irq_data;
101 readb(state->irq_port + ICS_ARCIN_V5_INTROFFSET);
104 static const expansioncard_ops_t icside_ops_arcin_v5 = {
105 .irqenable = icside_irqenable_arcin_v5,
106 .irqdisable = icside_irqdisable_arcin_v5,
110 /* ---------------- Version 6 PCB Support Functions --------------------- */
111 /* Prototype: icside_irqenable_arcin_v6 (struct expansion_card *ec, int irqnr)
112 * Purpose : enable interrupts from card
114 static void icside_irqenable_arcin_v6 (struct expansion_card *ec, int irqnr)
116 struct icside_state *state = ec->irq_data;
117 void __iomem *base = state->irq_port;
119 state->enabled = 1;
121 switch (state->channel) {
122 case 0:
123 writeb(0, base + ICS_ARCIN_V6_INTROFFSET_1);
124 readb(base + ICS_ARCIN_V6_INTROFFSET_2);
125 break;
126 case 1:
127 writeb(0, base + ICS_ARCIN_V6_INTROFFSET_2);
128 readb(base + ICS_ARCIN_V6_INTROFFSET_1);
129 break;
133 /* Prototype: icside_irqdisable_arcin_v6 (struct expansion_card *ec, int irqnr)
134 * Purpose : disable interrupts from card
136 static void icside_irqdisable_arcin_v6 (struct expansion_card *ec, int irqnr)
138 struct icside_state *state = ec->irq_data;
140 state->enabled = 0;
142 readb(state->irq_port + ICS_ARCIN_V6_INTROFFSET_1);
143 readb(state->irq_port + ICS_ARCIN_V6_INTROFFSET_2);
146 /* Prototype: icside_irqprobe(struct expansion_card *ec)
147 * Purpose : detect an active interrupt from card
149 static int icside_irqpending_arcin_v6(struct expansion_card *ec)
151 struct icside_state *state = ec->irq_data;
153 return readb(state->irq_port + ICS_ARCIN_V6_INTRSTAT_1) & 1 ||
154 readb(state->irq_port + ICS_ARCIN_V6_INTRSTAT_2) & 1;
157 static const expansioncard_ops_t icside_ops_arcin_v6 = {
158 .irqenable = icside_irqenable_arcin_v6,
159 .irqdisable = icside_irqdisable_arcin_v6,
160 .irqpending = icside_irqpending_arcin_v6,
164 * Handle routing of interrupts. This is called before
165 * we write the command to the drive.
167 static void icside_maskproc(ide_drive_t *drive, int mask)
169 ide_hwif_t *hwif = HWIF(drive);
170 struct expansion_card *ec = ECARD_DEV(hwif->dev);
171 struct icside_state *state = ecard_get_drvdata(ec);
172 unsigned long flags;
174 local_irq_save(flags);
176 state->channel = hwif->channel;
178 if (state->enabled && !mask) {
179 switch (hwif->channel) {
180 case 0:
181 writeb(0, state->irq_port + ICS_ARCIN_V6_INTROFFSET_1);
182 readb(state->irq_port + ICS_ARCIN_V6_INTROFFSET_2);
183 break;
184 case 1:
185 writeb(0, state->irq_port + ICS_ARCIN_V6_INTROFFSET_2);
186 readb(state->irq_port + ICS_ARCIN_V6_INTROFFSET_1);
187 break;
189 } else {
190 readb(state->irq_port + ICS_ARCIN_V6_INTROFFSET_2);
191 readb(state->irq_port + ICS_ARCIN_V6_INTROFFSET_1);
194 local_irq_restore(flags);
197 static const struct ide_port_ops icside_v6_no_dma_port_ops = {
198 .maskproc = icside_maskproc,
201 #ifdef CONFIG_BLK_DEV_IDEDMA_ICS
203 * SG-DMA support.
205 * Similar to the BM-DMA, but we use the RiscPCs IOMD DMA controllers.
206 * There is only one DMA controller per card, which means that only
207 * one drive can be accessed at one time. NOTE! We do not enforce that
208 * here, but we rely on the main IDE driver spotting that both
209 * interfaces use the same IRQ, which should guarantee this.
213 * Configure the IOMD to give the appropriate timings for the transfer
214 * mode being requested. We take the advice of the ATA standards, and
215 * calculate the cycle time based on the transfer mode, and the EIDE
216 * MW DMA specs that the drive provides in the IDENTIFY command.
218 * We have the following IOMD DMA modes to choose from:
220 * Type Active Recovery Cycle
221 * A 250 (250) 312 (550) 562 (800)
222 * B 187 250 437
223 * C 125 (125) 125 (375) 250 (500)
224 * D 62 125 187
226 * (figures in brackets are actual measured timings)
228 * However, we also need to take care of the read/write active and
229 * recovery timings:
231 * Read Write
232 * Mode Active -- Recovery -- Cycle IOMD type
233 * MW0 215 50 215 480 A
234 * MW1 80 50 50 150 C
235 * MW2 70 25 25 120 C
237 static void icside_set_dma_mode(ide_drive_t *drive, const u8 xfer_mode)
239 int cycle_time, use_dma_info = 0;
241 switch (xfer_mode) {
242 case XFER_MW_DMA_2:
243 cycle_time = 250;
244 use_dma_info = 1;
245 break;
247 case XFER_MW_DMA_1:
248 cycle_time = 250;
249 use_dma_info = 1;
250 break;
252 case XFER_MW_DMA_0:
253 cycle_time = 480;
254 break;
256 case XFER_SW_DMA_2:
257 case XFER_SW_DMA_1:
258 case XFER_SW_DMA_0:
259 cycle_time = 480;
260 break;
264 * If we're going to be doing MW_DMA_1 or MW_DMA_2, we should
265 * take care to note the values in the ID...
267 if (use_dma_info && drive->id[ATA_ID_EIDE_DMA_TIME] > cycle_time)
268 cycle_time = drive->id[ATA_ID_EIDE_DMA_TIME];
270 drive->drive_data = cycle_time;
272 printk("%s: %s selected (peak %dMB/s)\n", drive->name,
273 ide_xfer_verbose(xfer_mode), 2000 / drive->drive_data);
276 static const struct ide_port_ops icside_v6_port_ops = {
277 .set_dma_mode = icside_set_dma_mode,
278 .maskproc = icside_maskproc,
281 static void icside_dma_host_set(ide_drive_t *drive, int on)
285 static int icside_dma_end(ide_drive_t *drive)
287 ide_hwif_t *hwif = HWIF(drive);
288 struct expansion_card *ec = ECARD_DEV(hwif->dev);
290 drive->waiting_for_dma = 0;
292 disable_dma(ec->dma);
294 /* Teardown mappings after DMA has completed. */
295 ide_destroy_dmatable(drive);
297 return get_dma_residue(ec->dma) != 0;
300 static void icside_dma_start(ide_drive_t *drive)
302 ide_hwif_t *hwif = HWIF(drive);
303 struct expansion_card *ec = ECARD_DEV(hwif->dev);
305 /* We can not enable DMA on both channels simultaneously. */
306 BUG_ON(dma_channel_active(ec->dma));
307 enable_dma(ec->dma);
310 static int icside_dma_setup(ide_drive_t *drive)
312 ide_hwif_t *hwif = HWIF(drive);
313 struct expansion_card *ec = ECARD_DEV(hwif->dev);
314 struct icside_state *state = ecard_get_drvdata(ec);
315 struct request *rq = hwif->hwgroup->rq;
316 unsigned int dma_mode;
318 if (rq_data_dir(rq))
319 dma_mode = DMA_MODE_WRITE;
320 else
321 dma_mode = DMA_MODE_READ;
324 * We can not enable DMA on both channels.
326 BUG_ON(dma_channel_active(ec->dma));
328 hwif->sg_nents = ide_build_sglist(drive, rq);
331 * Ensure that we have the right interrupt routed.
333 icside_maskproc(drive, 0);
336 * Route the DMA signals to the correct interface.
338 writeb(state->sel | hwif->channel, state->ioc_base);
341 * Select the correct timing for this drive.
343 set_dma_speed(ec->dma, drive->drive_data);
346 * Tell the DMA engine about the SG table and
347 * data direction.
349 set_dma_sg(ec->dma, hwif->sg_table, hwif->sg_nents);
350 set_dma_mode(ec->dma, dma_mode);
352 drive->waiting_for_dma = 1;
354 return 0;
357 static void icside_dma_exec_cmd(ide_drive_t *drive, u8 cmd)
359 /* issue cmd to drive */
360 ide_execute_command(drive, cmd, ide_dma_intr, 2 * WAIT_CMD, NULL);
363 static int icside_dma_test_irq(ide_drive_t *drive)
365 ide_hwif_t *hwif = HWIF(drive);
366 struct expansion_card *ec = ECARD_DEV(hwif->dev);
367 struct icside_state *state = ecard_get_drvdata(ec);
369 return readb(state->irq_port +
370 (hwif->channel ?
371 ICS_ARCIN_V6_INTRSTAT_2 :
372 ICS_ARCIN_V6_INTRSTAT_1)) & 1;
375 static int icside_dma_init(ide_hwif_t *hwif, const struct ide_port_info *d)
377 hwif->dmatable_cpu = NULL;
378 hwif->dmatable_dma = 0;
380 return 0;
383 static const struct ide_dma_ops icside_v6_dma_ops = {
384 .dma_host_set = icside_dma_host_set,
385 .dma_setup = icside_dma_setup,
386 .dma_exec_cmd = icside_dma_exec_cmd,
387 .dma_start = icside_dma_start,
388 .dma_end = icside_dma_end,
389 .dma_test_irq = icside_dma_test_irq,
390 .dma_timeout = ide_dma_timeout,
391 .dma_lost_irq = ide_dma_lost_irq,
393 #else
394 #define icside_v6_dma_ops NULL
395 #endif
397 static int icside_dma_off_init(ide_hwif_t *hwif, const struct ide_port_info *d)
399 return -EOPNOTSUPP;
402 static void icside_setup_ports(hw_regs_t *hw, void __iomem *base,
403 struct cardinfo *info, struct expansion_card *ec)
405 unsigned long port = (unsigned long)base + info->dataoffset;
407 hw->io_ports.data_addr = port;
408 hw->io_ports.error_addr = port + (1 << info->stepping);
409 hw->io_ports.nsect_addr = port + (2 << info->stepping);
410 hw->io_ports.lbal_addr = port + (3 << info->stepping);
411 hw->io_ports.lbam_addr = port + (4 << info->stepping);
412 hw->io_ports.lbah_addr = port + (5 << info->stepping);
413 hw->io_ports.device_addr = port + (6 << info->stepping);
414 hw->io_ports.status_addr = port + (7 << info->stepping);
415 hw->io_ports.ctl_addr = (unsigned long)base + info->ctrloffset;
417 hw->irq = ec->irq;
418 hw->dev = &ec->dev;
419 hw->chipset = ide_acorn;
422 static int __init
423 icside_register_v5(struct icside_state *state, struct expansion_card *ec)
425 void __iomem *base;
426 struct ide_host *host;
427 hw_regs_t hw, *hws[] = { &hw, NULL, NULL, NULL };
428 int ret;
430 base = ecardm_iomap(ec, ECARD_RES_MEMC, 0, 0);
431 if (!base)
432 return -ENOMEM;
434 state->irq_port = base;
436 ec->irqaddr = base + ICS_ARCIN_V5_INTRSTAT;
437 ec->irqmask = 1;
439 ecard_setirq(ec, &icside_ops_arcin_v5, state);
442 * Be on the safe side - disable interrupts
444 icside_irqdisable_arcin_v5(ec, 0);
446 icside_setup_ports(&hw, base, &icside_cardinfo_v5, ec);
448 host = ide_host_alloc(NULL, hws);
449 if (host == NULL)
450 return -ENODEV;
452 state->host = host;
454 ecard_set_drvdata(ec, state);
456 ret = ide_host_register(host, NULL, hws);
457 if (ret)
458 goto err_free;
460 return 0;
461 err_free:
462 ide_host_free(host);
463 ecard_set_drvdata(ec, NULL);
464 return ret;
467 static const struct ide_port_info icside_v6_port_info __initdata = {
468 .init_dma = icside_dma_off_init,
469 .port_ops = &icside_v6_no_dma_port_ops,
470 .dma_ops = &icside_v6_dma_ops,
471 .host_flags = IDE_HFLAG_SERIALIZE | IDE_HFLAG_MMIO,
472 .mwdma_mask = ATA_MWDMA2,
473 .swdma_mask = ATA_SWDMA2,
476 static int __init
477 icside_register_v6(struct icside_state *state, struct expansion_card *ec)
479 void __iomem *ioc_base, *easi_base;
480 struct ide_host *host;
481 unsigned int sel = 0;
482 int ret;
483 hw_regs_t hw[2], *hws[] = { &hw[0], NULL, NULL, NULL };
484 struct ide_port_info d = icside_v6_port_info;
486 ioc_base = ecardm_iomap(ec, ECARD_RES_IOCFAST, 0, 0);
487 if (!ioc_base) {
488 ret = -ENOMEM;
489 goto out;
492 easi_base = ioc_base;
494 if (ecard_resource_flags(ec, ECARD_RES_EASI)) {
495 easi_base = ecardm_iomap(ec, ECARD_RES_EASI, 0, 0);
496 if (!easi_base) {
497 ret = -ENOMEM;
498 goto out;
502 * Enable access to the EASI region.
504 sel = 1 << 5;
507 writeb(sel, ioc_base);
509 ecard_setirq(ec, &icside_ops_arcin_v6, state);
511 state->irq_port = easi_base;
512 state->ioc_base = ioc_base;
513 state->sel = sel;
516 * Be on the safe side - disable interrupts
518 icside_irqdisable_arcin_v6(ec, 0);
520 icside_setup_ports(&hw[0], easi_base, &icside_cardinfo_v6_1, ec);
521 icside_setup_ports(&hw[1], easi_base, &icside_cardinfo_v6_2, ec);
523 host = ide_host_alloc(&d, hws);
524 if (host == NULL)
525 return -ENODEV;
527 state->host = host;
529 ecard_set_drvdata(ec, state);
531 if (ec->dma != NO_DMA && !request_dma(ec->dma, DRV_NAME)) {
532 d.init_dma = icside_dma_init;
533 d.port_ops = &icside_v6_port_ops;
534 d.dma_ops = NULL;
537 ret = ide_host_register(host, NULL, hws);
538 if (ret)
539 goto err_free;
541 return 0;
542 err_free:
543 ide_host_free(host);
544 if (d.dma_ops)
545 free_dma(ec->dma);
546 ecard_set_drvdata(ec, NULL);
547 out:
548 return ret;
551 static int __devinit
552 icside_probe(struct expansion_card *ec, const struct ecard_id *id)
554 struct icside_state *state;
555 void __iomem *idmem;
556 int ret;
558 ret = ecard_request_resources(ec);
559 if (ret)
560 goto out;
562 state = kzalloc(sizeof(struct icside_state), GFP_KERNEL);
563 if (!state) {
564 ret = -ENOMEM;
565 goto release;
568 state->type = ICS_TYPE_NOTYPE;
570 idmem = ecardm_iomap(ec, ECARD_RES_IOCFAST, 0, 0);
571 if (idmem) {
572 unsigned int type;
574 type = readb(idmem + ICS_IDENT_OFFSET) & 1;
575 type |= (readb(idmem + ICS_IDENT_OFFSET + 4) & 1) << 1;
576 type |= (readb(idmem + ICS_IDENT_OFFSET + 8) & 1) << 2;
577 type |= (readb(idmem + ICS_IDENT_OFFSET + 12) & 1) << 3;
578 ecardm_iounmap(ec, idmem);
580 state->type = type;
583 switch (state->type) {
584 case ICS_TYPE_A3IN:
585 dev_warn(&ec->dev, "A3IN unsupported\n");
586 ret = -ENODEV;
587 break;
589 case ICS_TYPE_A3USER:
590 dev_warn(&ec->dev, "A3USER unsupported\n");
591 ret = -ENODEV;
592 break;
594 case ICS_TYPE_V5:
595 ret = icside_register_v5(state, ec);
596 break;
598 case ICS_TYPE_V6:
599 ret = icside_register_v6(state, ec);
600 break;
602 default:
603 dev_warn(&ec->dev, "unknown interface type\n");
604 ret = -ENODEV;
605 break;
608 if (ret == 0)
609 goto out;
611 kfree(state);
612 release:
613 ecard_release_resources(ec);
614 out:
615 return ret;
618 static void __devexit icside_remove(struct expansion_card *ec)
620 struct icside_state *state = ecard_get_drvdata(ec);
622 switch (state->type) {
623 case ICS_TYPE_V5:
624 /* FIXME: tell IDE to stop using the interface */
626 /* Disable interrupts */
627 icside_irqdisable_arcin_v5(ec, 0);
628 break;
630 case ICS_TYPE_V6:
631 /* FIXME: tell IDE to stop using the interface */
632 if (ec->dma != NO_DMA)
633 free_dma(ec->dma);
635 /* Disable interrupts */
636 icside_irqdisable_arcin_v6(ec, 0);
638 /* Reset the ROM pointer/EASI selection */
639 writeb(0, state->ioc_base);
640 break;
643 ecard_set_drvdata(ec, NULL);
645 kfree(state);
646 ecard_release_resources(ec);
649 static void icside_shutdown(struct expansion_card *ec)
651 struct icside_state *state = ecard_get_drvdata(ec);
652 unsigned long flags;
655 * Disable interrupts from this card. We need to do
656 * this before disabling EASI since we may be accessing
657 * this register via that region.
659 local_irq_save(flags);
660 ec->ops->irqdisable(ec, 0);
661 local_irq_restore(flags);
664 * Reset the ROM pointer so that we can read the ROM
665 * after a soft reboot. This also disables access to
666 * the IDE taskfile via the EASI region.
668 if (state->ioc_base)
669 writeb(0, state->ioc_base);
672 static const struct ecard_id icside_ids[] = {
673 { MANU_ICS, PROD_ICS_IDE },
674 { MANU_ICS2, PROD_ICS2_IDE },
675 { 0xffff, 0xffff }
678 static struct ecard_driver icside_driver = {
679 .probe = icside_probe,
680 .remove = __devexit_p(icside_remove),
681 .shutdown = icside_shutdown,
682 .id_table = icside_ids,
683 .drv = {
684 .name = "icside",
688 static int __init icside_init(void)
690 return ecard_register_driver(&icside_driver);
693 static void __exit icside_exit(void)
695 ecard_remove_driver(&icside_driver);
698 MODULE_AUTHOR("Russell King <rmk@arm.linux.org.uk>");
699 MODULE_LICENSE("GPL");
700 MODULE_DESCRIPTION("ICS IDE driver");
702 module_init(icside_init);
703 module_exit(icside_exit);