nfs: Panic when commit fails
[linux-2.6/verdex.git] / drivers / ide / cmd64x.c
blob680e5975217fe59dce8febedc3389e1766704e27
1 /*
2 * cmd64x.c: Enable interrupts at initialization time on Ultra/PCI machines.
3 * Due to massive hardware bugs, UltraDMA is only supported
4 * on the 646U2 and not on the 646U.
6 * Copyright (C) 1998 Eddie C. Dost (ecd@skynet.be)
7 * Copyright (C) 1998 David S. Miller (davem@redhat.com)
9 * Copyright (C) 1999-2002 Andre Hedrick <andre@linux-ide.org>
10 * Copyright (C) 2007,2009 MontaVista Software, Inc. <source@mvista.com>
13 #include <linux/module.h>
14 #include <linux/types.h>
15 #include <linux/pci.h>
16 #include <linux/ide.h>
17 #include <linux/init.h>
19 #include <asm/io.h>
21 #define DRV_NAME "cmd64x"
23 #define CMD_DEBUG 0
25 #if CMD_DEBUG
26 #define cmdprintk(x...) printk(x)
27 #else
28 #define cmdprintk(x...)
29 #endif
32 * CMD64x specific registers definition.
34 #define CFR 0x50
35 #define CFR_INTR_CH0 0x04
37 #define CMDTIM 0x52
38 #define ARTTIM0 0x53
39 #define DRWTIM0 0x54
40 #define ARTTIM1 0x55
41 #define DRWTIM1 0x56
42 #define ARTTIM23 0x57
43 #define ARTTIM23_DIS_RA2 0x04
44 #define ARTTIM23_DIS_RA3 0x08
45 #define ARTTIM23_INTR_CH1 0x10
46 #define DRWTIM2 0x58
47 #define BRST 0x59
48 #define DRWTIM3 0x5b
50 #define BMIDECR0 0x70
51 #define MRDMODE 0x71
52 #define MRDMODE_INTR_CH0 0x04
53 #define MRDMODE_INTR_CH1 0x08
54 #define UDIDETCR0 0x73
55 #define DTPR0 0x74
56 #define BMIDECR1 0x78
57 #define BMIDECSR 0x79
58 #define UDIDETCR1 0x7B
59 #define DTPR1 0x7C
61 static u8 quantize_timing(int timing, int quant)
63 return (timing + quant - 1) / quant;
67 * This routine calculates active/recovery counts and then writes them into
68 * the chipset registers.
70 static void program_cycle_times (ide_drive_t *drive, int cycle_time, int active_time)
72 struct pci_dev *dev = to_pci_dev(drive->hwif->dev);
73 int clock_time = 1000 / (ide_pci_clk ? ide_pci_clk : 33);
74 u8 cycle_count, active_count, recovery_count, drwtim;
75 static const u8 recovery_values[] =
76 {15, 15, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 0};
77 static const u8 drwtim_regs[4] = {DRWTIM0, DRWTIM1, DRWTIM2, DRWTIM3};
79 cmdprintk("program_cycle_times parameters: total=%d, active=%d\n",
80 cycle_time, active_time);
82 cycle_count = quantize_timing( cycle_time, clock_time);
83 active_count = quantize_timing(active_time, clock_time);
84 recovery_count = cycle_count - active_count;
87 * In case we've got too long recovery phase, try to lengthen
88 * the active phase
90 if (recovery_count > 16) {
91 active_count += recovery_count - 16;
92 recovery_count = 16;
94 if (active_count > 16) /* shouldn't actually happen... */
95 active_count = 16;
97 cmdprintk("Final counts: total=%d, active=%d, recovery=%d\n",
98 cycle_count, active_count, recovery_count);
101 * Convert values to internal chipset representation
103 recovery_count = recovery_values[recovery_count];
104 active_count &= 0x0f;
106 /* Program the active/recovery counts into the DRWTIM register */
107 drwtim = (active_count << 4) | recovery_count;
108 (void) pci_write_config_byte(dev, drwtim_regs[drive->dn], drwtim);
109 cmdprintk("Write 0x%02x to reg 0x%x\n", drwtim, drwtim_regs[drive->dn]);
113 * This routine writes into the chipset registers
114 * PIO setup/active/recovery timings.
116 static void cmd64x_tune_pio(ide_drive_t *drive, const u8 pio)
118 ide_hwif_t *hwif = drive->hwif;
119 struct pci_dev *dev = to_pci_dev(hwif->dev);
120 struct ide_timing *t = ide_timing_find_mode(XFER_PIO_0 + pio);
121 unsigned long setup_count;
122 unsigned int cycle_time;
123 u8 arttim = 0;
125 static const u8 setup_values[] = {0x40, 0x40, 0x40, 0x80, 0, 0xc0};
126 static const u8 arttim_regs[4] = {ARTTIM0, ARTTIM1, ARTTIM23, ARTTIM23};
128 cycle_time = ide_pio_cycle_time(drive, pio);
130 program_cycle_times(drive, cycle_time, t->active);
132 setup_count = quantize_timing(t->setup,
133 1000 / (ide_pci_clk ? ide_pci_clk : 33));
136 * The primary channel has individual address setup timing registers
137 * for each drive and the hardware selects the slowest timing itself.
138 * The secondary channel has one common register and we have to select
139 * the slowest address setup timing ourselves.
141 if (hwif->channel) {
142 ide_drive_t *pair = ide_get_pair_dev(drive);
144 ide_set_drivedata(drive, (void *)setup_count);
146 if (pair)
147 setup_count = max_t(u8, setup_count,
148 (unsigned long)ide_get_drivedata(pair));
151 if (setup_count > 5) /* shouldn't actually happen... */
152 setup_count = 5;
153 cmdprintk("Final address setup count: %d\n", setup_count);
156 * Program the address setup clocks into the ARTTIM registers.
157 * Avoid clearing the secondary channel's interrupt bit.
159 (void) pci_read_config_byte (dev, arttim_regs[drive->dn], &arttim);
160 if (hwif->channel)
161 arttim &= ~ARTTIM23_INTR_CH1;
162 arttim &= ~0xc0;
163 arttim |= setup_values[setup_count];
164 (void) pci_write_config_byte(dev, arttim_regs[drive->dn], arttim);
165 cmdprintk("Write 0x%02x to reg 0x%x\n", arttim, arttim_regs[drive->dn]);
169 * Attempts to set drive's PIO mode.
170 * Special cases are 8: prefetch off, 9: prefetch on (both never worked)
173 static void cmd64x_set_pio_mode(ide_drive_t *drive, const u8 pio)
176 * Filter out the prefetch control values
177 * to prevent PIO5 from being programmed
179 if (pio == 8 || pio == 9)
180 return;
182 cmd64x_tune_pio(drive, pio);
185 static void cmd64x_set_dma_mode(ide_drive_t *drive, const u8 speed)
187 ide_hwif_t *hwif = drive->hwif;
188 struct pci_dev *dev = to_pci_dev(hwif->dev);
189 u8 unit = drive->dn & 0x01;
190 u8 regU = 0, pciU = hwif->channel ? UDIDETCR1 : UDIDETCR0;
192 if (speed >= XFER_SW_DMA_0) {
193 (void) pci_read_config_byte(dev, pciU, &regU);
194 regU &= ~(unit ? 0xCA : 0x35);
197 switch(speed) {
198 case XFER_UDMA_5:
199 regU |= unit ? 0x0A : 0x05;
200 break;
201 case XFER_UDMA_4:
202 regU |= unit ? 0x4A : 0x15;
203 break;
204 case XFER_UDMA_3:
205 regU |= unit ? 0x8A : 0x25;
206 break;
207 case XFER_UDMA_2:
208 regU |= unit ? 0x42 : 0x11;
209 break;
210 case XFER_UDMA_1:
211 regU |= unit ? 0x82 : 0x21;
212 break;
213 case XFER_UDMA_0:
214 regU |= unit ? 0xC2 : 0x31;
215 break;
216 case XFER_MW_DMA_2:
217 program_cycle_times(drive, 120, 70);
218 break;
219 case XFER_MW_DMA_1:
220 program_cycle_times(drive, 150, 80);
221 break;
222 case XFER_MW_DMA_0:
223 program_cycle_times(drive, 480, 215);
224 break;
227 if (speed >= XFER_SW_DMA_0)
228 (void) pci_write_config_byte(dev, pciU, regU);
231 static void cmd648_clear_irq(ide_drive_t *drive)
233 ide_hwif_t *hwif = drive->hwif;
234 struct pci_dev *dev = to_pci_dev(hwif->dev);
235 unsigned long base = pci_resource_start(dev, 4);
236 u8 irq_mask = hwif->channel ? MRDMODE_INTR_CH1 :
237 MRDMODE_INTR_CH0;
238 u8 mrdmode = inb(base + 1);
240 /* clear the interrupt bit */
241 outb((mrdmode & ~(MRDMODE_INTR_CH0 | MRDMODE_INTR_CH1)) | irq_mask,
242 base + 1);
245 static void cmd64x_clear_irq(ide_drive_t *drive)
247 ide_hwif_t *hwif = drive->hwif;
248 struct pci_dev *dev = to_pci_dev(hwif->dev);
249 int irq_reg = hwif->channel ? ARTTIM23 : CFR;
250 u8 irq_mask = hwif->channel ? ARTTIM23_INTR_CH1 :
251 CFR_INTR_CH0;
252 u8 irq_stat = 0;
254 (void) pci_read_config_byte(dev, irq_reg, &irq_stat);
255 /* clear the interrupt bit */
256 (void) pci_write_config_byte(dev, irq_reg, irq_stat | irq_mask);
259 static int cmd648_test_irq(ide_hwif_t *hwif)
261 struct pci_dev *dev = to_pci_dev(hwif->dev);
262 unsigned long base = pci_resource_start(dev, 4);
263 u8 irq_mask = hwif->channel ? MRDMODE_INTR_CH1 :
264 MRDMODE_INTR_CH0;
265 u8 mrdmode = inb(base + 1);
267 pr_debug("%s: mrdmode: 0x%02x irq_mask: 0x%02x\n",
268 hwif->name, mrdmode, irq_mask);
270 return (mrdmode & irq_mask) ? 1 : 0;
273 static int cmd64x_test_irq(ide_hwif_t *hwif)
275 struct pci_dev *dev = to_pci_dev(hwif->dev);
276 int irq_reg = hwif->channel ? ARTTIM23 : CFR;
277 u8 irq_mask = hwif->channel ? ARTTIM23_INTR_CH1 :
278 CFR_INTR_CH0;
279 u8 irq_stat = 0;
281 (void) pci_read_config_byte(dev, irq_reg, &irq_stat);
283 pr_debug("%s: irq_stat: 0x%02x irq_mask: 0x%02x\n",
284 hwif->name, irq_stat, irq_mask);
286 return (irq_stat & irq_mask) ? 1 : 0;
290 * ASUS P55T2P4D with CMD646 chipset revision 0x01 requires the old
291 * event order for DMA transfers.
294 static int cmd646_1_dma_end(ide_drive_t *drive)
296 ide_hwif_t *hwif = drive->hwif;
297 u8 dma_stat = 0, dma_cmd = 0;
299 /* get DMA status */
300 dma_stat = inb(hwif->dma_base + ATA_DMA_STATUS);
301 /* read DMA command state */
302 dma_cmd = inb(hwif->dma_base + ATA_DMA_CMD);
303 /* stop DMA */
304 outb(dma_cmd & ~1, hwif->dma_base + ATA_DMA_CMD);
305 /* clear the INTR & ERROR bits */
306 outb(dma_stat | 6, hwif->dma_base + ATA_DMA_STATUS);
307 /* verify good DMA status */
308 return (dma_stat & 7) != 4;
311 static int init_chipset_cmd64x(struct pci_dev *dev)
313 u8 mrdmode = 0;
315 /* Set a good latency timer and cache line size value. */
316 (void) pci_write_config_byte(dev, PCI_LATENCY_TIMER, 64);
317 /* FIXME: pci_set_master() to ensure a good latency timer value */
320 * Enable interrupts, select MEMORY READ LINE for reads.
322 * NOTE: although not mentioned in the PCI0646U specs,
323 * bits 0-1 are write only and won't be read back as
324 * set or not -- PCI0646U2 specs clarify this point.
326 (void) pci_read_config_byte (dev, MRDMODE, &mrdmode);
327 mrdmode &= ~0x30;
328 (void) pci_write_config_byte(dev, MRDMODE, (mrdmode | 0x02));
330 return 0;
333 static u8 cmd64x_cable_detect(ide_hwif_t *hwif)
335 struct pci_dev *dev = to_pci_dev(hwif->dev);
336 u8 bmidecsr = 0, mask = hwif->channel ? 0x02 : 0x01;
338 switch (dev->device) {
339 case PCI_DEVICE_ID_CMD_648:
340 case PCI_DEVICE_ID_CMD_649:
341 pci_read_config_byte(dev, BMIDECSR, &bmidecsr);
342 return (bmidecsr & mask) ? ATA_CBL_PATA80 : ATA_CBL_PATA40;
343 default:
344 return ATA_CBL_PATA40;
348 static const struct ide_port_ops cmd64x_port_ops = {
349 .set_pio_mode = cmd64x_set_pio_mode,
350 .set_dma_mode = cmd64x_set_dma_mode,
351 .clear_irq = cmd64x_clear_irq,
352 .test_irq = cmd64x_test_irq,
353 .cable_detect = cmd64x_cable_detect,
356 static const struct ide_port_ops cmd648_port_ops = {
357 .set_pio_mode = cmd64x_set_pio_mode,
358 .set_dma_mode = cmd64x_set_dma_mode,
359 .clear_irq = cmd648_clear_irq,
360 .test_irq = cmd648_test_irq,
361 .cable_detect = cmd64x_cable_detect,
364 static const struct ide_dma_ops cmd646_rev1_dma_ops = {
365 .dma_host_set = ide_dma_host_set,
366 .dma_setup = ide_dma_setup,
367 .dma_start = ide_dma_start,
368 .dma_end = cmd646_1_dma_end,
369 .dma_test_irq = ide_dma_test_irq,
370 .dma_lost_irq = ide_dma_lost_irq,
371 .dma_timer_expiry = ide_dma_sff_timer_expiry,
372 .dma_sff_read_status = ide_dma_sff_read_status,
375 static const struct ide_port_info cmd64x_chipsets[] __devinitdata = {
376 { /* 0: CMD643 */
377 .name = DRV_NAME,
378 .init_chipset = init_chipset_cmd64x,
379 .enablebits = {{0x00,0x00,0x00}, {0x51,0x08,0x08}},
380 .port_ops = &cmd64x_port_ops,
381 .host_flags = IDE_HFLAG_CLEAR_SIMPLEX |
382 IDE_HFLAG_ABUSE_PREFETCH,
383 .pio_mask = ATA_PIO5,
384 .mwdma_mask = ATA_MWDMA2,
385 .udma_mask = 0x00, /* no udma */
387 { /* 1: CMD646 */
388 .name = DRV_NAME,
389 .init_chipset = init_chipset_cmd64x,
390 .enablebits = {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
391 .port_ops = &cmd648_port_ops,
392 .host_flags = IDE_HFLAG_ABUSE_PREFETCH,
393 .pio_mask = ATA_PIO5,
394 .mwdma_mask = ATA_MWDMA2,
395 .udma_mask = ATA_UDMA2,
397 { /* 2: CMD648 */
398 .name = DRV_NAME,
399 .init_chipset = init_chipset_cmd64x,
400 .enablebits = {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
401 .port_ops = &cmd648_port_ops,
402 .host_flags = IDE_HFLAG_ABUSE_PREFETCH,
403 .pio_mask = ATA_PIO5,
404 .mwdma_mask = ATA_MWDMA2,
405 .udma_mask = ATA_UDMA4,
407 { /* 3: CMD649 */
408 .name = DRV_NAME,
409 .init_chipset = init_chipset_cmd64x,
410 .enablebits = {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
411 .port_ops = &cmd648_port_ops,
412 .host_flags = IDE_HFLAG_ABUSE_PREFETCH,
413 .pio_mask = ATA_PIO5,
414 .mwdma_mask = ATA_MWDMA2,
415 .udma_mask = ATA_UDMA5,
419 static int __devinit cmd64x_init_one(struct pci_dev *dev, const struct pci_device_id *id)
421 struct ide_port_info d;
422 u8 idx = id->driver_data;
424 d = cmd64x_chipsets[idx];
426 if (idx == 1) {
428 * UltraDMA only supported on PCI646U and PCI646U2, which
429 * correspond to revisions 0x03, 0x05 and 0x07 respectively.
430 * Actually, although the CMD tech support people won't
431 * tell me the details, the 0x03 revision cannot support
432 * UDMA correctly without hardware modifications, and even
433 * then it only works with Quantum disks due to some
434 * hold time assumptions in the 646U part which are fixed
435 * in the 646U2.
437 * So we only do UltraDMA on revision 0x05 and 0x07 chipsets.
439 if (dev->revision < 5) {
440 d.udma_mask = 0x00;
442 * The original PCI0646 didn't have the primary
443 * channel enable bit, it appeared starting with
444 * PCI0646U (i.e. revision ID 3).
446 if (dev->revision < 3) {
447 d.enablebits[0].reg = 0;
448 d.port_ops = &cmd64x_port_ops;
449 if (dev->revision == 1)
450 d.dma_ops = &cmd646_rev1_dma_ops;
455 return ide_pci_init_one(dev, &d, NULL);
458 static const struct pci_device_id cmd64x_pci_tbl[] = {
459 { PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_643), 0 },
460 { PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_646), 1 },
461 { PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_648), 2 },
462 { PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_649), 3 },
463 { 0, },
465 MODULE_DEVICE_TABLE(pci, cmd64x_pci_tbl);
467 static struct pci_driver cmd64x_pci_driver = {
468 .name = "CMD64x_IDE",
469 .id_table = cmd64x_pci_tbl,
470 .probe = cmd64x_init_one,
471 .remove = ide_pci_remove,
472 .suspend = ide_pci_suspend,
473 .resume = ide_pci_resume,
476 static int __init cmd64x_ide_init(void)
478 return ide_pci_register_driver(&cmd64x_pci_driver);
481 static void __exit cmd64x_ide_exit(void)
483 pci_unregister_driver(&cmd64x_pci_driver);
486 module_init(cmd64x_ide_init);
487 module_exit(cmd64x_ide_exit);
489 MODULE_AUTHOR("Eddie Dost, David Miller, Andre Hedrick");
490 MODULE_DESCRIPTION("PCI driver module for CMD64x IDE");
491 MODULE_LICENSE("GPL");