2 * cmd64x.c: Enable interrupts at initialization time on Ultra/PCI machines.
3 * Due to massive hardware bugs, UltraDMA is only supported
4 * on the 646U2 and not on the 646U.
6 * Copyright (C) 1998 Eddie C. Dost (ecd@skynet.be)
7 * Copyright (C) 1998 David S. Miller (davem@redhat.com)
9 * Copyright (C) 1999-2002 Andre Hedrick <andre@linux-ide.org>
10 * Copyright (C) 2007-2010 Bartlomiej Zolnierkiewicz
11 * Copyright (C) 2007,2009 MontaVista Software, Inc. <source@mvista.com>
14 #include <linux/module.h>
15 #include <linux/types.h>
16 #include <linux/pci.h>
17 #include <linux/ide.h>
18 #include <linux/init.h>
22 #define DRV_NAME "cmd64x"
25 * CMD64x specific registers definition.
28 #define CFR_INTR_CH0 0x04
36 #define ARTTIM23_DIS_RA2 0x04
37 #define ARTTIM23_DIS_RA3 0x08
38 #define ARTTIM23_INTR_CH1 0x10
45 #define MRDMODE_INTR_CH0 0x04
46 #define MRDMODE_INTR_CH1 0x08
47 #define UDIDETCR0 0x73
51 #define UDIDETCR1 0x7B
54 static void cmd64x_program_timings(ide_drive_t
*drive
, u8 mode
)
56 ide_hwif_t
*hwif
= drive
->hwif
;
57 struct pci_dev
*dev
= to_pci_dev(drive
->hwif
->dev
);
58 int bus_speed
= ide_pci_clk
? ide_pci_clk
: 33;
59 const unsigned long T
= 1000000 / bus_speed
;
60 static const u8 recovery_values
[] =
61 {15, 15, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 0};
62 static const u8 setup_values
[] = {0x40, 0x40, 0x40, 0x80, 0, 0xc0};
63 static const u8 arttim_regs
[4] = {ARTTIM0
, ARTTIM1
, ARTTIM23
, ARTTIM23
};
64 static const u8 drwtim_regs
[4] = {DRWTIM0
, DRWTIM1
, DRWTIM2
, DRWTIM3
};
68 ide_timing_compute(drive
, mode
, &t
, T
, 0);
71 * In case we've got too long recovery phase, try to lengthen
75 t
.active
+= t
.recover
- 16;
78 if (t
.active
> 16) /* shouldn't actually happen... */
82 * Convert values to internal chipset representation
84 t
.recover
= recovery_values
[t
.recover
];
87 /* Program the active/recovery counts into the DRWTIM register */
88 pci_write_config_byte(dev
, drwtim_regs
[drive
->dn
],
89 (t
.active
<< 4) | t
.recover
);
92 * The primary channel has individual address setup timing registers
93 * for each drive and the hardware selects the slowest timing itself.
94 * The secondary channel has one common register and we have to select
95 * the slowest address setup timing ourselves.
98 ide_drive_t
*pair
= ide_get_pair_dev(drive
);
101 struct ide_timing tp
;
103 ide_timing_compute(pair
, pair
->pio_mode
, &tp
, T
, 0);
104 ide_timing_merge(&t
, &tp
, &t
, IDE_TIMING_SETUP
);
105 if (pair
->dma_mode
) {
106 ide_timing_compute(pair
, pair
->dma_mode
,
108 ide_timing_merge(&tp
, &t
, &t
, IDE_TIMING_SETUP
);
113 if (t
.setup
> 5) /* shouldn't actually happen... */
117 * Program the address setup clocks into the ARTTIM registers.
118 * Avoid clearing the secondary channel's interrupt bit.
120 (void) pci_read_config_byte (dev
, arttim_regs
[drive
->dn
], &arttim
);
122 arttim
&= ~ARTTIM23_INTR_CH1
;
124 arttim
|= setup_values
[t
.setup
];
125 (void) pci_write_config_byte(dev
, arttim_regs
[drive
->dn
], arttim
);
129 * Attempts to set drive's PIO mode.
130 * Special cases are 8: prefetch off, 9: prefetch on (both never worked)
133 static void cmd64x_set_pio_mode(ide_hwif_t
*hwif
, ide_drive_t
*drive
)
135 const u8 pio
= drive
->pio_mode
- XFER_PIO_0
;
138 * Filter out the prefetch control values
139 * to prevent PIO5 from being programmed
141 if (pio
== 8 || pio
== 9)
144 cmd64x_program_timings(drive
, XFER_PIO_0
+ pio
);
147 static void cmd64x_set_dma_mode(ide_hwif_t
*hwif
, ide_drive_t
*drive
)
149 struct pci_dev
*dev
= to_pci_dev(hwif
->dev
);
150 u8 unit
= drive
->dn
& 0x01;
151 u8 regU
= 0, pciU
= hwif
->channel
? UDIDETCR1
: UDIDETCR0
;
152 const u8 speed
= drive
->dma_mode
;
154 pci_read_config_byte(dev
, pciU
, ®U
);
155 regU
&= ~(unit
? 0xCA : 0x35);
159 regU
|= unit
? 0x0A : 0x05;
162 regU
|= unit
? 0x4A : 0x15;
165 regU
|= unit
? 0x8A : 0x25;
168 regU
|= unit
? 0x42 : 0x11;
171 regU
|= unit
? 0x82 : 0x21;
174 regU
|= unit
? 0xC2 : 0x31;
179 cmd64x_program_timings(drive
, speed
);
183 pci_write_config_byte(dev
, pciU
, regU
);
186 static void cmd648_clear_irq(ide_drive_t
*drive
)
188 ide_hwif_t
*hwif
= drive
->hwif
;
189 struct pci_dev
*dev
= to_pci_dev(hwif
->dev
);
190 unsigned long base
= pci_resource_start(dev
, 4);
191 u8 irq_mask
= hwif
->channel
? MRDMODE_INTR_CH1
:
193 u8 mrdmode
= inb(base
+ 1);
195 /* clear the interrupt bit */
196 outb((mrdmode
& ~(MRDMODE_INTR_CH0
| MRDMODE_INTR_CH1
)) | irq_mask
,
200 static void cmd64x_clear_irq(ide_drive_t
*drive
)
202 ide_hwif_t
*hwif
= drive
->hwif
;
203 struct pci_dev
*dev
= to_pci_dev(hwif
->dev
);
204 int irq_reg
= hwif
->channel
? ARTTIM23
: CFR
;
205 u8 irq_mask
= hwif
->channel
? ARTTIM23_INTR_CH1
:
209 (void) pci_read_config_byte(dev
, irq_reg
, &irq_stat
);
210 /* clear the interrupt bit */
211 (void) pci_write_config_byte(dev
, irq_reg
, irq_stat
| irq_mask
);
214 static int cmd648_test_irq(ide_hwif_t
*hwif
)
216 struct pci_dev
*dev
= to_pci_dev(hwif
->dev
);
217 unsigned long base
= pci_resource_start(dev
, 4);
218 u8 irq_mask
= hwif
->channel
? MRDMODE_INTR_CH1
:
220 u8 mrdmode
= inb(base
+ 1);
222 pr_debug("%s: mrdmode: 0x%02x irq_mask: 0x%02x\n",
223 hwif
->name
, mrdmode
, irq_mask
);
225 return (mrdmode
& irq_mask
) ? 1 : 0;
228 static int cmd64x_test_irq(ide_hwif_t
*hwif
)
230 struct pci_dev
*dev
= to_pci_dev(hwif
->dev
);
231 int irq_reg
= hwif
->channel
? ARTTIM23
: CFR
;
232 u8 irq_mask
= hwif
->channel
? ARTTIM23_INTR_CH1
:
236 (void) pci_read_config_byte(dev
, irq_reg
, &irq_stat
);
238 pr_debug("%s: irq_stat: 0x%02x irq_mask: 0x%02x\n",
239 hwif
->name
, irq_stat
, irq_mask
);
241 return (irq_stat
& irq_mask
) ? 1 : 0;
245 * ASUS P55T2P4D with CMD646 chipset revision 0x01 requires the old
246 * event order for DMA transfers.
249 static int cmd646_1_dma_end(ide_drive_t
*drive
)
251 ide_hwif_t
*hwif
= drive
->hwif
;
252 u8 dma_stat
= 0, dma_cmd
= 0;
255 dma_stat
= inb(hwif
->dma_base
+ ATA_DMA_STATUS
);
256 /* read DMA command state */
257 dma_cmd
= inb(hwif
->dma_base
+ ATA_DMA_CMD
);
259 outb(dma_cmd
& ~1, hwif
->dma_base
+ ATA_DMA_CMD
);
260 /* clear the INTR & ERROR bits */
261 outb(dma_stat
| 6, hwif
->dma_base
+ ATA_DMA_STATUS
);
262 /* verify good DMA status */
263 return (dma_stat
& 7) != 4;
266 static int init_chipset_cmd64x(struct pci_dev
*dev
)
270 /* Set a good latency timer and cache line size value. */
271 (void) pci_write_config_byte(dev
, PCI_LATENCY_TIMER
, 64);
272 /* FIXME: pci_set_master() to ensure a good latency timer value */
275 * Enable interrupts, select MEMORY READ LINE for reads.
277 * NOTE: although not mentioned in the PCI0646U specs,
278 * bits 0-1 are write only and won't be read back as
279 * set or not -- PCI0646U2 specs clarify this point.
281 (void) pci_read_config_byte (dev
, MRDMODE
, &mrdmode
);
283 (void) pci_write_config_byte(dev
, MRDMODE
, (mrdmode
| 0x02));
288 static u8
cmd64x_cable_detect(ide_hwif_t
*hwif
)
290 struct pci_dev
*dev
= to_pci_dev(hwif
->dev
);
291 u8 bmidecsr
= 0, mask
= hwif
->channel
? 0x02 : 0x01;
293 switch (dev
->device
) {
294 case PCI_DEVICE_ID_CMD_648
:
295 case PCI_DEVICE_ID_CMD_649
:
296 pci_read_config_byte(dev
, BMIDECSR
, &bmidecsr
);
297 return (bmidecsr
& mask
) ? ATA_CBL_PATA80
: ATA_CBL_PATA40
;
299 return ATA_CBL_PATA40
;
303 static const struct ide_port_ops cmd64x_port_ops
= {
304 .set_pio_mode
= cmd64x_set_pio_mode
,
305 .set_dma_mode
= cmd64x_set_dma_mode
,
306 .clear_irq
= cmd64x_clear_irq
,
307 .test_irq
= cmd64x_test_irq
,
308 .cable_detect
= cmd64x_cable_detect
,
311 static const struct ide_port_ops cmd648_port_ops
= {
312 .set_pio_mode
= cmd64x_set_pio_mode
,
313 .set_dma_mode
= cmd64x_set_dma_mode
,
314 .clear_irq
= cmd648_clear_irq
,
315 .test_irq
= cmd648_test_irq
,
316 .cable_detect
= cmd64x_cable_detect
,
319 static const struct ide_dma_ops cmd646_rev1_dma_ops
= {
320 .dma_host_set
= ide_dma_host_set
,
321 .dma_setup
= ide_dma_setup
,
322 .dma_start
= ide_dma_start
,
323 .dma_end
= cmd646_1_dma_end
,
324 .dma_test_irq
= ide_dma_test_irq
,
325 .dma_lost_irq
= ide_dma_lost_irq
,
326 .dma_timer_expiry
= ide_dma_sff_timer_expiry
,
327 .dma_sff_read_status
= ide_dma_sff_read_status
,
330 static const struct ide_port_info cmd64x_chipsets
[] __devinitdata
= {
333 .init_chipset
= init_chipset_cmd64x
,
334 .enablebits
= {{0x00,0x00,0x00}, {0x51,0x08,0x08}},
335 .port_ops
= &cmd64x_port_ops
,
336 .host_flags
= IDE_HFLAG_CLEAR_SIMPLEX
|
337 IDE_HFLAG_ABUSE_PREFETCH
|
339 .pio_mask
= ATA_PIO5
,
340 .mwdma_mask
= ATA_MWDMA2
,
341 .udma_mask
= 0x00, /* no udma */
345 .init_chipset
= init_chipset_cmd64x
,
346 .enablebits
= {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
347 .port_ops
= &cmd648_port_ops
,
348 .host_flags
= IDE_HFLAG_ABUSE_PREFETCH
|
350 .pio_mask
= ATA_PIO5
,
351 .mwdma_mask
= ATA_MWDMA2
,
352 .udma_mask
= ATA_UDMA2
,
356 .init_chipset
= init_chipset_cmd64x
,
357 .enablebits
= {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
358 .port_ops
= &cmd648_port_ops
,
359 .host_flags
= IDE_HFLAG_ABUSE_PREFETCH
,
360 .pio_mask
= ATA_PIO5
,
361 .mwdma_mask
= ATA_MWDMA2
,
362 .udma_mask
= ATA_UDMA4
,
366 .init_chipset
= init_chipset_cmd64x
,
367 .enablebits
= {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
368 .port_ops
= &cmd648_port_ops
,
369 .host_flags
= IDE_HFLAG_ABUSE_PREFETCH
,
370 .pio_mask
= ATA_PIO5
,
371 .mwdma_mask
= ATA_MWDMA2
,
372 .udma_mask
= ATA_UDMA5
,
376 static int __devinit
cmd64x_init_one(struct pci_dev
*dev
, const struct pci_device_id
*id
)
378 struct ide_port_info d
;
379 u8 idx
= id
->driver_data
;
381 d
= cmd64x_chipsets
[idx
];
385 * UltraDMA only supported on PCI646U and PCI646U2, which
386 * correspond to revisions 0x03, 0x05 and 0x07 respectively.
387 * Actually, although the CMD tech support people won't
388 * tell me the details, the 0x03 revision cannot support
389 * UDMA correctly without hardware modifications, and even
390 * then it only works with Quantum disks due to some
391 * hold time assumptions in the 646U part which are fixed
394 * So we only do UltraDMA on revision 0x05 and 0x07 chipsets.
396 if (dev
->revision
< 5) {
399 * The original PCI0646 didn't have the primary
400 * channel enable bit, it appeared starting with
401 * PCI0646U (i.e. revision ID 3).
403 if (dev
->revision
< 3) {
404 d
.enablebits
[0].reg
= 0;
405 d
.port_ops
= &cmd64x_port_ops
;
406 if (dev
->revision
== 1)
407 d
.dma_ops
= &cmd646_rev1_dma_ops
;
412 return ide_pci_init_one(dev
, &d
, NULL
);
415 static const struct pci_device_id cmd64x_pci_tbl
[] = {
416 { PCI_VDEVICE(CMD
, PCI_DEVICE_ID_CMD_643
), 0 },
417 { PCI_VDEVICE(CMD
, PCI_DEVICE_ID_CMD_646
), 1 },
418 { PCI_VDEVICE(CMD
, PCI_DEVICE_ID_CMD_648
), 2 },
419 { PCI_VDEVICE(CMD
, PCI_DEVICE_ID_CMD_649
), 3 },
422 MODULE_DEVICE_TABLE(pci
, cmd64x_pci_tbl
);
424 static struct pci_driver cmd64x_pci_driver
= {
425 .name
= "CMD64x_IDE",
426 .id_table
= cmd64x_pci_tbl
,
427 .probe
= cmd64x_init_one
,
428 .remove
= ide_pci_remove
,
429 .suspend
= ide_pci_suspend
,
430 .resume
= ide_pci_resume
,
433 static int __init
cmd64x_ide_init(void)
435 return ide_pci_register_driver(&cmd64x_pci_driver
);
438 static void __exit
cmd64x_ide_exit(void)
440 pci_unregister_driver(&cmd64x_pci_driver
);
443 module_init(cmd64x_ide_init
);
444 module_exit(cmd64x_ide_exit
);
446 MODULE_AUTHOR("Eddie Dost, David Miller, Andre Hedrick, Bartlomiej Zolnierkiewicz");
447 MODULE_DESCRIPTION("PCI driver module for CMD64x IDE");
448 MODULE_LICENSE("GPL");