Loading...
1/*
2 * cmd64x.c: Enable interrupts at initialization time on Ultra/PCI machines.
3 * Due to massive hardware bugs, UltraDMA is only supported
4 * on the 646U2 and not on the 646U.
5 *
6 * Copyright (C) 1998 Eddie C. Dost (ecd@skynet.be)
7 * Copyright (C) 1998 David S. Miller (davem@redhat.com)
8 *
9 * Copyright (C) 1999-2002 Andre Hedrick <andre@linux-ide.org>
10 * Copyright (C) 2007-2010 Bartlomiej Zolnierkiewicz
11 * Copyright (C) 2007,2009 MontaVista Software, Inc. <source@mvista.com>
12 */
13
14#include <linux/module.h>
15#include <linux/types.h>
16#include <linux/pci.h>
17#include <linux/ide.h>
18#include <linux/init.h>
19
20#include <asm/io.h>
21
22#define DRV_NAME "cmd64x"
23
24/*
25 * CMD64x specific registers definition.
26 */
27#define CFR 0x50
28#define CFR_INTR_CH0 0x04
29
30#define CMDTIM 0x52
31#define ARTTIM0 0x53
32#define DRWTIM0 0x54
33#define ARTTIM1 0x55
34#define DRWTIM1 0x56
35#define ARTTIM23 0x57
36#define ARTTIM23_DIS_RA2 0x04
37#define ARTTIM23_DIS_RA3 0x08
38#define ARTTIM23_INTR_CH1 0x10
39#define DRWTIM2 0x58
40#define BRST 0x59
41#define DRWTIM3 0x5b
42
43#define BMIDECR0 0x70
44#define MRDMODE 0x71
45#define MRDMODE_INTR_CH0 0x04
46#define MRDMODE_INTR_CH1 0x08
47#define UDIDETCR0 0x73
48#define DTPR0 0x74
49#define BMIDECR1 0x78
50#define BMIDECSR 0x79
51#define UDIDETCR1 0x7B
52#define DTPR1 0x7C
53
54static void cmd64x_program_timings(ide_drive_t *drive, u8 mode)
55{
56 ide_hwif_t *hwif = drive->hwif;
57 struct pci_dev *dev = to_pci_dev(drive->hwif->dev);
58 int bus_speed = ide_pci_clk ? ide_pci_clk : 33;
59 const unsigned long T = 1000000 / bus_speed;
60 static const u8 recovery_values[] =
61 {15, 15, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 0};
62 static const u8 setup_values[] = {0x40, 0x40, 0x40, 0x80, 0, 0xc0};
63 static const u8 arttim_regs[4] = {ARTTIM0, ARTTIM1, ARTTIM23, ARTTIM23};
64 static const u8 drwtim_regs[4] = {DRWTIM0, DRWTIM1, DRWTIM2, DRWTIM3};
65 struct ide_timing t;
66 u8 arttim = 0;
67
68 ide_timing_compute(drive, mode, &t, T, 0);
69
70 /*
71 * In case we've got too long recovery phase, try to lengthen
72 * the active phase
73 */
74 if (t.recover > 16) {
75 t.active += t.recover - 16;
76 t.recover = 16;
77 }
78 if (t.active > 16) /* shouldn't actually happen... */
79 t.active = 16;
80
81 /*
82 * Convert values to internal chipset representation
83 */
84 t.recover = recovery_values[t.recover];
85 t.active &= 0x0f;
86
87 /* Program the active/recovery counts into the DRWTIM register */
88 pci_write_config_byte(dev, drwtim_regs[drive->dn],
89 (t.active << 4) | t.recover);
90
91 /*
92 * The primary channel has individual address setup timing registers
93 * for each drive and the hardware selects the slowest timing itself.
94 * The secondary channel has one common register and we have to select
95 * the slowest address setup timing ourselves.
96 */
97 if (hwif->channel) {
98 ide_drive_t *pair = ide_get_pair_dev(drive);
99
100 if (pair) {
101 struct ide_timing tp;
102
103 ide_timing_compute(pair, pair->pio_mode, &tp, T, 0);
104 ide_timing_merge(&t, &tp, &t, IDE_TIMING_SETUP);
105 if (pair->dma_mode) {
106 ide_timing_compute(pair, pair->dma_mode,
107 &tp, T, 0);
108 ide_timing_merge(&tp, &t, &t, IDE_TIMING_SETUP);
109 }
110 }
111 }
112
113 if (t.setup > 5) /* shouldn't actually happen... */
114 t.setup = 5;
115
116 /*
117 * Program the address setup clocks into the ARTTIM registers.
118 * Avoid clearing the secondary channel's interrupt bit.
119 */
120 (void) pci_read_config_byte (dev, arttim_regs[drive->dn], &arttim);
121 if (hwif->channel)
122 arttim &= ~ARTTIM23_INTR_CH1;
123 arttim &= ~0xc0;
124 arttim |= setup_values[t.setup];
125 (void) pci_write_config_byte(dev, arttim_regs[drive->dn], arttim);
126}
127
128/*
129 * Attempts to set drive's PIO mode.
130 * Special cases are 8: prefetch off, 9: prefetch on (both never worked)
131 */
132
133static void cmd64x_set_pio_mode(ide_hwif_t *hwif, ide_drive_t *drive)
134{
135 const u8 pio = drive->pio_mode - XFER_PIO_0;
136
137 /*
138 * Filter out the prefetch control values
139 * to prevent PIO5 from being programmed
140 */
141 if (pio == 8 || pio == 9)
142 return;
143
144 cmd64x_program_timings(drive, XFER_PIO_0 + pio);
145}
146
147static void cmd64x_set_dma_mode(ide_hwif_t *hwif, ide_drive_t *drive)
148{
149 struct pci_dev *dev = to_pci_dev(hwif->dev);
150 u8 unit = drive->dn & 0x01;
151 u8 regU = 0, pciU = hwif->channel ? UDIDETCR1 : UDIDETCR0;
152 const u8 speed = drive->dma_mode;
153
154 pci_read_config_byte(dev, pciU, ®U);
155 regU &= ~(unit ? 0xCA : 0x35);
156
157 switch(speed) {
158 case XFER_UDMA_5:
159 regU |= unit ? 0x0A : 0x05;
160 break;
161 case XFER_UDMA_4:
162 regU |= unit ? 0x4A : 0x15;
163 break;
164 case XFER_UDMA_3:
165 regU |= unit ? 0x8A : 0x25;
166 break;
167 case XFER_UDMA_2:
168 regU |= unit ? 0x42 : 0x11;
169 break;
170 case XFER_UDMA_1:
171 regU |= unit ? 0x82 : 0x21;
172 break;
173 case XFER_UDMA_0:
174 regU |= unit ? 0xC2 : 0x31;
175 break;
176 case XFER_MW_DMA_2:
177 case XFER_MW_DMA_1:
178 case XFER_MW_DMA_0:
179 cmd64x_program_timings(drive, speed);
180 break;
181 }
182
183 pci_write_config_byte(dev, pciU, regU);
184}
185
186static void cmd648_clear_irq(ide_drive_t *drive)
187{
188 ide_hwif_t *hwif = drive->hwif;
189 struct pci_dev *dev = to_pci_dev(hwif->dev);
190 unsigned long base = pci_resource_start(dev, 4);
191 u8 irq_mask = hwif->channel ? MRDMODE_INTR_CH1 :
192 MRDMODE_INTR_CH0;
193 u8 mrdmode = inb(base + 1);
194
195 /* clear the interrupt bit */
196 outb((mrdmode & ~(MRDMODE_INTR_CH0 | MRDMODE_INTR_CH1)) | irq_mask,
197 base + 1);
198}
199
200static void cmd64x_clear_irq(ide_drive_t *drive)
201{
202 ide_hwif_t *hwif = drive->hwif;
203 struct pci_dev *dev = to_pci_dev(hwif->dev);
204 int irq_reg = hwif->channel ? ARTTIM23 : CFR;
205 u8 irq_mask = hwif->channel ? ARTTIM23_INTR_CH1 :
206 CFR_INTR_CH0;
207 u8 irq_stat = 0;
208
209 (void) pci_read_config_byte(dev, irq_reg, &irq_stat);
210 /* clear the interrupt bit */
211 (void) pci_write_config_byte(dev, irq_reg, irq_stat | irq_mask);
212}
213
214static int cmd648_test_irq(ide_hwif_t *hwif)
215{
216 struct pci_dev *dev = to_pci_dev(hwif->dev);
217 unsigned long base = pci_resource_start(dev, 4);
218 u8 irq_mask = hwif->channel ? MRDMODE_INTR_CH1 :
219 MRDMODE_INTR_CH0;
220 u8 mrdmode = inb(base + 1);
221
222 pr_debug("%s: mrdmode: 0x%02x irq_mask: 0x%02x\n",
223 hwif->name, mrdmode, irq_mask);
224
225 return (mrdmode & irq_mask) ? 1 : 0;
226}
227
228static int cmd64x_test_irq(ide_hwif_t *hwif)
229{
230 struct pci_dev *dev = to_pci_dev(hwif->dev);
231 int irq_reg = hwif->channel ? ARTTIM23 : CFR;
232 u8 irq_mask = hwif->channel ? ARTTIM23_INTR_CH1 :
233 CFR_INTR_CH0;
234 u8 irq_stat = 0;
235
236 (void) pci_read_config_byte(dev, irq_reg, &irq_stat);
237
238 pr_debug("%s: irq_stat: 0x%02x irq_mask: 0x%02x\n",
239 hwif->name, irq_stat, irq_mask);
240
241 return (irq_stat & irq_mask) ? 1 : 0;
242}
243
244/*
245 * ASUS P55T2P4D with CMD646 chipset revision 0x01 requires the old
246 * event order for DMA transfers.
247 */
248
249static int cmd646_1_dma_end(ide_drive_t *drive)
250{
251 ide_hwif_t *hwif = drive->hwif;
252 u8 dma_stat = 0, dma_cmd = 0;
253
254 /* get DMA status */
255 dma_stat = inb(hwif->dma_base + ATA_DMA_STATUS);
256 /* read DMA command state */
257 dma_cmd = inb(hwif->dma_base + ATA_DMA_CMD);
258 /* stop DMA */
259 outb(dma_cmd & ~1, hwif->dma_base + ATA_DMA_CMD);
260 /* clear the INTR & ERROR bits */
261 outb(dma_stat | 6, hwif->dma_base + ATA_DMA_STATUS);
262 /* verify good DMA status */
263 return (dma_stat & 7) != 4;
264}
265
266static int init_chipset_cmd64x(struct pci_dev *dev)
267{
268 u8 mrdmode = 0;
269
270 /* Set a good latency timer and cache line size value. */
271 (void) pci_write_config_byte(dev, PCI_LATENCY_TIMER, 64);
272 /* FIXME: pci_set_master() to ensure a good latency timer value */
273
274 /*
275 * Enable interrupts, select MEMORY READ LINE for reads.
276 *
277 * NOTE: although not mentioned in the PCI0646U specs,
278 * bits 0-1 are write only and won't be read back as
279 * set or not -- PCI0646U2 specs clarify this point.
280 */
281 (void) pci_read_config_byte (dev, MRDMODE, &mrdmode);
282 mrdmode &= ~0x30;
283 (void) pci_write_config_byte(dev, MRDMODE, (mrdmode | 0x02));
284
285 return 0;
286}
287
288static u8 cmd64x_cable_detect(ide_hwif_t *hwif)
289{
290 struct pci_dev *dev = to_pci_dev(hwif->dev);
291 u8 bmidecsr = 0, mask = hwif->channel ? 0x02 : 0x01;
292
293 switch (dev->device) {
294 case PCI_DEVICE_ID_CMD_648:
295 case PCI_DEVICE_ID_CMD_649:
296 pci_read_config_byte(dev, BMIDECSR, &bmidecsr);
297 return (bmidecsr & mask) ? ATA_CBL_PATA80 : ATA_CBL_PATA40;
298 default:
299 return ATA_CBL_PATA40;
300 }
301}
302
303static const struct ide_port_ops cmd64x_port_ops = {
304 .set_pio_mode = cmd64x_set_pio_mode,
305 .set_dma_mode = cmd64x_set_dma_mode,
306 .clear_irq = cmd64x_clear_irq,
307 .test_irq = cmd64x_test_irq,
308 .cable_detect = cmd64x_cable_detect,
309};
310
311static const struct ide_port_ops cmd648_port_ops = {
312 .set_pio_mode = cmd64x_set_pio_mode,
313 .set_dma_mode = cmd64x_set_dma_mode,
314 .clear_irq = cmd648_clear_irq,
315 .test_irq = cmd648_test_irq,
316 .cable_detect = cmd64x_cable_detect,
317};
318
319static const struct ide_dma_ops cmd646_rev1_dma_ops = {
320 .dma_host_set = ide_dma_host_set,
321 .dma_setup = ide_dma_setup,
322 .dma_start = ide_dma_start,
323 .dma_end = cmd646_1_dma_end,
324 .dma_test_irq = ide_dma_test_irq,
325 .dma_lost_irq = ide_dma_lost_irq,
326 .dma_timer_expiry = ide_dma_sff_timer_expiry,
327 .dma_sff_read_status = ide_dma_sff_read_status,
328};
329
330static const struct ide_port_info cmd64x_chipsets[] __devinitdata = {
331 { /* 0: CMD643 */
332 .name = DRV_NAME,
333 .init_chipset = init_chipset_cmd64x,
334 .enablebits = {{0x00,0x00,0x00}, {0x51,0x08,0x08}},
335 .port_ops = &cmd64x_port_ops,
336 .host_flags = IDE_HFLAG_CLEAR_SIMPLEX |
337 IDE_HFLAG_ABUSE_PREFETCH |
338 IDE_HFLAG_SERIALIZE,
339 .pio_mask = ATA_PIO5,
340 .mwdma_mask = ATA_MWDMA2,
341 .udma_mask = 0x00, /* no udma */
342 },
343 { /* 1: CMD646 */
344 .name = DRV_NAME,
345 .init_chipset = init_chipset_cmd64x,
346 .enablebits = {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
347 .port_ops = &cmd648_port_ops,
348 .host_flags = IDE_HFLAG_ABUSE_PREFETCH |
349 IDE_HFLAG_SERIALIZE,
350 .pio_mask = ATA_PIO5,
351 .mwdma_mask = ATA_MWDMA2,
352 .udma_mask = ATA_UDMA2,
353 },
354 { /* 2: CMD648 */
355 .name = DRV_NAME,
356 .init_chipset = init_chipset_cmd64x,
357 .enablebits = {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
358 .port_ops = &cmd648_port_ops,
359 .host_flags = IDE_HFLAG_ABUSE_PREFETCH,
360 .pio_mask = ATA_PIO5,
361 .mwdma_mask = ATA_MWDMA2,
362 .udma_mask = ATA_UDMA4,
363 },
364 { /* 3: CMD649 */
365 .name = DRV_NAME,
366 .init_chipset = init_chipset_cmd64x,
367 .enablebits = {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
368 .port_ops = &cmd648_port_ops,
369 .host_flags = IDE_HFLAG_ABUSE_PREFETCH,
370 .pio_mask = ATA_PIO5,
371 .mwdma_mask = ATA_MWDMA2,
372 .udma_mask = ATA_UDMA5,
373 }
374};
375
376static int __devinit cmd64x_init_one(struct pci_dev *dev, const struct pci_device_id *id)
377{
378 struct ide_port_info d;
379 u8 idx = id->driver_data;
380
381 d = cmd64x_chipsets[idx];
382
383 if (idx == 1) {
384 /*
385 * UltraDMA only supported on PCI646U and PCI646U2, which
386 * correspond to revisions 0x03, 0x05 and 0x07 respectively.
387 * Actually, although the CMD tech support people won't
388 * tell me the details, the 0x03 revision cannot support
389 * UDMA correctly without hardware modifications, and even
390 * then it only works with Quantum disks due to some
391 * hold time assumptions in the 646U part which are fixed
392 * in the 646U2.
393 *
394 * So we only do UltraDMA on revision 0x05 and 0x07 chipsets.
395 */
396 if (dev->revision < 5) {
397 d.udma_mask = 0x00;
398 /*
399 * The original PCI0646 didn't have the primary
400 * channel enable bit, it appeared starting with
401 * PCI0646U (i.e. revision ID 3).
402 */
403 if (dev->revision < 3) {
404 d.enablebits[0].reg = 0;
405 d.port_ops = &cmd64x_port_ops;
406 if (dev->revision == 1)
407 d.dma_ops = &cmd646_rev1_dma_ops;
408 }
409 }
410 }
411
412 return ide_pci_init_one(dev, &d, NULL);
413}
414
415static const struct pci_device_id cmd64x_pci_tbl[] = {
416 { PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_643), 0 },
417 { PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_646), 1 },
418 { PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_648), 2 },
419 { PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_649), 3 },
420 { 0, },
421};
422MODULE_DEVICE_TABLE(pci, cmd64x_pci_tbl);
423
424static struct pci_driver cmd64x_pci_driver = {
425 .name = "CMD64x_IDE",
426 .id_table = cmd64x_pci_tbl,
427 .probe = cmd64x_init_one,
428 .remove = ide_pci_remove,
429 .suspend = ide_pci_suspend,
430 .resume = ide_pci_resume,
431};
432
433static int __init cmd64x_ide_init(void)
434{
435 return ide_pci_register_driver(&cmd64x_pci_driver);
436}
437
438static void __exit cmd64x_ide_exit(void)
439{
440 pci_unregister_driver(&cmd64x_pci_driver);
441}
442
443module_init(cmd64x_ide_init);
444module_exit(cmd64x_ide_exit);
445
446MODULE_AUTHOR("Eddie Dost, David Miller, Andre Hedrick, Bartlomiej Zolnierkiewicz");
447MODULE_DESCRIPTION("PCI driver module for CMD64x IDE");
448MODULE_LICENSE("GPL");
1// SPDX-License-Identifier: GPL-2.0-only
2/*
3 * cmd64x.c: Enable interrupts at initialization time on Ultra/PCI machines.
4 * Due to massive hardware bugs, UltraDMA is only supported
5 * on the 646U2 and not on the 646U.
6 *
7 * Copyright (C) 1998 Eddie C. Dost (ecd@skynet.be)
8 * Copyright (C) 1998 David S. Miller (davem@redhat.com)
9 *
10 * Copyright (C) 1999-2002 Andre Hedrick <andre@linux-ide.org>
11 * Copyright (C) 2007-2010 Bartlomiej Zolnierkiewicz
12 * Copyright (C) 2007,2009 MontaVista Software, Inc. <source@mvista.com>
13 */
14
15#include <linux/module.h>
16#include <linux/types.h>
17#include <linux/pci.h>
18#include <linux/ide.h>
19#include <linux/init.h>
20
21#include <asm/io.h>
22
23#define DRV_NAME "cmd64x"
24
25/*
26 * CMD64x specific registers definition.
27 */
28#define CFR 0x50
29#define CFR_INTR_CH0 0x04
30
31#define CMDTIM 0x52
32#define ARTTIM0 0x53
33#define DRWTIM0 0x54
34#define ARTTIM1 0x55
35#define DRWTIM1 0x56
36#define ARTTIM23 0x57
37#define ARTTIM23_DIS_RA2 0x04
38#define ARTTIM23_DIS_RA3 0x08
39#define ARTTIM23_INTR_CH1 0x10
40#define DRWTIM2 0x58
41#define BRST 0x59
42#define DRWTIM3 0x5b
43
44#define BMIDECR0 0x70
45#define MRDMODE 0x71
46#define MRDMODE_INTR_CH0 0x04
47#define MRDMODE_INTR_CH1 0x08
48#define UDIDETCR0 0x73
49#define DTPR0 0x74
50#define BMIDECR1 0x78
51#define BMIDECSR 0x79
52#define UDIDETCR1 0x7B
53#define DTPR1 0x7C
54
55static void cmd64x_program_timings(ide_drive_t *drive, u8 mode)
56{
57 ide_hwif_t *hwif = drive->hwif;
58 struct pci_dev *dev = to_pci_dev(drive->hwif->dev);
59 int bus_speed = ide_pci_clk ? ide_pci_clk : 33;
60 const unsigned long T = 1000000 / bus_speed;
61 static const u8 recovery_values[] =
62 {15, 15, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 0};
63 static const u8 setup_values[] = {0x40, 0x40, 0x40, 0x80, 0, 0xc0};
64 static const u8 arttim_regs[4] = {ARTTIM0, ARTTIM1, ARTTIM23, ARTTIM23};
65 static const u8 drwtim_regs[4] = {DRWTIM0, DRWTIM1, DRWTIM2, DRWTIM3};
66 struct ide_timing t;
67 u8 arttim = 0;
68
69 ide_timing_compute(drive, mode, &t, T, 0);
70
71 /*
72 * In case we've got too long recovery phase, try to lengthen
73 * the active phase
74 */
75 if (t.recover > 16) {
76 t.active += t.recover - 16;
77 t.recover = 16;
78 }
79 if (t.active > 16) /* shouldn't actually happen... */
80 t.active = 16;
81
82 /*
83 * Convert values to internal chipset representation
84 */
85 t.recover = recovery_values[t.recover];
86 t.active &= 0x0f;
87
88 /* Program the active/recovery counts into the DRWTIM register */
89 pci_write_config_byte(dev, drwtim_regs[drive->dn],
90 (t.active << 4) | t.recover);
91
92 /*
93 * The primary channel has individual address setup timing registers
94 * for each drive and the hardware selects the slowest timing itself.
95 * The secondary channel has one common register and we have to select
96 * the slowest address setup timing ourselves.
97 */
98 if (hwif->channel) {
99 ide_drive_t *pair = ide_get_pair_dev(drive);
100
101 if (pair) {
102 struct ide_timing tp;
103
104 ide_timing_compute(pair, pair->pio_mode, &tp, T, 0);
105 ide_timing_merge(&t, &tp, &t, IDE_TIMING_SETUP);
106 if (pair->dma_mode) {
107 ide_timing_compute(pair, pair->dma_mode,
108 &tp, T, 0);
109 ide_timing_merge(&tp, &t, &t, IDE_TIMING_SETUP);
110 }
111 }
112 }
113
114 if (t.setup > 5) /* shouldn't actually happen... */
115 t.setup = 5;
116
117 /*
118 * Program the address setup clocks into the ARTTIM registers.
119 * Avoid clearing the secondary channel's interrupt bit.
120 */
121 (void) pci_read_config_byte (dev, arttim_regs[drive->dn], &arttim);
122 if (hwif->channel)
123 arttim &= ~ARTTIM23_INTR_CH1;
124 arttim &= ~0xc0;
125 arttim |= setup_values[t.setup];
126 (void) pci_write_config_byte(dev, arttim_regs[drive->dn], arttim);
127}
128
129/*
130 * Attempts to set drive's PIO mode.
131 * Special cases are 8: prefetch off, 9: prefetch on (both never worked)
132 */
133
134static void cmd64x_set_pio_mode(ide_hwif_t *hwif, ide_drive_t *drive)
135{
136 const u8 pio = drive->pio_mode - XFER_PIO_0;
137
138 /*
139 * Filter out the prefetch control values
140 * to prevent PIO5 from being programmed
141 */
142 if (pio == 8 || pio == 9)
143 return;
144
145 cmd64x_program_timings(drive, XFER_PIO_0 + pio);
146}
147
148static void cmd64x_set_dma_mode(ide_hwif_t *hwif, ide_drive_t *drive)
149{
150 struct pci_dev *dev = to_pci_dev(hwif->dev);
151 u8 unit = drive->dn & 0x01;
152 u8 regU = 0, pciU = hwif->channel ? UDIDETCR1 : UDIDETCR0;
153 const u8 speed = drive->dma_mode;
154
155 pci_read_config_byte(dev, pciU, ®U);
156 regU &= ~(unit ? 0xCA : 0x35);
157
158 switch(speed) {
159 case XFER_UDMA_5:
160 regU |= unit ? 0x0A : 0x05;
161 break;
162 case XFER_UDMA_4:
163 regU |= unit ? 0x4A : 0x15;
164 break;
165 case XFER_UDMA_3:
166 regU |= unit ? 0x8A : 0x25;
167 break;
168 case XFER_UDMA_2:
169 regU |= unit ? 0x42 : 0x11;
170 break;
171 case XFER_UDMA_1:
172 regU |= unit ? 0x82 : 0x21;
173 break;
174 case XFER_UDMA_0:
175 regU |= unit ? 0xC2 : 0x31;
176 break;
177 case XFER_MW_DMA_2:
178 case XFER_MW_DMA_1:
179 case XFER_MW_DMA_0:
180 cmd64x_program_timings(drive, speed);
181 break;
182 }
183
184 pci_write_config_byte(dev, pciU, regU);
185}
186
187static void cmd648_clear_irq(ide_drive_t *drive)
188{
189 ide_hwif_t *hwif = drive->hwif;
190 struct pci_dev *dev = to_pci_dev(hwif->dev);
191 unsigned long base = pci_resource_start(dev, 4);
192 u8 irq_mask = hwif->channel ? MRDMODE_INTR_CH1 :
193 MRDMODE_INTR_CH0;
194 u8 mrdmode = inb(base + 1);
195
196 /* clear the interrupt bit */
197 outb((mrdmode & ~(MRDMODE_INTR_CH0 | MRDMODE_INTR_CH1)) | irq_mask,
198 base + 1);
199}
200
201static void cmd64x_clear_irq(ide_drive_t *drive)
202{
203 ide_hwif_t *hwif = drive->hwif;
204 struct pci_dev *dev = to_pci_dev(hwif->dev);
205 int irq_reg = hwif->channel ? ARTTIM23 : CFR;
206 u8 irq_mask = hwif->channel ? ARTTIM23_INTR_CH1 :
207 CFR_INTR_CH0;
208 u8 irq_stat = 0;
209
210 (void) pci_read_config_byte(dev, irq_reg, &irq_stat);
211 /* clear the interrupt bit */
212 (void) pci_write_config_byte(dev, irq_reg, irq_stat | irq_mask);
213}
214
215static int cmd648_test_irq(ide_hwif_t *hwif)
216{
217 struct pci_dev *dev = to_pci_dev(hwif->dev);
218 unsigned long base = pci_resource_start(dev, 4);
219 u8 irq_mask = hwif->channel ? MRDMODE_INTR_CH1 :
220 MRDMODE_INTR_CH0;
221 u8 mrdmode = inb(base + 1);
222
223 pr_debug("%s: mrdmode: 0x%02x irq_mask: 0x%02x\n",
224 hwif->name, mrdmode, irq_mask);
225
226 return (mrdmode & irq_mask) ? 1 : 0;
227}
228
229static int cmd64x_test_irq(ide_hwif_t *hwif)
230{
231 struct pci_dev *dev = to_pci_dev(hwif->dev);
232 int irq_reg = hwif->channel ? ARTTIM23 : CFR;
233 u8 irq_mask = hwif->channel ? ARTTIM23_INTR_CH1 :
234 CFR_INTR_CH0;
235 u8 irq_stat = 0;
236
237 (void) pci_read_config_byte(dev, irq_reg, &irq_stat);
238
239 pr_debug("%s: irq_stat: 0x%02x irq_mask: 0x%02x\n",
240 hwif->name, irq_stat, irq_mask);
241
242 return (irq_stat & irq_mask) ? 1 : 0;
243}
244
245/*
246 * ASUS P55T2P4D with CMD646 chipset revision 0x01 requires the old
247 * event order for DMA transfers.
248 */
249
250static int cmd646_1_dma_end(ide_drive_t *drive)
251{
252 ide_hwif_t *hwif = drive->hwif;
253 u8 dma_stat = 0, dma_cmd = 0;
254
255 /* get DMA status */
256 dma_stat = inb(hwif->dma_base + ATA_DMA_STATUS);
257 /* read DMA command state */
258 dma_cmd = inb(hwif->dma_base + ATA_DMA_CMD);
259 /* stop DMA */
260 outb(dma_cmd & ~1, hwif->dma_base + ATA_DMA_CMD);
261 /* clear the INTR & ERROR bits */
262 outb(dma_stat | 6, hwif->dma_base + ATA_DMA_STATUS);
263 /* verify good DMA status */
264 return (dma_stat & 7) != 4;
265}
266
267static int init_chipset_cmd64x(struct pci_dev *dev)
268{
269 u8 mrdmode = 0;
270
271 /* Set a good latency timer and cache line size value. */
272 (void) pci_write_config_byte(dev, PCI_LATENCY_TIMER, 64);
273 /* FIXME: pci_set_master() to ensure a good latency timer value */
274
275 /*
276 * Enable interrupts, select MEMORY READ LINE for reads.
277 *
278 * NOTE: although not mentioned in the PCI0646U specs,
279 * bits 0-1 are write only and won't be read back as
280 * set or not -- PCI0646U2 specs clarify this point.
281 */
282 (void) pci_read_config_byte (dev, MRDMODE, &mrdmode);
283 mrdmode &= ~0x30;
284 (void) pci_write_config_byte(dev, MRDMODE, (mrdmode | 0x02));
285
286 return 0;
287}
288
289static u8 cmd64x_cable_detect(ide_hwif_t *hwif)
290{
291 struct pci_dev *dev = to_pci_dev(hwif->dev);
292 u8 bmidecsr = 0, mask = hwif->channel ? 0x02 : 0x01;
293
294 switch (dev->device) {
295 case PCI_DEVICE_ID_CMD_648:
296 case PCI_DEVICE_ID_CMD_649:
297 pci_read_config_byte(dev, BMIDECSR, &bmidecsr);
298 return (bmidecsr & mask) ? ATA_CBL_PATA80 : ATA_CBL_PATA40;
299 default:
300 return ATA_CBL_PATA40;
301 }
302}
303
304static const struct ide_port_ops cmd64x_port_ops = {
305 .set_pio_mode = cmd64x_set_pio_mode,
306 .set_dma_mode = cmd64x_set_dma_mode,
307 .clear_irq = cmd64x_clear_irq,
308 .test_irq = cmd64x_test_irq,
309 .cable_detect = cmd64x_cable_detect,
310};
311
312static const struct ide_port_ops cmd648_port_ops = {
313 .set_pio_mode = cmd64x_set_pio_mode,
314 .set_dma_mode = cmd64x_set_dma_mode,
315 .clear_irq = cmd648_clear_irq,
316 .test_irq = cmd648_test_irq,
317 .cable_detect = cmd64x_cable_detect,
318};
319
320static const struct ide_dma_ops cmd646_rev1_dma_ops = {
321 .dma_host_set = ide_dma_host_set,
322 .dma_setup = ide_dma_setup,
323 .dma_start = ide_dma_start,
324 .dma_end = cmd646_1_dma_end,
325 .dma_test_irq = ide_dma_test_irq,
326 .dma_lost_irq = ide_dma_lost_irq,
327 .dma_timer_expiry = ide_dma_sff_timer_expiry,
328 .dma_sff_read_status = ide_dma_sff_read_status,
329};
330
331static const struct ide_port_info cmd64x_chipsets[] = {
332 { /* 0: CMD643 */
333 .name = DRV_NAME,
334 .init_chipset = init_chipset_cmd64x,
335 .enablebits = {{0x00,0x00,0x00}, {0x51,0x08,0x08}},
336 .port_ops = &cmd64x_port_ops,
337 .host_flags = IDE_HFLAG_CLEAR_SIMPLEX |
338 IDE_HFLAG_ABUSE_PREFETCH |
339 IDE_HFLAG_SERIALIZE,
340 .pio_mask = ATA_PIO5,
341 .mwdma_mask = ATA_MWDMA2,
342 .udma_mask = 0x00, /* no udma */
343 },
344 { /* 1: CMD646 */
345 .name = DRV_NAME,
346 .init_chipset = init_chipset_cmd64x,
347 .enablebits = {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
348 .port_ops = &cmd648_port_ops,
349 .host_flags = IDE_HFLAG_ABUSE_PREFETCH |
350 IDE_HFLAG_SERIALIZE,
351 .pio_mask = ATA_PIO5,
352 .mwdma_mask = ATA_MWDMA2,
353 .udma_mask = ATA_UDMA2,
354 },
355 { /* 2: CMD648 */
356 .name = DRV_NAME,
357 .init_chipset = init_chipset_cmd64x,
358 .enablebits = {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
359 .port_ops = &cmd648_port_ops,
360 .host_flags = IDE_HFLAG_ABUSE_PREFETCH,
361 .pio_mask = ATA_PIO5,
362 .mwdma_mask = ATA_MWDMA2,
363 .udma_mask = ATA_UDMA4,
364 },
365 { /* 3: CMD649 */
366 .name = DRV_NAME,
367 .init_chipset = init_chipset_cmd64x,
368 .enablebits = {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
369 .port_ops = &cmd648_port_ops,
370 .host_flags = IDE_HFLAG_ABUSE_PREFETCH,
371 .pio_mask = ATA_PIO5,
372 .mwdma_mask = ATA_MWDMA2,
373 .udma_mask = ATA_UDMA5,
374 }
375};
376
377static int cmd64x_init_one(struct pci_dev *dev, const struct pci_device_id *id)
378{
379 struct ide_port_info d;
380 u8 idx = id->driver_data;
381
382 d = cmd64x_chipsets[idx];
383
384 if (idx == 1) {
385 /*
386 * UltraDMA only supported on PCI646U and PCI646U2, which
387 * correspond to revisions 0x03, 0x05 and 0x07 respectively.
388 * Actually, although the CMD tech support people won't
389 * tell me the details, the 0x03 revision cannot support
390 * UDMA correctly without hardware modifications, and even
391 * then it only works with Quantum disks due to some
392 * hold time assumptions in the 646U part which are fixed
393 * in the 646U2.
394 *
395 * So we only do UltraDMA on revision 0x05 and 0x07 chipsets.
396 */
397 if (dev->revision < 5) {
398 d.udma_mask = 0x00;
399 /*
400 * The original PCI0646 didn't have the primary
401 * channel enable bit, it appeared starting with
402 * PCI0646U (i.e. revision ID 3).
403 */
404 if (dev->revision < 3) {
405 d.enablebits[0].reg = 0;
406 d.port_ops = &cmd64x_port_ops;
407 if (dev->revision == 1)
408 d.dma_ops = &cmd646_rev1_dma_ops;
409 }
410 }
411 }
412
413 return ide_pci_init_one(dev, &d, NULL);
414}
415
416static const struct pci_device_id cmd64x_pci_tbl[] = {
417 { PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_643), 0 },
418 { PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_646), 1 },
419 { PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_648), 2 },
420 { PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_649), 3 },
421 { 0, },
422};
423MODULE_DEVICE_TABLE(pci, cmd64x_pci_tbl);
424
425static struct pci_driver cmd64x_pci_driver = {
426 .name = "CMD64x_IDE",
427 .id_table = cmd64x_pci_tbl,
428 .probe = cmd64x_init_one,
429 .remove = ide_pci_remove,
430 .suspend = ide_pci_suspend,
431 .resume = ide_pci_resume,
432};
433
434static int __init cmd64x_ide_init(void)
435{
436 return ide_pci_register_driver(&cmd64x_pci_driver);
437}
438
439static void __exit cmd64x_ide_exit(void)
440{
441 pci_unregister_driver(&cmd64x_pci_driver);
442}
443
444module_init(cmd64x_ide_init);
445module_exit(cmd64x_ide_exit);
446
447MODULE_AUTHOR("Eddie Dost, David Miller, Andre Hedrick, Bartlomiej Zolnierkiewicz");
448MODULE_DESCRIPTION("PCI driver module for CMD64x IDE");
449MODULE_LICENSE("GPL");