Linux Audio

Check our new training course

Loading...
Note: File does not exist in v6.8.
  1// SPDX-License-Identifier: GPL-2.0-only
  2/*
  3 * cmd64x.c: Enable interrupts at initialization time on Ultra/PCI machines.
  4 *           Due to massive hardware bugs, UltraDMA is only supported
  5 *           on the 646U2 and not on the 646U.
  6 *
  7 * Copyright (C) 1998		Eddie C. Dost  (ecd@skynet.be)
  8 * Copyright (C) 1998		David S. Miller (davem@redhat.com)
  9 *
 10 * Copyright (C) 1999-2002	Andre Hedrick <andre@linux-ide.org>
 11 * Copyright (C) 2007-2010	Bartlomiej Zolnierkiewicz
 12 * Copyright (C) 2007,2009	MontaVista Software, Inc. <source@mvista.com>
 13 */
 14
 15#include <linux/module.h>
 16#include <linux/types.h>
 17#include <linux/pci.h>
 18#include <linux/ide.h>
 19#include <linux/init.h>
 20
 21#include <asm/io.h>
 22
 23#define DRV_NAME "cmd64x"
 24
 25/*
 26 * CMD64x specific registers definition.
 27 */
 28#define CFR		0x50
 29#define   CFR_INTR_CH0		0x04
 30
 31#define	CMDTIM		0x52
 32#define	ARTTIM0		0x53
 33#define	DRWTIM0		0x54
 34#define ARTTIM1 	0x55
 35#define DRWTIM1		0x56
 36#define ARTTIM23	0x57
 37#define   ARTTIM23_DIS_RA2	0x04
 38#define   ARTTIM23_DIS_RA3	0x08
 39#define   ARTTIM23_INTR_CH1	0x10
 40#define DRWTIM2		0x58
 41#define BRST		0x59
 42#define DRWTIM3		0x5b
 43
 44#define BMIDECR0	0x70
 45#define MRDMODE		0x71
 46#define   MRDMODE_INTR_CH0	0x04
 47#define   MRDMODE_INTR_CH1	0x08
 48#define UDIDETCR0	0x73
 49#define DTPR0		0x74
 50#define BMIDECR1	0x78
 51#define BMIDECSR	0x79
 52#define UDIDETCR1	0x7B
 53#define DTPR1		0x7C
 54
 55static void cmd64x_program_timings(ide_drive_t *drive, u8 mode)
 56{
 57	ide_hwif_t *hwif = drive->hwif;
 58	struct pci_dev *dev = to_pci_dev(drive->hwif->dev);
 59	int bus_speed = ide_pci_clk ? ide_pci_clk : 33;
 60	const unsigned long T = 1000000 / bus_speed;
 61	static const u8 recovery_values[] =
 62		{15, 15, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 0};
 63	static const u8 setup_values[] = {0x40, 0x40, 0x40, 0x80, 0, 0xc0};
 64	static const u8 arttim_regs[4] = {ARTTIM0, ARTTIM1, ARTTIM23, ARTTIM23};
 65	static const u8 drwtim_regs[4] = {DRWTIM0, DRWTIM1, DRWTIM2, DRWTIM3};
 66	struct ide_timing t;
 67	u8 arttim = 0;
 68
 69	ide_timing_compute(drive, mode, &t, T, 0);
 70
 71	/*
 72	 * In case we've got too long recovery phase, try to lengthen
 73	 * the active phase
 74	 */
 75	if (t.recover > 16) {
 76		t.active += t.recover - 16;
 77		t.recover = 16;
 78	}
 79	if (t.active > 16)		/* shouldn't actually happen... */
 80		t.active = 16;
 81
 82	/*
 83	 * Convert values to internal chipset representation
 84	 */
 85	t.recover = recovery_values[t.recover];
 86	t.active &= 0x0f;
 87
 88	/* Program the active/recovery counts into the DRWTIM register */
 89	pci_write_config_byte(dev, drwtim_regs[drive->dn],
 90			      (t.active << 4) | t.recover);
 91
 92	/*
 93	 * The primary channel has individual address setup timing registers
 94	 * for each drive and the hardware selects the slowest timing itself.
 95	 * The secondary channel has one common register and we have to select
 96	 * the slowest address setup timing ourselves.
 97	 */
 98	if (hwif->channel) {
 99		ide_drive_t *pair = ide_get_pair_dev(drive);
100
101		if (pair) {
102			struct ide_timing tp;
103
104			ide_timing_compute(pair, pair->pio_mode, &tp, T, 0);
105			ide_timing_merge(&t, &tp, &t, IDE_TIMING_SETUP);
106			if (pair->dma_mode) {
107				ide_timing_compute(pair, pair->dma_mode,
108						&tp, T, 0);
109				ide_timing_merge(&tp, &t, &t, IDE_TIMING_SETUP);
110			}
111		}
112	}
113
114	if (t.setup > 5)		/* shouldn't actually happen... */
115		t.setup = 5;
116
117	/*
118	 * Program the address setup clocks into the ARTTIM registers.
119	 * Avoid clearing the secondary channel's interrupt bit.
120	 */
121	(void) pci_read_config_byte (dev, arttim_regs[drive->dn], &arttim);
122	if (hwif->channel)
123		arttim &= ~ARTTIM23_INTR_CH1;
124	arttim &= ~0xc0;
125	arttim |= setup_values[t.setup];
126	(void) pci_write_config_byte(dev, arttim_regs[drive->dn], arttim);
127}
128
129/*
130 * Attempts to set drive's PIO mode.
131 * Special cases are 8: prefetch off, 9: prefetch on (both never worked)
132 */
133
134static void cmd64x_set_pio_mode(ide_hwif_t *hwif, ide_drive_t *drive)
135{
136	const u8 pio = drive->pio_mode - XFER_PIO_0;
137
138	/*
139	 * Filter out the prefetch control values
140	 * to prevent PIO5 from being programmed
141	 */
142	if (pio == 8 || pio == 9)
143		return;
144
145	cmd64x_program_timings(drive, XFER_PIO_0 + pio);
146}
147
148static void cmd64x_set_dma_mode(ide_hwif_t *hwif, ide_drive_t *drive)
149{
150	struct pci_dev *dev	= to_pci_dev(hwif->dev);
151	u8 unit			= drive->dn & 0x01;
152	u8 regU = 0, pciU	= hwif->channel ? UDIDETCR1 : UDIDETCR0;
153	const u8 speed		= drive->dma_mode;
154
155	pci_read_config_byte(dev, pciU, &regU);
156	regU &= ~(unit ? 0xCA : 0x35);
157
158	switch(speed) {
159	case XFER_UDMA_5:
160		regU |= unit ? 0x0A : 0x05;
161		break;
162	case XFER_UDMA_4:
163		regU |= unit ? 0x4A : 0x15;
164		break;
165	case XFER_UDMA_3:
166		regU |= unit ? 0x8A : 0x25;
167		break;
168	case XFER_UDMA_2:
169		regU |= unit ? 0x42 : 0x11;
170		break;
171	case XFER_UDMA_1:
172		regU |= unit ? 0x82 : 0x21;
173		break;
174	case XFER_UDMA_0:
175		regU |= unit ? 0xC2 : 0x31;
176		break;
177	case XFER_MW_DMA_2:
178	case XFER_MW_DMA_1:
179	case XFER_MW_DMA_0:
180		cmd64x_program_timings(drive, speed);
181		break;
182	}
183
184	pci_write_config_byte(dev, pciU, regU);
185}
186
187static void cmd648_clear_irq(ide_drive_t *drive)
188{
189	ide_hwif_t *hwif	= drive->hwif;
190	struct pci_dev *dev	= to_pci_dev(hwif->dev);
191	unsigned long base	= pci_resource_start(dev, 4);
192	u8  irq_mask		= hwif->channel ? MRDMODE_INTR_CH1 :
193						  MRDMODE_INTR_CH0;
194	u8  mrdmode		= inb(base + 1);
195
196	/* clear the interrupt bit */
197	outb((mrdmode & ~(MRDMODE_INTR_CH0 | MRDMODE_INTR_CH1)) | irq_mask,
198	     base + 1);
199}
200
201static void cmd64x_clear_irq(ide_drive_t *drive)
202{
203	ide_hwif_t *hwif	= drive->hwif;
204	struct pci_dev *dev	= to_pci_dev(hwif->dev);
205	int irq_reg		= hwif->channel ? ARTTIM23 : CFR;
206	u8  irq_mask		= hwif->channel ? ARTTIM23_INTR_CH1 :
207						  CFR_INTR_CH0;
208	u8  irq_stat		= 0;
209
210	(void) pci_read_config_byte(dev, irq_reg, &irq_stat);
211	/* clear the interrupt bit */
212	(void) pci_write_config_byte(dev, irq_reg, irq_stat | irq_mask);
213}
214
215static int cmd648_test_irq(ide_hwif_t *hwif)
216{
217	struct pci_dev *dev	= to_pci_dev(hwif->dev);
218	unsigned long base	= pci_resource_start(dev, 4);
219	u8 irq_mask		= hwif->channel ? MRDMODE_INTR_CH1 :
220						  MRDMODE_INTR_CH0;
221	u8 mrdmode		= inb(base + 1);
222
223	pr_debug("%s: mrdmode: 0x%02x irq_mask: 0x%02x\n",
224		 hwif->name, mrdmode, irq_mask);
225
226	return (mrdmode & irq_mask) ? 1 : 0;
227}
228
229static int cmd64x_test_irq(ide_hwif_t *hwif)
230{
231	struct pci_dev *dev	= to_pci_dev(hwif->dev);
232	int irq_reg		= hwif->channel ? ARTTIM23 : CFR;
233	u8  irq_mask		= hwif->channel ? ARTTIM23_INTR_CH1 :
234						  CFR_INTR_CH0;
235	u8  irq_stat		= 0;
236
237	(void) pci_read_config_byte(dev, irq_reg, &irq_stat);
238
239	pr_debug("%s: irq_stat: 0x%02x irq_mask: 0x%02x\n",
240		 hwif->name, irq_stat, irq_mask);
241
242	return (irq_stat & irq_mask) ? 1 : 0;
243}
244
245/*
246 * ASUS P55T2P4D with CMD646 chipset revision 0x01 requires the old
247 * event order for DMA transfers.
248 */
249
250static int cmd646_1_dma_end(ide_drive_t *drive)
251{
252	ide_hwif_t *hwif = drive->hwif;
253	u8 dma_stat = 0, dma_cmd = 0;
254
255	/* get DMA status */
256	dma_stat = inb(hwif->dma_base + ATA_DMA_STATUS);
257	/* read DMA command state */
258	dma_cmd = inb(hwif->dma_base + ATA_DMA_CMD);
259	/* stop DMA */
260	outb(dma_cmd & ~1, hwif->dma_base + ATA_DMA_CMD);
261	/* clear the INTR & ERROR bits */
262	outb(dma_stat | 6, hwif->dma_base + ATA_DMA_STATUS);
263	/* verify good DMA status */
264	return (dma_stat & 7) != 4;
265}
266
267static int init_chipset_cmd64x(struct pci_dev *dev)
268{
269	u8 mrdmode = 0;
270
271	/* Set a good latency timer and cache line size value. */
272	(void) pci_write_config_byte(dev, PCI_LATENCY_TIMER, 64);
273	/* FIXME: pci_set_master() to ensure a good latency timer value */
274
275	/*
276	 * Enable interrupts, select MEMORY READ LINE for reads.
277	 *
278	 * NOTE: although not mentioned in the PCI0646U specs,
279	 * bits 0-1 are write only and won't be read back as
280	 * set or not -- PCI0646U2 specs clarify this point.
281	 */
282	(void) pci_read_config_byte (dev, MRDMODE, &mrdmode);
283	mrdmode &= ~0x30;
284	(void) pci_write_config_byte(dev, MRDMODE, (mrdmode | 0x02));
285
286	return 0;
287}
288
289static u8 cmd64x_cable_detect(ide_hwif_t *hwif)
290{
291	struct pci_dev  *dev	= to_pci_dev(hwif->dev);
292	u8 bmidecsr = 0, mask	= hwif->channel ? 0x02 : 0x01;
293
294	switch (dev->device) {
295	case PCI_DEVICE_ID_CMD_648:
296	case PCI_DEVICE_ID_CMD_649:
297 		pci_read_config_byte(dev, BMIDECSR, &bmidecsr);
298		return (bmidecsr & mask) ? ATA_CBL_PATA80 : ATA_CBL_PATA40;
299	default:
300		return ATA_CBL_PATA40;
301	}
302}
303
304static const struct ide_port_ops cmd64x_port_ops = {
305	.set_pio_mode		= cmd64x_set_pio_mode,
306	.set_dma_mode		= cmd64x_set_dma_mode,
307	.clear_irq		= cmd64x_clear_irq,
308	.test_irq		= cmd64x_test_irq,
309	.cable_detect		= cmd64x_cable_detect,
310};
311
312static const struct ide_port_ops cmd648_port_ops = {
313	.set_pio_mode		= cmd64x_set_pio_mode,
314	.set_dma_mode		= cmd64x_set_dma_mode,
315	.clear_irq		= cmd648_clear_irq,
316	.test_irq		= cmd648_test_irq,
317	.cable_detect		= cmd64x_cable_detect,
318};
319
320static const struct ide_dma_ops cmd646_rev1_dma_ops = {
321	.dma_host_set		= ide_dma_host_set,
322	.dma_setup		= ide_dma_setup,
323	.dma_start		= ide_dma_start,
324	.dma_end		= cmd646_1_dma_end,
325	.dma_test_irq		= ide_dma_test_irq,
326	.dma_lost_irq		= ide_dma_lost_irq,
327	.dma_timer_expiry	= ide_dma_sff_timer_expiry,
328	.dma_sff_read_status	= ide_dma_sff_read_status,
329};
330
331static const struct ide_port_info cmd64x_chipsets[] = {
332	{	/* 0: CMD643 */
333		.name		= DRV_NAME,
334		.init_chipset	= init_chipset_cmd64x,
335		.enablebits	= {{0x00,0x00,0x00}, {0x51,0x08,0x08}},
336		.port_ops	= &cmd64x_port_ops,
337		.host_flags	= IDE_HFLAG_CLEAR_SIMPLEX |
338				  IDE_HFLAG_ABUSE_PREFETCH |
339				  IDE_HFLAG_SERIALIZE,
340		.pio_mask	= ATA_PIO5,
341		.mwdma_mask	= ATA_MWDMA2,
342		.udma_mask	= 0x00, /* no udma */
343	},
344	{	/* 1: CMD646 */
345		.name		= DRV_NAME,
346		.init_chipset	= init_chipset_cmd64x,
347		.enablebits	= {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
348		.port_ops	= &cmd648_port_ops,
349		.host_flags	= IDE_HFLAG_ABUSE_PREFETCH |
350				  IDE_HFLAG_SERIALIZE,
351		.pio_mask	= ATA_PIO5,
352		.mwdma_mask	= ATA_MWDMA2,
353		.udma_mask	= ATA_UDMA2,
354	},
355	{	/* 2: CMD648 */
356		.name		= DRV_NAME,
357		.init_chipset	= init_chipset_cmd64x,
358		.enablebits	= {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
359		.port_ops	= &cmd648_port_ops,
360		.host_flags	= IDE_HFLAG_ABUSE_PREFETCH,
361		.pio_mask	= ATA_PIO5,
362		.mwdma_mask	= ATA_MWDMA2,
363		.udma_mask	= ATA_UDMA4,
364	},
365	{	/* 3: CMD649 */
366		.name		= DRV_NAME,
367		.init_chipset	= init_chipset_cmd64x,
368		.enablebits	= {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
369		.port_ops	= &cmd648_port_ops,
370		.host_flags	= IDE_HFLAG_ABUSE_PREFETCH,
371		.pio_mask	= ATA_PIO5,
372		.mwdma_mask	= ATA_MWDMA2,
373		.udma_mask	= ATA_UDMA5,
374	}
375};
376
377static int cmd64x_init_one(struct pci_dev *dev, const struct pci_device_id *id)
378{
379	struct ide_port_info d;
380	u8 idx = id->driver_data;
381
382	d = cmd64x_chipsets[idx];
383
384	if (idx == 1) {
385		/*
386		 * UltraDMA only supported on PCI646U and PCI646U2, which
387		 * correspond to revisions 0x03, 0x05 and 0x07 respectively.
388		 * Actually, although the CMD tech support people won't
389		 * tell me the details, the 0x03 revision cannot support
390		 * UDMA correctly without hardware modifications, and even
391		 * then it only works with Quantum disks due to some
392		 * hold time assumptions in the 646U part which are fixed
393		 * in the 646U2.
394		 *
395		 * So we only do UltraDMA on revision 0x05 and 0x07 chipsets.
396		 */
397		if (dev->revision < 5) {
398			d.udma_mask = 0x00;
399			/*
400			 * The original PCI0646 didn't have the primary
401			 * channel enable bit, it appeared starting with
402			 * PCI0646U (i.e. revision ID 3).
403			 */
404			if (dev->revision < 3) {
405				d.enablebits[0].reg = 0;
406				d.port_ops = &cmd64x_port_ops;
407				if (dev->revision == 1)
408					d.dma_ops = &cmd646_rev1_dma_ops;
409			}
410		}
411	}
412
413	return ide_pci_init_one(dev, &d, NULL);
414}
415
416static const struct pci_device_id cmd64x_pci_tbl[] = {
417	{ PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_643), 0 },
418	{ PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_646), 1 },
419	{ PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_648), 2 },
420	{ PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_649), 3 },
421	{ 0, },
422};
423MODULE_DEVICE_TABLE(pci, cmd64x_pci_tbl);
424
425static struct pci_driver cmd64x_pci_driver = {
426	.name		= "CMD64x_IDE",
427	.id_table	= cmd64x_pci_tbl,
428	.probe		= cmd64x_init_one,
429	.remove		= ide_pci_remove,
430	.suspend	= ide_pci_suspend,
431	.resume		= ide_pci_resume,
432};
433
434static int __init cmd64x_ide_init(void)
435{
436	return ide_pci_register_driver(&cmd64x_pci_driver);
437}
438
439static void __exit cmd64x_ide_exit(void)
440{
441	pci_unregister_driver(&cmd64x_pci_driver);
442}
443
444module_init(cmd64x_ide_init);
445module_exit(cmd64x_ide_exit);
446
447MODULE_AUTHOR("Eddie Dost, David Miller, Andre Hedrick, Bartlomiej Zolnierkiewicz");
448MODULE_DESCRIPTION("PCI driver module for CMD64x IDE");
449MODULE_LICENSE("GPL");