• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * cmd64x.c: Enable interrupts at initialization time on Ultra/PCI machines.
4  *           Due to massive hardware bugs, UltraDMA is only supported
5  *           on the 646U2 and not on the 646U.
6  *
7  * Copyright (C) 1998		Eddie C. Dost  (ecd@skynet.be)
8  * Copyright (C) 1998		David S. Miller (davem@redhat.com)
9  *
10  * Copyright (C) 1999-2002	Andre Hedrick <andre@linux-ide.org>
11  * Copyright (C) 2007-2010	Bartlomiej Zolnierkiewicz
12  * Copyright (C) 2007,2009	MontaVista Software, Inc. <source@mvista.com>
13  */
14 
15 #include <linux/module.h>
16 #include <linux/types.h>
17 #include <linux/pci.h>
18 #include <linux/ide.h>
19 #include <linux/init.h>
20 
21 #include <asm/io.h>
22 
23 #define DRV_NAME "cmd64x"
24 
25 /*
26  * CMD64x specific registers definition.
27  */
28 #define CFR		0x50
29 #define   CFR_INTR_CH0		0x04
30 
31 #define	CMDTIM		0x52
32 #define	ARTTIM0		0x53
33 #define	DRWTIM0		0x54
34 #define ARTTIM1 	0x55
35 #define DRWTIM1		0x56
36 #define ARTTIM23	0x57
37 #define   ARTTIM23_DIS_RA2	0x04
38 #define   ARTTIM23_DIS_RA3	0x08
39 #define   ARTTIM23_INTR_CH1	0x10
40 #define DRWTIM2		0x58
41 #define BRST		0x59
42 #define DRWTIM3		0x5b
43 
44 #define BMIDECR0	0x70
45 #define MRDMODE		0x71
46 #define   MRDMODE_INTR_CH0	0x04
47 #define   MRDMODE_INTR_CH1	0x08
48 #define UDIDETCR0	0x73
49 #define DTPR0		0x74
50 #define BMIDECR1	0x78
51 #define BMIDECSR	0x79
52 #define UDIDETCR1	0x7B
53 #define DTPR1		0x7C
54 
cmd64x_program_timings(ide_drive_t * drive,u8 mode)55 static void cmd64x_program_timings(ide_drive_t *drive, u8 mode)
56 {
57 	ide_hwif_t *hwif = drive->hwif;
58 	struct pci_dev *dev = to_pci_dev(drive->hwif->dev);
59 	int bus_speed = ide_pci_clk ? ide_pci_clk : 33;
60 	const unsigned long T = 1000000 / bus_speed;
61 	static const u8 recovery_values[] =
62 		{15, 15, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 0};
63 	static const u8 setup_values[] = {0x40, 0x40, 0x40, 0x80, 0, 0xc0};
64 	static const u8 arttim_regs[4] = {ARTTIM0, ARTTIM1, ARTTIM23, ARTTIM23};
65 	static const u8 drwtim_regs[4] = {DRWTIM0, DRWTIM1, DRWTIM2, DRWTIM3};
66 	struct ide_timing t;
67 	u8 arttim = 0;
68 
69 	ide_timing_compute(drive, mode, &t, T, 0);
70 
71 	/*
72 	 * In case we've got too long recovery phase, try to lengthen
73 	 * the active phase
74 	 */
75 	if (t.recover > 16) {
76 		t.active += t.recover - 16;
77 		t.recover = 16;
78 	}
79 	if (t.active > 16)		/* shouldn't actually happen... */
80 		t.active = 16;
81 
82 	/*
83 	 * Convert values to internal chipset representation
84 	 */
85 	t.recover = recovery_values[t.recover];
86 	t.active &= 0x0f;
87 
88 	/* Program the active/recovery counts into the DRWTIM register */
89 	pci_write_config_byte(dev, drwtim_regs[drive->dn],
90 			      (t.active << 4) | t.recover);
91 
92 	/*
93 	 * The primary channel has individual address setup timing registers
94 	 * for each drive and the hardware selects the slowest timing itself.
95 	 * The secondary channel has one common register and we have to select
96 	 * the slowest address setup timing ourselves.
97 	 */
98 	if (hwif->channel) {
99 		ide_drive_t *pair = ide_get_pair_dev(drive);
100 
101 		if (pair) {
102 			struct ide_timing tp;
103 
104 			ide_timing_compute(pair, pair->pio_mode, &tp, T, 0);
105 			ide_timing_merge(&t, &tp, &t, IDE_TIMING_SETUP);
106 			if (pair->dma_mode) {
107 				ide_timing_compute(pair, pair->dma_mode,
108 						&tp, T, 0);
109 				ide_timing_merge(&tp, &t, &t, IDE_TIMING_SETUP);
110 			}
111 		}
112 	}
113 
114 	if (t.setup > 5)		/* shouldn't actually happen... */
115 		t.setup = 5;
116 
117 	/*
118 	 * Program the address setup clocks into the ARTTIM registers.
119 	 * Avoid clearing the secondary channel's interrupt bit.
120 	 */
121 	(void) pci_read_config_byte (dev, arttim_regs[drive->dn], &arttim);
122 	if (hwif->channel)
123 		arttim &= ~ARTTIM23_INTR_CH1;
124 	arttim &= ~0xc0;
125 	arttim |= setup_values[t.setup];
126 	(void) pci_write_config_byte(dev, arttim_regs[drive->dn], arttim);
127 }
128 
129 /*
130  * Attempts to set drive's PIO mode.
131  * Special cases are 8: prefetch off, 9: prefetch on (both never worked)
132  */
133 
cmd64x_set_pio_mode(ide_hwif_t * hwif,ide_drive_t * drive)134 static void cmd64x_set_pio_mode(ide_hwif_t *hwif, ide_drive_t *drive)
135 {
136 	const u8 pio = drive->pio_mode - XFER_PIO_0;
137 
138 	/*
139 	 * Filter out the prefetch control values
140 	 * to prevent PIO5 from being programmed
141 	 */
142 	if (pio == 8 || pio == 9)
143 		return;
144 
145 	cmd64x_program_timings(drive, XFER_PIO_0 + pio);
146 }
147 
cmd64x_set_dma_mode(ide_hwif_t * hwif,ide_drive_t * drive)148 static void cmd64x_set_dma_mode(ide_hwif_t *hwif, ide_drive_t *drive)
149 {
150 	struct pci_dev *dev	= to_pci_dev(hwif->dev);
151 	u8 unit			= drive->dn & 0x01;
152 	u8 regU = 0, pciU	= hwif->channel ? UDIDETCR1 : UDIDETCR0;
153 	const u8 speed		= drive->dma_mode;
154 
155 	pci_read_config_byte(dev, pciU, &regU);
156 	regU &= ~(unit ? 0xCA : 0x35);
157 
158 	switch(speed) {
159 	case XFER_UDMA_5:
160 		regU |= unit ? 0x0A : 0x05;
161 		break;
162 	case XFER_UDMA_4:
163 		regU |= unit ? 0x4A : 0x15;
164 		break;
165 	case XFER_UDMA_3:
166 		regU |= unit ? 0x8A : 0x25;
167 		break;
168 	case XFER_UDMA_2:
169 		regU |= unit ? 0x42 : 0x11;
170 		break;
171 	case XFER_UDMA_1:
172 		regU |= unit ? 0x82 : 0x21;
173 		break;
174 	case XFER_UDMA_0:
175 		regU |= unit ? 0xC2 : 0x31;
176 		break;
177 	case XFER_MW_DMA_2:
178 	case XFER_MW_DMA_1:
179 	case XFER_MW_DMA_0:
180 		cmd64x_program_timings(drive, speed);
181 		break;
182 	}
183 
184 	pci_write_config_byte(dev, pciU, regU);
185 }
186 
cmd648_clear_irq(ide_drive_t * drive)187 static void cmd648_clear_irq(ide_drive_t *drive)
188 {
189 	ide_hwif_t *hwif	= drive->hwif;
190 	struct pci_dev *dev	= to_pci_dev(hwif->dev);
191 	unsigned long base	= pci_resource_start(dev, 4);
192 	u8  irq_mask		= hwif->channel ? MRDMODE_INTR_CH1 :
193 						  MRDMODE_INTR_CH0;
194 	u8  mrdmode		= inb(base + 1);
195 
196 	/* clear the interrupt bit */
197 	outb((mrdmode & ~(MRDMODE_INTR_CH0 | MRDMODE_INTR_CH1)) | irq_mask,
198 	     base + 1);
199 }
200 
cmd64x_clear_irq(ide_drive_t * drive)201 static void cmd64x_clear_irq(ide_drive_t *drive)
202 {
203 	ide_hwif_t *hwif	= drive->hwif;
204 	struct pci_dev *dev	= to_pci_dev(hwif->dev);
205 	int irq_reg		= hwif->channel ? ARTTIM23 : CFR;
206 	u8  irq_mask		= hwif->channel ? ARTTIM23_INTR_CH1 :
207 						  CFR_INTR_CH0;
208 	u8  irq_stat		= 0;
209 
210 	(void) pci_read_config_byte(dev, irq_reg, &irq_stat);
211 	/* clear the interrupt bit */
212 	(void) pci_write_config_byte(dev, irq_reg, irq_stat | irq_mask);
213 }
214 
cmd648_test_irq(ide_hwif_t * hwif)215 static int cmd648_test_irq(ide_hwif_t *hwif)
216 {
217 	struct pci_dev *dev	= to_pci_dev(hwif->dev);
218 	unsigned long base	= pci_resource_start(dev, 4);
219 	u8 irq_mask		= hwif->channel ? MRDMODE_INTR_CH1 :
220 						  MRDMODE_INTR_CH0;
221 	u8 mrdmode		= inb(base + 1);
222 
223 	pr_debug("%s: mrdmode: 0x%02x irq_mask: 0x%02x\n",
224 		 hwif->name, mrdmode, irq_mask);
225 
226 	return (mrdmode & irq_mask) ? 1 : 0;
227 }
228 
cmd64x_test_irq(ide_hwif_t * hwif)229 static int cmd64x_test_irq(ide_hwif_t *hwif)
230 {
231 	struct pci_dev *dev	= to_pci_dev(hwif->dev);
232 	int irq_reg		= hwif->channel ? ARTTIM23 : CFR;
233 	u8  irq_mask		= hwif->channel ? ARTTIM23_INTR_CH1 :
234 						  CFR_INTR_CH0;
235 	u8  irq_stat		= 0;
236 
237 	(void) pci_read_config_byte(dev, irq_reg, &irq_stat);
238 
239 	pr_debug("%s: irq_stat: 0x%02x irq_mask: 0x%02x\n",
240 		 hwif->name, irq_stat, irq_mask);
241 
242 	return (irq_stat & irq_mask) ? 1 : 0;
243 }
244 
245 /*
246  * ASUS P55T2P4D with CMD646 chipset revision 0x01 requires the old
247  * event order for DMA transfers.
248  */
249 
cmd646_1_dma_end(ide_drive_t * drive)250 static int cmd646_1_dma_end(ide_drive_t *drive)
251 {
252 	ide_hwif_t *hwif = drive->hwif;
253 	u8 dma_stat = 0, dma_cmd = 0;
254 
255 	/* get DMA status */
256 	dma_stat = inb(hwif->dma_base + ATA_DMA_STATUS);
257 	/* read DMA command state */
258 	dma_cmd = inb(hwif->dma_base + ATA_DMA_CMD);
259 	/* stop DMA */
260 	outb(dma_cmd & ~1, hwif->dma_base + ATA_DMA_CMD);
261 	/* clear the INTR & ERROR bits */
262 	outb(dma_stat | 6, hwif->dma_base + ATA_DMA_STATUS);
263 	/* verify good DMA status */
264 	return (dma_stat & 7) != 4;
265 }
266 
init_chipset_cmd64x(struct pci_dev * dev)267 static int init_chipset_cmd64x(struct pci_dev *dev)
268 {
269 	u8 mrdmode = 0;
270 
271 	/* Set a good latency timer and cache line size value. */
272 	(void) pci_write_config_byte(dev, PCI_LATENCY_TIMER, 64);
273 	/* FIXME: pci_set_master() to ensure a good latency timer value */
274 
275 	/*
276 	 * Enable interrupts, select MEMORY READ LINE for reads.
277 	 *
278 	 * NOTE: although not mentioned in the PCI0646U specs,
279 	 * bits 0-1 are write only and won't be read back as
280 	 * set or not -- PCI0646U2 specs clarify this point.
281 	 */
282 	(void) pci_read_config_byte (dev, MRDMODE, &mrdmode);
283 	mrdmode &= ~0x30;
284 	(void) pci_write_config_byte(dev, MRDMODE, (mrdmode | 0x02));
285 
286 	return 0;
287 }
288 
cmd64x_cable_detect(ide_hwif_t * hwif)289 static u8 cmd64x_cable_detect(ide_hwif_t *hwif)
290 {
291 	struct pci_dev  *dev	= to_pci_dev(hwif->dev);
292 	u8 bmidecsr = 0, mask	= hwif->channel ? 0x02 : 0x01;
293 
294 	switch (dev->device) {
295 	case PCI_DEVICE_ID_CMD_648:
296 	case PCI_DEVICE_ID_CMD_649:
297  		pci_read_config_byte(dev, BMIDECSR, &bmidecsr);
298 		return (bmidecsr & mask) ? ATA_CBL_PATA80 : ATA_CBL_PATA40;
299 	default:
300 		return ATA_CBL_PATA40;
301 	}
302 }
303 
304 static const struct ide_port_ops cmd64x_port_ops = {
305 	.set_pio_mode		= cmd64x_set_pio_mode,
306 	.set_dma_mode		= cmd64x_set_dma_mode,
307 	.clear_irq		= cmd64x_clear_irq,
308 	.test_irq		= cmd64x_test_irq,
309 	.cable_detect		= cmd64x_cable_detect,
310 };
311 
312 static const struct ide_port_ops cmd648_port_ops = {
313 	.set_pio_mode		= cmd64x_set_pio_mode,
314 	.set_dma_mode		= cmd64x_set_dma_mode,
315 	.clear_irq		= cmd648_clear_irq,
316 	.test_irq		= cmd648_test_irq,
317 	.cable_detect		= cmd64x_cable_detect,
318 };
319 
320 static const struct ide_dma_ops cmd646_rev1_dma_ops = {
321 	.dma_host_set		= ide_dma_host_set,
322 	.dma_setup		= ide_dma_setup,
323 	.dma_start		= ide_dma_start,
324 	.dma_end		= cmd646_1_dma_end,
325 	.dma_test_irq		= ide_dma_test_irq,
326 	.dma_lost_irq		= ide_dma_lost_irq,
327 	.dma_timer_expiry	= ide_dma_sff_timer_expiry,
328 	.dma_sff_read_status	= ide_dma_sff_read_status,
329 };
330 
331 static const struct ide_port_info cmd64x_chipsets[] = {
332 	{	/* 0: CMD643 */
333 		.name		= DRV_NAME,
334 		.init_chipset	= init_chipset_cmd64x,
335 		.enablebits	= {{0x00,0x00,0x00}, {0x51,0x08,0x08}},
336 		.port_ops	= &cmd64x_port_ops,
337 		.host_flags	= IDE_HFLAG_CLEAR_SIMPLEX |
338 				  IDE_HFLAG_ABUSE_PREFETCH |
339 				  IDE_HFLAG_SERIALIZE,
340 		.pio_mask	= ATA_PIO5,
341 		.mwdma_mask	= ATA_MWDMA2,
342 		.udma_mask	= 0x00, /* no udma */
343 	},
344 	{	/* 1: CMD646 */
345 		.name		= DRV_NAME,
346 		.init_chipset	= init_chipset_cmd64x,
347 		.enablebits	= {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
348 		.port_ops	= &cmd648_port_ops,
349 		.host_flags	= IDE_HFLAG_ABUSE_PREFETCH |
350 				  IDE_HFLAG_SERIALIZE,
351 		.pio_mask	= ATA_PIO5,
352 		.mwdma_mask	= ATA_MWDMA2,
353 		.udma_mask	= ATA_UDMA2,
354 	},
355 	{	/* 2: CMD648 */
356 		.name		= DRV_NAME,
357 		.init_chipset	= init_chipset_cmd64x,
358 		.enablebits	= {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
359 		.port_ops	= &cmd648_port_ops,
360 		.host_flags	= IDE_HFLAG_ABUSE_PREFETCH,
361 		.pio_mask	= ATA_PIO5,
362 		.mwdma_mask	= ATA_MWDMA2,
363 		.udma_mask	= ATA_UDMA4,
364 	},
365 	{	/* 3: CMD649 */
366 		.name		= DRV_NAME,
367 		.init_chipset	= init_chipset_cmd64x,
368 		.enablebits	= {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
369 		.port_ops	= &cmd648_port_ops,
370 		.host_flags	= IDE_HFLAG_ABUSE_PREFETCH,
371 		.pio_mask	= ATA_PIO5,
372 		.mwdma_mask	= ATA_MWDMA2,
373 		.udma_mask	= ATA_UDMA5,
374 	}
375 };
376 
cmd64x_init_one(struct pci_dev * dev,const struct pci_device_id * id)377 static int cmd64x_init_one(struct pci_dev *dev, const struct pci_device_id *id)
378 {
379 	struct ide_port_info d;
380 	u8 idx = id->driver_data;
381 
382 	d = cmd64x_chipsets[idx];
383 
384 	if (idx == 1) {
385 		/*
386 		 * UltraDMA only supported on PCI646U and PCI646U2, which
387 		 * correspond to revisions 0x03, 0x05 and 0x07 respectively.
388 		 * Actually, although the CMD tech support people won't
389 		 * tell me the details, the 0x03 revision cannot support
390 		 * UDMA correctly without hardware modifications, and even
391 		 * then it only works with Quantum disks due to some
392 		 * hold time assumptions in the 646U part which are fixed
393 		 * in the 646U2.
394 		 *
395 		 * So we only do UltraDMA on revision 0x05 and 0x07 chipsets.
396 		 */
397 		if (dev->revision < 5) {
398 			d.udma_mask = 0x00;
399 			/*
400 			 * The original PCI0646 didn't have the primary
401 			 * channel enable bit, it appeared starting with
402 			 * PCI0646U (i.e. revision ID 3).
403 			 */
404 			if (dev->revision < 3) {
405 				d.enablebits[0].reg = 0;
406 				d.port_ops = &cmd64x_port_ops;
407 				if (dev->revision == 1)
408 					d.dma_ops = &cmd646_rev1_dma_ops;
409 			}
410 		}
411 	}
412 
413 	return ide_pci_init_one(dev, &d, NULL);
414 }
415 
416 static const struct pci_device_id cmd64x_pci_tbl[] = {
417 	{ PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_643), 0 },
418 	{ PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_646), 1 },
419 	{ PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_648), 2 },
420 	{ PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_649), 3 },
421 	{ 0, },
422 };
423 MODULE_DEVICE_TABLE(pci, cmd64x_pci_tbl);
424 
425 static struct pci_driver cmd64x_pci_driver = {
426 	.name		= "CMD64x_IDE",
427 	.id_table	= cmd64x_pci_tbl,
428 	.probe		= cmd64x_init_one,
429 	.remove		= ide_pci_remove,
430 	.suspend	= ide_pci_suspend,
431 	.resume		= ide_pci_resume,
432 };
433 
cmd64x_ide_init(void)434 static int __init cmd64x_ide_init(void)
435 {
436 	return ide_pci_register_driver(&cmd64x_pci_driver);
437 }
438 
cmd64x_ide_exit(void)439 static void __exit cmd64x_ide_exit(void)
440 {
441 	pci_unregister_driver(&cmd64x_pci_driver);
442 }
443 
444 module_init(cmd64x_ide_init);
445 module_exit(cmd64x_ide_exit);
446 
447 MODULE_AUTHOR("Eddie Dost, David Miller, Andre Hedrick, Bartlomiej Zolnierkiewicz");
448 MODULE_DESCRIPTION("PCI driver module for CMD64x IDE");
449 MODULE_LICENSE("GPL");
450