cmd64x.c 12 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452
  1. /*
  2. * cmd64x.c: Enable interrupts at initialization time on Ultra/PCI machines.
  3. * Due to massive hardware bugs, UltraDMA is only supported
  4. * on the 646U2 and not on the 646U.
  5. *
  6. * Copyright (C) 1998 Eddie C. Dost (ecd@skynet.be)
  7. * Copyright (C) 1998 David S. Miller (davem@redhat.com)
  8. *
  9. * Copyright (C) 1999-2002 Andre Hedrick <andre@linux-ide.org>
  10. * Copyright (C) 2007-2010 Bartlomiej Zolnierkiewicz
  11. * Copyright (C) 2007,2009 MontaVista Software, Inc. <source@mvista.com>
  12. */
  13. #include <linux/module.h>
  14. #include <linux/types.h>
  15. #include <linux/pci.h>
  16. #include <linux/ide.h>
  17. #include <linux/init.h>
  18. #include <asm/io.h>
  19. #define DRV_NAME "cmd64x"
  20. /*
  21. * CMD64x specific registers definition.
  22. */
  23. #define CFR 0x50
  24. #define CFR_INTR_CH0 0x04
  25. #define CMDTIM 0x52
  26. #define ARTTIM0 0x53
  27. #define DRWTIM0 0x54
  28. #define ARTTIM1 0x55
  29. #define DRWTIM1 0x56
  30. #define ARTTIM23 0x57
  31. #define ARTTIM23_DIS_RA2 0x04
  32. #define ARTTIM23_DIS_RA3 0x08
  33. #define ARTTIM23_INTR_CH1 0x10
  34. #define DRWTIM2 0x58
  35. #define BRST 0x59
  36. #define DRWTIM3 0x5b
  37. #define BMIDECR0 0x70
  38. #define MRDMODE 0x71
  39. #define MRDMODE_INTR_CH0 0x04
  40. #define MRDMODE_INTR_CH1 0x08
  41. #define UDIDETCR0 0x73
  42. #define DTPR0 0x74
  43. #define BMIDECR1 0x78
  44. #define BMIDECSR 0x79
  45. #define UDIDETCR1 0x7B
  46. #define DTPR1 0x7C
  47. static void cmd64x_program_timings(ide_drive_t *drive, u8 mode)
  48. {
  49. ide_hwif_t *hwif = drive->hwif;
  50. struct pci_dev *dev = to_pci_dev(drive->hwif->dev);
  51. int bus_speed = ide_pci_clk ? ide_pci_clk : 33;
  52. const unsigned long T = 1000000 / bus_speed;
  53. static const u8 recovery_values[] =
  54. {15, 15, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 0};
  55. static const u8 setup_values[] = {0x40, 0x40, 0x40, 0x80, 0, 0xc0};
  56. static const u8 arttim_regs[4] = {ARTTIM0, ARTTIM1, ARTTIM23, ARTTIM23};
  57. static const u8 drwtim_regs[4] = {DRWTIM0, DRWTIM1, DRWTIM2, DRWTIM3};
  58. struct ide_timing t;
  59. u8 arttim = 0;
  60. if (drive->dn >= ARRAY_SIZE(drwtim_regs))
  61. return;
  62. ide_timing_compute(drive, mode, &t, T, 0);
  63. /*
  64. * In case we've got too long recovery phase, try to lengthen
  65. * the active phase
  66. */
  67. if (t.recover > 16) {
  68. t.active += t.recover - 16;
  69. t.recover = 16;
  70. }
  71. if (t.active > 16) /* shouldn't actually happen... */
  72. t.active = 16;
  73. /*
  74. * Convert values to internal chipset representation
  75. */
  76. t.recover = recovery_values[t.recover];
  77. t.active &= 0x0f;
  78. /* Program the active/recovery counts into the DRWTIM register */
  79. pci_write_config_byte(dev, drwtim_regs[drive->dn],
  80. (t.active << 4) | t.recover);
  81. /*
  82. * The primary channel has individual address setup timing registers
  83. * for each drive and the hardware selects the slowest timing itself.
  84. * The secondary channel has one common register and we have to select
  85. * the slowest address setup timing ourselves.
  86. */
  87. if (hwif->channel) {
  88. ide_drive_t *pair = ide_get_pair_dev(drive);
  89. if (pair) {
  90. struct ide_timing tp;
  91. ide_timing_compute(pair, pair->pio_mode, &tp, T, 0);
  92. ide_timing_merge(&t, &tp, &t, IDE_TIMING_SETUP);
  93. if (pair->dma_mode) {
  94. ide_timing_compute(pair, pair->dma_mode,
  95. &tp, T, 0);
  96. ide_timing_merge(&tp, &t, &t, IDE_TIMING_SETUP);
  97. }
  98. }
  99. }
  100. if (t.setup > 5) /* shouldn't actually happen... */
  101. t.setup = 5;
  102. /*
  103. * Program the address setup clocks into the ARTTIM registers.
  104. * Avoid clearing the secondary channel's interrupt bit.
  105. */
  106. (void) pci_read_config_byte (dev, arttim_regs[drive->dn], &arttim);
  107. if (hwif->channel)
  108. arttim &= ~ARTTIM23_INTR_CH1;
  109. arttim &= ~0xc0;
  110. arttim |= setup_values[t.setup];
  111. (void) pci_write_config_byte(dev, arttim_regs[drive->dn], arttim);
  112. }
  113. /*
  114. * Attempts to set drive's PIO mode.
  115. * Special cases are 8: prefetch off, 9: prefetch on (both never worked)
  116. */
  117. static void cmd64x_set_pio_mode(ide_hwif_t *hwif, ide_drive_t *drive)
  118. {
  119. const u8 pio = drive->pio_mode - XFER_PIO_0;
  120. /*
  121. * Filter out the prefetch control values
  122. * to prevent PIO5 from being programmed
  123. */
  124. if (pio == 8 || pio == 9)
  125. return;
  126. cmd64x_program_timings(drive, XFER_PIO_0 + pio);
  127. }
  128. static void cmd64x_set_dma_mode(ide_hwif_t *hwif, ide_drive_t *drive)
  129. {
  130. struct pci_dev *dev = to_pci_dev(hwif->dev);
  131. u8 unit = drive->dn & 0x01;
  132. u8 regU = 0, pciU = hwif->channel ? UDIDETCR1 : UDIDETCR0;
  133. const u8 speed = drive->dma_mode;
  134. pci_read_config_byte(dev, pciU, &regU);
  135. regU &= ~(unit ? 0xCA : 0x35);
  136. switch(speed) {
  137. case XFER_UDMA_5:
  138. regU |= unit ? 0x0A : 0x05;
  139. break;
  140. case XFER_UDMA_4:
  141. regU |= unit ? 0x4A : 0x15;
  142. break;
  143. case XFER_UDMA_3:
  144. regU |= unit ? 0x8A : 0x25;
  145. break;
  146. case XFER_UDMA_2:
  147. regU |= unit ? 0x42 : 0x11;
  148. break;
  149. case XFER_UDMA_1:
  150. regU |= unit ? 0x82 : 0x21;
  151. break;
  152. case XFER_UDMA_0:
  153. regU |= unit ? 0xC2 : 0x31;
  154. break;
  155. case XFER_MW_DMA_2:
  156. case XFER_MW_DMA_1:
  157. case XFER_MW_DMA_0:
  158. cmd64x_program_timings(drive, speed);
  159. break;
  160. }
  161. pci_write_config_byte(dev, pciU, regU);
  162. }
  163. static void cmd648_clear_irq(ide_drive_t *drive)
  164. {
  165. ide_hwif_t *hwif = drive->hwif;
  166. struct pci_dev *dev = to_pci_dev(hwif->dev);
  167. unsigned long base = pci_resource_start(dev, 4);
  168. u8 irq_mask = hwif->channel ? MRDMODE_INTR_CH1 :
  169. MRDMODE_INTR_CH0;
  170. u8 mrdmode = inb(base + 1);
  171. /* clear the interrupt bit */
  172. outb((mrdmode & ~(MRDMODE_INTR_CH0 | MRDMODE_INTR_CH1)) | irq_mask,
  173. base + 1);
  174. }
  175. static void cmd64x_clear_irq(ide_drive_t *drive)
  176. {
  177. ide_hwif_t *hwif = drive->hwif;
  178. struct pci_dev *dev = to_pci_dev(hwif->dev);
  179. int irq_reg = hwif->channel ? ARTTIM23 : CFR;
  180. u8 irq_mask = hwif->channel ? ARTTIM23_INTR_CH1 :
  181. CFR_INTR_CH0;
  182. u8 irq_stat = 0;
  183. (void) pci_read_config_byte(dev, irq_reg, &irq_stat);
  184. /* clear the interrupt bit */
  185. (void) pci_write_config_byte(dev, irq_reg, irq_stat | irq_mask);
  186. }
  187. static int cmd648_test_irq(ide_hwif_t *hwif)
  188. {
  189. struct pci_dev *dev = to_pci_dev(hwif->dev);
  190. unsigned long base = pci_resource_start(dev, 4);
  191. u8 irq_mask = hwif->channel ? MRDMODE_INTR_CH1 :
  192. MRDMODE_INTR_CH0;
  193. u8 mrdmode = inb(base + 1);
  194. pr_debug("%s: mrdmode: 0x%02x irq_mask: 0x%02x\n",
  195. hwif->name, mrdmode, irq_mask);
  196. return (mrdmode & irq_mask) ? 1 : 0;
  197. }
  198. static int cmd64x_test_irq(ide_hwif_t *hwif)
  199. {
  200. struct pci_dev *dev = to_pci_dev(hwif->dev);
  201. int irq_reg = hwif->channel ? ARTTIM23 : CFR;
  202. u8 irq_mask = hwif->channel ? ARTTIM23_INTR_CH1 :
  203. CFR_INTR_CH0;
  204. u8 irq_stat = 0;
  205. (void) pci_read_config_byte(dev, irq_reg, &irq_stat);
  206. pr_debug("%s: irq_stat: 0x%02x irq_mask: 0x%02x\n",
  207. hwif->name, irq_stat, irq_mask);
  208. return (irq_stat & irq_mask) ? 1 : 0;
  209. }
  210. /*
  211. * ASUS P55T2P4D with CMD646 chipset revision 0x01 requires the old
  212. * event order for DMA transfers.
  213. */
  214. static int cmd646_1_dma_end(ide_drive_t *drive)
  215. {
  216. ide_hwif_t *hwif = drive->hwif;
  217. u8 dma_stat = 0, dma_cmd = 0;
  218. /* get DMA status */
  219. dma_stat = inb(hwif->dma_base + ATA_DMA_STATUS);
  220. /* read DMA command state */
  221. dma_cmd = inb(hwif->dma_base + ATA_DMA_CMD);
  222. /* stop DMA */
  223. outb(dma_cmd & ~1, hwif->dma_base + ATA_DMA_CMD);
  224. /* clear the INTR & ERROR bits */
  225. outb(dma_stat | 6, hwif->dma_base + ATA_DMA_STATUS);
  226. /* verify good DMA status */
  227. return (dma_stat & 7) != 4;
  228. }
  229. static int init_chipset_cmd64x(struct pci_dev *dev)
  230. {
  231. u8 mrdmode = 0;
  232. /* Set a good latency timer and cache line size value. */
  233. (void) pci_write_config_byte(dev, PCI_LATENCY_TIMER, 64);
  234. /* FIXME: pci_set_master() to ensure a good latency timer value */
  235. /*
  236. * Enable interrupts, select MEMORY READ LINE for reads.
  237. *
  238. * NOTE: although not mentioned in the PCI0646U specs,
  239. * bits 0-1 are write only and won't be read back as
  240. * set or not -- PCI0646U2 specs clarify this point.
  241. */
  242. (void) pci_read_config_byte (dev, MRDMODE, &mrdmode);
  243. mrdmode &= ~0x30;
  244. (void) pci_write_config_byte(dev, MRDMODE, (mrdmode | 0x02));
  245. return 0;
  246. }
  247. static u8 cmd64x_cable_detect(ide_hwif_t *hwif)
  248. {
  249. struct pci_dev *dev = to_pci_dev(hwif->dev);
  250. u8 bmidecsr = 0, mask = hwif->channel ? 0x02 : 0x01;
  251. switch (dev->device) {
  252. case PCI_DEVICE_ID_CMD_648:
  253. case PCI_DEVICE_ID_CMD_649:
  254. pci_read_config_byte(dev, BMIDECSR, &bmidecsr);
  255. return (bmidecsr & mask) ? ATA_CBL_PATA80 : ATA_CBL_PATA40;
  256. default:
  257. return ATA_CBL_PATA40;
  258. }
  259. }
  260. static const struct ide_port_ops cmd64x_port_ops = {
  261. .set_pio_mode = cmd64x_set_pio_mode,
  262. .set_dma_mode = cmd64x_set_dma_mode,
  263. .clear_irq = cmd64x_clear_irq,
  264. .test_irq = cmd64x_test_irq,
  265. .cable_detect = cmd64x_cable_detect,
  266. };
  267. static const struct ide_port_ops cmd648_port_ops = {
  268. .set_pio_mode = cmd64x_set_pio_mode,
  269. .set_dma_mode = cmd64x_set_dma_mode,
  270. .clear_irq = cmd648_clear_irq,
  271. .test_irq = cmd648_test_irq,
  272. .cable_detect = cmd64x_cable_detect,
  273. };
  274. static const struct ide_dma_ops cmd646_rev1_dma_ops = {
  275. .dma_host_set = ide_dma_host_set,
  276. .dma_setup = ide_dma_setup,
  277. .dma_start = ide_dma_start,
  278. .dma_end = cmd646_1_dma_end,
  279. .dma_test_irq = ide_dma_test_irq,
  280. .dma_lost_irq = ide_dma_lost_irq,
  281. .dma_timer_expiry = ide_dma_sff_timer_expiry,
  282. .dma_sff_read_status = ide_dma_sff_read_status,
  283. };
  284. static const struct ide_port_info cmd64x_chipsets[] = {
  285. { /* 0: CMD643 */
  286. .name = DRV_NAME,
  287. .init_chipset = init_chipset_cmd64x,
  288. .enablebits = {{0x00,0x00,0x00}, {0x51,0x08,0x08}},
  289. .port_ops = &cmd64x_port_ops,
  290. .host_flags = IDE_HFLAG_CLEAR_SIMPLEX |
  291. IDE_HFLAG_ABUSE_PREFETCH |
  292. IDE_HFLAG_SERIALIZE,
  293. .pio_mask = ATA_PIO5,
  294. .mwdma_mask = ATA_MWDMA2,
  295. .udma_mask = 0x00, /* no udma */
  296. },
  297. { /* 1: CMD646 */
  298. .name = DRV_NAME,
  299. .init_chipset = init_chipset_cmd64x,
  300. .enablebits = {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
  301. .port_ops = &cmd648_port_ops,
  302. .host_flags = IDE_HFLAG_ABUSE_PREFETCH |
  303. IDE_HFLAG_SERIALIZE,
  304. .pio_mask = ATA_PIO5,
  305. .mwdma_mask = ATA_MWDMA2,
  306. .udma_mask = ATA_UDMA2,
  307. },
  308. { /* 2: CMD648 */
  309. .name = DRV_NAME,
  310. .init_chipset = init_chipset_cmd64x,
  311. .enablebits = {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
  312. .port_ops = &cmd648_port_ops,
  313. .host_flags = IDE_HFLAG_ABUSE_PREFETCH,
  314. .pio_mask = ATA_PIO5,
  315. .mwdma_mask = ATA_MWDMA2,
  316. .udma_mask = ATA_UDMA4,
  317. },
  318. { /* 3: CMD649 */
  319. .name = DRV_NAME,
  320. .init_chipset = init_chipset_cmd64x,
  321. .enablebits = {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
  322. .port_ops = &cmd648_port_ops,
  323. .host_flags = IDE_HFLAG_ABUSE_PREFETCH,
  324. .pio_mask = ATA_PIO5,
  325. .mwdma_mask = ATA_MWDMA2,
  326. .udma_mask = ATA_UDMA5,
  327. }
  328. };
  329. static int cmd64x_init_one(struct pci_dev *dev, const struct pci_device_id *id)
  330. {
  331. struct ide_port_info d;
  332. u8 idx = id->driver_data;
  333. d = cmd64x_chipsets[idx];
  334. if (idx == 1) {
  335. /*
  336. * UltraDMA only supported on PCI646U and PCI646U2, which
  337. * correspond to revisions 0x03, 0x05 and 0x07 respectively.
  338. * Actually, although the CMD tech support people won't
  339. * tell me the details, the 0x03 revision cannot support
  340. * UDMA correctly without hardware modifications, and even
  341. * then it only works with Quantum disks due to some
  342. * hold time assumptions in the 646U part which are fixed
  343. * in the 646U2.
  344. *
  345. * So we only do UltraDMA on revision 0x05 and 0x07 chipsets.
  346. */
  347. if (dev->revision < 5) {
  348. d.udma_mask = 0x00;
  349. /*
  350. * The original PCI0646 didn't have the primary
  351. * channel enable bit, it appeared starting with
  352. * PCI0646U (i.e. revision ID 3).
  353. */
  354. if (dev->revision < 3) {
  355. d.enablebits[0].reg = 0;
  356. d.port_ops = &cmd64x_port_ops;
  357. if (dev->revision == 1)
  358. d.dma_ops = &cmd646_rev1_dma_ops;
  359. }
  360. }
  361. }
  362. return ide_pci_init_one(dev, &d, NULL);
  363. }
  364. static const struct pci_device_id cmd64x_pci_tbl[] = {
  365. { PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_643), 0 },
  366. { PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_646), 1 },
  367. { PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_648), 2 },
  368. { PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_649), 3 },
  369. { 0, },
  370. };
  371. MODULE_DEVICE_TABLE(pci, cmd64x_pci_tbl);
  372. static struct pci_driver cmd64x_pci_driver = {
  373. .name = "CMD64x_IDE",
  374. .id_table = cmd64x_pci_tbl,
  375. .probe = cmd64x_init_one,
  376. .remove = ide_pci_remove,
  377. .suspend = ide_pci_suspend,
  378. .resume = ide_pci_resume,
  379. };
  380. static int __init cmd64x_ide_init(void)
  381. {
  382. return ide_pci_register_driver(&cmd64x_pci_driver);
  383. }
  384. static void __exit cmd64x_ide_exit(void)
  385. {
  386. pci_unregister_driver(&cmd64x_pci_driver);
  387. }
  388. module_init(cmd64x_ide_init);
  389. module_exit(cmd64x_ide_exit);
  390. MODULE_AUTHOR("Eddie Dost, David Miller, Andre Hedrick, Bartlomiej Zolnierkiewicz");
  391. MODULE_DESCRIPTION("PCI driver module for CMD64x IDE");
  392. MODULE_LICENSE("GPL");