pata_arasan_cf.c 26 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984
  1. /*
  2. * drivers/ata/pata_arasan_cf.c
  3. *
  4. * Arasan Compact Flash host controller source file
  5. *
  6. * Copyright (C) 2011 ST Microelectronics
  7. * Viresh Kumar <viresh.kumar@st.com>
  8. *
  9. * This file is licensed under the terms of the GNU General Public
  10. * License version 2. This program is licensed "as is" without any
  11. * warranty of any kind, whether express or implied.
  12. */
  13. /*
  14. * The Arasan CompactFlash Device Controller IP core has three basic modes of
  15. * operation: PC card ATA using I/O mode, PC card ATA using memory mode, PC card
  16. * ATA using true IDE modes. This driver supports only True IDE mode currently.
  17. *
  18. * Arasan CF Controller shares global irq register with Arasan XD Controller.
  19. *
  20. * Tested on arch/arm/mach-spear13xx
  21. */
  22. #include <linux/ata.h>
  23. #include <linux/clk.h>
  24. #include <linux/completion.h>
  25. #include <linux/delay.h>
  26. #include <linux/dmaengine.h>
  27. #include <linux/io.h>
  28. #include <linux/irq.h>
  29. #include <linux/kernel.h>
  30. #include <linux/libata.h>
  31. #include <linux/module.h>
  32. #include <linux/pata_arasan_cf_data.h>
  33. #include <linux/platform_device.h>
  34. #include <linux/pm.h>
  35. #include <linux/slab.h>
  36. #include <linux/spinlock.h>
  37. #include <linux/types.h>
  38. #include <linux/workqueue.h>
  39. #define DRIVER_NAME "arasan_cf"
  40. #define TIMEOUT msecs_to_jiffies(3000)
  41. /* Registers */
  42. /* CompactFlash Interface Status */
  43. #define CFI_STS 0x000
  44. #define STS_CHG (1)
  45. #define BIN_AUDIO_OUT (1 << 1)
  46. #define CARD_DETECT1 (1 << 2)
  47. #define CARD_DETECT2 (1 << 3)
  48. #define INP_ACK (1 << 4)
  49. #define CARD_READY (1 << 5)
  50. #define IO_READY (1 << 6)
  51. #define B16_IO_PORT_SEL (1 << 7)
  52. /* IRQ */
  53. #define IRQ_STS 0x004
  54. /* Interrupt Enable */
  55. #define IRQ_EN 0x008
  56. #define CARD_DETECT_IRQ (1)
  57. #define STATUS_CHNG_IRQ (1 << 1)
  58. #define MEM_MODE_IRQ (1 << 2)
  59. #define IO_MODE_IRQ (1 << 3)
  60. #define TRUE_IDE_MODE_IRQ (1 << 8)
  61. #define PIO_XFER_ERR_IRQ (1 << 9)
  62. #define BUF_AVAIL_IRQ (1 << 10)
  63. #define XFER_DONE_IRQ (1 << 11)
  64. #define IGNORED_IRQS (STATUS_CHNG_IRQ | MEM_MODE_IRQ | IO_MODE_IRQ |\
  65. TRUE_IDE_MODE_IRQ)
  66. #define TRUE_IDE_IRQS (CARD_DETECT_IRQ | PIO_XFER_ERR_IRQ |\
  67. BUF_AVAIL_IRQ | XFER_DONE_IRQ)
  68. /* Operation Mode */
  69. #define OP_MODE 0x00C
  70. #define CARD_MODE_MASK (0x3)
  71. #define MEM_MODE (0x0)
  72. #define IO_MODE (0x1)
  73. #define TRUE_IDE_MODE (0x2)
  74. #define CARD_TYPE_MASK (1 << 2)
  75. #define CF_CARD (0)
  76. #define CF_PLUS_CARD (1 << 2)
  77. #define CARD_RESET (1 << 3)
  78. #define CFHOST_ENB (1 << 4)
  79. #define OUTPUTS_TRISTATE (1 << 5)
  80. #define ULTRA_DMA_ENB (1 << 8)
  81. #define MULTI_WORD_DMA_ENB (1 << 9)
  82. #define DRQ_BLOCK_SIZE_MASK (0x3 << 11)
  83. #define DRQ_BLOCK_SIZE_512 (0)
  84. #define DRQ_BLOCK_SIZE_1024 (1 << 11)
  85. #define DRQ_BLOCK_SIZE_2048 (2 << 11)
  86. #define DRQ_BLOCK_SIZE_4096 (3 << 11)
  87. /* CF Interface Clock Configuration */
  88. #define CLK_CFG 0x010
  89. #define CF_IF_CLK_MASK (0XF)
  90. /* CF Timing Mode Configuration */
  91. #define TM_CFG 0x014
  92. #define MEM_MODE_TIMING_MASK (0x3)
  93. #define MEM_MODE_TIMING_250NS (0x0)
  94. #define MEM_MODE_TIMING_120NS (0x1)
  95. #define MEM_MODE_TIMING_100NS (0x2)
  96. #define MEM_MODE_TIMING_80NS (0x3)
  97. #define IO_MODE_TIMING_MASK (0x3 << 2)
  98. #define IO_MODE_TIMING_250NS (0x0 << 2)
  99. #define IO_MODE_TIMING_120NS (0x1 << 2)
  100. #define IO_MODE_TIMING_100NS (0x2 << 2)
  101. #define IO_MODE_TIMING_80NS (0x3 << 2)
  102. #define TRUEIDE_PIO_TIMING_MASK (0x7 << 4)
  103. #define TRUEIDE_PIO_TIMING_SHIFT 4
  104. #define TRUEIDE_MWORD_DMA_TIMING_MASK (0x7 << 7)
  105. #define TRUEIDE_MWORD_DMA_TIMING_SHIFT 7
  106. #define ULTRA_DMA_TIMING_MASK (0x7 << 10)
  107. #define ULTRA_DMA_TIMING_SHIFT 10
  108. /* CF Transfer Address */
  109. #define XFER_ADDR 0x014
  110. #define XFER_ADDR_MASK (0x7FF)
  111. #define MAX_XFER_COUNT 0x20000u
  112. /* Transfer Control */
  113. #define XFER_CTR 0x01C
  114. #define XFER_COUNT_MASK (0x3FFFF)
  115. #define ADDR_INC_DISABLE (1 << 24)
  116. #define XFER_WIDTH_MASK (1 << 25)
  117. #define XFER_WIDTH_8B (0)
  118. #define XFER_WIDTH_16B (1 << 25)
  119. #define MEM_TYPE_MASK (1 << 26)
  120. #define MEM_TYPE_COMMON (0)
  121. #define MEM_TYPE_ATTRIBUTE (1 << 26)
  122. #define MEM_IO_XFER_MASK (1 << 27)
  123. #define MEM_XFER (0)
  124. #define IO_XFER (1 << 27)
  125. #define DMA_XFER_MODE (1 << 28)
  126. #define AHB_BUS_NORMAL_PIO_OPRTN (~(1 << 29))
  127. #define XFER_DIR_MASK (1 << 30)
  128. #define XFER_READ (0)
  129. #define XFER_WRITE (1 << 30)
  130. #define XFER_START (1 << 31)
  131. /* Write Data Port */
  132. #define WRITE_PORT 0x024
  133. /* Read Data Port */
  134. #define READ_PORT 0x028
  135. /* ATA Data Port */
  136. #define ATA_DATA_PORT 0x030
  137. #define ATA_DATA_PORT_MASK (0xFFFF)
  138. /* ATA Error/Features */
  139. #define ATA_ERR_FTR 0x034
  140. /* ATA Sector Count */
  141. #define ATA_SC 0x038
  142. /* ATA Sector Number */
  143. #define ATA_SN 0x03C
  144. /* ATA Cylinder Low */
  145. #define ATA_CL 0x040
  146. /* ATA Cylinder High */
  147. #define ATA_CH 0x044
  148. /* ATA Select Card/Head */
  149. #define ATA_SH 0x048
  150. /* ATA Status-Command */
  151. #define ATA_STS_CMD 0x04C
  152. /* ATA Alternate Status/Device Control */
  153. #define ATA_ASTS_DCTR 0x050
  154. /* Extended Write Data Port 0x200-0x3FC */
  155. #define EXT_WRITE_PORT 0x200
  156. /* Extended Read Data Port 0x400-0x5FC */
  157. #define EXT_READ_PORT 0x400
  158. #define FIFO_SIZE 0x200u
  159. /* Global Interrupt Status */
  160. #define GIRQ_STS 0x800
  161. /* Global Interrupt Status enable */
  162. #define GIRQ_STS_EN 0x804
  163. /* Global Interrupt Signal enable */
  164. #define GIRQ_SGN_EN 0x808
  165. #define GIRQ_CF (1)
  166. #define GIRQ_XD (1 << 1)
  167. /* Compact Flash Controller Dev Structure */
  168. struct arasan_cf_dev {
  169. /* pointer to ata_host structure */
  170. struct ata_host *host;
  171. /* clk structure, only if HAVE_CLK is defined */
  172. #ifdef CONFIG_HAVE_CLK
  173. struct clk *clk;
  174. #endif
  175. /* physical base address of controller */
  176. dma_addr_t pbase;
  177. /* virtual base address of controller */
  178. void __iomem *vbase;
  179. /* irq number*/
  180. int irq;
  181. /* status to be updated to framework regarding DMA transfer */
  182. u8 dma_status;
  183. /* Card is present or Not */
  184. u8 card_present;
  185. /* dma specific */
  186. /* Completion for transfer complete interrupt from controller */
  187. struct completion cf_completion;
  188. /* Completion for DMA transfer complete. */
  189. struct completion dma_completion;
  190. /* Dma channel allocated */
  191. struct dma_chan *dma_chan;
  192. /* Mask for DMA transfers */
  193. dma_cap_mask_t mask;
  194. /* dma channel private data */
  195. void *dma_priv;
  196. /* DMA transfer work */
  197. struct work_struct work;
  198. /* DMA delayed finish work */
  199. struct delayed_work dwork;
  200. /* qc to be transferred using DMA */
  201. struct ata_queued_cmd *qc;
  202. };
  203. static struct scsi_host_template arasan_cf_sht = {
  204. ATA_BASE_SHT(DRIVER_NAME),
  205. .sg_tablesize = SG_NONE,
  206. .dma_boundary = 0xFFFFFFFFUL,
  207. };
  208. static void cf_dumpregs(struct arasan_cf_dev *acdev)
  209. {
  210. struct device *dev = acdev->host->dev;
  211. dev_dbg(dev, ": =========== REGISTER DUMP ===========");
  212. dev_dbg(dev, ": CFI_STS: %x", readl(acdev->vbase + CFI_STS));
  213. dev_dbg(dev, ": IRQ_STS: %x", readl(acdev->vbase + IRQ_STS));
  214. dev_dbg(dev, ": IRQ_EN: %x", readl(acdev->vbase + IRQ_EN));
  215. dev_dbg(dev, ": OP_MODE: %x", readl(acdev->vbase + OP_MODE));
  216. dev_dbg(dev, ": CLK_CFG: %x", readl(acdev->vbase + CLK_CFG));
  217. dev_dbg(dev, ": TM_CFG: %x", readl(acdev->vbase + TM_CFG));
  218. dev_dbg(dev, ": XFER_CTR: %x", readl(acdev->vbase + XFER_CTR));
  219. dev_dbg(dev, ": GIRQ_STS: %x", readl(acdev->vbase + GIRQ_STS));
  220. dev_dbg(dev, ": GIRQ_STS_EN: %x", readl(acdev->vbase + GIRQ_STS_EN));
  221. dev_dbg(dev, ": GIRQ_SGN_EN: %x", readl(acdev->vbase + GIRQ_SGN_EN));
  222. dev_dbg(dev, ": =====================================");
  223. }
  224. /* Enable/Disable global interrupts shared between CF and XD ctrlr. */
  225. static void cf_ginterrupt_enable(struct arasan_cf_dev *acdev, bool enable)
  226. {
  227. /* enable should be 0 or 1 */
  228. writel(enable, acdev->vbase + GIRQ_STS_EN);
  229. writel(enable, acdev->vbase + GIRQ_SGN_EN);
  230. }
  231. /* Enable/Disable CF interrupts */
  232. static inline void
  233. cf_interrupt_enable(struct arasan_cf_dev *acdev, u32 mask, bool enable)
  234. {
  235. u32 val = readl(acdev->vbase + IRQ_EN);
  236. /* clear & enable/disable irqs */
  237. if (enable) {
  238. writel(mask, acdev->vbase + IRQ_STS);
  239. writel(val | mask, acdev->vbase + IRQ_EN);
  240. } else
  241. writel(val & ~mask, acdev->vbase + IRQ_EN);
  242. }
  243. static inline void cf_card_reset(struct arasan_cf_dev *acdev)
  244. {
  245. u32 val = readl(acdev->vbase + OP_MODE);
  246. writel(val | CARD_RESET, acdev->vbase + OP_MODE);
  247. udelay(200);
  248. writel(val & ~CARD_RESET, acdev->vbase + OP_MODE);
  249. }
  250. static inline void cf_ctrl_reset(struct arasan_cf_dev *acdev)
  251. {
  252. writel(readl(acdev->vbase + OP_MODE) & ~CFHOST_ENB,
  253. acdev->vbase + OP_MODE);
  254. writel(readl(acdev->vbase + OP_MODE) | CFHOST_ENB,
  255. acdev->vbase + OP_MODE);
  256. }
  257. static void cf_card_detect(struct arasan_cf_dev *acdev, bool hotplugged)
  258. {
  259. struct ata_port *ap = acdev->host->ports[0];
  260. struct ata_eh_info *ehi = &ap->link.eh_info;
  261. u32 val = readl(acdev->vbase + CFI_STS);
  262. /* Both CD1 & CD2 should be low if card inserted completely */
  263. if (!(val & (CARD_DETECT1 | CARD_DETECT2))) {
  264. if (acdev->card_present)
  265. return;
  266. acdev->card_present = 1;
  267. cf_card_reset(acdev);
  268. } else {
  269. if (!acdev->card_present)
  270. return;
  271. acdev->card_present = 0;
  272. }
  273. if (hotplugged) {
  274. ata_ehi_hotplugged(ehi);
  275. ata_port_freeze(ap);
  276. }
  277. }
  278. static int cf_init(struct arasan_cf_dev *acdev)
  279. {
  280. struct arasan_cf_pdata *pdata = dev_get_platdata(acdev->host->dev);
  281. unsigned long flags;
  282. int ret = 0;
  283. #ifdef CONFIG_HAVE_CLK
  284. ret = clk_enable(acdev->clk);
  285. if (ret) {
  286. dev_dbg(acdev->host->dev, "clock enable failed");
  287. return ret;
  288. }
  289. #endif
  290. spin_lock_irqsave(&acdev->host->lock, flags);
  291. /* configure CF interface clock */
  292. writel((pdata->cf_if_clk <= CF_IF_CLK_200M) ? pdata->cf_if_clk :
  293. CF_IF_CLK_166M, acdev->vbase + CLK_CFG);
  294. writel(TRUE_IDE_MODE | CFHOST_ENB, acdev->vbase + OP_MODE);
  295. cf_interrupt_enable(acdev, CARD_DETECT_IRQ, 1);
  296. cf_ginterrupt_enable(acdev, 1);
  297. spin_unlock_irqrestore(&acdev->host->lock, flags);
  298. return ret;
  299. }
  300. static void cf_exit(struct arasan_cf_dev *acdev)
  301. {
  302. unsigned long flags;
  303. spin_lock_irqsave(&acdev->host->lock, flags);
  304. cf_ginterrupt_enable(acdev, 0);
  305. cf_interrupt_enable(acdev, TRUE_IDE_IRQS, 0);
  306. cf_card_reset(acdev);
  307. writel(readl(acdev->vbase + OP_MODE) & ~CFHOST_ENB,
  308. acdev->vbase + OP_MODE);
  309. spin_unlock_irqrestore(&acdev->host->lock, flags);
  310. #ifdef CONFIG_HAVE_CLK
  311. clk_disable(acdev->clk);
  312. #endif
  313. }
  314. static void dma_callback(void *dev)
  315. {
  316. struct arasan_cf_dev *acdev = (struct arasan_cf_dev *) dev;
  317. complete(&acdev->dma_completion);
  318. }
  319. static bool filter(struct dma_chan *chan, void *slave)
  320. {
  321. chan->private = slave;
  322. return true;
  323. }
  324. static inline void dma_complete(struct arasan_cf_dev *acdev)
  325. {
  326. struct ata_queued_cmd *qc = acdev->qc;
  327. unsigned long flags;
  328. acdev->qc = NULL;
  329. ata_sff_interrupt(acdev->irq, acdev->host);
  330. spin_lock_irqsave(&acdev->host->lock, flags);
  331. if (unlikely(qc->err_mask) && ata_is_dma(qc->tf.protocol))
  332. ata_ehi_push_desc(&qc->ap->link.eh_info, "DMA Failed: Timeout");
  333. spin_unlock_irqrestore(&acdev->host->lock, flags);
  334. }
  335. static inline int wait4buf(struct arasan_cf_dev *acdev)
  336. {
  337. if (!wait_for_completion_timeout(&acdev->cf_completion, TIMEOUT)) {
  338. u32 rw = acdev->qc->tf.flags & ATA_TFLAG_WRITE;
  339. dev_err(acdev->host->dev, "%s TimeOut", rw ? "write" : "read");
  340. return -ETIMEDOUT;
  341. }
  342. /* Check if PIO Error interrupt has occurred */
  343. if (acdev->dma_status & ATA_DMA_ERR)
  344. return -EAGAIN;
  345. return 0;
  346. }
  347. static int
  348. dma_xfer(struct arasan_cf_dev *acdev, dma_addr_t src, dma_addr_t dest, u32 len)
  349. {
  350. struct dma_async_tx_descriptor *tx;
  351. struct dma_chan *chan = acdev->dma_chan;
  352. dma_cookie_t cookie;
  353. unsigned long flags = DMA_PREP_INTERRUPT | DMA_COMPL_SKIP_SRC_UNMAP |
  354. DMA_COMPL_SKIP_DEST_UNMAP;
  355. int ret = 0;
  356. tx = chan->device->device_prep_dma_memcpy(chan, dest, src, len, flags);
  357. if (!tx) {
  358. dev_err(acdev->host->dev, "device_prep_dma_memcpy failed\n");
  359. return -EAGAIN;
  360. }
  361. tx->callback = dma_callback;
  362. tx->callback_param = acdev;
  363. cookie = tx->tx_submit(tx);
  364. ret = dma_submit_error(cookie);
  365. if (ret) {
  366. dev_err(acdev->host->dev, "dma_submit_error\n");
  367. return ret;
  368. }
  369. chan->device->device_issue_pending(chan);
  370. /* Wait for DMA to complete */
  371. if (!wait_for_completion_timeout(&acdev->dma_completion, TIMEOUT)) {
  372. chan->device->device_control(chan, DMA_TERMINATE_ALL, 0);
  373. dev_err(acdev->host->dev, "wait_for_completion_timeout\n");
  374. return -ETIMEDOUT;
  375. }
  376. return ret;
  377. }
  378. static int sg_xfer(struct arasan_cf_dev *acdev, struct scatterlist *sg)
  379. {
  380. dma_addr_t dest = 0, src = 0;
  381. u32 xfer_cnt, sglen, dma_len, xfer_ctr;
  382. u32 write = acdev->qc->tf.flags & ATA_TFLAG_WRITE;
  383. unsigned long flags;
  384. int ret = 0;
  385. sglen = sg_dma_len(sg);
  386. if (write) {
  387. src = sg_dma_address(sg);
  388. dest = acdev->pbase + EXT_WRITE_PORT;
  389. } else {
  390. dest = sg_dma_address(sg);
  391. src = acdev->pbase + EXT_READ_PORT;
  392. }
  393. /*
  394. * For each sg:
  395. * MAX_XFER_COUNT data will be transferred before we get transfer
  396. * complete interrupt. Between after FIFO_SIZE data
  397. * buffer available interrupt will be generated. At this time we will
  398. * fill FIFO again: max FIFO_SIZE data.
  399. */
  400. while (sglen) {
  401. xfer_cnt = min(sglen, MAX_XFER_COUNT);
  402. spin_lock_irqsave(&acdev->host->lock, flags);
  403. xfer_ctr = readl(acdev->vbase + XFER_CTR) &
  404. ~XFER_COUNT_MASK;
  405. writel(xfer_ctr | xfer_cnt | XFER_START,
  406. acdev->vbase + XFER_CTR);
  407. spin_unlock_irqrestore(&acdev->host->lock, flags);
  408. /* continue dma xfers until current sg is completed */
  409. while (xfer_cnt) {
  410. /* wait for read to complete */
  411. if (!write) {
  412. ret = wait4buf(acdev);
  413. if (ret)
  414. goto fail;
  415. }
  416. /* read/write FIFO in chunk of FIFO_SIZE */
  417. dma_len = min(xfer_cnt, FIFO_SIZE);
  418. ret = dma_xfer(acdev, src, dest, dma_len);
  419. if (ret) {
  420. dev_err(acdev->host->dev, "dma failed");
  421. goto fail;
  422. }
  423. if (write)
  424. src += dma_len;
  425. else
  426. dest += dma_len;
  427. sglen -= dma_len;
  428. xfer_cnt -= dma_len;
  429. /* wait for write to complete */
  430. if (write) {
  431. ret = wait4buf(acdev);
  432. if (ret)
  433. goto fail;
  434. }
  435. }
  436. }
  437. fail:
  438. spin_lock_irqsave(&acdev->host->lock, flags);
  439. writel(readl(acdev->vbase + XFER_CTR) & ~XFER_START,
  440. acdev->vbase + XFER_CTR);
  441. spin_unlock_irqrestore(&acdev->host->lock, flags);
  442. return ret;
  443. }
  444. /*
  445. * This routine uses External DMA controller to read/write data to FIFO of CF
  446. * controller. There are two xfer related interrupt supported by CF controller:
  447. * - buf_avail: This interrupt is generated as soon as we have buffer of 512
  448. * bytes available for reading or empty buffer available for writing.
  449. * - xfer_done: This interrupt is generated on transfer of "xfer_size" amount of
  450. * data to/from FIFO. xfer_size is programmed in XFER_CTR register.
  451. *
  452. * Max buffer size = FIFO_SIZE = 512 Bytes.
  453. * Max xfer_size = MAX_XFER_COUNT = 256 KB.
  454. */
  455. static void data_xfer(struct work_struct *work)
  456. {
  457. struct arasan_cf_dev *acdev = container_of(work, struct arasan_cf_dev,
  458. work);
  459. struct ata_queued_cmd *qc = acdev->qc;
  460. struct scatterlist *sg;
  461. unsigned long flags;
  462. u32 temp;
  463. int ret = 0;
  464. /* request dma channels */
  465. /* dma_request_channel may sleep, so calling from process context */
  466. acdev->dma_chan = dma_request_channel(acdev->mask, filter,
  467. acdev->dma_priv);
  468. if (!acdev->dma_chan) {
  469. dev_err(acdev->host->dev, "Unable to get dma_chan\n");
  470. goto chan_request_fail;
  471. }
  472. for_each_sg(qc->sg, sg, qc->n_elem, temp) {
  473. ret = sg_xfer(acdev, sg);
  474. if (ret)
  475. break;
  476. }
  477. dma_release_channel(acdev->dma_chan);
  478. /* data xferred successfully */
  479. if (!ret) {
  480. u32 status;
  481. spin_lock_irqsave(&acdev->host->lock, flags);
  482. status = ioread8(qc->ap->ioaddr.altstatus_addr);
  483. spin_unlock_irqrestore(&acdev->host->lock, flags);
  484. if (status & (ATA_BUSY | ATA_DRQ)) {
  485. ata_sff_queue_delayed_work(&acdev->dwork, 1);
  486. return;
  487. }
  488. goto sff_intr;
  489. }
  490. cf_dumpregs(acdev);
  491. chan_request_fail:
  492. spin_lock_irqsave(&acdev->host->lock, flags);
  493. /* error when transferring data to/from memory */
  494. qc->err_mask |= AC_ERR_HOST_BUS;
  495. qc->ap->hsm_task_state = HSM_ST_ERR;
  496. cf_ctrl_reset(acdev);
  497. spin_unlock_irqrestore(qc->ap->lock, flags);
  498. sff_intr:
  499. dma_complete(acdev);
  500. }
  501. static void delayed_finish(struct work_struct *work)
  502. {
  503. struct arasan_cf_dev *acdev = container_of(work, struct arasan_cf_dev,
  504. dwork.work);
  505. struct ata_queued_cmd *qc = acdev->qc;
  506. unsigned long flags;
  507. u8 status;
  508. spin_lock_irqsave(&acdev->host->lock, flags);
  509. status = ioread8(qc->ap->ioaddr.altstatus_addr);
  510. spin_unlock_irqrestore(&acdev->host->lock, flags);
  511. if (status & (ATA_BUSY | ATA_DRQ))
  512. ata_sff_queue_delayed_work(&acdev->dwork, 1);
  513. else
  514. dma_complete(acdev);
  515. }
  516. static irqreturn_t arasan_cf_interrupt(int irq, void *dev)
  517. {
  518. struct arasan_cf_dev *acdev = ((struct ata_host *)dev)->private_data;
  519. unsigned long flags;
  520. u32 irqsts;
  521. irqsts = readl(acdev->vbase + GIRQ_STS);
  522. if (!(irqsts & GIRQ_CF))
  523. return IRQ_NONE;
  524. spin_lock_irqsave(&acdev->host->lock, flags);
  525. irqsts = readl(acdev->vbase + IRQ_STS);
  526. writel(irqsts, acdev->vbase + IRQ_STS); /* clear irqs */
  527. writel(GIRQ_CF, acdev->vbase + GIRQ_STS); /* clear girqs */
  528. /* handle only relevant interrupts */
  529. irqsts &= ~IGNORED_IRQS;
  530. if (irqsts & CARD_DETECT_IRQ) {
  531. cf_card_detect(acdev, 1);
  532. spin_unlock_irqrestore(&acdev->host->lock, flags);
  533. return IRQ_HANDLED;
  534. }
  535. if (irqsts & PIO_XFER_ERR_IRQ) {
  536. acdev->dma_status = ATA_DMA_ERR;
  537. writel(readl(acdev->vbase + XFER_CTR) & ~XFER_START,
  538. acdev->vbase + XFER_CTR);
  539. spin_unlock_irqrestore(&acdev->host->lock, flags);
  540. complete(&acdev->cf_completion);
  541. dev_err(acdev->host->dev, "pio xfer err irq\n");
  542. return IRQ_HANDLED;
  543. }
  544. spin_unlock_irqrestore(&acdev->host->lock, flags);
  545. if (irqsts & BUF_AVAIL_IRQ) {
  546. complete(&acdev->cf_completion);
  547. return IRQ_HANDLED;
  548. }
  549. if (irqsts & XFER_DONE_IRQ) {
  550. struct ata_queued_cmd *qc = acdev->qc;
  551. /* Send Complete only for write */
  552. if (qc->tf.flags & ATA_TFLAG_WRITE)
  553. complete(&acdev->cf_completion);
  554. }
  555. return IRQ_HANDLED;
  556. }
  557. static void arasan_cf_freeze(struct ata_port *ap)
  558. {
  559. struct arasan_cf_dev *acdev = ap->host->private_data;
  560. /* stop transfer and reset controller */
  561. writel(readl(acdev->vbase + XFER_CTR) & ~XFER_START,
  562. acdev->vbase + XFER_CTR);
  563. cf_ctrl_reset(acdev);
  564. acdev->dma_status = ATA_DMA_ERR;
  565. ata_sff_dma_pause(ap);
  566. ata_sff_freeze(ap);
  567. }
  568. void arasan_cf_error_handler(struct ata_port *ap)
  569. {
  570. struct arasan_cf_dev *acdev = ap->host->private_data;
  571. /*
  572. * DMA transfers using an external DMA controller may be scheduled.
  573. * Abort them before handling error. Refer data_xfer() for further
  574. * details.
  575. */
  576. cancel_work_sync(&acdev->work);
  577. cancel_delayed_work_sync(&acdev->dwork);
  578. return ata_sff_error_handler(ap);
  579. }
  580. static void arasan_cf_dma_start(struct arasan_cf_dev *acdev)
  581. {
  582. u32 xfer_ctr = readl(acdev->vbase + XFER_CTR) & ~XFER_DIR_MASK;
  583. u32 write = acdev->qc->tf.flags & ATA_TFLAG_WRITE;
  584. xfer_ctr |= write ? XFER_WRITE : XFER_READ;
  585. writel(xfer_ctr, acdev->vbase + XFER_CTR);
  586. acdev->qc->ap->ops->sff_exec_command(acdev->qc->ap, &acdev->qc->tf);
  587. ata_sff_queue_work(&acdev->work);
  588. }
  589. unsigned int arasan_cf_qc_issue(struct ata_queued_cmd *qc)
  590. {
  591. struct ata_port *ap = qc->ap;
  592. struct arasan_cf_dev *acdev = ap->host->private_data;
  593. /* defer PIO handling to sff_qc_issue */
  594. if (!ata_is_dma(qc->tf.protocol))
  595. return ata_sff_qc_issue(qc);
  596. /* select the device */
  597. ata_wait_idle(ap);
  598. ata_sff_dev_select(ap, qc->dev->devno);
  599. ata_wait_idle(ap);
  600. /* start the command */
  601. switch (qc->tf.protocol) {
  602. case ATA_PROT_DMA:
  603. WARN_ON_ONCE(qc->tf.flags & ATA_TFLAG_POLLING);
  604. ap->ops->sff_tf_load(ap, &qc->tf);
  605. acdev->dma_status = 0;
  606. acdev->qc = qc;
  607. arasan_cf_dma_start(acdev);
  608. ap->hsm_task_state = HSM_ST_LAST;
  609. break;
  610. default:
  611. WARN_ON(1);
  612. return AC_ERR_SYSTEM;
  613. }
  614. return 0;
  615. }
  616. static void arasan_cf_set_piomode(struct ata_port *ap, struct ata_device *adev)
  617. {
  618. struct arasan_cf_dev *acdev = ap->host->private_data;
  619. u8 pio = adev->pio_mode - XFER_PIO_0;
  620. unsigned long flags;
  621. u32 val;
  622. /* Arasan ctrl supports Mode0 -> Mode6 */
  623. if (pio > 6) {
  624. dev_err(ap->dev, "Unknown PIO mode\n");
  625. return;
  626. }
  627. spin_lock_irqsave(&acdev->host->lock, flags);
  628. val = readl(acdev->vbase + OP_MODE) &
  629. ~(ULTRA_DMA_ENB | MULTI_WORD_DMA_ENB | DRQ_BLOCK_SIZE_MASK);
  630. writel(val, acdev->vbase + OP_MODE);
  631. val = readl(acdev->vbase + TM_CFG) & ~TRUEIDE_PIO_TIMING_MASK;
  632. val |= pio << TRUEIDE_PIO_TIMING_SHIFT;
  633. writel(val, acdev->vbase + TM_CFG);
  634. cf_interrupt_enable(acdev, BUF_AVAIL_IRQ | XFER_DONE_IRQ, 0);
  635. cf_interrupt_enable(acdev, PIO_XFER_ERR_IRQ, 1);
  636. spin_unlock_irqrestore(&acdev->host->lock, flags);
  637. }
  638. static void arasan_cf_set_dmamode(struct ata_port *ap, struct ata_device *adev)
  639. {
  640. struct arasan_cf_dev *acdev = ap->host->private_data;
  641. u32 opmode, tmcfg, dma_mode = adev->dma_mode;
  642. unsigned long flags;
  643. spin_lock_irqsave(&acdev->host->lock, flags);
  644. opmode = readl(acdev->vbase + OP_MODE) &
  645. ~(MULTI_WORD_DMA_ENB | ULTRA_DMA_ENB);
  646. tmcfg = readl(acdev->vbase + TM_CFG);
  647. if ((dma_mode >= XFER_UDMA_0) && (dma_mode <= XFER_UDMA_6)) {
  648. opmode |= ULTRA_DMA_ENB;
  649. tmcfg &= ~ULTRA_DMA_TIMING_MASK;
  650. tmcfg |= (dma_mode - XFER_UDMA_0) << ULTRA_DMA_TIMING_SHIFT;
  651. } else if ((dma_mode >= XFER_MW_DMA_0) && (dma_mode <= XFER_MW_DMA_4)) {
  652. opmode |= MULTI_WORD_DMA_ENB;
  653. tmcfg &= ~TRUEIDE_MWORD_DMA_TIMING_MASK;
  654. tmcfg |= (dma_mode - XFER_MW_DMA_0) <<
  655. TRUEIDE_MWORD_DMA_TIMING_SHIFT;
  656. } else {
  657. dev_err(ap->dev, "Unknown DMA mode\n");
  658. spin_unlock_irqrestore(&acdev->host->lock, flags);
  659. return;
  660. }
  661. writel(opmode, acdev->vbase + OP_MODE);
  662. writel(tmcfg, acdev->vbase + TM_CFG);
  663. writel(DMA_XFER_MODE, acdev->vbase + XFER_CTR);
  664. cf_interrupt_enable(acdev, PIO_XFER_ERR_IRQ, 0);
  665. cf_interrupt_enable(acdev, BUF_AVAIL_IRQ | XFER_DONE_IRQ, 1);
  666. spin_unlock_irqrestore(&acdev->host->lock, flags);
  667. }
  668. static struct ata_port_operations arasan_cf_ops = {
  669. .inherits = &ata_sff_port_ops,
  670. .freeze = arasan_cf_freeze,
  671. .error_handler = arasan_cf_error_handler,
  672. .qc_issue = arasan_cf_qc_issue,
  673. .set_piomode = arasan_cf_set_piomode,
  674. .set_dmamode = arasan_cf_set_dmamode,
  675. };
  676. static int __devinit arasan_cf_probe(struct platform_device *pdev)
  677. {
  678. struct arasan_cf_dev *acdev;
  679. struct arasan_cf_pdata *pdata = dev_get_platdata(&pdev->dev);
  680. struct ata_host *host;
  681. struct ata_port *ap;
  682. struct resource *res;
  683. irq_handler_t irq_handler = NULL;
  684. int ret = 0;
  685. res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
  686. if (!res)
  687. return -EINVAL;
  688. if (!devm_request_mem_region(&pdev->dev, res->start, resource_size(res),
  689. DRIVER_NAME)) {
  690. dev_warn(&pdev->dev, "Failed to get memory region resource\n");
  691. return -ENOENT;
  692. }
  693. acdev = devm_kzalloc(&pdev->dev, sizeof(*acdev), GFP_KERNEL);
  694. if (!acdev) {
  695. dev_warn(&pdev->dev, "kzalloc fail\n");
  696. return -ENOMEM;
  697. }
  698. /* if irq is 0, support only PIO */
  699. acdev->irq = platform_get_irq(pdev, 0);
  700. if (acdev->irq)
  701. irq_handler = arasan_cf_interrupt;
  702. else
  703. pdata->quirk |= CF_BROKEN_MWDMA | CF_BROKEN_UDMA;
  704. acdev->pbase = res->start;
  705. acdev->vbase = devm_ioremap_nocache(&pdev->dev, res->start,
  706. resource_size(res));
  707. if (!acdev->vbase) {
  708. dev_warn(&pdev->dev, "ioremap fail\n");
  709. return -ENOMEM;
  710. }
  711. #ifdef CONFIG_HAVE_CLK
  712. acdev->clk = clk_get(&pdev->dev, NULL);
  713. if (IS_ERR(acdev->clk)) {
  714. dev_warn(&pdev->dev, "Clock not found\n");
  715. return PTR_ERR(acdev->clk);
  716. }
  717. #endif
  718. /* allocate host */
  719. host = ata_host_alloc(&pdev->dev, 1);
  720. if (!host) {
  721. ret = -ENOMEM;
  722. dev_warn(&pdev->dev, "alloc host fail\n");
  723. goto free_clk;
  724. }
  725. ap = host->ports[0];
  726. host->private_data = acdev;
  727. acdev->host = host;
  728. ap->ops = &arasan_cf_ops;
  729. ap->pio_mask = ATA_PIO6;
  730. ap->mwdma_mask = ATA_MWDMA4;
  731. ap->udma_mask = ATA_UDMA6;
  732. init_completion(&acdev->cf_completion);
  733. init_completion(&acdev->dma_completion);
  734. INIT_WORK(&acdev->work, data_xfer);
  735. INIT_DELAYED_WORK(&acdev->dwork, delayed_finish);
  736. dma_cap_set(DMA_MEMCPY, acdev->mask);
  737. acdev->dma_priv = pdata->dma_priv;
  738. /* Handle platform specific quirks */
  739. if (pdata->quirk) {
  740. if (pdata->quirk & CF_BROKEN_PIO) {
  741. ap->ops->set_piomode = NULL;
  742. ap->pio_mask = 0;
  743. }
  744. if (pdata->quirk & CF_BROKEN_MWDMA)
  745. ap->mwdma_mask = 0;
  746. if (pdata->quirk & CF_BROKEN_UDMA)
  747. ap->udma_mask = 0;
  748. }
  749. ap->flags |= ATA_FLAG_PIO_POLLING | ATA_FLAG_NO_ATAPI;
  750. ap->ioaddr.cmd_addr = acdev->vbase + ATA_DATA_PORT;
  751. ap->ioaddr.data_addr = acdev->vbase + ATA_DATA_PORT;
  752. ap->ioaddr.error_addr = acdev->vbase + ATA_ERR_FTR;
  753. ap->ioaddr.feature_addr = acdev->vbase + ATA_ERR_FTR;
  754. ap->ioaddr.nsect_addr = acdev->vbase + ATA_SC;
  755. ap->ioaddr.lbal_addr = acdev->vbase + ATA_SN;
  756. ap->ioaddr.lbam_addr = acdev->vbase + ATA_CL;
  757. ap->ioaddr.lbah_addr = acdev->vbase + ATA_CH;
  758. ap->ioaddr.device_addr = acdev->vbase + ATA_SH;
  759. ap->ioaddr.status_addr = acdev->vbase + ATA_STS_CMD;
  760. ap->ioaddr.command_addr = acdev->vbase + ATA_STS_CMD;
  761. ap->ioaddr.altstatus_addr = acdev->vbase + ATA_ASTS_DCTR;
  762. ap->ioaddr.ctl_addr = acdev->vbase + ATA_ASTS_DCTR;
  763. ata_port_desc(ap, "phy_addr %llx virt_addr %p",
  764. (unsigned long long) res->start, acdev->vbase);
  765. ret = cf_init(acdev);
  766. if (ret)
  767. goto free_clk;
  768. cf_card_detect(acdev, 0);
  769. return ata_host_activate(host, acdev->irq, irq_handler, 0,
  770. &arasan_cf_sht);
  771. free_clk:
  772. #ifdef CONFIG_HAVE_CLK
  773. clk_put(acdev->clk);
  774. #endif
  775. return ret;
  776. }
  777. static int __devexit arasan_cf_remove(struct platform_device *pdev)
  778. {
  779. struct ata_host *host = dev_get_drvdata(&pdev->dev);
  780. struct arasan_cf_dev *acdev = host->ports[0]->private_data;
  781. ata_host_detach(host);
  782. cf_exit(acdev);
  783. #ifdef CONFIG_HAVE_CLK
  784. clk_put(acdev->clk);
  785. #endif
  786. return 0;
  787. }
  788. #ifdef CONFIG_PM
  789. static int arasan_cf_suspend(struct device *dev)
  790. {
  791. struct platform_device *pdev = to_platform_device(dev);
  792. struct ata_host *host = dev_get_drvdata(&pdev->dev);
  793. struct arasan_cf_dev *acdev = host->ports[0]->private_data;
  794. if (acdev->dma_chan) {
  795. acdev->dma_chan->device->device_control(acdev->dma_chan,
  796. DMA_TERMINATE_ALL, 0);
  797. dma_release_channel(acdev->dma_chan);
  798. }
  799. cf_exit(acdev);
  800. return ata_host_suspend(host, PMSG_SUSPEND);
  801. }
  802. static int arasan_cf_resume(struct device *dev)
  803. {
  804. struct platform_device *pdev = to_platform_device(dev);
  805. struct ata_host *host = dev_get_drvdata(&pdev->dev);
  806. struct arasan_cf_dev *acdev = host->ports[0]->private_data;
  807. cf_init(acdev);
  808. ata_host_resume(host);
  809. return 0;
  810. }
  811. static const struct dev_pm_ops arasan_cf_pm_ops = {
  812. .suspend = arasan_cf_suspend,
  813. .resume = arasan_cf_resume,
  814. };
  815. #endif
  816. static struct platform_driver arasan_cf_driver = {
  817. .probe = arasan_cf_probe,
  818. .remove = __devexit_p(arasan_cf_remove),
  819. .driver = {
  820. .name = DRIVER_NAME,
  821. .owner = THIS_MODULE,
  822. #ifdef CONFIG_PM
  823. .pm = &arasan_cf_pm_ops,
  824. #endif
  825. },
  826. };
  827. static int __init arasan_cf_init(void)
  828. {
  829. return platform_driver_register(&arasan_cf_driver);
  830. }
  831. module_init(arasan_cf_init);
  832. static void __exit arasan_cf_exit(void)
  833. {
  834. platform_driver_unregister(&arasan_cf_driver);
  835. }
  836. module_exit(arasan_cf_exit);
  837. MODULE_AUTHOR("Viresh Kumar <viresh.kumar@st.com>");
  838. MODULE_DESCRIPTION("Arasan ATA Compact Flash driver");
  839. MODULE_LICENSE("GPL");
  840. MODULE_ALIAS("platform:" DRIVER_NAME);