zynqmp_dma.c 29 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062
  1. /*
  2. * DMA driver for Xilinx ZynqMP DMA Engine
  3. *
  4. * Copyright (C) 2016 Xilinx, Inc. All rights reserved.
  5. *
  6. * This program is free software: you can redistribute it and/or modify
  7. * it under the terms of the GNU General Public License as published by
  8. * the Free Software Foundation, either version 2 of the License, or
  9. * (at your option) any later version.
  10. */
  11. #include <linux/bitops.h>
  12. #include <linux/dmapool.h>
  13. #include <linux/dma/xilinx_dma.h>
  14. #include <linux/init.h>
  15. #include <linux/interrupt.h>
  16. #include <linux/io.h>
  17. #include <linux/module.h>
  18. #include <linux/of_address.h>
  19. #include <linux/of_dma.h>
  20. #include <linux/of_irq.h>
  21. #include <linux/of_platform.h>
  22. #include <linux/slab.h>
  23. #include <linux/clk.h>
  24. #include <linux/io-64-nonatomic-lo-hi.h>
  25. #include "../dmaengine.h"
  26. /* Register Offsets */
  27. #define ZYNQMP_DMA_ISR 0x100
  28. #define ZYNQMP_DMA_IMR 0x104
  29. #define ZYNQMP_DMA_IER 0x108
  30. #define ZYNQMP_DMA_IDS 0x10C
  31. #define ZYNQMP_DMA_CTRL0 0x110
  32. #define ZYNQMP_DMA_CTRL1 0x114
  33. #define ZYNQMP_DMA_DATA_ATTR 0x120
  34. #define ZYNQMP_DMA_DSCR_ATTR 0x124
  35. #define ZYNQMP_DMA_SRC_DSCR_WRD0 0x128
  36. #define ZYNQMP_DMA_SRC_DSCR_WRD1 0x12C
  37. #define ZYNQMP_DMA_SRC_DSCR_WRD2 0x130
  38. #define ZYNQMP_DMA_SRC_DSCR_WRD3 0x134
  39. #define ZYNQMP_DMA_DST_DSCR_WRD0 0x138
  40. #define ZYNQMP_DMA_DST_DSCR_WRD1 0x13C
  41. #define ZYNQMP_DMA_DST_DSCR_WRD2 0x140
  42. #define ZYNQMP_DMA_DST_DSCR_WRD3 0x144
  43. #define ZYNQMP_DMA_SRC_START_LSB 0x158
  44. #define ZYNQMP_DMA_SRC_START_MSB 0x15C
  45. #define ZYNQMP_DMA_DST_START_LSB 0x160
  46. #define ZYNQMP_DMA_DST_START_MSB 0x164
  47. #define ZYNQMP_DMA_RATE_CTRL 0x18C
  48. #define ZYNQMP_DMA_IRQ_SRC_ACCT 0x190
  49. #define ZYNQMP_DMA_IRQ_DST_ACCT 0x194
  50. #define ZYNQMP_DMA_CTRL2 0x200
  51. /* Interrupt registers bit field definitions */
  52. #define ZYNQMP_DMA_DONE BIT(10)
  53. #define ZYNQMP_DMA_AXI_WR_DATA BIT(9)
  54. #define ZYNQMP_DMA_AXI_RD_DATA BIT(8)
  55. #define ZYNQMP_DMA_AXI_RD_DST_DSCR BIT(7)
  56. #define ZYNQMP_DMA_AXI_RD_SRC_DSCR BIT(6)
  57. #define ZYNQMP_DMA_IRQ_DST_ACCT_ERR BIT(5)
  58. #define ZYNQMP_DMA_IRQ_SRC_ACCT_ERR BIT(4)
  59. #define ZYNQMP_DMA_BYTE_CNT_OVRFL BIT(3)
  60. #define ZYNQMP_DMA_DST_DSCR_DONE BIT(2)
  61. #define ZYNQMP_DMA_INV_APB BIT(0)
  62. /* Control 0 register bit field definitions */
  63. #define ZYNQMP_DMA_OVR_FETCH BIT(7)
  64. #define ZYNQMP_DMA_POINT_TYPE_SG BIT(6)
  65. #define ZYNQMP_DMA_RATE_CTRL_EN BIT(3)
  66. /* Control 1 register bit field definitions */
  67. #define ZYNQMP_DMA_SRC_ISSUE GENMASK(4, 0)
  68. /* Data Attribute register bit field definitions */
  69. #define ZYNQMP_DMA_ARBURST GENMASK(27, 26)
  70. #define ZYNQMP_DMA_ARCACHE GENMASK(25, 22)
  71. #define ZYNQMP_DMA_ARCACHE_OFST 22
  72. #define ZYNQMP_DMA_ARQOS GENMASK(21, 18)
  73. #define ZYNQMP_DMA_ARQOS_OFST 18
  74. #define ZYNQMP_DMA_ARLEN GENMASK(17, 14)
  75. #define ZYNQMP_DMA_ARLEN_OFST 14
  76. #define ZYNQMP_DMA_AWBURST GENMASK(13, 12)
  77. #define ZYNQMP_DMA_AWCACHE GENMASK(11, 8)
  78. #define ZYNQMP_DMA_AWCACHE_OFST 8
  79. #define ZYNQMP_DMA_AWQOS GENMASK(7, 4)
  80. #define ZYNQMP_DMA_AWQOS_OFST 4
  81. #define ZYNQMP_DMA_AWLEN GENMASK(3, 0)
  82. #define ZYNQMP_DMA_AWLEN_OFST 0
  83. /* Descriptor Attribute register bit field definitions */
  84. #define ZYNQMP_DMA_AXCOHRNT BIT(8)
  85. #define ZYNQMP_DMA_AXCACHE GENMASK(7, 4)
  86. #define ZYNQMP_DMA_AXCACHE_OFST 4
  87. #define ZYNQMP_DMA_AXQOS GENMASK(3, 0)
  88. #define ZYNQMP_DMA_AXQOS_OFST 0
  89. /* Control register 2 bit field definitions */
  90. #define ZYNQMP_DMA_ENABLE BIT(0)
  91. /* Buffer Descriptor definitions */
  92. #define ZYNQMP_DMA_DESC_CTRL_STOP 0x10
  93. #define ZYNQMP_DMA_DESC_CTRL_COMP_INT 0x4
  94. #define ZYNQMP_DMA_DESC_CTRL_SIZE_256 0x2
  95. #define ZYNQMP_DMA_DESC_CTRL_COHRNT 0x1
  96. /* Interrupt Mask specific definitions */
  97. #define ZYNQMP_DMA_INT_ERR (ZYNQMP_DMA_AXI_RD_DATA | \
  98. ZYNQMP_DMA_AXI_WR_DATA | \
  99. ZYNQMP_DMA_AXI_RD_DST_DSCR | \
  100. ZYNQMP_DMA_AXI_RD_SRC_DSCR | \
  101. ZYNQMP_DMA_INV_APB)
  102. #define ZYNQMP_DMA_INT_OVRFL (ZYNQMP_DMA_BYTE_CNT_OVRFL | \
  103. ZYNQMP_DMA_IRQ_SRC_ACCT_ERR | \
  104. ZYNQMP_DMA_IRQ_DST_ACCT_ERR)
  105. #define ZYNQMP_DMA_INT_DONE (ZYNQMP_DMA_DONE | ZYNQMP_DMA_DST_DSCR_DONE)
  106. #define ZYNQMP_DMA_INT_EN_DEFAULT_MASK (ZYNQMP_DMA_INT_DONE | \
  107. ZYNQMP_DMA_INT_ERR | \
  108. ZYNQMP_DMA_INT_OVRFL | \
  109. ZYNQMP_DMA_DST_DSCR_DONE)
  110. /* Max number of descriptors per channel */
  111. #define ZYNQMP_DMA_NUM_DESCS 32
  112. /* Max transfer size per descriptor */
  113. #define ZYNQMP_DMA_MAX_TRANS_LEN 0x40000000
  114. /* Max burst lengths */
  115. #define ZYNQMP_DMA_MAX_DST_BURST_LEN 32768U
  116. #define ZYNQMP_DMA_MAX_SRC_BURST_LEN 32768U
  117. /* Reset values for data attributes */
  118. #define ZYNQMP_DMA_AXCACHE_VAL 0xF
  119. #define ZYNQMP_DMA_SRC_ISSUE_RST_VAL 0x1F
  120. #define ZYNQMP_DMA_IDS_DEFAULT_MASK 0xFFF
  121. /* Bus width in bits */
  122. #define ZYNQMP_DMA_BUS_WIDTH_64 64
  123. #define ZYNQMP_DMA_BUS_WIDTH_128 128
  124. #define ZYNQMP_DMA_DESC_SIZE(chan) (chan->desc_size)
  125. #define to_chan(chan) container_of(chan, struct zynqmp_dma_chan, \
  126. common)
  127. #define tx_to_desc(tx) container_of(tx, struct zynqmp_dma_desc_sw, \
  128. async_tx)
  129. /**
  130. * struct zynqmp_dma_desc_ll - Hw linked list descriptor
  131. * @addr: Buffer address
  132. * @size: Size of the buffer
  133. * @ctrl: Control word
  134. * @nxtdscraddr: Next descriptor base address
  135. * @rsvd: Reserved field and for Hw internal use.
  136. */
  137. struct zynqmp_dma_desc_ll {
  138. u64 addr;
  139. u32 size;
  140. u32 ctrl;
  141. u64 nxtdscraddr;
  142. u64 rsvd;
  143. };
  144. /**
  145. * struct zynqmp_dma_desc_sw - Per Transaction structure
  146. * @src: Source address for simple mode dma
  147. * @dst: Destination address for simple mode dma
  148. * @len: Transfer length for simple mode dma
  149. * @node: Node in the channel descriptor list
  150. * @tx_list: List head for the current transfer
  151. * @async_tx: Async transaction descriptor
  152. * @src_v: Virtual address of the src descriptor
  153. * @src_p: Physical address of the src descriptor
  154. * @dst_v: Virtual address of the dst descriptor
  155. * @dst_p: Physical address of the dst descriptor
  156. */
  157. struct zynqmp_dma_desc_sw {
  158. u64 src;
  159. u64 dst;
  160. u32 len;
  161. struct list_head node;
  162. struct list_head tx_list;
  163. struct dma_async_tx_descriptor async_tx;
  164. struct zynqmp_dma_desc_ll *src_v;
  165. dma_addr_t src_p;
  166. struct zynqmp_dma_desc_ll *dst_v;
  167. dma_addr_t dst_p;
  168. };
  169. /**
  170. * struct zynqmp_dma_chan - Driver specific DMA channel structure
  171. * @zdev: Driver specific device structure
  172. * @regs: Control registers offset
  173. * @lock: Descriptor operation lock
  174. * @pending_list: Descriptors waiting
  175. * @free_list: Descriptors free
  176. * @active_list: Descriptors active
  177. * @sw_desc_pool: SW descriptor pool
  178. * @done_list: Complete descriptors
  179. * @common: DMA common channel
  180. * @desc_pool_v: Statically allocated descriptor base
  181. * @desc_pool_p: Physical allocated descriptor base
  182. * @desc_free_cnt: Descriptor available count
  183. * @dev: The dma device
  184. * @irq: Channel IRQ
  185. * @is_dmacoherent: Tells whether dma operations are coherent or not
  186. * @tasklet: Cleanup work after irq
  187. * @idle : Channel status;
  188. * @desc_size: Size of the low level descriptor
  189. * @err: Channel has errors
  190. * @bus_width: Bus width
  191. * @src_burst_len: Source burst length
  192. * @dst_burst_len: Dest burst length
  193. * @clk_main: Pointer to main clock
  194. * @clk_apb: Pointer to apb clock
  195. */
  196. struct zynqmp_dma_chan {
  197. struct zynqmp_dma_device *zdev;
  198. void __iomem *regs;
  199. spinlock_t lock;
  200. struct list_head pending_list;
  201. struct list_head free_list;
  202. struct list_head active_list;
  203. struct zynqmp_dma_desc_sw *sw_desc_pool;
  204. struct list_head done_list;
  205. struct dma_chan common;
  206. void *desc_pool_v;
  207. dma_addr_t desc_pool_p;
  208. u32 desc_free_cnt;
  209. struct device *dev;
  210. int irq;
  211. bool is_dmacoherent;
  212. struct tasklet_struct tasklet;
  213. bool idle;
  214. u32 desc_size;
  215. bool err;
  216. u32 bus_width;
  217. u32 src_burst_len;
  218. u32 dst_burst_len;
  219. struct clk *clk_main;
  220. struct clk *clk_apb;
  221. };
  222. /**
  223. * struct zynqmp_dma_device - DMA device structure
  224. * @dev: Device Structure
  225. * @common: DMA device structure
  226. * @chan: Driver specific DMA channel
  227. */
  228. struct zynqmp_dma_device {
  229. struct device *dev;
  230. struct dma_device common;
  231. struct zynqmp_dma_chan *chan;
  232. };
  233. static inline void zynqmp_dma_writeq(struct zynqmp_dma_chan *chan, u32 reg,
  234. u64 value)
  235. {
  236. lo_hi_writeq(value, chan->regs + reg);
  237. }
  238. /**
  239. * zynqmp_dma_update_desc_to_ctrlr - Updates descriptor to the controller
  240. * @chan: ZynqMP DMA DMA channel pointer
  241. * @desc: Transaction descriptor pointer
  242. */
  243. static void zynqmp_dma_update_desc_to_ctrlr(struct zynqmp_dma_chan *chan,
  244. struct zynqmp_dma_desc_sw *desc)
  245. {
  246. dma_addr_t addr;
  247. addr = desc->src_p;
  248. zynqmp_dma_writeq(chan, ZYNQMP_DMA_SRC_START_LSB, addr);
  249. addr = desc->dst_p;
  250. zynqmp_dma_writeq(chan, ZYNQMP_DMA_DST_START_LSB, addr);
  251. }
  252. /**
  253. * zynqmp_dma_desc_config_eod - Mark the descriptor as end descriptor
  254. * @chan: ZynqMP DMA channel pointer
  255. * @desc: Hw descriptor pointer
  256. */
  257. static void zynqmp_dma_desc_config_eod(struct zynqmp_dma_chan *chan,
  258. void *desc)
  259. {
  260. struct zynqmp_dma_desc_ll *hw = (struct zynqmp_dma_desc_ll *)desc;
  261. hw->ctrl |= ZYNQMP_DMA_DESC_CTRL_STOP;
  262. hw++;
  263. hw->ctrl |= ZYNQMP_DMA_DESC_CTRL_COMP_INT | ZYNQMP_DMA_DESC_CTRL_STOP;
  264. }
  265. /**
  266. * zynqmp_dma_config_sg_ll_desc - Configure the linked list descriptor
  267. * @chan: ZynqMP DMA channel pointer
  268. * @sdesc: Hw descriptor pointer
  269. * @src: Source buffer address
  270. * @dst: Destination buffer address
  271. * @len: Transfer length
  272. * @prev: Previous hw descriptor pointer
  273. */
  274. static void zynqmp_dma_config_sg_ll_desc(struct zynqmp_dma_chan *chan,
  275. struct zynqmp_dma_desc_ll *sdesc,
  276. dma_addr_t src, dma_addr_t dst, size_t len,
  277. struct zynqmp_dma_desc_ll *prev)
  278. {
  279. struct zynqmp_dma_desc_ll *ddesc = sdesc + 1;
  280. sdesc->size = ddesc->size = len;
  281. sdesc->addr = src;
  282. ddesc->addr = dst;
  283. sdesc->ctrl = ddesc->ctrl = ZYNQMP_DMA_DESC_CTRL_SIZE_256;
  284. if (chan->is_dmacoherent) {
  285. sdesc->ctrl |= ZYNQMP_DMA_DESC_CTRL_COHRNT;
  286. ddesc->ctrl |= ZYNQMP_DMA_DESC_CTRL_COHRNT;
  287. }
  288. if (prev) {
  289. dma_addr_t addr = chan->desc_pool_p +
  290. ((uintptr_t)sdesc - (uintptr_t)chan->desc_pool_v);
  291. ddesc = prev + 1;
  292. prev->nxtdscraddr = addr;
  293. ddesc->nxtdscraddr = addr + ZYNQMP_DMA_DESC_SIZE(chan);
  294. }
  295. }
  296. /**
  297. * zynqmp_dma_init - Initialize the channel
  298. * @chan: ZynqMP DMA channel pointer
  299. */
  300. static void zynqmp_dma_init(struct zynqmp_dma_chan *chan)
  301. {
  302. u32 val;
  303. writel(ZYNQMP_DMA_IDS_DEFAULT_MASK, chan->regs + ZYNQMP_DMA_IDS);
  304. val = readl(chan->regs + ZYNQMP_DMA_ISR);
  305. writel(val, chan->regs + ZYNQMP_DMA_ISR);
  306. if (chan->is_dmacoherent) {
  307. val = ZYNQMP_DMA_AXCOHRNT;
  308. val = (val & ~ZYNQMP_DMA_AXCACHE) |
  309. (ZYNQMP_DMA_AXCACHE_VAL << ZYNQMP_DMA_AXCACHE_OFST);
  310. writel(val, chan->regs + ZYNQMP_DMA_DSCR_ATTR);
  311. }
  312. val = readl(chan->regs + ZYNQMP_DMA_DATA_ATTR);
  313. if (chan->is_dmacoherent) {
  314. val = (val & ~ZYNQMP_DMA_ARCACHE) |
  315. (ZYNQMP_DMA_AXCACHE_VAL << ZYNQMP_DMA_ARCACHE_OFST);
  316. val = (val & ~ZYNQMP_DMA_AWCACHE) |
  317. (ZYNQMP_DMA_AXCACHE_VAL << ZYNQMP_DMA_AWCACHE_OFST);
  318. }
  319. writel(val, chan->regs + ZYNQMP_DMA_DATA_ATTR);
  320. /* Clearing the interrupt account rgisters */
  321. val = readl(chan->regs + ZYNQMP_DMA_IRQ_SRC_ACCT);
  322. val = readl(chan->regs + ZYNQMP_DMA_IRQ_DST_ACCT);
  323. chan->idle = true;
  324. }
  325. /**
  326. * zynqmp_dma_tx_submit - Submit DMA transaction
  327. * @tx: Async transaction descriptor pointer
  328. *
  329. * Return: cookie value
  330. */
  331. static dma_cookie_t zynqmp_dma_tx_submit(struct dma_async_tx_descriptor *tx)
  332. {
  333. struct zynqmp_dma_chan *chan = to_chan(tx->chan);
  334. struct zynqmp_dma_desc_sw *desc, *new;
  335. dma_cookie_t cookie;
  336. new = tx_to_desc(tx);
  337. spin_lock_bh(&chan->lock);
  338. cookie = dma_cookie_assign(tx);
  339. if (!list_empty(&chan->pending_list)) {
  340. desc = list_last_entry(&chan->pending_list,
  341. struct zynqmp_dma_desc_sw, node);
  342. if (!list_empty(&desc->tx_list))
  343. desc = list_last_entry(&desc->tx_list,
  344. struct zynqmp_dma_desc_sw, node);
  345. desc->src_v->nxtdscraddr = new->src_p;
  346. desc->src_v->ctrl &= ~ZYNQMP_DMA_DESC_CTRL_STOP;
  347. desc->dst_v->nxtdscraddr = new->dst_p;
  348. desc->dst_v->ctrl &= ~ZYNQMP_DMA_DESC_CTRL_STOP;
  349. }
  350. list_add_tail(&new->node, &chan->pending_list);
  351. spin_unlock_bh(&chan->lock);
  352. return cookie;
  353. }
  354. /**
  355. * zynqmp_dma_get_descriptor - Get the sw descriptor from the pool
  356. * @chan: ZynqMP DMA channel pointer
  357. *
  358. * Return: The sw descriptor
  359. */
  360. static struct zynqmp_dma_desc_sw *
  361. zynqmp_dma_get_descriptor(struct zynqmp_dma_chan *chan)
  362. {
  363. struct zynqmp_dma_desc_sw *desc;
  364. spin_lock_bh(&chan->lock);
  365. desc = list_first_entry(&chan->free_list,
  366. struct zynqmp_dma_desc_sw, node);
  367. list_del(&desc->node);
  368. spin_unlock_bh(&chan->lock);
  369. INIT_LIST_HEAD(&desc->tx_list);
  370. /* Clear the src and dst descriptor memory */
  371. memset((void *)desc->src_v, 0, ZYNQMP_DMA_DESC_SIZE(chan));
  372. memset((void *)desc->dst_v, 0, ZYNQMP_DMA_DESC_SIZE(chan));
  373. return desc;
  374. }
  375. /**
  376. * zynqmp_dma_free_descriptor - Issue pending transactions
  377. * @chan: ZynqMP DMA channel pointer
  378. * @sdesc: Transaction descriptor pointer
  379. */
  380. static void zynqmp_dma_free_descriptor(struct zynqmp_dma_chan *chan,
  381. struct zynqmp_dma_desc_sw *sdesc)
  382. {
  383. struct zynqmp_dma_desc_sw *child, *next;
  384. chan->desc_free_cnt++;
  385. list_add_tail(&sdesc->node, &chan->free_list);
  386. list_for_each_entry_safe(child, next, &sdesc->tx_list, node) {
  387. chan->desc_free_cnt++;
  388. list_move_tail(&child->node, &chan->free_list);
  389. }
  390. }
  391. /**
  392. * zynqmp_dma_free_desc_list - Free descriptors list
  393. * @chan: ZynqMP DMA channel pointer
  394. * @list: List to parse and delete the descriptor
  395. */
  396. static void zynqmp_dma_free_desc_list(struct zynqmp_dma_chan *chan,
  397. struct list_head *list)
  398. {
  399. struct zynqmp_dma_desc_sw *desc, *next;
  400. list_for_each_entry_safe(desc, next, list, node)
  401. zynqmp_dma_free_descriptor(chan, desc);
  402. }
  403. /**
  404. * zynqmp_dma_alloc_chan_resources - Allocate channel resources
  405. * @dchan: DMA channel
  406. *
  407. * Return: Number of descriptors on success and failure value on error
  408. */
  409. static int zynqmp_dma_alloc_chan_resources(struct dma_chan *dchan)
  410. {
  411. struct zynqmp_dma_chan *chan = to_chan(dchan);
  412. struct zynqmp_dma_desc_sw *desc;
  413. int i;
  414. chan->sw_desc_pool = kzalloc(sizeof(*desc) * ZYNQMP_DMA_NUM_DESCS,
  415. GFP_KERNEL);
  416. if (!chan->sw_desc_pool)
  417. return -ENOMEM;
  418. chan->idle = true;
  419. chan->desc_free_cnt = ZYNQMP_DMA_NUM_DESCS;
  420. INIT_LIST_HEAD(&chan->free_list);
  421. for (i = 0; i < ZYNQMP_DMA_NUM_DESCS; i++) {
  422. desc = chan->sw_desc_pool + i;
  423. dma_async_tx_descriptor_init(&desc->async_tx, &chan->common);
  424. desc->async_tx.tx_submit = zynqmp_dma_tx_submit;
  425. list_add_tail(&desc->node, &chan->free_list);
  426. }
  427. chan->desc_pool_v = dma_zalloc_coherent(chan->dev,
  428. (2 * chan->desc_size * ZYNQMP_DMA_NUM_DESCS),
  429. &chan->desc_pool_p, GFP_KERNEL);
  430. if (!chan->desc_pool_v)
  431. return -ENOMEM;
  432. for (i = 0; i < ZYNQMP_DMA_NUM_DESCS; i++) {
  433. desc = chan->sw_desc_pool + i;
  434. desc->src_v = (struct zynqmp_dma_desc_ll *) (chan->desc_pool_v +
  435. (i * ZYNQMP_DMA_DESC_SIZE(chan) * 2));
  436. desc->dst_v = (struct zynqmp_dma_desc_ll *) (desc->src_v + 1);
  437. desc->src_p = chan->desc_pool_p +
  438. (i * ZYNQMP_DMA_DESC_SIZE(chan) * 2);
  439. desc->dst_p = desc->src_p + ZYNQMP_DMA_DESC_SIZE(chan);
  440. }
  441. return ZYNQMP_DMA_NUM_DESCS;
  442. }
  443. /**
  444. * zynqmp_dma_start - Start DMA channel
  445. * @chan: ZynqMP DMA channel pointer
  446. */
  447. static void zynqmp_dma_start(struct zynqmp_dma_chan *chan)
  448. {
  449. writel(ZYNQMP_DMA_INT_EN_DEFAULT_MASK, chan->regs + ZYNQMP_DMA_IER);
  450. chan->idle = false;
  451. writel(ZYNQMP_DMA_ENABLE, chan->regs + ZYNQMP_DMA_CTRL2);
  452. }
  453. /**
  454. * zynqmp_dma_handle_ovfl_int - Process the overflow interrupt
  455. * @chan: ZynqMP DMA channel pointer
  456. * @status: Interrupt status value
  457. */
  458. static void zynqmp_dma_handle_ovfl_int(struct zynqmp_dma_chan *chan, u32 status)
  459. {
  460. u32 val;
  461. if (status & ZYNQMP_DMA_IRQ_DST_ACCT_ERR)
  462. val = readl(chan->regs + ZYNQMP_DMA_IRQ_DST_ACCT);
  463. if (status & ZYNQMP_DMA_IRQ_SRC_ACCT_ERR)
  464. val = readl(chan->regs + ZYNQMP_DMA_IRQ_SRC_ACCT);
  465. }
  466. static void zynqmp_dma_config(struct zynqmp_dma_chan *chan)
  467. {
  468. u32 val, burst_val;
  469. val = readl(chan->regs + ZYNQMP_DMA_CTRL0);
  470. val |= ZYNQMP_DMA_POINT_TYPE_SG;
  471. writel(val, chan->regs + ZYNQMP_DMA_CTRL0);
  472. val = readl(chan->regs + ZYNQMP_DMA_DATA_ATTR);
  473. burst_val = __ilog2_u32(chan->src_burst_len);
  474. val = (val & ~ZYNQMP_DMA_ARLEN) |
  475. ((burst_val << ZYNQMP_DMA_ARLEN_OFST) & ZYNQMP_DMA_ARLEN);
  476. burst_val = __ilog2_u32(chan->dst_burst_len);
  477. val = (val & ~ZYNQMP_DMA_AWLEN) |
  478. ((burst_val << ZYNQMP_DMA_AWLEN_OFST) & ZYNQMP_DMA_AWLEN);
  479. writel(val, chan->regs + ZYNQMP_DMA_DATA_ATTR);
  480. }
  481. /**
  482. * zynqmp_dma_device_config - Zynqmp dma device configuration
  483. * @dchan: DMA channel
  484. * @config: DMA device config
  485. */
  486. static int zynqmp_dma_device_config(struct dma_chan *dchan,
  487. struct dma_slave_config *config)
  488. {
  489. struct zynqmp_dma_chan *chan = to_chan(dchan);
  490. chan->src_burst_len = clamp(config->src_maxburst, 1U,
  491. ZYNQMP_DMA_MAX_SRC_BURST_LEN);
  492. chan->dst_burst_len = clamp(config->dst_maxburst, 1U,
  493. ZYNQMP_DMA_MAX_DST_BURST_LEN);
  494. return 0;
  495. }
  496. /**
  497. * zynqmp_dma_start_transfer - Initiate the new transfer
  498. * @chan: ZynqMP DMA channel pointer
  499. */
  500. static void zynqmp_dma_start_transfer(struct zynqmp_dma_chan *chan)
  501. {
  502. struct zynqmp_dma_desc_sw *desc;
  503. if (!chan->idle)
  504. return;
  505. zynqmp_dma_config(chan);
  506. desc = list_first_entry_or_null(&chan->pending_list,
  507. struct zynqmp_dma_desc_sw, node);
  508. if (!desc)
  509. return;
  510. list_splice_tail_init(&chan->pending_list, &chan->active_list);
  511. zynqmp_dma_update_desc_to_ctrlr(chan, desc);
  512. zynqmp_dma_start(chan);
  513. }
  514. /**
  515. * zynqmp_dma_chan_desc_cleanup - Cleanup the completed descriptors
  516. * @chan: ZynqMP DMA channel
  517. */
  518. static void zynqmp_dma_chan_desc_cleanup(struct zynqmp_dma_chan *chan)
  519. {
  520. struct zynqmp_dma_desc_sw *desc, *next;
  521. list_for_each_entry_safe(desc, next, &chan->done_list, node) {
  522. dma_async_tx_callback callback;
  523. void *callback_param;
  524. list_del(&desc->node);
  525. callback = desc->async_tx.callback;
  526. callback_param = desc->async_tx.callback_param;
  527. if (callback) {
  528. spin_unlock(&chan->lock);
  529. callback(callback_param);
  530. spin_lock(&chan->lock);
  531. }
  532. /* Run any dependencies, then free the descriptor */
  533. zynqmp_dma_free_descriptor(chan, desc);
  534. }
  535. }
  536. /**
  537. * zynqmp_dma_complete_descriptor - Mark the active descriptor as complete
  538. * @chan: ZynqMP DMA channel pointer
  539. */
  540. static void zynqmp_dma_complete_descriptor(struct zynqmp_dma_chan *chan)
  541. {
  542. struct zynqmp_dma_desc_sw *desc;
  543. desc = list_first_entry_or_null(&chan->active_list,
  544. struct zynqmp_dma_desc_sw, node);
  545. if (!desc)
  546. return;
  547. list_del(&desc->node);
  548. dma_cookie_complete(&desc->async_tx);
  549. list_add_tail(&desc->node, &chan->done_list);
  550. }
  551. /**
  552. * zynqmp_dma_issue_pending - Issue pending transactions
  553. * @dchan: DMA channel pointer
  554. */
  555. static void zynqmp_dma_issue_pending(struct dma_chan *dchan)
  556. {
  557. struct zynqmp_dma_chan *chan = to_chan(dchan);
  558. spin_lock_bh(&chan->lock);
  559. zynqmp_dma_start_transfer(chan);
  560. spin_unlock_bh(&chan->lock);
  561. }
  562. /**
  563. * zynqmp_dma_free_descriptors - Free channel descriptors
  564. * @dchan: DMA channel pointer
  565. */
  566. static void zynqmp_dma_free_descriptors(struct zynqmp_dma_chan *chan)
  567. {
  568. zynqmp_dma_free_desc_list(chan, &chan->active_list);
  569. zynqmp_dma_free_desc_list(chan, &chan->pending_list);
  570. zynqmp_dma_free_desc_list(chan, &chan->done_list);
  571. }
  572. /**
  573. * zynqmp_dma_free_chan_resources - Free channel resources
  574. * @dchan: DMA channel pointer
  575. */
  576. static void zynqmp_dma_free_chan_resources(struct dma_chan *dchan)
  577. {
  578. struct zynqmp_dma_chan *chan = to_chan(dchan);
  579. spin_lock_bh(&chan->lock);
  580. zynqmp_dma_free_descriptors(chan);
  581. spin_unlock_bh(&chan->lock);
  582. dma_free_coherent(chan->dev,
  583. (2 * ZYNQMP_DMA_DESC_SIZE(chan) * ZYNQMP_DMA_NUM_DESCS),
  584. chan->desc_pool_v, chan->desc_pool_p);
  585. kfree(chan->sw_desc_pool);
  586. }
  587. /**
  588. * zynqmp_dma_reset - Reset the channel
  589. * @chan: ZynqMP DMA channel pointer
  590. */
  591. static void zynqmp_dma_reset(struct zynqmp_dma_chan *chan)
  592. {
  593. writel(ZYNQMP_DMA_IDS_DEFAULT_MASK, chan->regs + ZYNQMP_DMA_IDS);
  594. zynqmp_dma_complete_descriptor(chan);
  595. zynqmp_dma_chan_desc_cleanup(chan);
  596. zynqmp_dma_free_descriptors(chan);
  597. zynqmp_dma_init(chan);
  598. }
  599. /**
  600. * zynqmp_dma_irq_handler - ZynqMP DMA Interrupt handler
  601. * @irq: IRQ number
  602. * @data: Pointer to the ZynqMP DMA channel structure
  603. *
  604. * Return: IRQ_HANDLED/IRQ_NONE
  605. */
  606. static irqreturn_t zynqmp_dma_irq_handler(int irq, void *data)
  607. {
  608. struct zynqmp_dma_chan *chan = (struct zynqmp_dma_chan *)data;
  609. u32 isr, imr, status;
  610. irqreturn_t ret = IRQ_NONE;
  611. isr = readl(chan->regs + ZYNQMP_DMA_ISR);
  612. imr = readl(chan->regs + ZYNQMP_DMA_IMR);
  613. status = isr & ~imr;
  614. writel(isr, chan->regs + ZYNQMP_DMA_ISR);
  615. if (status & ZYNQMP_DMA_INT_DONE) {
  616. tasklet_schedule(&chan->tasklet);
  617. ret = IRQ_HANDLED;
  618. }
  619. if (status & ZYNQMP_DMA_DONE)
  620. chan->idle = true;
  621. if (status & ZYNQMP_DMA_INT_ERR) {
  622. chan->err = true;
  623. tasklet_schedule(&chan->tasklet);
  624. dev_err(chan->dev, "Channel %p has errors\n", chan);
  625. ret = IRQ_HANDLED;
  626. }
  627. if (status & ZYNQMP_DMA_INT_OVRFL) {
  628. zynqmp_dma_handle_ovfl_int(chan, status);
  629. dev_info(chan->dev, "Channel %p overflow interrupt\n", chan);
  630. ret = IRQ_HANDLED;
  631. }
  632. return ret;
  633. }
  634. /**
  635. * zynqmp_dma_do_tasklet - Schedule completion tasklet
  636. * @data: Pointer to the ZynqMP DMA channel structure
  637. */
  638. static void zynqmp_dma_do_tasklet(unsigned long data)
  639. {
  640. struct zynqmp_dma_chan *chan = (struct zynqmp_dma_chan *)data;
  641. u32 count;
  642. spin_lock(&chan->lock);
  643. if (chan->err) {
  644. zynqmp_dma_reset(chan);
  645. chan->err = false;
  646. goto unlock;
  647. }
  648. count = readl(chan->regs + ZYNQMP_DMA_IRQ_DST_ACCT);
  649. while (count) {
  650. zynqmp_dma_complete_descriptor(chan);
  651. zynqmp_dma_chan_desc_cleanup(chan);
  652. count--;
  653. }
  654. if (chan->idle)
  655. zynqmp_dma_start_transfer(chan);
  656. unlock:
  657. spin_unlock(&chan->lock);
  658. }
  659. /**
  660. * zynqmp_dma_device_terminate_all - Aborts all transfers on a channel
  661. * @dchan: DMA channel pointer
  662. *
  663. * Return: Always '0'
  664. */
  665. static int zynqmp_dma_device_terminate_all(struct dma_chan *dchan)
  666. {
  667. struct zynqmp_dma_chan *chan = to_chan(dchan);
  668. spin_lock_bh(&chan->lock);
  669. writel(ZYNQMP_DMA_IDS_DEFAULT_MASK, chan->regs + ZYNQMP_DMA_IDS);
  670. zynqmp_dma_free_descriptors(chan);
  671. spin_unlock_bh(&chan->lock);
  672. return 0;
  673. }
  674. /**
  675. * zynqmp_dma_prep_memcpy - prepare descriptors for memcpy transaction
  676. * @dchan: DMA channel
  677. * @dma_dst: Destination buffer address
  678. * @dma_src: Source buffer address
  679. * @len: Transfer length
  680. * @flags: transfer ack flags
  681. *
  682. * Return: Async transaction descriptor on success and NULL on failure
  683. */
  684. static struct dma_async_tx_descriptor *zynqmp_dma_prep_memcpy(
  685. struct dma_chan *dchan, dma_addr_t dma_dst,
  686. dma_addr_t dma_src, size_t len, ulong flags)
  687. {
  688. struct zynqmp_dma_chan *chan;
  689. struct zynqmp_dma_desc_sw *new, *first = NULL;
  690. void *desc = NULL, *prev = NULL;
  691. size_t copy;
  692. u32 desc_cnt;
  693. chan = to_chan(dchan);
  694. desc_cnt = DIV_ROUND_UP(len, ZYNQMP_DMA_MAX_TRANS_LEN);
  695. spin_lock_bh(&chan->lock);
  696. if (desc_cnt > chan->desc_free_cnt) {
  697. spin_unlock_bh(&chan->lock);
  698. dev_dbg(chan->dev, "chan %p descs are not available\n", chan);
  699. return NULL;
  700. }
  701. chan->desc_free_cnt = chan->desc_free_cnt - desc_cnt;
  702. spin_unlock_bh(&chan->lock);
  703. do {
  704. /* Allocate and populate the descriptor */
  705. new = zynqmp_dma_get_descriptor(chan);
  706. copy = min_t(size_t, len, ZYNQMP_DMA_MAX_TRANS_LEN);
  707. desc = (struct zynqmp_dma_desc_ll *)new->src_v;
  708. zynqmp_dma_config_sg_ll_desc(chan, desc, dma_src,
  709. dma_dst, copy, prev);
  710. prev = desc;
  711. len -= copy;
  712. dma_src += copy;
  713. dma_dst += copy;
  714. if (!first)
  715. first = new;
  716. else
  717. list_add_tail(&new->node, &first->tx_list);
  718. } while (len);
  719. zynqmp_dma_desc_config_eod(chan, desc);
  720. async_tx_ack(&first->async_tx);
  721. first->async_tx.flags = flags;
  722. return &first->async_tx;
  723. }
  724. /**
  725. * zynqmp_dma_chan_remove - Channel remove function
  726. * @chan: ZynqMP DMA channel pointer
  727. */
  728. static void zynqmp_dma_chan_remove(struct zynqmp_dma_chan *chan)
  729. {
  730. if (!chan)
  731. return;
  732. if (chan->irq)
  733. devm_free_irq(chan->zdev->dev, chan->irq, chan);
  734. tasklet_kill(&chan->tasklet);
  735. list_del(&chan->common.device_node);
  736. clk_disable_unprepare(chan->clk_apb);
  737. clk_disable_unprepare(chan->clk_main);
  738. }
  739. /**
  740. * zynqmp_dma_chan_probe - Per Channel Probing
  741. * @zdev: Driver specific device structure
  742. * @pdev: Pointer to the platform_device structure
  743. *
  744. * Return: '0' on success and failure value on error
  745. */
  746. static int zynqmp_dma_chan_probe(struct zynqmp_dma_device *zdev,
  747. struct platform_device *pdev)
  748. {
  749. struct zynqmp_dma_chan *chan;
  750. struct resource *res;
  751. struct device_node *node = pdev->dev.of_node;
  752. int err;
  753. chan = devm_kzalloc(zdev->dev, sizeof(*chan), GFP_KERNEL);
  754. if (!chan)
  755. return -ENOMEM;
  756. chan->dev = zdev->dev;
  757. chan->zdev = zdev;
  758. res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
  759. chan->regs = devm_ioremap_resource(&pdev->dev, res);
  760. if (IS_ERR(chan->regs))
  761. return PTR_ERR(chan->regs);
  762. chan->bus_width = ZYNQMP_DMA_BUS_WIDTH_64;
  763. chan->dst_burst_len = ZYNQMP_DMA_MAX_DST_BURST_LEN;
  764. chan->src_burst_len = ZYNQMP_DMA_MAX_SRC_BURST_LEN;
  765. err = of_property_read_u32(node, "xlnx,bus-width", &chan->bus_width);
  766. if (err < 0) {
  767. dev_err(&pdev->dev, "missing xlnx,bus-width property\n");
  768. return err;
  769. }
  770. if (chan->bus_width != ZYNQMP_DMA_BUS_WIDTH_64 &&
  771. chan->bus_width != ZYNQMP_DMA_BUS_WIDTH_128) {
  772. dev_err(zdev->dev, "invalid bus-width value");
  773. return -EINVAL;
  774. }
  775. chan->is_dmacoherent = of_property_read_bool(node, "dma-coherent");
  776. zdev->chan = chan;
  777. tasklet_init(&chan->tasklet, zynqmp_dma_do_tasklet, (ulong)chan);
  778. spin_lock_init(&chan->lock);
  779. INIT_LIST_HEAD(&chan->active_list);
  780. INIT_LIST_HEAD(&chan->pending_list);
  781. INIT_LIST_HEAD(&chan->done_list);
  782. INIT_LIST_HEAD(&chan->free_list);
  783. dma_cookie_init(&chan->common);
  784. chan->common.device = &zdev->common;
  785. list_add_tail(&chan->common.device_node, &zdev->common.channels);
  786. zynqmp_dma_init(chan);
  787. chan->irq = platform_get_irq(pdev, 0);
  788. if (chan->irq < 0)
  789. return -ENXIO;
  790. err = devm_request_irq(&pdev->dev, chan->irq, zynqmp_dma_irq_handler, 0,
  791. "zynqmp-dma", chan);
  792. if (err)
  793. return err;
  794. chan->clk_main = devm_clk_get(&pdev->dev, "clk_main");
  795. if (IS_ERR(chan->clk_main)) {
  796. dev_err(&pdev->dev, "main clock not found.\n");
  797. return PTR_ERR(chan->clk_main);
  798. }
  799. chan->clk_apb = devm_clk_get(&pdev->dev, "clk_apb");
  800. if (IS_ERR(chan->clk_apb)) {
  801. dev_err(&pdev->dev, "apb clock not found.\n");
  802. return PTR_ERR(chan->clk_apb);
  803. }
  804. err = clk_prepare_enable(chan->clk_main);
  805. if (err) {
  806. dev_err(&pdev->dev, "Unable to enable main clock.\n");
  807. return err;
  808. }
  809. err = clk_prepare_enable(chan->clk_apb);
  810. if (err) {
  811. clk_disable_unprepare(chan->clk_main);
  812. dev_err(&pdev->dev, "Unable to enable apb clock.\n");
  813. return err;
  814. }
  815. chan->desc_size = sizeof(struct zynqmp_dma_desc_ll);
  816. chan->idle = true;
  817. return 0;
  818. }
  819. /**
  820. * of_zynqmp_dma_xlate - Translation function
  821. * @dma_spec: Pointer to DMA specifier as found in the device tree
  822. * @ofdma: Pointer to DMA controller data
  823. *
  824. * Return: DMA channel pointer on success and NULL on error
  825. */
  826. static struct dma_chan *of_zynqmp_dma_xlate(struct of_phandle_args *dma_spec,
  827. struct of_dma *ofdma)
  828. {
  829. struct zynqmp_dma_device *zdev = ofdma->of_dma_data;
  830. return dma_get_slave_channel(&zdev->chan->common);
  831. }
  832. /**
  833. * zynqmp_dma_probe - Driver probe function
  834. * @pdev: Pointer to the platform_device structure
  835. *
  836. * Return: '0' on success and failure value on error
  837. */
  838. static int zynqmp_dma_probe(struct platform_device *pdev)
  839. {
  840. struct zynqmp_dma_device *zdev;
  841. struct dma_device *p;
  842. int ret;
  843. zdev = devm_kzalloc(&pdev->dev, sizeof(*zdev), GFP_KERNEL);
  844. if (!zdev)
  845. return -ENOMEM;
  846. zdev->dev = &pdev->dev;
  847. INIT_LIST_HEAD(&zdev->common.channels);
  848. dma_set_mask(&pdev->dev, DMA_BIT_MASK(44));
  849. dma_cap_set(DMA_MEMCPY, zdev->common.cap_mask);
  850. p = &zdev->common;
  851. p->device_prep_dma_memcpy = zynqmp_dma_prep_memcpy;
  852. p->device_terminate_all = zynqmp_dma_device_terminate_all;
  853. p->device_issue_pending = zynqmp_dma_issue_pending;
  854. p->device_alloc_chan_resources = zynqmp_dma_alloc_chan_resources;
  855. p->device_free_chan_resources = zynqmp_dma_free_chan_resources;
  856. p->device_tx_status = dma_cookie_status;
  857. p->device_config = zynqmp_dma_device_config;
  858. p->dev = &pdev->dev;
  859. platform_set_drvdata(pdev, zdev);
  860. ret = zynqmp_dma_chan_probe(zdev, pdev);
  861. if (ret) {
  862. dev_err(&pdev->dev, "Probing channel failed\n");
  863. goto free_chan_resources;
  864. }
  865. p->dst_addr_widths = BIT(zdev->chan->bus_width / 8);
  866. p->src_addr_widths = BIT(zdev->chan->bus_width / 8);
  867. dma_async_device_register(&zdev->common);
  868. ret = of_dma_controller_register(pdev->dev.of_node,
  869. of_zynqmp_dma_xlate, zdev);
  870. if (ret) {
  871. dev_err(&pdev->dev, "Unable to register DMA to DT\n");
  872. dma_async_device_unregister(&zdev->common);
  873. goto free_chan_resources;
  874. }
  875. dev_info(&pdev->dev, "ZynqMP DMA driver Probe success\n");
  876. return 0;
  877. free_chan_resources:
  878. zynqmp_dma_chan_remove(zdev->chan);
  879. return ret;
  880. }
  881. /**
  882. * zynqmp_dma_remove - Driver remove function
  883. * @pdev: Pointer to the platform_device structure
  884. *
  885. * Return: Always '0'
  886. */
  887. static int zynqmp_dma_remove(struct platform_device *pdev)
  888. {
  889. struct zynqmp_dma_device *zdev = platform_get_drvdata(pdev);
  890. of_dma_controller_free(pdev->dev.of_node);
  891. dma_async_device_unregister(&zdev->common);
  892. zynqmp_dma_chan_remove(zdev->chan);
  893. return 0;
  894. }
  895. static const struct of_device_id zynqmp_dma_of_match[] = {
  896. { .compatible = "xlnx,zynqmp-dma-1.0", },
  897. {}
  898. };
  899. MODULE_DEVICE_TABLE(of, zynqmp_dma_of_match);
  900. static struct platform_driver zynqmp_dma_driver = {
  901. .driver = {
  902. .name = "xilinx-zynqmp-dma",
  903. .of_match_table = zynqmp_dma_of_match,
  904. },
  905. .probe = zynqmp_dma_probe,
  906. .remove = zynqmp_dma_remove,
  907. };
  908. module_platform_driver(zynqmp_dma_driver);
  909. MODULE_LICENSE("GPL");
  910. MODULE_AUTHOR("Xilinx, Inc.");
  911. MODULE_DESCRIPTION("Xilinx ZynqMP DMA driver");