zynqmp_dma.c 31 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161
  1. /*
  2. * DMA driver for Xilinx ZynqMP DMA Engine
  3. *
  4. * Copyright (C) 2016 Xilinx, Inc. All rights reserved.
  5. *
  6. * This program is free software: you can redistribute it and/or modify
  7. * it under the terms of the GNU General Public License as published by
  8. * the Free Software Foundation, either version 2 of the License, or
  9. * (at your option) any later version.
  10. */
  11. #include <linux/bitops.h>
  12. #include <linux/dmapool.h>
  13. #include <linux/dma/xilinx_dma.h>
  14. #include <linux/init.h>
  15. #include <linux/interrupt.h>
  16. #include <linux/io.h>
  17. #include <linux/module.h>
  18. #include <linux/of_address.h>
  19. #include <linux/of_dma.h>
  20. #include <linux/of_irq.h>
  21. #include <linux/of_platform.h>
  22. #include <linux/slab.h>
  23. #include <linux/clk.h>
  24. #include <linux/io-64-nonatomic-lo-hi.h>
  25. #include <linux/pm_runtime.h>
  26. #include "../dmaengine.h"
  27. /* Register Offsets */
  28. #define ZYNQMP_DMA_ISR 0x100
  29. #define ZYNQMP_DMA_IMR 0x104
  30. #define ZYNQMP_DMA_IER 0x108
  31. #define ZYNQMP_DMA_IDS 0x10C
  32. #define ZYNQMP_DMA_CTRL0 0x110
  33. #define ZYNQMP_DMA_CTRL1 0x114
  34. #define ZYNQMP_DMA_DATA_ATTR 0x120
  35. #define ZYNQMP_DMA_DSCR_ATTR 0x124
  36. #define ZYNQMP_DMA_SRC_DSCR_WRD0 0x128
  37. #define ZYNQMP_DMA_SRC_DSCR_WRD1 0x12C
  38. #define ZYNQMP_DMA_SRC_DSCR_WRD2 0x130
  39. #define ZYNQMP_DMA_SRC_DSCR_WRD3 0x134
  40. #define ZYNQMP_DMA_DST_DSCR_WRD0 0x138
  41. #define ZYNQMP_DMA_DST_DSCR_WRD1 0x13C
  42. #define ZYNQMP_DMA_DST_DSCR_WRD2 0x140
  43. #define ZYNQMP_DMA_DST_DSCR_WRD3 0x144
  44. #define ZYNQMP_DMA_SRC_START_LSB 0x158
  45. #define ZYNQMP_DMA_SRC_START_MSB 0x15C
  46. #define ZYNQMP_DMA_DST_START_LSB 0x160
  47. #define ZYNQMP_DMA_DST_START_MSB 0x164
  48. #define ZYNQMP_DMA_TOTAL_BYTE 0x188
  49. #define ZYNQMP_DMA_RATE_CTRL 0x18C
  50. #define ZYNQMP_DMA_IRQ_SRC_ACCT 0x190
  51. #define ZYNQMP_DMA_IRQ_DST_ACCT 0x194
  52. #define ZYNQMP_DMA_CTRL2 0x200
  53. /* Interrupt registers bit field definitions */
  54. #define ZYNQMP_DMA_DONE BIT(10)
  55. #define ZYNQMP_DMA_AXI_WR_DATA BIT(9)
  56. #define ZYNQMP_DMA_AXI_RD_DATA BIT(8)
  57. #define ZYNQMP_DMA_AXI_RD_DST_DSCR BIT(7)
  58. #define ZYNQMP_DMA_AXI_RD_SRC_DSCR BIT(6)
  59. #define ZYNQMP_DMA_IRQ_DST_ACCT_ERR BIT(5)
  60. #define ZYNQMP_DMA_IRQ_SRC_ACCT_ERR BIT(4)
  61. #define ZYNQMP_DMA_BYTE_CNT_OVRFL BIT(3)
  62. #define ZYNQMP_DMA_DST_DSCR_DONE BIT(2)
  63. #define ZYNQMP_DMA_INV_APB BIT(0)
  64. /* Control 0 register bit field definitions */
  65. #define ZYNQMP_DMA_OVR_FETCH BIT(7)
  66. #define ZYNQMP_DMA_POINT_TYPE_SG BIT(6)
  67. #define ZYNQMP_DMA_RATE_CTRL_EN BIT(3)
  68. /* Control 1 register bit field definitions */
  69. #define ZYNQMP_DMA_SRC_ISSUE GENMASK(4, 0)
  70. /* Data Attribute register bit field definitions */
  71. #define ZYNQMP_DMA_ARBURST GENMASK(27, 26)
  72. #define ZYNQMP_DMA_ARCACHE GENMASK(25, 22)
  73. #define ZYNQMP_DMA_ARCACHE_OFST 22
  74. #define ZYNQMP_DMA_ARQOS GENMASK(21, 18)
  75. #define ZYNQMP_DMA_ARQOS_OFST 18
  76. #define ZYNQMP_DMA_ARLEN GENMASK(17, 14)
  77. #define ZYNQMP_DMA_ARLEN_OFST 14
  78. #define ZYNQMP_DMA_AWBURST GENMASK(13, 12)
  79. #define ZYNQMP_DMA_AWCACHE GENMASK(11, 8)
  80. #define ZYNQMP_DMA_AWCACHE_OFST 8
  81. #define ZYNQMP_DMA_AWQOS GENMASK(7, 4)
  82. #define ZYNQMP_DMA_AWQOS_OFST 4
  83. #define ZYNQMP_DMA_AWLEN GENMASK(3, 0)
  84. #define ZYNQMP_DMA_AWLEN_OFST 0
  85. /* Descriptor Attribute register bit field definitions */
  86. #define ZYNQMP_DMA_AXCOHRNT BIT(8)
  87. #define ZYNQMP_DMA_AXCACHE GENMASK(7, 4)
  88. #define ZYNQMP_DMA_AXCACHE_OFST 4
  89. #define ZYNQMP_DMA_AXQOS GENMASK(3, 0)
  90. #define ZYNQMP_DMA_AXQOS_OFST 0
  91. /* Control register 2 bit field definitions */
  92. #define ZYNQMP_DMA_ENABLE BIT(0)
  93. /* Buffer Descriptor definitions */
  94. #define ZYNQMP_DMA_DESC_CTRL_STOP 0x10
  95. #define ZYNQMP_DMA_DESC_CTRL_COMP_INT 0x4
  96. #define ZYNQMP_DMA_DESC_CTRL_SIZE_256 0x2
  97. #define ZYNQMP_DMA_DESC_CTRL_COHRNT 0x1
  98. /* Interrupt Mask specific definitions */
  99. #define ZYNQMP_DMA_INT_ERR (ZYNQMP_DMA_AXI_RD_DATA | \
  100. ZYNQMP_DMA_AXI_WR_DATA | \
  101. ZYNQMP_DMA_AXI_RD_DST_DSCR | \
  102. ZYNQMP_DMA_AXI_RD_SRC_DSCR | \
  103. ZYNQMP_DMA_INV_APB)
  104. #define ZYNQMP_DMA_INT_OVRFL (ZYNQMP_DMA_BYTE_CNT_OVRFL | \
  105. ZYNQMP_DMA_IRQ_SRC_ACCT_ERR | \
  106. ZYNQMP_DMA_IRQ_DST_ACCT_ERR)
  107. #define ZYNQMP_DMA_INT_DONE (ZYNQMP_DMA_DONE | ZYNQMP_DMA_DST_DSCR_DONE)
  108. #define ZYNQMP_DMA_INT_EN_DEFAULT_MASK (ZYNQMP_DMA_INT_DONE | \
  109. ZYNQMP_DMA_INT_ERR | \
  110. ZYNQMP_DMA_INT_OVRFL | \
  111. ZYNQMP_DMA_DST_DSCR_DONE)
  112. /* Max number of descriptors per channel */
  113. #define ZYNQMP_DMA_NUM_DESCS 32
  114. /* Max transfer size per descriptor */
  115. #define ZYNQMP_DMA_MAX_TRANS_LEN 0x40000000
  116. /* Max burst lengths */
  117. #define ZYNQMP_DMA_MAX_DST_BURST_LEN 32768U
  118. #define ZYNQMP_DMA_MAX_SRC_BURST_LEN 32768U
  119. /* Reset values for data attributes */
  120. #define ZYNQMP_DMA_AXCACHE_VAL 0xF
  121. #define ZYNQMP_DMA_SRC_ISSUE_RST_VAL 0x1F
  122. #define ZYNQMP_DMA_IDS_DEFAULT_MASK 0xFFF
  123. /* Bus width in bits */
  124. #define ZYNQMP_DMA_BUS_WIDTH_64 64
  125. #define ZYNQMP_DMA_BUS_WIDTH_128 128
  126. #define ZDMA_PM_TIMEOUT 100
  127. #define ZYNQMP_DMA_DESC_SIZE(chan) (chan->desc_size)
  128. #define to_chan(chan) container_of(chan, struct zynqmp_dma_chan, \
  129. common)
  130. #define tx_to_desc(tx) container_of(tx, struct zynqmp_dma_desc_sw, \
  131. async_tx)
  132. /**
  133. * struct zynqmp_dma_desc_ll - Hw linked list descriptor
  134. * @addr: Buffer address
  135. * @size: Size of the buffer
  136. * @ctrl: Control word
  137. * @nxtdscraddr: Next descriptor base address
  138. * @rsvd: Reserved field and for Hw internal use.
  139. */
  140. struct zynqmp_dma_desc_ll {
  141. u64 addr;
  142. u32 size;
  143. u32 ctrl;
  144. u64 nxtdscraddr;
  145. u64 rsvd;
  146. };
  147. /**
  148. * struct zynqmp_dma_desc_sw - Per Transaction structure
  149. * @src: Source address for simple mode dma
  150. * @dst: Destination address for simple mode dma
  151. * @len: Transfer length for simple mode dma
  152. * @node: Node in the channel descriptor list
  153. * @tx_list: List head for the current transfer
  154. * @async_tx: Async transaction descriptor
  155. * @src_v: Virtual address of the src descriptor
  156. * @src_p: Physical address of the src descriptor
  157. * @dst_v: Virtual address of the dst descriptor
  158. * @dst_p: Physical address of the dst descriptor
  159. */
  160. struct zynqmp_dma_desc_sw {
  161. u64 src;
  162. u64 dst;
  163. u32 len;
  164. struct list_head node;
  165. struct list_head tx_list;
  166. struct dma_async_tx_descriptor async_tx;
  167. struct zynqmp_dma_desc_ll *src_v;
  168. dma_addr_t src_p;
  169. struct zynqmp_dma_desc_ll *dst_v;
  170. dma_addr_t dst_p;
  171. };
  172. /**
  173. * struct zynqmp_dma_chan - Driver specific DMA channel structure
  174. * @zdev: Driver specific device structure
  175. * @regs: Control registers offset
  176. * @lock: Descriptor operation lock
  177. * @pending_list: Descriptors waiting
  178. * @free_list: Descriptors free
  179. * @active_list: Descriptors active
  180. * @sw_desc_pool: SW descriptor pool
  181. * @done_list: Complete descriptors
  182. * @common: DMA common channel
  183. * @desc_pool_v: Statically allocated descriptor base
  184. * @desc_pool_p: Physical allocated descriptor base
  185. * @desc_free_cnt: Descriptor available count
  186. * @dev: The dma device
  187. * @irq: Channel IRQ
  188. * @is_dmacoherent: Tells whether dma operations are coherent or not
  189. * @tasklet: Cleanup work after irq
  190. * @idle : Channel status;
  191. * @desc_size: Size of the low level descriptor
  192. * @err: Channel has errors
  193. * @bus_width: Bus width
  194. * @src_burst_len: Source burst length
  195. * @dst_burst_len: Dest burst length
  196. */
  197. struct zynqmp_dma_chan {
  198. struct zynqmp_dma_device *zdev;
  199. void __iomem *regs;
  200. spinlock_t lock;
  201. struct list_head pending_list;
  202. struct list_head free_list;
  203. struct list_head active_list;
  204. struct zynqmp_dma_desc_sw *sw_desc_pool;
  205. struct list_head done_list;
  206. struct dma_chan common;
  207. void *desc_pool_v;
  208. dma_addr_t desc_pool_p;
  209. u32 desc_free_cnt;
  210. struct device *dev;
  211. int irq;
  212. bool is_dmacoherent;
  213. struct tasklet_struct tasklet;
  214. bool idle;
  215. u32 desc_size;
  216. bool err;
  217. u32 bus_width;
  218. u32 src_burst_len;
  219. u32 dst_burst_len;
  220. };
  221. /**
  222. * struct zynqmp_dma_device - DMA device structure
  223. * @dev: Device Structure
  224. * @common: DMA device structure
  225. * @chan: Driver specific DMA channel
  226. * @clk_main: Pointer to main clock
  227. * @clk_apb: Pointer to apb clock
  228. */
  229. struct zynqmp_dma_device {
  230. struct device *dev;
  231. struct dma_device common;
  232. struct zynqmp_dma_chan *chan;
  233. struct clk *clk_main;
  234. struct clk *clk_apb;
  235. };
  236. static inline void zynqmp_dma_writeq(struct zynqmp_dma_chan *chan, u32 reg,
  237. u64 value)
  238. {
  239. lo_hi_writeq(value, chan->regs + reg);
  240. }
  241. /**
  242. * zynqmp_dma_update_desc_to_ctrlr - Updates descriptor to the controller
  243. * @chan: ZynqMP DMA DMA channel pointer
  244. * @desc: Transaction descriptor pointer
  245. */
  246. static void zynqmp_dma_update_desc_to_ctrlr(struct zynqmp_dma_chan *chan,
  247. struct zynqmp_dma_desc_sw *desc)
  248. {
  249. dma_addr_t addr;
  250. addr = desc->src_p;
  251. zynqmp_dma_writeq(chan, ZYNQMP_DMA_SRC_START_LSB, addr);
  252. addr = desc->dst_p;
  253. zynqmp_dma_writeq(chan, ZYNQMP_DMA_DST_START_LSB, addr);
  254. }
  255. /**
  256. * zynqmp_dma_desc_config_eod - Mark the descriptor as end descriptor
  257. * @chan: ZynqMP DMA channel pointer
  258. * @desc: Hw descriptor pointer
  259. */
  260. static void zynqmp_dma_desc_config_eod(struct zynqmp_dma_chan *chan,
  261. void *desc)
  262. {
  263. struct zynqmp_dma_desc_ll *hw = (struct zynqmp_dma_desc_ll *)desc;
  264. hw->ctrl |= ZYNQMP_DMA_DESC_CTRL_STOP;
  265. hw++;
  266. hw->ctrl |= ZYNQMP_DMA_DESC_CTRL_COMP_INT | ZYNQMP_DMA_DESC_CTRL_STOP;
  267. }
  268. /**
  269. * zynqmp_dma_config_sg_ll_desc - Configure the linked list descriptor
  270. * @chan: ZynqMP DMA channel pointer
  271. * @sdesc: Hw descriptor pointer
  272. * @src: Source buffer address
  273. * @dst: Destination buffer address
  274. * @len: Transfer length
  275. * @prev: Previous hw descriptor pointer
  276. */
  277. static void zynqmp_dma_config_sg_ll_desc(struct zynqmp_dma_chan *chan,
  278. struct zynqmp_dma_desc_ll *sdesc,
  279. dma_addr_t src, dma_addr_t dst, size_t len,
  280. struct zynqmp_dma_desc_ll *prev)
  281. {
  282. struct zynqmp_dma_desc_ll *ddesc = sdesc + 1;
  283. sdesc->size = ddesc->size = len;
  284. sdesc->addr = src;
  285. ddesc->addr = dst;
  286. sdesc->ctrl = ddesc->ctrl = ZYNQMP_DMA_DESC_CTRL_SIZE_256;
  287. if (chan->is_dmacoherent) {
  288. sdesc->ctrl |= ZYNQMP_DMA_DESC_CTRL_COHRNT;
  289. ddesc->ctrl |= ZYNQMP_DMA_DESC_CTRL_COHRNT;
  290. }
  291. if (prev) {
  292. dma_addr_t addr = chan->desc_pool_p +
  293. ((uintptr_t)sdesc - (uintptr_t)chan->desc_pool_v);
  294. ddesc = prev + 1;
  295. prev->nxtdscraddr = addr;
  296. ddesc->nxtdscraddr = addr + ZYNQMP_DMA_DESC_SIZE(chan);
  297. }
  298. }
  299. /**
  300. * zynqmp_dma_init - Initialize the channel
  301. * @chan: ZynqMP DMA channel pointer
  302. */
  303. static void zynqmp_dma_init(struct zynqmp_dma_chan *chan)
  304. {
  305. u32 val;
  306. writel(ZYNQMP_DMA_IDS_DEFAULT_MASK, chan->regs + ZYNQMP_DMA_IDS);
  307. val = readl(chan->regs + ZYNQMP_DMA_ISR);
  308. writel(val, chan->regs + ZYNQMP_DMA_ISR);
  309. if (chan->is_dmacoherent) {
  310. val = ZYNQMP_DMA_AXCOHRNT;
  311. val = (val & ~ZYNQMP_DMA_AXCACHE) |
  312. (ZYNQMP_DMA_AXCACHE_VAL << ZYNQMP_DMA_AXCACHE_OFST);
  313. writel(val, chan->regs + ZYNQMP_DMA_DSCR_ATTR);
  314. }
  315. val = readl(chan->regs + ZYNQMP_DMA_DATA_ATTR);
  316. if (chan->is_dmacoherent) {
  317. val = (val & ~ZYNQMP_DMA_ARCACHE) |
  318. (ZYNQMP_DMA_AXCACHE_VAL << ZYNQMP_DMA_ARCACHE_OFST);
  319. val = (val & ~ZYNQMP_DMA_AWCACHE) |
  320. (ZYNQMP_DMA_AXCACHE_VAL << ZYNQMP_DMA_AWCACHE_OFST);
  321. }
  322. writel(val, chan->regs + ZYNQMP_DMA_DATA_ATTR);
  323. /* Clearing the interrupt account rgisters */
  324. val = readl(chan->regs + ZYNQMP_DMA_IRQ_SRC_ACCT);
  325. val = readl(chan->regs + ZYNQMP_DMA_IRQ_DST_ACCT);
  326. chan->idle = true;
  327. }
  328. /**
  329. * zynqmp_dma_tx_submit - Submit DMA transaction
  330. * @tx: Async transaction descriptor pointer
  331. *
  332. * Return: cookie value
  333. */
  334. static dma_cookie_t zynqmp_dma_tx_submit(struct dma_async_tx_descriptor *tx)
  335. {
  336. struct zynqmp_dma_chan *chan = to_chan(tx->chan);
  337. struct zynqmp_dma_desc_sw *desc, *new;
  338. dma_cookie_t cookie;
  339. new = tx_to_desc(tx);
  340. spin_lock_bh(&chan->lock);
  341. cookie = dma_cookie_assign(tx);
  342. if (!list_empty(&chan->pending_list)) {
  343. desc = list_last_entry(&chan->pending_list,
  344. struct zynqmp_dma_desc_sw, node);
  345. if (!list_empty(&desc->tx_list))
  346. desc = list_last_entry(&desc->tx_list,
  347. struct zynqmp_dma_desc_sw, node);
  348. desc->src_v->nxtdscraddr = new->src_p;
  349. desc->src_v->ctrl &= ~ZYNQMP_DMA_DESC_CTRL_STOP;
  350. desc->dst_v->nxtdscraddr = new->dst_p;
  351. desc->dst_v->ctrl &= ~ZYNQMP_DMA_DESC_CTRL_STOP;
  352. }
  353. list_add_tail(&new->node, &chan->pending_list);
  354. spin_unlock_bh(&chan->lock);
  355. return cookie;
  356. }
  357. /**
  358. * zynqmp_dma_get_descriptor - Get the sw descriptor from the pool
  359. * @chan: ZynqMP DMA channel pointer
  360. *
  361. * Return: The sw descriptor
  362. */
  363. static struct zynqmp_dma_desc_sw *
  364. zynqmp_dma_get_descriptor(struct zynqmp_dma_chan *chan)
  365. {
  366. struct zynqmp_dma_desc_sw *desc;
  367. spin_lock_bh(&chan->lock);
  368. desc = list_first_entry(&chan->free_list,
  369. struct zynqmp_dma_desc_sw, node);
  370. list_del(&desc->node);
  371. spin_unlock_bh(&chan->lock);
  372. INIT_LIST_HEAD(&desc->tx_list);
  373. /* Clear the src and dst descriptor memory */
  374. memset((void *)desc->src_v, 0, ZYNQMP_DMA_DESC_SIZE(chan));
  375. memset((void *)desc->dst_v, 0, ZYNQMP_DMA_DESC_SIZE(chan));
  376. return desc;
  377. }
  378. /**
  379. * zynqmp_dma_free_descriptor - Issue pending transactions
  380. * @chan: ZynqMP DMA channel pointer
  381. * @sdesc: Transaction descriptor pointer
  382. */
  383. static void zynqmp_dma_free_descriptor(struct zynqmp_dma_chan *chan,
  384. struct zynqmp_dma_desc_sw *sdesc)
  385. {
  386. struct zynqmp_dma_desc_sw *child, *next;
  387. chan->desc_free_cnt++;
  388. list_add_tail(&sdesc->node, &chan->free_list);
  389. list_for_each_entry_safe(child, next, &sdesc->tx_list, node) {
  390. chan->desc_free_cnt++;
  391. list_move_tail(&child->node, &chan->free_list);
  392. }
  393. }
  394. /**
  395. * zynqmp_dma_free_desc_list - Free descriptors list
  396. * @chan: ZynqMP DMA channel pointer
  397. * @list: List to parse and delete the descriptor
  398. */
  399. static void zynqmp_dma_free_desc_list(struct zynqmp_dma_chan *chan,
  400. struct list_head *list)
  401. {
  402. struct zynqmp_dma_desc_sw *desc, *next;
  403. list_for_each_entry_safe(desc, next, list, node)
  404. zynqmp_dma_free_descriptor(chan, desc);
  405. }
  406. /**
  407. * zynqmp_dma_alloc_chan_resources - Allocate channel resources
  408. * @dchan: DMA channel
  409. *
  410. * Return: Number of descriptors on success and failure value on error
  411. */
  412. static int zynqmp_dma_alloc_chan_resources(struct dma_chan *dchan)
  413. {
  414. struct zynqmp_dma_chan *chan = to_chan(dchan);
  415. struct zynqmp_dma_desc_sw *desc;
  416. int i, ret;
  417. ret = pm_runtime_get_sync(chan->dev);
  418. if (ret < 0)
  419. return ret;
  420. chan->sw_desc_pool = kcalloc(ZYNQMP_DMA_NUM_DESCS, sizeof(*desc),
  421. GFP_KERNEL);
  422. if (!chan->sw_desc_pool)
  423. return -ENOMEM;
  424. chan->idle = true;
  425. chan->desc_free_cnt = ZYNQMP_DMA_NUM_DESCS;
  426. INIT_LIST_HEAD(&chan->free_list);
  427. for (i = 0; i < ZYNQMP_DMA_NUM_DESCS; i++) {
  428. desc = chan->sw_desc_pool + i;
  429. dma_async_tx_descriptor_init(&desc->async_tx, &chan->common);
  430. desc->async_tx.tx_submit = zynqmp_dma_tx_submit;
  431. list_add_tail(&desc->node, &chan->free_list);
  432. }
  433. chan->desc_pool_v = dma_zalloc_coherent(chan->dev,
  434. (2 * chan->desc_size * ZYNQMP_DMA_NUM_DESCS),
  435. &chan->desc_pool_p, GFP_KERNEL);
  436. if (!chan->desc_pool_v)
  437. return -ENOMEM;
  438. for (i = 0; i < ZYNQMP_DMA_NUM_DESCS; i++) {
  439. desc = chan->sw_desc_pool + i;
  440. desc->src_v = (struct zynqmp_dma_desc_ll *) (chan->desc_pool_v +
  441. (i * ZYNQMP_DMA_DESC_SIZE(chan) * 2));
  442. desc->dst_v = (struct zynqmp_dma_desc_ll *) (desc->src_v + 1);
  443. desc->src_p = chan->desc_pool_p +
  444. (i * ZYNQMP_DMA_DESC_SIZE(chan) * 2);
  445. desc->dst_p = desc->src_p + ZYNQMP_DMA_DESC_SIZE(chan);
  446. }
  447. return ZYNQMP_DMA_NUM_DESCS;
  448. }
  449. /**
  450. * zynqmp_dma_start - Start DMA channel
  451. * @chan: ZynqMP DMA channel pointer
  452. */
  453. static void zynqmp_dma_start(struct zynqmp_dma_chan *chan)
  454. {
  455. writel(ZYNQMP_DMA_INT_EN_DEFAULT_MASK, chan->regs + ZYNQMP_DMA_IER);
  456. writel(0, chan->regs + ZYNQMP_DMA_TOTAL_BYTE);
  457. chan->idle = false;
  458. writel(ZYNQMP_DMA_ENABLE, chan->regs + ZYNQMP_DMA_CTRL2);
  459. }
  460. /**
  461. * zynqmp_dma_handle_ovfl_int - Process the overflow interrupt
  462. * @chan: ZynqMP DMA channel pointer
  463. * @status: Interrupt status value
  464. */
  465. static void zynqmp_dma_handle_ovfl_int(struct zynqmp_dma_chan *chan, u32 status)
  466. {
  467. if (status & ZYNQMP_DMA_BYTE_CNT_OVRFL)
  468. writel(0, chan->regs + ZYNQMP_DMA_TOTAL_BYTE);
  469. if (status & ZYNQMP_DMA_IRQ_DST_ACCT_ERR)
  470. readl(chan->regs + ZYNQMP_DMA_IRQ_DST_ACCT);
  471. if (status & ZYNQMP_DMA_IRQ_SRC_ACCT_ERR)
  472. readl(chan->regs + ZYNQMP_DMA_IRQ_SRC_ACCT);
  473. }
  474. static void zynqmp_dma_config(struct zynqmp_dma_chan *chan)
  475. {
  476. u32 val, burst_val;
  477. val = readl(chan->regs + ZYNQMP_DMA_CTRL0);
  478. val |= ZYNQMP_DMA_POINT_TYPE_SG;
  479. writel(val, chan->regs + ZYNQMP_DMA_CTRL0);
  480. val = readl(chan->regs + ZYNQMP_DMA_DATA_ATTR);
  481. burst_val = __ilog2_u32(chan->src_burst_len);
  482. val = (val & ~ZYNQMP_DMA_ARLEN) |
  483. ((burst_val << ZYNQMP_DMA_ARLEN_OFST) & ZYNQMP_DMA_ARLEN);
  484. burst_val = __ilog2_u32(chan->dst_burst_len);
  485. val = (val & ~ZYNQMP_DMA_AWLEN) |
  486. ((burst_val << ZYNQMP_DMA_AWLEN_OFST) & ZYNQMP_DMA_AWLEN);
  487. writel(val, chan->regs + ZYNQMP_DMA_DATA_ATTR);
  488. }
  489. /**
  490. * zynqmp_dma_device_config - Zynqmp dma device configuration
  491. * @dchan: DMA channel
  492. * @config: DMA device config
  493. *
  494. * Return: 0 always
  495. */
  496. static int zynqmp_dma_device_config(struct dma_chan *dchan,
  497. struct dma_slave_config *config)
  498. {
  499. struct zynqmp_dma_chan *chan = to_chan(dchan);
  500. chan->src_burst_len = clamp(config->src_maxburst, 1U,
  501. ZYNQMP_DMA_MAX_SRC_BURST_LEN);
  502. chan->dst_burst_len = clamp(config->dst_maxburst, 1U,
  503. ZYNQMP_DMA_MAX_DST_BURST_LEN);
  504. return 0;
  505. }
  506. /**
  507. * zynqmp_dma_start_transfer - Initiate the new transfer
  508. * @chan: ZynqMP DMA channel pointer
  509. */
  510. static void zynqmp_dma_start_transfer(struct zynqmp_dma_chan *chan)
  511. {
  512. struct zynqmp_dma_desc_sw *desc;
  513. if (!chan->idle)
  514. return;
  515. zynqmp_dma_config(chan);
  516. desc = list_first_entry_or_null(&chan->pending_list,
  517. struct zynqmp_dma_desc_sw, node);
  518. if (!desc)
  519. return;
  520. list_splice_tail_init(&chan->pending_list, &chan->active_list);
  521. zynqmp_dma_update_desc_to_ctrlr(chan, desc);
  522. zynqmp_dma_start(chan);
  523. }
  524. /**
  525. * zynqmp_dma_chan_desc_cleanup - Cleanup the completed descriptors
  526. * @chan: ZynqMP DMA channel
  527. */
  528. static void zynqmp_dma_chan_desc_cleanup(struct zynqmp_dma_chan *chan)
  529. {
  530. struct zynqmp_dma_desc_sw *desc, *next;
  531. list_for_each_entry_safe(desc, next, &chan->done_list, node) {
  532. dma_async_tx_callback callback;
  533. void *callback_param;
  534. list_del(&desc->node);
  535. callback = desc->async_tx.callback;
  536. callback_param = desc->async_tx.callback_param;
  537. if (callback) {
  538. spin_unlock(&chan->lock);
  539. callback(callback_param);
  540. spin_lock(&chan->lock);
  541. }
  542. /* Run any dependencies, then free the descriptor */
  543. zynqmp_dma_free_descriptor(chan, desc);
  544. }
  545. }
  546. /**
  547. * zynqmp_dma_complete_descriptor - Mark the active descriptor as complete
  548. * @chan: ZynqMP DMA channel pointer
  549. */
  550. static void zynqmp_dma_complete_descriptor(struct zynqmp_dma_chan *chan)
  551. {
  552. struct zynqmp_dma_desc_sw *desc;
  553. desc = list_first_entry_or_null(&chan->active_list,
  554. struct zynqmp_dma_desc_sw, node);
  555. if (!desc)
  556. return;
  557. list_del(&desc->node);
  558. dma_cookie_complete(&desc->async_tx);
  559. list_add_tail(&desc->node, &chan->done_list);
  560. }
  561. /**
  562. * zynqmp_dma_issue_pending - Issue pending transactions
  563. * @dchan: DMA channel pointer
  564. */
  565. static void zynqmp_dma_issue_pending(struct dma_chan *dchan)
  566. {
  567. struct zynqmp_dma_chan *chan = to_chan(dchan);
  568. spin_lock_bh(&chan->lock);
  569. zynqmp_dma_start_transfer(chan);
  570. spin_unlock_bh(&chan->lock);
  571. }
  572. /**
  573. * zynqmp_dma_free_descriptors - Free channel descriptors
  574. * @chan: ZynqMP DMA channel pointer
  575. */
  576. static void zynqmp_dma_free_descriptors(struct zynqmp_dma_chan *chan)
  577. {
  578. zynqmp_dma_free_desc_list(chan, &chan->active_list);
  579. zynqmp_dma_free_desc_list(chan, &chan->pending_list);
  580. zynqmp_dma_free_desc_list(chan, &chan->done_list);
  581. }
  582. /**
  583. * zynqmp_dma_free_chan_resources - Free channel resources
  584. * @dchan: DMA channel pointer
  585. */
  586. static void zynqmp_dma_free_chan_resources(struct dma_chan *dchan)
  587. {
  588. struct zynqmp_dma_chan *chan = to_chan(dchan);
  589. spin_lock_bh(&chan->lock);
  590. zynqmp_dma_free_descriptors(chan);
  591. spin_unlock_bh(&chan->lock);
  592. dma_free_coherent(chan->dev,
  593. (2 * ZYNQMP_DMA_DESC_SIZE(chan) * ZYNQMP_DMA_NUM_DESCS),
  594. chan->desc_pool_v, chan->desc_pool_p);
  595. kfree(chan->sw_desc_pool);
  596. pm_runtime_mark_last_busy(chan->dev);
  597. pm_runtime_put_autosuspend(chan->dev);
  598. }
  599. /**
  600. * zynqmp_dma_reset - Reset the channel
  601. * @chan: ZynqMP DMA channel pointer
  602. */
  603. static void zynqmp_dma_reset(struct zynqmp_dma_chan *chan)
  604. {
  605. writel(ZYNQMP_DMA_IDS_DEFAULT_MASK, chan->regs + ZYNQMP_DMA_IDS);
  606. zynqmp_dma_complete_descriptor(chan);
  607. zynqmp_dma_chan_desc_cleanup(chan);
  608. zynqmp_dma_free_descriptors(chan);
  609. zynqmp_dma_init(chan);
  610. }
  611. /**
  612. * zynqmp_dma_irq_handler - ZynqMP DMA Interrupt handler
  613. * @irq: IRQ number
  614. * @data: Pointer to the ZynqMP DMA channel structure
  615. *
  616. * Return: IRQ_HANDLED/IRQ_NONE
  617. */
  618. static irqreturn_t zynqmp_dma_irq_handler(int irq, void *data)
  619. {
  620. struct zynqmp_dma_chan *chan = (struct zynqmp_dma_chan *)data;
  621. u32 isr, imr, status;
  622. irqreturn_t ret = IRQ_NONE;
  623. isr = readl(chan->regs + ZYNQMP_DMA_ISR);
  624. imr = readl(chan->regs + ZYNQMP_DMA_IMR);
  625. status = isr & ~imr;
  626. writel(isr, chan->regs + ZYNQMP_DMA_ISR);
  627. if (status & ZYNQMP_DMA_INT_DONE) {
  628. tasklet_schedule(&chan->tasklet);
  629. ret = IRQ_HANDLED;
  630. }
  631. if (status & ZYNQMP_DMA_DONE)
  632. chan->idle = true;
  633. if (status & ZYNQMP_DMA_INT_ERR) {
  634. chan->err = true;
  635. tasklet_schedule(&chan->tasklet);
  636. dev_err(chan->dev, "Channel %p has errors\n", chan);
  637. ret = IRQ_HANDLED;
  638. }
  639. if (status & ZYNQMP_DMA_INT_OVRFL) {
  640. zynqmp_dma_handle_ovfl_int(chan, status);
  641. dev_dbg(chan->dev, "Channel %p overflow interrupt\n", chan);
  642. ret = IRQ_HANDLED;
  643. }
  644. return ret;
  645. }
  646. /**
  647. * zynqmp_dma_do_tasklet - Schedule completion tasklet
  648. * @data: Pointer to the ZynqMP DMA channel structure
  649. */
  650. static void zynqmp_dma_do_tasklet(unsigned long data)
  651. {
  652. struct zynqmp_dma_chan *chan = (struct zynqmp_dma_chan *)data;
  653. u32 count;
  654. spin_lock(&chan->lock);
  655. if (chan->err) {
  656. zynqmp_dma_reset(chan);
  657. chan->err = false;
  658. goto unlock;
  659. }
  660. count = readl(chan->regs + ZYNQMP_DMA_IRQ_DST_ACCT);
  661. while (count) {
  662. zynqmp_dma_complete_descriptor(chan);
  663. zynqmp_dma_chan_desc_cleanup(chan);
  664. count--;
  665. }
  666. if (chan->idle)
  667. zynqmp_dma_start_transfer(chan);
  668. unlock:
  669. spin_unlock(&chan->lock);
  670. }
  671. /**
  672. * zynqmp_dma_device_terminate_all - Aborts all transfers on a channel
  673. * @dchan: DMA channel pointer
  674. *
  675. * Return: Always '0'
  676. */
  677. static int zynqmp_dma_device_terminate_all(struct dma_chan *dchan)
  678. {
  679. struct zynqmp_dma_chan *chan = to_chan(dchan);
  680. spin_lock_bh(&chan->lock);
  681. writel(ZYNQMP_DMA_IDS_DEFAULT_MASK, chan->regs + ZYNQMP_DMA_IDS);
  682. zynqmp_dma_free_descriptors(chan);
  683. spin_unlock_bh(&chan->lock);
  684. return 0;
  685. }
  686. /**
  687. * zynqmp_dma_prep_memcpy - prepare descriptors for memcpy transaction
  688. * @dchan: DMA channel
  689. * @dma_dst: Destination buffer address
  690. * @dma_src: Source buffer address
  691. * @len: Transfer length
  692. * @flags: transfer ack flags
  693. *
  694. * Return: Async transaction descriptor on success and NULL on failure
  695. */
  696. static struct dma_async_tx_descriptor *zynqmp_dma_prep_memcpy(
  697. struct dma_chan *dchan, dma_addr_t dma_dst,
  698. dma_addr_t dma_src, size_t len, ulong flags)
  699. {
  700. struct zynqmp_dma_chan *chan;
  701. struct zynqmp_dma_desc_sw *new, *first = NULL;
  702. void *desc = NULL, *prev = NULL;
  703. size_t copy;
  704. u32 desc_cnt;
  705. chan = to_chan(dchan);
  706. desc_cnt = DIV_ROUND_UP(len, ZYNQMP_DMA_MAX_TRANS_LEN);
  707. spin_lock_bh(&chan->lock);
  708. if (desc_cnt > chan->desc_free_cnt) {
  709. spin_unlock_bh(&chan->lock);
  710. dev_dbg(chan->dev, "chan %p descs are not available\n", chan);
  711. return NULL;
  712. }
  713. chan->desc_free_cnt = chan->desc_free_cnt - desc_cnt;
  714. spin_unlock_bh(&chan->lock);
  715. do {
  716. /* Allocate and populate the descriptor */
  717. new = zynqmp_dma_get_descriptor(chan);
  718. copy = min_t(size_t, len, ZYNQMP_DMA_MAX_TRANS_LEN);
  719. desc = (struct zynqmp_dma_desc_ll *)new->src_v;
  720. zynqmp_dma_config_sg_ll_desc(chan, desc, dma_src,
  721. dma_dst, copy, prev);
  722. prev = desc;
  723. len -= copy;
  724. dma_src += copy;
  725. dma_dst += copy;
  726. if (!first)
  727. first = new;
  728. else
  729. list_add_tail(&new->node, &first->tx_list);
  730. } while (len);
  731. zynqmp_dma_desc_config_eod(chan, desc);
  732. async_tx_ack(&first->async_tx);
  733. first->async_tx.flags = flags;
  734. return &first->async_tx;
  735. }
  736. /**
  737. * zynqmp_dma_chan_remove - Channel remove function
  738. * @chan: ZynqMP DMA channel pointer
  739. */
  740. static void zynqmp_dma_chan_remove(struct zynqmp_dma_chan *chan)
  741. {
  742. if (!chan)
  743. return;
  744. if (chan->irq)
  745. devm_free_irq(chan->zdev->dev, chan->irq, chan);
  746. tasklet_kill(&chan->tasklet);
  747. list_del(&chan->common.device_node);
  748. }
  749. /**
  750. * zynqmp_dma_chan_probe - Per Channel Probing
  751. * @zdev: Driver specific device structure
  752. * @pdev: Pointer to the platform_device structure
  753. *
  754. * Return: '0' on success and failure value on error
  755. */
  756. static int zynqmp_dma_chan_probe(struct zynqmp_dma_device *zdev,
  757. struct platform_device *pdev)
  758. {
  759. struct zynqmp_dma_chan *chan;
  760. struct resource *res;
  761. struct device_node *node = pdev->dev.of_node;
  762. int err;
  763. chan = devm_kzalloc(zdev->dev, sizeof(*chan), GFP_KERNEL);
  764. if (!chan)
  765. return -ENOMEM;
  766. chan->dev = zdev->dev;
  767. chan->zdev = zdev;
  768. res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
  769. chan->regs = devm_ioremap_resource(&pdev->dev, res);
  770. if (IS_ERR(chan->regs))
  771. return PTR_ERR(chan->regs);
  772. chan->bus_width = ZYNQMP_DMA_BUS_WIDTH_64;
  773. chan->dst_burst_len = ZYNQMP_DMA_MAX_DST_BURST_LEN;
  774. chan->src_burst_len = ZYNQMP_DMA_MAX_SRC_BURST_LEN;
  775. err = of_property_read_u32(node, "xlnx,bus-width", &chan->bus_width);
  776. if (err < 0) {
  777. dev_err(&pdev->dev, "missing xlnx,bus-width property\n");
  778. return err;
  779. }
  780. if (chan->bus_width != ZYNQMP_DMA_BUS_WIDTH_64 &&
  781. chan->bus_width != ZYNQMP_DMA_BUS_WIDTH_128) {
  782. dev_err(zdev->dev, "invalid bus-width value");
  783. return -EINVAL;
  784. }
  785. chan->is_dmacoherent = of_property_read_bool(node, "dma-coherent");
  786. zdev->chan = chan;
  787. tasklet_init(&chan->tasklet, zynqmp_dma_do_tasklet, (ulong)chan);
  788. spin_lock_init(&chan->lock);
  789. INIT_LIST_HEAD(&chan->active_list);
  790. INIT_LIST_HEAD(&chan->pending_list);
  791. INIT_LIST_HEAD(&chan->done_list);
  792. INIT_LIST_HEAD(&chan->free_list);
  793. dma_cookie_init(&chan->common);
  794. chan->common.device = &zdev->common;
  795. list_add_tail(&chan->common.device_node, &zdev->common.channels);
  796. zynqmp_dma_init(chan);
  797. chan->irq = platform_get_irq(pdev, 0);
  798. if (chan->irq < 0)
  799. return -ENXIO;
  800. err = devm_request_irq(&pdev->dev, chan->irq, zynqmp_dma_irq_handler, 0,
  801. "zynqmp-dma", chan);
  802. if (err)
  803. return err;
  804. chan->desc_size = sizeof(struct zynqmp_dma_desc_ll);
  805. chan->idle = true;
  806. return 0;
  807. }
  808. /**
  809. * of_zynqmp_dma_xlate - Translation function
  810. * @dma_spec: Pointer to DMA specifier as found in the device tree
  811. * @ofdma: Pointer to DMA controller data
  812. *
  813. * Return: DMA channel pointer on success and NULL on error
  814. */
  815. static struct dma_chan *of_zynqmp_dma_xlate(struct of_phandle_args *dma_spec,
  816. struct of_dma *ofdma)
  817. {
  818. struct zynqmp_dma_device *zdev = ofdma->of_dma_data;
  819. return dma_get_slave_channel(&zdev->chan->common);
  820. }
  821. /**
  822. * zynqmp_dma_suspend - Suspend method for the driver
  823. * @dev: Address of the device structure
  824. *
  825. * Put the driver into low power mode.
  826. * Return: 0 on success and failure value on error
  827. */
  828. static int __maybe_unused zynqmp_dma_suspend(struct device *dev)
  829. {
  830. if (!device_may_wakeup(dev))
  831. return pm_runtime_force_suspend(dev);
  832. return 0;
  833. }
  834. /**
  835. * zynqmp_dma_resume - Resume from suspend
  836. * @dev: Address of the device structure
  837. *
  838. * Resume operation after suspend.
  839. * Return: 0 on success and failure value on error
  840. */
  841. static int __maybe_unused zynqmp_dma_resume(struct device *dev)
  842. {
  843. if (!device_may_wakeup(dev))
  844. return pm_runtime_force_resume(dev);
  845. return 0;
  846. }
  847. /**
  848. * zynqmp_dma_runtime_suspend - Runtime suspend method for the driver
  849. * @dev: Address of the device structure
  850. *
  851. * Put the driver into low power mode.
  852. * Return: 0 always
  853. */
  854. static int __maybe_unused zynqmp_dma_runtime_suspend(struct device *dev)
  855. {
  856. struct zynqmp_dma_device *zdev = dev_get_drvdata(dev);
  857. clk_disable_unprepare(zdev->clk_main);
  858. clk_disable_unprepare(zdev->clk_apb);
  859. return 0;
  860. }
  861. /**
  862. * zynqmp_dma_runtime_resume - Runtime suspend method for the driver
  863. * @dev: Address of the device structure
  864. *
  865. * Put the driver into low power mode.
  866. * Return: 0 always
  867. */
  868. static int __maybe_unused zynqmp_dma_runtime_resume(struct device *dev)
  869. {
  870. struct zynqmp_dma_device *zdev = dev_get_drvdata(dev);
  871. int err;
  872. err = clk_prepare_enable(zdev->clk_main);
  873. if (err) {
  874. dev_err(dev, "Unable to enable main clock.\n");
  875. return err;
  876. }
  877. err = clk_prepare_enable(zdev->clk_apb);
  878. if (err) {
  879. dev_err(dev, "Unable to enable apb clock.\n");
  880. clk_disable_unprepare(zdev->clk_main);
  881. return err;
  882. }
  883. return 0;
  884. }
  885. static const struct dev_pm_ops zynqmp_dma_dev_pm_ops = {
  886. SET_SYSTEM_SLEEP_PM_OPS(zynqmp_dma_suspend, zynqmp_dma_resume)
  887. SET_RUNTIME_PM_OPS(zynqmp_dma_runtime_suspend,
  888. zynqmp_dma_runtime_resume, NULL)
  889. };
  890. /**
  891. * zynqmp_dma_probe - Driver probe function
  892. * @pdev: Pointer to the platform_device structure
  893. *
  894. * Return: '0' on success and failure value on error
  895. */
  896. static int zynqmp_dma_probe(struct platform_device *pdev)
  897. {
  898. struct zynqmp_dma_device *zdev;
  899. struct dma_device *p;
  900. int ret;
  901. zdev = devm_kzalloc(&pdev->dev, sizeof(*zdev), GFP_KERNEL);
  902. if (!zdev)
  903. return -ENOMEM;
  904. zdev->dev = &pdev->dev;
  905. INIT_LIST_HEAD(&zdev->common.channels);
  906. dma_set_mask(&pdev->dev, DMA_BIT_MASK(44));
  907. dma_cap_set(DMA_MEMCPY, zdev->common.cap_mask);
  908. p = &zdev->common;
  909. p->device_prep_dma_memcpy = zynqmp_dma_prep_memcpy;
  910. p->device_terminate_all = zynqmp_dma_device_terminate_all;
  911. p->device_issue_pending = zynqmp_dma_issue_pending;
  912. p->device_alloc_chan_resources = zynqmp_dma_alloc_chan_resources;
  913. p->device_free_chan_resources = zynqmp_dma_free_chan_resources;
  914. p->device_tx_status = dma_cookie_status;
  915. p->device_config = zynqmp_dma_device_config;
  916. p->dev = &pdev->dev;
  917. zdev->clk_main = devm_clk_get(&pdev->dev, "clk_main");
  918. if (IS_ERR(zdev->clk_main)) {
  919. dev_err(&pdev->dev, "main clock not found.\n");
  920. return PTR_ERR(zdev->clk_main);
  921. }
  922. zdev->clk_apb = devm_clk_get(&pdev->dev, "clk_apb");
  923. if (IS_ERR(zdev->clk_apb)) {
  924. dev_err(&pdev->dev, "apb clock not found.\n");
  925. return PTR_ERR(zdev->clk_apb);
  926. }
  927. platform_set_drvdata(pdev, zdev);
  928. pm_runtime_set_autosuspend_delay(zdev->dev, ZDMA_PM_TIMEOUT);
  929. pm_runtime_use_autosuspend(zdev->dev);
  930. pm_runtime_enable(zdev->dev);
  931. pm_runtime_get_sync(zdev->dev);
  932. if (!pm_runtime_enabled(zdev->dev)) {
  933. ret = zynqmp_dma_runtime_resume(zdev->dev);
  934. if (ret)
  935. return ret;
  936. }
  937. ret = zynqmp_dma_chan_probe(zdev, pdev);
  938. if (ret) {
  939. dev_err(&pdev->dev, "Probing channel failed\n");
  940. goto err_disable_pm;
  941. }
  942. p->dst_addr_widths = BIT(zdev->chan->bus_width / 8);
  943. p->src_addr_widths = BIT(zdev->chan->bus_width / 8);
  944. dma_async_device_register(&zdev->common);
  945. ret = of_dma_controller_register(pdev->dev.of_node,
  946. of_zynqmp_dma_xlate, zdev);
  947. if (ret) {
  948. dev_err(&pdev->dev, "Unable to register DMA to DT\n");
  949. dma_async_device_unregister(&zdev->common);
  950. goto free_chan_resources;
  951. }
  952. pm_runtime_mark_last_busy(zdev->dev);
  953. pm_runtime_put_sync_autosuspend(zdev->dev);
  954. dev_info(&pdev->dev, "ZynqMP DMA driver Probe success\n");
  955. return 0;
  956. free_chan_resources:
  957. zynqmp_dma_chan_remove(zdev->chan);
  958. err_disable_pm:
  959. if (!pm_runtime_enabled(zdev->dev))
  960. zynqmp_dma_runtime_suspend(zdev->dev);
  961. pm_runtime_disable(zdev->dev);
  962. return ret;
  963. }
  964. /**
  965. * zynqmp_dma_remove - Driver remove function
  966. * @pdev: Pointer to the platform_device structure
  967. *
  968. * Return: Always '0'
  969. */
  970. static int zynqmp_dma_remove(struct platform_device *pdev)
  971. {
  972. struct zynqmp_dma_device *zdev = platform_get_drvdata(pdev);
  973. of_dma_controller_free(pdev->dev.of_node);
  974. dma_async_device_unregister(&zdev->common);
  975. zynqmp_dma_chan_remove(zdev->chan);
  976. pm_runtime_disable(zdev->dev);
  977. if (!pm_runtime_enabled(zdev->dev))
  978. zynqmp_dma_runtime_suspend(zdev->dev);
  979. return 0;
  980. }
  981. static const struct of_device_id zynqmp_dma_of_match[] = {
  982. { .compatible = "xlnx,zynqmp-dma-1.0", },
  983. {}
  984. };
  985. MODULE_DEVICE_TABLE(of, zynqmp_dma_of_match);
  986. static struct platform_driver zynqmp_dma_driver = {
  987. .driver = {
  988. .name = "xilinx-zynqmp-dma",
  989. .of_match_table = zynqmp_dma_of_match,
  990. .pm = &zynqmp_dma_dev_pm_ops,
  991. },
  992. .probe = zynqmp_dma_probe,
  993. .remove = zynqmp_dma_remove,
  994. };
  995. module_platform_driver(zynqmp_dma_driver);
  996. MODULE_LICENSE("GPL");
  997. MODULE_AUTHOR("Xilinx, Inc.");
  998. MODULE_DESCRIPTION("Xilinx ZynqMP DMA driver");