dma.c 12 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547
  1. /*
  2. * Copyright (C) 2015 Jakub Kicinski <kubakici@wp.pl>
  3. *
  4. * This program is free software; you can redistribute it and/or modify
  5. * it under the terms of the GNU General Public License version 2
  6. * as published by the Free Software Foundation
  7. *
  8. * This program is distributed in the hope that it will be useful,
  9. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  10. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
  11. * GNU General Public License for more details.
  12. */
  13. #include "mt7601u.h"
  14. #include "dma.h"
  15. #include "usb.h"
  16. #include "trace.h"
  17. static int mt7601u_submit_rx_buf(struct mt7601u_dev *dev,
  18. struct mt7601u_dma_buf_rx *e, gfp_t gfp);
  19. static unsigned int ieee80211_get_hdrlen_from_buf(const u8 *data, unsigned len)
  20. {
  21. const struct ieee80211_hdr *hdr = (const struct ieee80211_hdr *)data;
  22. unsigned int hdrlen;
  23. if (unlikely(len < 10))
  24. return 0;
  25. hdrlen = ieee80211_hdrlen(hdr->frame_control);
  26. if (unlikely(hdrlen > len))
  27. return 0;
  28. return hdrlen;
  29. }
  30. static struct sk_buff *
  31. mt7601u_rx_skb_from_seg(struct mt7601u_dev *dev, struct mt7601u_rxwi *rxwi,
  32. void *data, u32 seg_len, u32 truesize, struct page *p)
  33. {
  34. struct sk_buff *skb;
  35. u32 true_len, hdr_len = 0, copy, frag;
  36. skb = alloc_skb(p ? 128 : seg_len, GFP_ATOMIC);
  37. if (!skb)
  38. return NULL;
  39. true_len = mt76_mac_process_rx(dev, skb, data, rxwi);
  40. if (!true_len || true_len > seg_len)
  41. goto bad_frame;
  42. hdr_len = ieee80211_get_hdrlen_from_buf(data, true_len);
  43. if (!hdr_len)
  44. goto bad_frame;
  45. if (rxwi->rxinfo & cpu_to_le32(MT_RXINFO_L2PAD)) {
  46. skb_put_data(skb, data, hdr_len);
  47. data += hdr_len + 2;
  48. true_len -= hdr_len;
  49. hdr_len = 0;
  50. }
  51. /* If not doing paged RX allocated skb will always have enough space */
  52. copy = (true_len <= skb_tailroom(skb)) ? true_len : hdr_len + 8;
  53. frag = true_len - copy;
  54. skb_put_data(skb, data, copy);
  55. data += copy;
  56. if (frag) {
  57. skb_add_rx_frag(skb, 0, p, data - page_address(p),
  58. frag, truesize);
  59. get_page(p);
  60. }
  61. return skb;
  62. bad_frame:
  63. dev_err_ratelimited(dev->dev, "Error: incorrect frame len:%u hdr:%u\n",
  64. true_len, hdr_len);
  65. dev_kfree_skb(skb);
  66. return NULL;
  67. }
  68. static void mt7601u_rx_process_seg(struct mt7601u_dev *dev, u8 *data,
  69. u32 seg_len, struct page *p)
  70. {
  71. struct sk_buff *skb;
  72. struct mt7601u_rxwi *rxwi;
  73. u32 fce_info, truesize = seg_len;
  74. /* DMA_INFO field at the beginning of the segment contains only some of
  75. * the information, we need to read the FCE descriptor from the end.
  76. */
  77. fce_info = get_unaligned_le32(data + seg_len - MT_FCE_INFO_LEN);
  78. seg_len -= MT_FCE_INFO_LEN;
  79. data += MT_DMA_HDR_LEN;
  80. seg_len -= MT_DMA_HDR_LEN;
  81. rxwi = (struct mt7601u_rxwi *) data;
  82. data += sizeof(struct mt7601u_rxwi);
  83. seg_len -= sizeof(struct mt7601u_rxwi);
  84. if (unlikely(rxwi->zero[0] || rxwi->zero[1] || rxwi->zero[2]))
  85. dev_err_once(dev->dev, "Error: RXWI zero fields are set\n");
  86. if (unlikely(FIELD_GET(MT_RXD_INFO_TYPE, fce_info)))
  87. dev_err_once(dev->dev, "Error: RX path seen a non-pkt urb\n");
  88. trace_mt_rx(dev, rxwi, fce_info);
  89. skb = mt7601u_rx_skb_from_seg(dev, rxwi, data, seg_len, truesize, p);
  90. if (!skb)
  91. return;
  92. spin_lock(&dev->mac_lock);
  93. ieee80211_rx(dev->hw, skb);
  94. spin_unlock(&dev->mac_lock);
  95. }
  96. static u16 mt7601u_rx_next_seg_len(u8 *data, u32 data_len)
  97. {
  98. u32 min_seg_len = MT_DMA_HDR_LEN + MT_RX_INFO_LEN +
  99. sizeof(struct mt7601u_rxwi) + MT_FCE_INFO_LEN;
  100. u16 dma_len = get_unaligned_le16(data);
  101. if (data_len < min_seg_len ||
  102. WARN_ON(!dma_len) ||
  103. WARN_ON(dma_len + MT_DMA_HDRS > data_len) ||
  104. WARN_ON(dma_len & 0x3))
  105. return 0;
  106. return MT_DMA_HDRS + dma_len;
  107. }
  108. static void
  109. mt7601u_rx_process_entry(struct mt7601u_dev *dev, struct mt7601u_dma_buf_rx *e)
  110. {
  111. u32 seg_len, data_len = e->urb->actual_length;
  112. u8 *data = page_address(e->p);
  113. struct page *new_p = NULL;
  114. int cnt = 0;
  115. if (!test_bit(MT7601U_STATE_INITIALIZED, &dev->state))
  116. return;
  117. /* Copy if there is very little data in the buffer. */
  118. if (data_len > 512)
  119. new_p = dev_alloc_pages(MT_RX_ORDER);
  120. while ((seg_len = mt7601u_rx_next_seg_len(data, data_len))) {
  121. mt7601u_rx_process_seg(dev, data, seg_len, new_p ? e->p : NULL);
  122. data_len -= seg_len;
  123. data += seg_len;
  124. cnt++;
  125. }
  126. if (cnt > 1)
  127. trace_mt_rx_dma_aggr(dev, cnt, !!new_p);
  128. if (new_p) {
  129. /* we have one extra ref from the allocator */
  130. put_page(e->p);
  131. e->p = new_p;
  132. }
  133. }
  134. static struct mt7601u_dma_buf_rx *
  135. mt7601u_rx_get_pending_entry(struct mt7601u_dev *dev)
  136. {
  137. struct mt7601u_rx_queue *q = &dev->rx_q;
  138. struct mt7601u_dma_buf_rx *buf = NULL;
  139. unsigned long flags;
  140. spin_lock_irqsave(&dev->rx_lock, flags);
  141. if (!q->pending)
  142. goto out;
  143. buf = &q->e[q->start];
  144. q->pending--;
  145. q->start = (q->start + 1) % q->entries;
  146. out:
  147. spin_unlock_irqrestore(&dev->rx_lock, flags);
  148. return buf;
  149. }
  150. static void mt7601u_complete_rx(struct urb *urb)
  151. {
  152. struct mt7601u_dev *dev = urb->context;
  153. struct mt7601u_rx_queue *q = &dev->rx_q;
  154. unsigned long flags;
  155. /* do no schedule rx tasklet if urb has been unlinked
  156. * or the device has been removed
  157. */
  158. switch (urb->status) {
  159. case -ECONNRESET:
  160. case -ESHUTDOWN:
  161. case -ENOENT:
  162. return;
  163. default:
  164. dev_err_ratelimited(dev->dev, "rx urb failed: %d\n",
  165. urb->status);
  166. /* fall through */
  167. case 0:
  168. break;
  169. }
  170. spin_lock_irqsave(&dev->rx_lock, flags);
  171. if (WARN_ONCE(q->e[q->end].urb != urb, "RX urb mismatch"))
  172. goto out;
  173. q->end = (q->end + 1) % q->entries;
  174. q->pending++;
  175. tasklet_schedule(&dev->rx_tasklet);
  176. out:
  177. spin_unlock_irqrestore(&dev->rx_lock, flags);
  178. }
  179. static void mt7601u_rx_tasklet(unsigned long data)
  180. {
  181. struct mt7601u_dev *dev = (struct mt7601u_dev *) data;
  182. struct mt7601u_dma_buf_rx *e;
  183. while ((e = mt7601u_rx_get_pending_entry(dev))) {
  184. if (e->urb->status)
  185. continue;
  186. mt7601u_rx_process_entry(dev, e);
  187. mt7601u_submit_rx_buf(dev, e, GFP_ATOMIC);
  188. }
  189. }
  190. static void mt7601u_complete_tx(struct urb *urb)
  191. {
  192. struct mt7601u_tx_queue *q = urb->context;
  193. struct mt7601u_dev *dev = q->dev;
  194. struct sk_buff *skb;
  195. unsigned long flags;
  196. switch (urb->status) {
  197. case -ECONNRESET:
  198. case -ESHUTDOWN:
  199. case -ENOENT:
  200. return;
  201. default:
  202. dev_err_ratelimited(dev->dev, "tx urb failed: %d\n",
  203. urb->status);
  204. /* fall through */
  205. case 0:
  206. break;
  207. }
  208. spin_lock_irqsave(&dev->tx_lock, flags);
  209. if (WARN_ONCE(q->e[q->start].urb != urb, "TX urb mismatch"))
  210. goto out;
  211. skb = q->e[q->start].skb;
  212. q->e[q->start].skb = NULL;
  213. trace_mt_tx_dma_done(dev, skb);
  214. __skb_queue_tail(&dev->tx_skb_done, skb);
  215. tasklet_schedule(&dev->tx_tasklet);
  216. if (q->used == q->entries - q->entries / 8)
  217. ieee80211_wake_queue(dev->hw, skb_get_queue_mapping(skb));
  218. q->start = (q->start + 1) % q->entries;
  219. q->used--;
  220. out:
  221. spin_unlock_irqrestore(&dev->tx_lock, flags);
  222. }
  223. static void mt7601u_tx_tasklet(unsigned long data)
  224. {
  225. struct mt7601u_dev *dev = (struct mt7601u_dev *) data;
  226. struct sk_buff_head skbs;
  227. unsigned long flags;
  228. __skb_queue_head_init(&skbs);
  229. spin_lock_irqsave(&dev->tx_lock, flags);
  230. set_bit(MT7601U_STATE_MORE_STATS, &dev->state);
  231. if (!test_and_set_bit(MT7601U_STATE_READING_STATS, &dev->state))
  232. queue_delayed_work(dev->stat_wq, &dev->stat_work,
  233. msecs_to_jiffies(10));
  234. skb_queue_splice_init(&dev->tx_skb_done, &skbs);
  235. spin_unlock_irqrestore(&dev->tx_lock, flags);
  236. while (!skb_queue_empty(&skbs)) {
  237. struct sk_buff *skb = __skb_dequeue(&skbs);
  238. mt7601u_tx_status(dev, skb);
  239. }
  240. }
  241. static int mt7601u_dma_submit_tx(struct mt7601u_dev *dev,
  242. struct sk_buff *skb, u8 ep)
  243. {
  244. struct usb_device *usb_dev = mt7601u_to_usb_dev(dev);
  245. unsigned snd_pipe = usb_sndbulkpipe(usb_dev, dev->out_eps[ep]);
  246. struct mt7601u_dma_buf_tx *e;
  247. struct mt7601u_tx_queue *q = &dev->tx_q[ep];
  248. unsigned long flags;
  249. int ret;
  250. spin_lock_irqsave(&dev->tx_lock, flags);
  251. if (WARN_ON(q->entries <= q->used)) {
  252. ret = -ENOSPC;
  253. goto out;
  254. }
  255. e = &q->e[q->end];
  256. usb_fill_bulk_urb(e->urb, usb_dev, snd_pipe, skb->data, skb->len,
  257. mt7601u_complete_tx, q);
  258. ret = usb_submit_urb(e->urb, GFP_ATOMIC);
  259. if (ret) {
  260. /* Special-handle ENODEV from TX urb submission because it will
  261. * often be the first ENODEV we see after device is removed.
  262. */
  263. if (ret == -ENODEV)
  264. set_bit(MT7601U_STATE_REMOVED, &dev->state);
  265. else
  266. dev_err(dev->dev, "Error: TX urb submit failed:%d\n",
  267. ret);
  268. goto out;
  269. }
  270. q->end = (q->end + 1) % q->entries;
  271. q->used++;
  272. e->skb = skb;
  273. if (q->used >= q->entries)
  274. ieee80211_stop_queue(dev->hw, skb_get_queue_mapping(skb));
  275. out:
  276. spin_unlock_irqrestore(&dev->tx_lock, flags);
  277. return ret;
  278. }
  279. /* Map hardware Q to USB endpoint number */
  280. static u8 q2ep(u8 qid)
  281. {
  282. /* TODO: take management packets to queue 5 */
  283. return qid + 1;
  284. }
  285. /* Map USB endpoint number to Q id in the DMA engine */
  286. static enum mt76_qsel ep2dmaq(u8 ep)
  287. {
  288. if (ep == 5)
  289. return MT_QSEL_MGMT;
  290. return MT_QSEL_EDCA;
  291. }
  292. int mt7601u_dma_enqueue_tx(struct mt7601u_dev *dev, struct sk_buff *skb,
  293. struct mt76_wcid *wcid, int hw_q)
  294. {
  295. u8 ep = q2ep(hw_q);
  296. u32 dma_flags;
  297. int ret;
  298. dma_flags = MT_TXD_PKT_INFO_80211;
  299. if (wcid->hw_key_idx == 0xff)
  300. dma_flags |= MT_TXD_PKT_INFO_WIV;
  301. ret = mt7601u_dma_skb_wrap_pkt(skb, ep2dmaq(ep), dma_flags);
  302. if (ret)
  303. return ret;
  304. ret = mt7601u_dma_submit_tx(dev, skb, ep);
  305. if (ret) {
  306. ieee80211_free_txskb(dev->hw, skb);
  307. return ret;
  308. }
  309. return 0;
  310. }
  311. static void mt7601u_kill_rx(struct mt7601u_dev *dev)
  312. {
  313. int i;
  314. for (i = 0; i < dev->rx_q.entries; i++)
  315. usb_poison_urb(dev->rx_q.e[i].urb);
  316. }
  317. static int mt7601u_submit_rx_buf(struct mt7601u_dev *dev,
  318. struct mt7601u_dma_buf_rx *e, gfp_t gfp)
  319. {
  320. struct usb_device *usb_dev = mt7601u_to_usb_dev(dev);
  321. u8 *buf = page_address(e->p);
  322. unsigned pipe;
  323. int ret;
  324. pipe = usb_rcvbulkpipe(usb_dev, dev->in_eps[MT_EP_IN_PKT_RX]);
  325. usb_fill_bulk_urb(e->urb, usb_dev, pipe, buf, MT_RX_URB_SIZE,
  326. mt7601u_complete_rx, dev);
  327. trace_mt_submit_urb(dev, e->urb);
  328. ret = usb_submit_urb(e->urb, gfp);
  329. if (ret)
  330. dev_err(dev->dev, "Error: submit RX URB failed:%d\n", ret);
  331. return ret;
  332. }
  333. static int mt7601u_submit_rx(struct mt7601u_dev *dev)
  334. {
  335. int i, ret;
  336. for (i = 0; i < dev->rx_q.entries; i++) {
  337. ret = mt7601u_submit_rx_buf(dev, &dev->rx_q.e[i], GFP_KERNEL);
  338. if (ret)
  339. return ret;
  340. }
  341. return 0;
  342. }
  343. static void mt7601u_free_rx(struct mt7601u_dev *dev)
  344. {
  345. int i;
  346. for (i = 0; i < dev->rx_q.entries; i++) {
  347. __free_pages(dev->rx_q.e[i].p, MT_RX_ORDER);
  348. usb_free_urb(dev->rx_q.e[i].urb);
  349. }
  350. }
  351. static int mt7601u_alloc_rx(struct mt7601u_dev *dev)
  352. {
  353. int i;
  354. memset(&dev->rx_q, 0, sizeof(dev->rx_q));
  355. dev->rx_q.dev = dev;
  356. dev->rx_q.entries = N_RX_ENTRIES;
  357. for (i = 0; i < N_RX_ENTRIES; i++) {
  358. dev->rx_q.e[i].urb = usb_alloc_urb(0, GFP_KERNEL);
  359. dev->rx_q.e[i].p = dev_alloc_pages(MT_RX_ORDER);
  360. if (!dev->rx_q.e[i].urb || !dev->rx_q.e[i].p)
  361. return -ENOMEM;
  362. }
  363. return 0;
  364. }
  365. static void mt7601u_free_tx_queue(struct mt7601u_tx_queue *q)
  366. {
  367. int i;
  368. for (i = 0; i < q->entries; i++) {
  369. usb_poison_urb(q->e[i].urb);
  370. if (q->e[i].skb)
  371. mt7601u_tx_status(q->dev, q->e[i].skb);
  372. usb_free_urb(q->e[i].urb);
  373. }
  374. }
  375. static void mt7601u_free_tx(struct mt7601u_dev *dev)
  376. {
  377. int i;
  378. if (!dev->tx_q)
  379. return;
  380. for (i = 0; i < __MT_EP_OUT_MAX; i++)
  381. mt7601u_free_tx_queue(&dev->tx_q[i]);
  382. }
  383. static int mt7601u_alloc_tx_queue(struct mt7601u_dev *dev,
  384. struct mt7601u_tx_queue *q)
  385. {
  386. int i;
  387. q->dev = dev;
  388. q->entries = N_TX_ENTRIES;
  389. for (i = 0; i < N_TX_ENTRIES; i++) {
  390. q->e[i].urb = usb_alloc_urb(0, GFP_KERNEL);
  391. if (!q->e[i].urb)
  392. return -ENOMEM;
  393. }
  394. return 0;
  395. }
  396. static int mt7601u_alloc_tx(struct mt7601u_dev *dev)
  397. {
  398. int i;
  399. dev->tx_q = devm_kcalloc(dev->dev, __MT_EP_OUT_MAX,
  400. sizeof(*dev->tx_q), GFP_KERNEL);
  401. if (!dev->tx_q)
  402. return -ENOMEM;
  403. for (i = 0; i < __MT_EP_OUT_MAX; i++)
  404. if (mt7601u_alloc_tx_queue(dev, &dev->tx_q[i]))
  405. return -ENOMEM;
  406. return 0;
  407. }
  408. int mt7601u_dma_init(struct mt7601u_dev *dev)
  409. {
  410. int ret = -ENOMEM;
  411. tasklet_init(&dev->tx_tasklet, mt7601u_tx_tasklet, (unsigned long) dev);
  412. tasklet_init(&dev->rx_tasklet, mt7601u_rx_tasklet, (unsigned long) dev);
  413. ret = mt7601u_alloc_tx(dev);
  414. if (ret)
  415. goto err;
  416. ret = mt7601u_alloc_rx(dev);
  417. if (ret)
  418. goto err;
  419. ret = mt7601u_submit_rx(dev);
  420. if (ret)
  421. goto err;
  422. return 0;
  423. err:
  424. mt7601u_dma_cleanup(dev);
  425. return ret;
  426. }
  427. void mt7601u_dma_cleanup(struct mt7601u_dev *dev)
  428. {
  429. mt7601u_kill_rx(dev);
  430. tasklet_kill(&dev->rx_tasklet);
  431. mt7601u_free_rx(dev);
  432. mt7601u_free_tx(dev);
  433. tasklet_kill(&dev->tx_tasklet);
  434. }