nitrox_hal.c 10 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402
  1. // SPDX-License-Identifier: GPL-2.0
  2. #include <linux/delay.h>
  3. #include "nitrox_dev.h"
  4. #include "nitrox_csr.h"
  5. /**
  6. * emu_enable_cores - Enable EMU cluster cores.
  7. * @ndev: N5 device
  8. */
  9. static void emu_enable_cores(struct nitrox_device *ndev)
  10. {
  11. union emu_se_enable emu_se;
  12. union emu_ae_enable emu_ae;
  13. int i;
  14. /* AE cores 20 per cluster */
  15. emu_ae.value = 0;
  16. emu_ae.s.enable = 0xfffff;
  17. /* SE cores 16 per cluster */
  18. emu_se.value = 0;
  19. emu_se.s.enable = 0xffff;
  20. /* enable per cluster cores */
  21. for (i = 0; i < NR_CLUSTERS; i++) {
  22. nitrox_write_csr(ndev, EMU_AE_ENABLEX(i), emu_ae.value);
  23. nitrox_write_csr(ndev, EMU_SE_ENABLEX(i), emu_se.value);
  24. }
  25. }
  26. /**
  27. * nitrox_config_emu_unit - configure EMU unit.
  28. * @ndev: N5 device
  29. */
  30. void nitrox_config_emu_unit(struct nitrox_device *ndev)
  31. {
  32. union emu_wd_int_ena_w1s emu_wd_int;
  33. union emu_ge_int_ena_w1s emu_ge_int;
  34. u64 offset;
  35. int i;
  36. /* enable cores */
  37. emu_enable_cores(ndev);
  38. /* enable general error and watch dog interrupts */
  39. emu_ge_int.value = 0;
  40. emu_ge_int.s.se_ge = 0xffff;
  41. emu_ge_int.s.ae_ge = 0xfffff;
  42. emu_wd_int.value = 0;
  43. emu_wd_int.s.se_wd = 1;
  44. for (i = 0; i < NR_CLUSTERS; i++) {
  45. offset = EMU_WD_INT_ENA_W1SX(i);
  46. nitrox_write_csr(ndev, offset, emu_wd_int.value);
  47. offset = EMU_GE_INT_ENA_W1SX(i);
  48. nitrox_write_csr(ndev, offset, emu_ge_int.value);
  49. }
  50. }
  51. static void reset_pkt_input_ring(struct nitrox_device *ndev, int ring)
  52. {
  53. union nps_pkt_in_instr_ctl pkt_in_ctl;
  54. union nps_pkt_in_instr_baoff_dbell pkt_in_dbell;
  55. union nps_pkt_in_done_cnts pkt_in_cnts;
  56. u64 offset;
  57. offset = NPS_PKT_IN_INSTR_CTLX(ring);
  58. /* disable the ring */
  59. pkt_in_ctl.value = nitrox_read_csr(ndev, offset);
  60. pkt_in_ctl.s.enb = 0;
  61. nitrox_write_csr(ndev, offset, pkt_in_ctl.value);
  62. usleep_range(100, 150);
  63. /* wait to clear [ENB] */
  64. do {
  65. pkt_in_ctl.value = nitrox_read_csr(ndev, offset);
  66. } while (pkt_in_ctl.s.enb);
  67. /* clear off door bell counts */
  68. offset = NPS_PKT_IN_INSTR_BAOFF_DBELLX(ring);
  69. pkt_in_dbell.value = 0;
  70. pkt_in_dbell.s.dbell = 0xffffffff;
  71. nitrox_write_csr(ndev, offset, pkt_in_dbell.value);
  72. /* clear done counts */
  73. offset = NPS_PKT_IN_DONE_CNTSX(ring);
  74. pkt_in_cnts.value = nitrox_read_csr(ndev, offset);
  75. nitrox_write_csr(ndev, offset, pkt_in_cnts.value);
  76. usleep_range(50, 100);
  77. }
  78. void enable_pkt_input_ring(struct nitrox_device *ndev, int ring)
  79. {
  80. union nps_pkt_in_instr_ctl pkt_in_ctl;
  81. u64 offset;
  82. /* 64-byte instruction size */
  83. offset = NPS_PKT_IN_INSTR_CTLX(ring);
  84. pkt_in_ctl.value = nitrox_read_csr(ndev, offset);
  85. pkt_in_ctl.s.is64b = 1;
  86. pkt_in_ctl.s.enb = 1;
  87. nitrox_write_csr(ndev, offset, pkt_in_ctl.value);
  88. /* wait for set [ENB] */
  89. do {
  90. pkt_in_ctl.value = nitrox_read_csr(ndev, offset);
  91. } while (!pkt_in_ctl.s.enb);
  92. }
  93. /**
  94. * nitrox_config_pkt_input_rings - configure Packet Input Rings
  95. * @ndev: N5 device
  96. */
  97. void nitrox_config_pkt_input_rings(struct nitrox_device *ndev)
  98. {
  99. int i;
  100. for (i = 0; i < ndev->nr_queues; i++) {
  101. struct nitrox_cmdq *cmdq = &ndev->pkt_cmdqs[i];
  102. union nps_pkt_in_instr_rsize pkt_in_rsize;
  103. u64 offset;
  104. reset_pkt_input_ring(ndev, i);
  105. /* configure ring base address 16-byte aligned,
  106. * size and interrupt threshold.
  107. */
  108. offset = NPS_PKT_IN_INSTR_BADDRX(i);
  109. nitrox_write_csr(ndev, offset, cmdq->dma);
  110. /* configure ring size */
  111. offset = NPS_PKT_IN_INSTR_RSIZEX(i);
  112. pkt_in_rsize.value = 0;
  113. pkt_in_rsize.s.rsize = ndev->qlen;
  114. nitrox_write_csr(ndev, offset, pkt_in_rsize.value);
  115. /* set high threshold for pkt input ring interrupts */
  116. offset = NPS_PKT_IN_INT_LEVELSX(i);
  117. nitrox_write_csr(ndev, offset, 0xffffffff);
  118. enable_pkt_input_ring(ndev, i);
  119. }
  120. }
  121. static void reset_pkt_solicit_port(struct nitrox_device *ndev, int port)
  122. {
  123. union nps_pkt_slc_ctl pkt_slc_ctl;
  124. union nps_pkt_slc_cnts pkt_slc_cnts;
  125. u64 offset;
  126. /* disable slc port */
  127. offset = NPS_PKT_SLC_CTLX(port);
  128. pkt_slc_ctl.value = nitrox_read_csr(ndev, offset);
  129. pkt_slc_ctl.s.enb = 0;
  130. nitrox_write_csr(ndev, offset, pkt_slc_ctl.value);
  131. usleep_range(100, 150);
  132. /* wait to clear [ENB] */
  133. do {
  134. pkt_slc_ctl.value = nitrox_read_csr(ndev, offset);
  135. } while (pkt_slc_ctl.s.enb);
  136. /* clear slc counters */
  137. offset = NPS_PKT_SLC_CNTSX(port);
  138. pkt_slc_cnts.value = nitrox_read_csr(ndev, offset);
  139. nitrox_write_csr(ndev, offset, pkt_slc_cnts.value);
  140. usleep_range(50, 100);
  141. }
  142. void enable_pkt_solicit_port(struct nitrox_device *ndev, int port)
  143. {
  144. union nps_pkt_slc_ctl pkt_slc_ctl;
  145. u64 offset;
  146. offset = NPS_PKT_SLC_CTLX(port);
  147. pkt_slc_ctl.value = 0;
  148. pkt_slc_ctl.s.enb = 1;
  149. /*
  150. * 8 trailing 0x00 bytes will be added
  151. * to the end of the outgoing packet.
  152. */
  153. pkt_slc_ctl.s.z = 1;
  154. /* enable response header */
  155. pkt_slc_ctl.s.rh = 1;
  156. nitrox_write_csr(ndev, offset, pkt_slc_ctl.value);
  157. /* wait to set [ENB] */
  158. do {
  159. pkt_slc_ctl.value = nitrox_read_csr(ndev, offset);
  160. } while (!pkt_slc_ctl.s.enb);
  161. }
  162. static void config_single_pkt_solicit_port(struct nitrox_device *ndev,
  163. int port)
  164. {
  165. union nps_pkt_slc_int_levels pkt_slc_int;
  166. u64 offset;
  167. reset_pkt_solicit_port(ndev, port);
  168. offset = NPS_PKT_SLC_INT_LEVELSX(port);
  169. pkt_slc_int.value = 0;
  170. /* time interrupt threshold */
  171. pkt_slc_int.s.timet = 0x3fffff;
  172. nitrox_write_csr(ndev, offset, pkt_slc_int.value);
  173. enable_pkt_solicit_port(ndev, port);
  174. }
  175. void nitrox_config_pkt_solicit_ports(struct nitrox_device *ndev)
  176. {
  177. int i;
  178. for (i = 0; i < ndev->nr_queues; i++)
  179. config_single_pkt_solicit_port(ndev, i);
  180. }
  181. /**
  182. * enable_nps_interrupts - enable NPS interrutps
  183. * @ndev: N5 device.
  184. *
  185. * This includes NPS core, packet in and slc interrupts.
  186. */
  187. static void enable_nps_interrupts(struct nitrox_device *ndev)
  188. {
  189. union nps_core_int_ena_w1s core_int;
  190. /* NPS core interrutps */
  191. core_int.value = 0;
  192. core_int.s.host_wr_err = 1;
  193. core_int.s.host_wr_timeout = 1;
  194. core_int.s.exec_wr_timeout = 1;
  195. core_int.s.npco_dma_malform = 1;
  196. core_int.s.host_nps_wr_err = 1;
  197. nitrox_write_csr(ndev, NPS_CORE_INT_ENA_W1S, core_int.value);
  198. /* NPS packet in ring interrupts */
  199. nitrox_write_csr(ndev, NPS_PKT_IN_RERR_LO_ENA_W1S, (~0ULL));
  200. nitrox_write_csr(ndev, NPS_PKT_IN_RERR_HI_ENA_W1S, (~0ULL));
  201. nitrox_write_csr(ndev, NPS_PKT_IN_ERR_TYPE_ENA_W1S, (~0ULL));
  202. /* NPS packet slc port interrupts */
  203. nitrox_write_csr(ndev, NPS_PKT_SLC_RERR_HI_ENA_W1S, (~0ULL));
  204. nitrox_write_csr(ndev, NPS_PKT_SLC_RERR_LO_ENA_W1S, (~0ULL));
  205. nitrox_write_csr(ndev, NPS_PKT_SLC_ERR_TYPE_ENA_W1S, (~0uLL));
  206. }
  207. void nitrox_config_nps_unit(struct nitrox_device *ndev)
  208. {
  209. union nps_core_gbl_vfcfg core_gbl_vfcfg;
  210. /* endian control information */
  211. nitrox_write_csr(ndev, NPS_CORE_CONTROL, 1ULL);
  212. /* disable ILK interface */
  213. core_gbl_vfcfg.value = 0;
  214. core_gbl_vfcfg.s.ilk_disable = 1;
  215. core_gbl_vfcfg.s.cfg = PF_MODE;
  216. nitrox_write_csr(ndev, NPS_CORE_GBL_VFCFG, core_gbl_vfcfg.value);
  217. /* config input and solicit ports */
  218. nitrox_config_pkt_input_rings(ndev);
  219. nitrox_config_pkt_solicit_ports(ndev);
  220. /* enable interrupts */
  221. enable_nps_interrupts(ndev);
  222. }
  223. void nitrox_config_pom_unit(struct nitrox_device *ndev)
  224. {
  225. union pom_int_ena_w1s pom_int;
  226. int i;
  227. /* enable pom interrupts */
  228. pom_int.value = 0;
  229. pom_int.s.illegal_dport = 1;
  230. nitrox_write_csr(ndev, POM_INT_ENA_W1S, pom_int.value);
  231. /* enable perf counters */
  232. for (i = 0; i < ndev->hw.se_cores; i++)
  233. nitrox_write_csr(ndev, POM_PERF_CTL, BIT_ULL(i));
  234. }
  235. /**
  236. * nitrox_config_rand_unit - enable N5 random number unit
  237. * @ndev: N5 device
  238. */
  239. void nitrox_config_rand_unit(struct nitrox_device *ndev)
  240. {
  241. union efl_rnm_ctl_status efl_rnm_ctl;
  242. u64 offset;
  243. offset = EFL_RNM_CTL_STATUS;
  244. efl_rnm_ctl.value = nitrox_read_csr(ndev, offset);
  245. efl_rnm_ctl.s.ent_en = 1;
  246. efl_rnm_ctl.s.rng_en = 1;
  247. nitrox_write_csr(ndev, offset, efl_rnm_ctl.value);
  248. }
  249. void nitrox_config_efl_unit(struct nitrox_device *ndev)
  250. {
  251. int i;
  252. for (i = 0; i < NR_CLUSTERS; i++) {
  253. union efl_core_int_ena_w1s efl_core_int;
  254. u64 offset;
  255. /* EFL core interrupts */
  256. offset = EFL_CORE_INT_ENA_W1SX(i);
  257. efl_core_int.value = 0;
  258. efl_core_int.s.len_ovr = 1;
  259. efl_core_int.s.d_left = 1;
  260. efl_core_int.s.epci_decode_err = 1;
  261. nitrox_write_csr(ndev, offset, efl_core_int.value);
  262. offset = EFL_CORE_VF_ERR_INT0_ENA_W1SX(i);
  263. nitrox_write_csr(ndev, offset, (~0ULL));
  264. offset = EFL_CORE_VF_ERR_INT1_ENA_W1SX(i);
  265. nitrox_write_csr(ndev, offset, (~0ULL));
  266. }
  267. }
  268. void nitrox_config_bmi_unit(struct nitrox_device *ndev)
  269. {
  270. union bmi_ctl bmi_ctl;
  271. union bmi_int_ena_w1s bmi_int_ena;
  272. u64 offset;
  273. /* no threshold limits for PCIe */
  274. offset = BMI_CTL;
  275. bmi_ctl.value = nitrox_read_csr(ndev, offset);
  276. bmi_ctl.s.max_pkt_len = 0xff;
  277. bmi_ctl.s.nps_free_thrsh = 0xff;
  278. bmi_ctl.s.nps_hdrq_thrsh = 0x7a;
  279. nitrox_write_csr(ndev, offset, bmi_ctl.value);
  280. /* enable interrupts */
  281. offset = BMI_INT_ENA_W1S;
  282. bmi_int_ena.value = 0;
  283. bmi_int_ena.s.max_len_err_nps = 1;
  284. bmi_int_ena.s.pkt_rcv_err_nps = 1;
  285. bmi_int_ena.s.fpf_undrrn = 1;
  286. nitrox_write_csr(ndev, offset, bmi_int_ena.value);
  287. }
  288. void nitrox_config_bmo_unit(struct nitrox_device *ndev)
  289. {
  290. union bmo_ctl2 bmo_ctl2;
  291. u64 offset;
  292. /* no threshold limits for PCIe */
  293. offset = BMO_CTL2;
  294. bmo_ctl2.value = nitrox_read_csr(ndev, offset);
  295. bmo_ctl2.s.nps_slc_buf_thrsh = 0xff;
  296. nitrox_write_csr(ndev, offset, bmo_ctl2.value);
  297. }
  298. void invalidate_lbc(struct nitrox_device *ndev)
  299. {
  300. union lbc_inval_ctl lbc_ctl;
  301. union lbc_inval_status lbc_stat;
  302. u64 offset;
  303. /* invalidate LBC */
  304. offset = LBC_INVAL_CTL;
  305. lbc_ctl.value = nitrox_read_csr(ndev, offset);
  306. lbc_ctl.s.cam_inval_start = 1;
  307. nitrox_write_csr(ndev, offset, lbc_ctl.value);
  308. offset = LBC_INVAL_STATUS;
  309. do {
  310. lbc_stat.value = nitrox_read_csr(ndev, offset);
  311. } while (!lbc_stat.s.done);
  312. }
  313. void nitrox_config_lbc_unit(struct nitrox_device *ndev)
  314. {
  315. union lbc_int_ena_w1s lbc_int_ena;
  316. u64 offset;
  317. invalidate_lbc(ndev);
  318. /* enable interrupts */
  319. offset = LBC_INT_ENA_W1S;
  320. lbc_int_ena.value = 0;
  321. lbc_int_ena.s.dma_rd_err = 1;
  322. lbc_int_ena.s.over_fetch_err = 1;
  323. lbc_int_ena.s.cam_inval_abort = 1;
  324. lbc_int_ena.s.cam_hard_err = 1;
  325. nitrox_write_csr(ndev, offset, lbc_int_ena.value);
  326. offset = LBC_PLM_VF1_64_INT_ENA_W1S;
  327. nitrox_write_csr(ndev, offset, (~0ULL));
  328. offset = LBC_PLM_VF65_128_INT_ENA_W1S;
  329. nitrox_write_csr(ndev, offset, (~0ULL));
  330. offset = LBC_ELM_VF1_64_INT_ENA_W1S;
  331. nitrox_write_csr(ndev, offset, (~0ULL));
  332. offset = LBC_ELM_VF65_128_INT_ENA_W1S;
  333. nitrox_write_csr(ndev, offset, (~0ULL));
  334. }