dfl.c 50 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323132413251326132713281329133013311332133313341335133613371338133913401341134213431344134513461347134813491350135113521353135413551356135713581359136013611362136313641365136613671368136913701371137213731374137513761377137813791380138113821383138413851386138713881389139013911392139313941395139613971398139914001401140214031404140514061407140814091410141114121413141414151416141714181419142014211422142314241425142614271428142914301431143214331434143514361437143814391440144114421443144414451446144714481449145014511452145314541455145614571458145914601461146214631464146514661467146814691470147114721473147414751476147714781479148014811482148314841485148614871488148914901491149214931494149514961497149814991500150115021503150415051506150715081509151015111512151315141515151615171518151915201521152215231524152515261527152815291530153115321533153415351536153715381539154015411542154315441545154615471548154915501551155215531554155515561557155815591560156115621563156415651566156715681569157015711572157315741575157615771578157915801581158215831584158515861587158815891590159115921593159415951596159715981599160016011602160316041605160616071608160916101611161216131614161516161617161816191620162116221623162416251626162716281629163016311632163316341635163616371638163916401641164216431644164516461647164816491650165116521653165416551656165716581659166016611662166316641665166616671668166916701671167216731674167516761677167816791680168116821683168416851686168716881689169016911692169316941695169616971698169917001701170217031704170517061707170817091710171117121713171417151716171717181719172017211722172317241725172617271728172917301731173217331734173517361737173817391740174117421743174417451746174717481749175017511752175317541755175617571758175917601761176217631764176517661767176817691770177117721773177417751776177717781779178017811782178317841785178617871788178917901791179217931794179517961797179817991800180118021803180418051806180718081809181018111812181318141815181618171818181918201821182218231824182518261827182818291830183118321833183418351836183718381839184018411842184318441845184618471848184918501851185218531854185518561857185818591860186118621863186418651866186718681869187018711872187318741875187618771878187918801881188218831884188518861887188818891890189118921893189418951896189718981899190019011902190319041905190619071908190919101911191219131914191519161917191819191920192119221923192419251926192719281929193019311932193319341935193619371938193919401941194219431944194519461947194819491950195119521953195419551956195719581959196019611962196319641965196619671968196919701971197219731974197519761977197819791980198119821983198419851986198719881989199019911992199319941995199619971998199920002001200220032004200520062007200820092010201120122013201420152016201720182019202020212022202320242025202620272028202920302031203220332034203520362037
  1. // SPDX-License-Identifier: GPL-2.0
  2. /*
  3. * Driver for FPGA Device Feature List (DFL) Support
  4. *
  5. * Copyright (C) 2017-2018 Intel Corporation, Inc.
  6. *
  7. * Authors:
  8. * Kang Luwei <luwei.kang@intel.com>
  9. * Zhang Yi <yi.z.zhang@intel.com>
  10. * Wu Hao <hao.wu@intel.com>
  11. * Xiao Guangrong <guangrong.xiao@linux.intel.com>
  12. */
  13. #include <linux/dfl.h>
  14. #include <linux/fpga-dfl.h>
  15. #include <linux/module.h>
  16. #include <linux/overflow.h>
  17. #include <linux/uaccess.h>
  18. #include "dfl.h"
  19. static DEFINE_MUTEX(dfl_id_mutex);
  20. /*
  21. * when adding a new feature dev support in DFL framework, it's required to
  22. * add a new item in enum dfl_id_type and provide related information in below
  23. * dfl_devs table which is indexed by dfl_id_type, e.g. name string used for
  24. * platform device creation (define name strings in dfl.h, as they could be
  25. * reused by platform device drivers).
  26. *
  27. * if the new feature dev needs chardev support, then it's required to add
  28. * a new item in dfl_chardevs table and configure dfl_devs[i].devt_type as
  29. * index to dfl_chardevs table. If no chardev support just set devt_type
  30. * as one invalid index (DFL_FPGA_DEVT_MAX).
  31. */
  32. enum dfl_fpga_devt_type {
  33. DFL_FPGA_DEVT_FME,
  34. DFL_FPGA_DEVT_PORT,
  35. DFL_FPGA_DEVT_MAX,
  36. };
  37. static struct lock_class_key dfl_pdata_keys[DFL_ID_MAX];
  38. static const char *dfl_pdata_key_strings[DFL_ID_MAX] = {
  39. "dfl-fme-pdata",
  40. "dfl-port-pdata",
  41. };
  42. /**
  43. * struct dfl_dev_info - dfl feature device information.
  44. * @name: name string of the feature platform device.
  45. * @dfh_id: id value in Device Feature Header (DFH) register by DFL spec.
  46. * @id: idr id of the feature dev.
  47. * @devt_type: index to dfl_chrdevs[].
  48. */
  49. struct dfl_dev_info {
  50. const char *name;
  51. u16 dfh_id;
  52. struct idr id;
  53. enum dfl_fpga_devt_type devt_type;
  54. };
  55. /* it is indexed by dfl_id_type */
  56. static struct dfl_dev_info dfl_devs[] = {
  57. {.name = DFL_FPGA_FEATURE_DEV_FME, .dfh_id = DFH_ID_FIU_FME,
  58. .devt_type = DFL_FPGA_DEVT_FME},
  59. {.name = DFL_FPGA_FEATURE_DEV_PORT, .dfh_id = DFH_ID_FIU_PORT,
  60. .devt_type = DFL_FPGA_DEVT_PORT},
  61. };
  62. /**
  63. * struct dfl_chardev_info - chardev information of dfl feature device
  64. * @name: nmae string of the char device.
  65. * @devt: devt of the char device.
  66. */
  67. struct dfl_chardev_info {
  68. const char *name;
  69. dev_t devt;
  70. };
  71. /* indexed by enum dfl_fpga_devt_type */
  72. static struct dfl_chardev_info dfl_chrdevs[] = {
  73. {.name = DFL_FPGA_FEATURE_DEV_FME},
  74. {.name = DFL_FPGA_FEATURE_DEV_PORT},
  75. };
  76. static void dfl_ids_init(void)
  77. {
  78. int i;
  79. for (i = 0; i < ARRAY_SIZE(dfl_devs); i++)
  80. idr_init(&dfl_devs[i].id);
  81. }
  82. static void dfl_ids_destroy(void)
  83. {
  84. int i;
  85. for (i = 0; i < ARRAY_SIZE(dfl_devs); i++)
  86. idr_destroy(&dfl_devs[i].id);
  87. }
  88. static int dfl_id_alloc(enum dfl_id_type type, struct device *dev)
  89. {
  90. int id;
  91. WARN_ON(type >= DFL_ID_MAX);
  92. mutex_lock(&dfl_id_mutex);
  93. id = idr_alloc(&dfl_devs[type].id, dev, 0, 0, GFP_KERNEL);
  94. mutex_unlock(&dfl_id_mutex);
  95. return id;
  96. }
  97. static void dfl_id_free(enum dfl_id_type type, int id)
  98. {
  99. WARN_ON(type >= DFL_ID_MAX);
  100. mutex_lock(&dfl_id_mutex);
  101. idr_remove(&dfl_devs[type].id, id);
  102. mutex_unlock(&dfl_id_mutex);
  103. }
  104. static enum dfl_id_type feature_dev_id_type(struct platform_device *pdev)
  105. {
  106. int i;
  107. for (i = 0; i < ARRAY_SIZE(dfl_devs); i++)
  108. if (!strcmp(dfl_devs[i].name, pdev->name))
  109. return i;
  110. return DFL_ID_MAX;
  111. }
  112. static enum dfl_id_type dfh_id_to_type(u16 id)
  113. {
  114. int i;
  115. for (i = 0; i < ARRAY_SIZE(dfl_devs); i++)
  116. if (dfl_devs[i].dfh_id == id)
  117. return i;
  118. return DFL_ID_MAX;
  119. }
  120. /*
  121. * introduce a global port_ops list, it allows port drivers to register ops
  122. * in such list, then other feature devices (e.g. FME), could use the port
  123. * functions even related port platform device is hidden. Below is one example,
  124. * in virtualization case of PCIe-based FPGA DFL device, when SRIOV is
  125. * enabled, port (and it's AFU) is turned into VF and port platform device
  126. * is hidden from system but it's still required to access port to finish FPGA
  127. * reconfiguration function in FME.
  128. */
  129. static DEFINE_MUTEX(dfl_port_ops_mutex);
  130. static LIST_HEAD(dfl_port_ops_list);
  131. /**
  132. * dfl_fpga_port_ops_get - get matched port ops from the global list
  133. * @pdev: platform device to match with associated port ops.
  134. * Return: matched port ops on success, NULL otherwise.
  135. *
  136. * Please note that must dfl_fpga_port_ops_put after use the port_ops.
  137. */
  138. struct dfl_fpga_port_ops *dfl_fpga_port_ops_get(struct platform_device *pdev)
  139. {
  140. struct dfl_fpga_port_ops *ops = NULL;
  141. mutex_lock(&dfl_port_ops_mutex);
  142. if (list_empty(&dfl_port_ops_list))
  143. goto done;
  144. list_for_each_entry(ops, &dfl_port_ops_list, node) {
  145. /* match port_ops using the name of platform device */
  146. if (!strcmp(pdev->name, ops->name)) {
  147. if (!try_module_get(ops->owner))
  148. ops = NULL;
  149. goto done;
  150. }
  151. }
  152. ops = NULL;
  153. done:
  154. mutex_unlock(&dfl_port_ops_mutex);
  155. return ops;
  156. }
  157. EXPORT_SYMBOL_GPL(dfl_fpga_port_ops_get);
  158. /**
  159. * dfl_fpga_port_ops_put - put port ops
  160. * @ops: port ops.
  161. */
  162. void dfl_fpga_port_ops_put(struct dfl_fpga_port_ops *ops)
  163. {
  164. if (ops && ops->owner)
  165. module_put(ops->owner);
  166. }
  167. EXPORT_SYMBOL_GPL(dfl_fpga_port_ops_put);
  168. /**
  169. * dfl_fpga_port_ops_add - add port_ops to global list
  170. * @ops: port ops to add.
  171. */
  172. void dfl_fpga_port_ops_add(struct dfl_fpga_port_ops *ops)
  173. {
  174. mutex_lock(&dfl_port_ops_mutex);
  175. list_add_tail(&ops->node, &dfl_port_ops_list);
  176. mutex_unlock(&dfl_port_ops_mutex);
  177. }
  178. EXPORT_SYMBOL_GPL(dfl_fpga_port_ops_add);
  179. /**
  180. * dfl_fpga_port_ops_del - remove port_ops from global list
  181. * @ops: port ops to del.
  182. */
  183. void dfl_fpga_port_ops_del(struct dfl_fpga_port_ops *ops)
  184. {
  185. mutex_lock(&dfl_port_ops_mutex);
  186. list_del(&ops->node);
  187. mutex_unlock(&dfl_port_ops_mutex);
  188. }
  189. EXPORT_SYMBOL_GPL(dfl_fpga_port_ops_del);
  190. /**
  191. * dfl_fpga_check_port_id - check the port id
  192. * @pdev: port platform device.
  193. * @pport_id: port id to compare.
  194. *
  195. * Return: 1 if port device matches with given port id, otherwise 0.
  196. */
  197. int dfl_fpga_check_port_id(struct platform_device *pdev, void *pport_id)
  198. {
  199. struct dfl_feature_platform_data *pdata = dev_get_platdata(&pdev->dev);
  200. struct dfl_fpga_port_ops *port_ops;
  201. if (pdata->id != FEATURE_DEV_ID_UNUSED)
  202. return pdata->id == *(int *)pport_id;
  203. port_ops = dfl_fpga_port_ops_get(pdev);
  204. if (!port_ops || !port_ops->get_id)
  205. return 0;
  206. pdata->id = port_ops->get_id(pdev);
  207. dfl_fpga_port_ops_put(port_ops);
  208. return pdata->id == *(int *)pport_id;
  209. }
  210. EXPORT_SYMBOL_GPL(dfl_fpga_check_port_id);
  211. static DEFINE_IDA(dfl_device_ida);
  212. static const struct dfl_device_id *
  213. dfl_match_one_device(const struct dfl_device_id *id, struct dfl_device *ddev)
  214. {
  215. if (id->type == ddev->type && id->feature_id == ddev->feature_id)
  216. return id;
  217. return NULL;
  218. }
  219. static int dfl_bus_match(struct device *dev, const struct device_driver *drv)
  220. {
  221. struct dfl_device *ddev = to_dfl_dev(dev);
  222. const struct dfl_driver *ddrv = to_dfl_drv(drv);
  223. const struct dfl_device_id *id_entry;
  224. id_entry = ddrv->id_table;
  225. if (id_entry) {
  226. while (id_entry->feature_id) {
  227. if (dfl_match_one_device(id_entry, ddev)) {
  228. ddev->id_entry = id_entry;
  229. return 1;
  230. }
  231. id_entry++;
  232. }
  233. }
  234. return 0;
  235. }
  236. static int dfl_bus_probe(struct device *dev)
  237. {
  238. struct dfl_driver *ddrv = to_dfl_drv(dev->driver);
  239. struct dfl_device *ddev = to_dfl_dev(dev);
  240. return ddrv->probe(ddev);
  241. }
  242. static void dfl_bus_remove(struct device *dev)
  243. {
  244. struct dfl_driver *ddrv = to_dfl_drv(dev->driver);
  245. struct dfl_device *ddev = to_dfl_dev(dev);
  246. if (ddrv->remove)
  247. ddrv->remove(ddev);
  248. }
  249. static int dfl_bus_uevent(const struct device *dev, struct kobj_uevent_env *env)
  250. {
  251. const struct dfl_device *ddev = to_dfl_dev(dev);
  252. return add_uevent_var(env, "MODALIAS=dfl:t%04Xf%04X",
  253. ddev->type, ddev->feature_id);
  254. }
  255. static ssize_t
  256. type_show(struct device *dev, struct device_attribute *attr, char *buf)
  257. {
  258. struct dfl_device *ddev = to_dfl_dev(dev);
  259. return sprintf(buf, "0x%x\n", ddev->type);
  260. }
  261. static DEVICE_ATTR_RO(type);
  262. static ssize_t
  263. feature_id_show(struct device *dev, struct device_attribute *attr, char *buf)
  264. {
  265. struct dfl_device *ddev = to_dfl_dev(dev);
  266. return sprintf(buf, "0x%x\n", ddev->feature_id);
  267. }
  268. static DEVICE_ATTR_RO(feature_id);
  269. static struct attribute *dfl_dev_attrs[] = {
  270. &dev_attr_type.attr,
  271. &dev_attr_feature_id.attr,
  272. NULL,
  273. };
  274. ATTRIBUTE_GROUPS(dfl_dev);
  275. static const struct bus_type dfl_bus_type = {
  276. .name = "dfl",
  277. .match = dfl_bus_match,
  278. .probe = dfl_bus_probe,
  279. .remove = dfl_bus_remove,
  280. .uevent = dfl_bus_uevent,
  281. .dev_groups = dfl_dev_groups,
  282. };
  283. static void release_dfl_dev(struct device *dev)
  284. {
  285. struct dfl_device *ddev = to_dfl_dev(dev);
  286. if (ddev->mmio_res.parent)
  287. release_resource(&ddev->mmio_res);
  288. kfree(ddev->params);
  289. ida_free(&dfl_device_ida, ddev->id);
  290. kfree(ddev->irqs);
  291. kfree(ddev);
  292. }
  293. static struct dfl_device *
  294. dfl_dev_add(struct dfl_feature_platform_data *pdata,
  295. struct dfl_feature *feature)
  296. {
  297. struct platform_device *pdev = pdata->dev;
  298. struct resource *parent_res;
  299. struct dfl_device *ddev;
  300. int id, i, ret;
  301. ddev = kzalloc(sizeof(*ddev), GFP_KERNEL);
  302. if (!ddev)
  303. return ERR_PTR(-ENOMEM);
  304. id = ida_alloc(&dfl_device_ida, GFP_KERNEL);
  305. if (id < 0) {
  306. dev_err(&pdev->dev, "unable to get id\n");
  307. kfree(ddev);
  308. return ERR_PTR(id);
  309. }
  310. /* freeing resources by put_device() after device_initialize() */
  311. device_initialize(&ddev->dev);
  312. ddev->dev.parent = &pdev->dev;
  313. ddev->dev.bus = &dfl_bus_type;
  314. ddev->dev.release = release_dfl_dev;
  315. ddev->id = id;
  316. ret = dev_set_name(&ddev->dev, "dfl_dev.%d", id);
  317. if (ret)
  318. goto put_dev;
  319. ddev->type = feature_dev_id_type(pdev);
  320. ddev->feature_id = feature->id;
  321. ddev->revision = feature->revision;
  322. ddev->dfh_version = feature->dfh_version;
  323. ddev->cdev = pdata->dfl_cdev;
  324. if (feature->param_size) {
  325. ddev->params = kmemdup(feature->params, feature->param_size, GFP_KERNEL);
  326. if (!ddev->params) {
  327. ret = -ENOMEM;
  328. goto put_dev;
  329. }
  330. ddev->param_size = feature->param_size;
  331. }
  332. /* add mmio resource */
  333. parent_res = &pdev->resource[feature->resource_index];
  334. ddev->mmio_res.flags = IORESOURCE_MEM;
  335. ddev->mmio_res.start = parent_res->start;
  336. ddev->mmio_res.end = parent_res->end;
  337. ddev->mmio_res.name = dev_name(&ddev->dev);
  338. ret = insert_resource(parent_res, &ddev->mmio_res);
  339. if (ret) {
  340. dev_err(&pdev->dev, "%s failed to claim resource: %pR\n",
  341. dev_name(&ddev->dev), &ddev->mmio_res);
  342. goto put_dev;
  343. }
  344. /* then add irq resource */
  345. if (feature->nr_irqs) {
  346. ddev->irqs = kcalloc(feature->nr_irqs,
  347. sizeof(*ddev->irqs), GFP_KERNEL);
  348. if (!ddev->irqs) {
  349. ret = -ENOMEM;
  350. goto put_dev;
  351. }
  352. for (i = 0; i < feature->nr_irqs; i++)
  353. ddev->irqs[i] = feature->irq_ctx[i].irq;
  354. ddev->num_irqs = feature->nr_irqs;
  355. }
  356. ret = device_add(&ddev->dev);
  357. if (ret)
  358. goto put_dev;
  359. dev_dbg(&pdev->dev, "add dfl_dev: %s\n", dev_name(&ddev->dev));
  360. return ddev;
  361. put_dev:
  362. /* calls release_dfl_dev() which does the clean up */
  363. put_device(&ddev->dev);
  364. return ERR_PTR(ret);
  365. }
  366. static void dfl_devs_remove(struct dfl_feature_platform_data *pdata)
  367. {
  368. struct dfl_feature *feature;
  369. dfl_fpga_dev_for_each_feature(pdata, feature) {
  370. if (feature->ddev) {
  371. device_unregister(&feature->ddev->dev);
  372. feature->ddev = NULL;
  373. }
  374. }
  375. }
  376. static int dfl_devs_add(struct dfl_feature_platform_data *pdata)
  377. {
  378. struct dfl_feature *feature;
  379. struct dfl_device *ddev;
  380. int ret;
  381. dfl_fpga_dev_for_each_feature(pdata, feature) {
  382. if (feature->ioaddr)
  383. continue;
  384. if (feature->ddev) {
  385. ret = -EEXIST;
  386. goto err;
  387. }
  388. ddev = dfl_dev_add(pdata, feature);
  389. if (IS_ERR(ddev)) {
  390. ret = PTR_ERR(ddev);
  391. goto err;
  392. }
  393. feature->ddev = ddev;
  394. }
  395. return 0;
  396. err:
  397. dfl_devs_remove(pdata);
  398. return ret;
  399. }
  400. int __dfl_driver_register(struct dfl_driver *dfl_drv, struct module *owner)
  401. {
  402. if (!dfl_drv || !dfl_drv->probe || !dfl_drv->id_table)
  403. return -EINVAL;
  404. dfl_drv->drv.owner = owner;
  405. dfl_drv->drv.bus = &dfl_bus_type;
  406. return driver_register(&dfl_drv->drv);
  407. }
  408. EXPORT_SYMBOL(__dfl_driver_register);
  409. void dfl_driver_unregister(struct dfl_driver *dfl_drv)
  410. {
  411. driver_unregister(&dfl_drv->drv);
  412. }
  413. EXPORT_SYMBOL(dfl_driver_unregister);
  414. #define is_header_feature(feature) ((feature)->id == FEATURE_ID_FIU_HEADER)
  415. /**
  416. * dfl_fpga_dev_feature_uinit - uinit for sub features of dfl feature device
  417. * @pdev: feature device.
  418. */
  419. void dfl_fpga_dev_feature_uinit(struct platform_device *pdev)
  420. {
  421. struct dfl_feature_platform_data *pdata = dev_get_platdata(&pdev->dev);
  422. struct dfl_feature *feature;
  423. dfl_devs_remove(pdata);
  424. dfl_fpga_dev_for_each_feature(pdata, feature) {
  425. if (feature->ops) {
  426. if (feature->ops->uinit)
  427. feature->ops->uinit(pdev, feature);
  428. feature->ops = NULL;
  429. }
  430. }
  431. }
  432. EXPORT_SYMBOL_GPL(dfl_fpga_dev_feature_uinit);
  433. static int dfl_feature_instance_init(struct platform_device *pdev,
  434. struct dfl_feature_platform_data *pdata,
  435. struct dfl_feature *feature,
  436. struct dfl_feature_driver *drv)
  437. {
  438. void __iomem *base;
  439. int ret = 0;
  440. if (!is_header_feature(feature)) {
  441. base = devm_platform_ioremap_resource(pdev,
  442. feature->resource_index);
  443. if (IS_ERR(base)) {
  444. dev_err(&pdev->dev,
  445. "ioremap failed for feature 0x%x!\n",
  446. feature->id);
  447. return PTR_ERR(base);
  448. }
  449. feature->ioaddr = base;
  450. }
  451. if (drv->ops->init) {
  452. ret = drv->ops->init(pdev, feature);
  453. if (ret)
  454. return ret;
  455. }
  456. feature->ops = drv->ops;
  457. return ret;
  458. }
  459. static bool dfl_feature_drv_match(struct dfl_feature *feature,
  460. struct dfl_feature_driver *driver)
  461. {
  462. const struct dfl_feature_id *ids = driver->id_table;
  463. if (ids) {
  464. while (ids->id) {
  465. if (ids->id == feature->id)
  466. return true;
  467. ids++;
  468. }
  469. }
  470. return false;
  471. }
  472. /**
  473. * dfl_fpga_dev_feature_init - init for sub features of dfl feature device
  474. * @pdev: feature device.
  475. * @feature_drvs: drvs for sub features.
  476. *
  477. * This function will match sub features with given feature drvs list and
  478. * use matched drv to init related sub feature.
  479. *
  480. * Return: 0 on success, negative error code otherwise.
  481. */
  482. int dfl_fpga_dev_feature_init(struct platform_device *pdev,
  483. struct dfl_feature_driver *feature_drvs)
  484. {
  485. struct dfl_feature_platform_data *pdata = dev_get_platdata(&pdev->dev);
  486. struct dfl_feature_driver *drv = feature_drvs;
  487. struct dfl_feature *feature;
  488. int ret;
  489. while (drv->ops) {
  490. dfl_fpga_dev_for_each_feature(pdata, feature) {
  491. if (dfl_feature_drv_match(feature, drv)) {
  492. ret = dfl_feature_instance_init(pdev, pdata,
  493. feature, drv);
  494. if (ret)
  495. goto exit;
  496. }
  497. }
  498. drv++;
  499. }
  500. ret = dfl_devs_add(pdata);
  501. if (ret)
  502. goto exit;
  503. return 0;
  504. exit:
  505. dfl_fpga_dev_feature_uinit(pdev);
  506. return ret;
  507. }
  508. EXPORT_SYMBOL_GPL(dfl_fpga_dev_feature_init);
  509. static void dfl_chardev_uinit(void)
  510. {
  511. int i;
  512. for (i = 0; i < DFL_FPGA_DEVT_MAX; i++)
  513. if (MAJOR(dfl_chrdevs[i].devt)) {
  514. unregister_chrdev_region(dfl_chrdevs[i].devt,
  515. MINORMASK + 1);
  516. dfl_chrdevs[i].devt = MKDEV(0, 0);
  517. }
  518. }
  519. static int dfl_chardev_init(void)
  520. {
  521. int i, ret;
  522. for (i = 0; i < DFL_FPGA_DEVT_MAX; i++) {
  523. ret = alloc_chrdev_region(&dfl_chrdevs[i].devt, 0,
  524. MINORMASK + 1, dfl_chrdevs[i].name);
  525. if (ret)
  526. goto exit;
  527. }
  528. return 0;
  529. exit:
  530. dfl_chardev_uinit();
  531. return ret;
  532. }
  533. static dev_t dfl_get_devt(enum dfl_fpga_devt_type type, int id)
  534. {
  535. if (type >= DFL_FPGA_DEVT_MAX)
  536. return 0;
  537. return MKDEV(MAJOR(dfl_chrdevs[type].devt), id);
  538. }
  539. /**
  540. * dfl_fpga_dev_ops_register - register cdev ops for feature dev
  541. *
  542. * @pdev: feature dev.
  543. * @fops: file operations for feature dev's cdev.
  544. * @owner: owning module/driver.
  545. *
  546. * Return: 0 on success, negative error code otherwise.
  547. */
  548. int dfl_fpga_dev_ops_register(struct platform_device *pdev,
  549. const struct file_operations *fops,
  550. struct module *owner)
  551. {
  552. struct dfl_feature_platform_data *pdata = dev_get_platdata(&pdev->dev);
  553. cdev_init(&pdata->cdev, fops);
  554. pdata->cdev.owner = owner;
  555. /*
  556. * set parent to the feature device so that its refcount is
  557. * decreased after the last refcount of cdev is gone, that
  558. * makes sure the feature device is valid during device
  559. * file's life-cycle.
  560. */
  561. pdata->cdev.kobj.parent = &pdev->dev.kobj;
  562. return cdev_add(&pdata->cdev, pdev->dev.devt, 1);
  563. }
  564. EXPORT_SYMBOL_GPL(dfl_fpga_dev_ops_register);
  565. /**
  566. * dfl_fpga_dev_ops_unregister - unregister cdev ops for feature dev
  567. * @pdev: feature dev.
  568. */
  569. void dfl_fpga_dev_ops_unregister(struct platform_device *pdev)
  570. {
  571. struct dfl_feature_platform_data *pdata = dev_get_platdata(&pdev->dev);
  572. cdev_del(&pdata->cdev);
  573. }
  574. EXPORT_SYMBOL_GPL(dfl_fpga_dev_ops_unregister);
  575. /**
  576. * struct build_feature_devs_info - info collected during feature dev build.
  577. *
  578. * @dev: device to enumerate.
  579. * @cdev: the container device for all feature devices.
  580. * @nr_irqs: number of irqs for all feature devices.
  581. * @irq_table: Linux IRQ numbers for all irqs, indexed by local irq index of
  582. * this device.
  583. * @feature_dev: current feature device.
  584. * @ioaddr: header register region address of current FIU in enumeration.
  585. * @start: register resource start of current FIU.
  586. * @len: max register resource length of current FIU.
  587. * @sub_features: a sub features linked list for feature device in enumeration.
  588. * @feature_num: number of sub features for feature device in enumeration.
  589. */
  590. struct build_feature_devs_info {
  591. struct device *dev;
  592. struct dfl_fpga_cdev *cdev;
  593. unsigned int nr_irqs;
  594. int *irq_table;
  595. struct platform_device *feature_dev;
  596. void __iomem *ioaddr;
  597. resource_size_t start;
  598. resource_size_t len;
  599. struct list_head sub_features;
  600. int feature_num;
  601. };
  602. /**
  603. * struct dfl_feature_info - sub feature info collected during feature dev build
  604. *
  605. * @fid: id of this sub feature.
  606. * @revision: revision of this sub feature
  607. * @dfh_version: version of Device Feature Header (DFH)
  608. * @mmio_res: mmio resource of this sub feature.
  609. * @ioaddr: mapped base address of mmio resource.
  610. * @node: node in sub_features linked list.
  611. * @irq_base: start of irq index in this sub feature.
  612. * @nr_irqs: number of irqs of this sub feature.
  613. * @param_size: size DFH parameters.
  614. * @params: DFH parameter data.
  615. */
  616. struct dfl_feature_info {
  617. u16 fid;
  618. u8 revision;
  619. u8 dfh_version;
  620. struct resource mmio_res;
  621. void __iomem *ioaddr;
  622. struct list_head node;
  623. unsigned int irq_base;
  624. unsigned int nr_irqs;
  625. unsigned int param_size;
  626. u64 params[];
  627. };
  628. static void dfl_fpga_cdev_add_port_dev(struct dfl_fpga_cdev *cdev,
  629. struct platform_device *port)
  630. {
  631. struct dfl_feature_platform_data *pdata = dev_get_platdata(&port->dev);
  632. mutex_lock(&cdev->lock);
  633. list_add(&pdata->node, &cdev->port_dev_list);
  634. get_device(&pdata->dev->dev);
  635. mutex_unlock(&cdev->lock);
  636. }
  637. /*
  638. * register current feature device, it is called when we need to switch to
  639. * another feature parsing or we have parsed all features on given device
  640. * feature list.
  641. */
  642. static int build_info_commit_dev(struct build_feature_devs_info *binfo)
  643. {
  644. struct platform_device *fdev = binfo->feature_dev;
  645. struct dfl_feature_platform_data *pdata;
  646. struct dfl_feature_info *finfo, *p;
  647. enum dfl_id_type type;
  648. int ret, index = 0, res_idx = 0;
  649. type = feature_dev_id_type(fdev);
  650. if (WARN_ON_ONCE(type >= DFL_ID_MAX))
  651. return -EINVAL;
  652. /*
  653. * we do not need to care for the memory which is associated with
  654. * the platform device. After calling platform_device_unregister(),
  655. * it will be automatically freed by device's release() callback,
  656. * platform_device_release().
  657. */
  658. pdata = kzalloc(struct_size(pdata, features, binfo->feature_num), GFP_KERNEL);
  659. if (!pdata)
  660. return -ENOMEM;
  661. pdata->dev = fdev;
  662. pdata->num = binfo->feature_num;
  663. pdata->dfl_cdev = binfo->cdev;
  664. pdata->id = FEATURE_DEV_ID_UNUSED;
  665. mutex_init(&pdata->lock);
  666. lockdep_set_class_and_name(&pdata->lock, &dfl_pdata_keys[type],
  667. dfl_pdata_key_strings[type]);
  668. /*
  669. * the count should be initialized to 0 to make sure
  670. *__fpga_port_enable() following __fpga_port_disable()
  671. * works properly for port device.
  672. * and it should always be 0 for fme device.
  673. */
  674. WARN_ON(pdata->disable_count);
  675. fdev->dev.platform_data = pdata;
  676. /* each sub feature has one MMIO resource */
  677. fdev->num_resources = binfo->feature_num;
  678. fdev->resource = kcalloc(binfo->feature_num, sizeof(*fdev->resource),
  679. GFP_KERNEL);
  680. if (!fdev->resource)
  681. return -ENOMEM;
  682. /* fill features and resource information for feature dev */
  683. list_for_each_entry_safe(finfo, p, &binfo->sub_features, node) {
  684. struct dfl_feature *feature = &pdata->features[index++];
  685. struct dfl_feature_irq_ctx *ctx;
  686. unsigned int i;
  687. /* save resource information for each feature */
  688. feature->dev = fdev;
  689. feature->id = finfo->fid;
  690. feature->revision = finfo->revision;
  691. feature->dfh_version = finfo->dfh_version;
  692. if (finfo->param_size) {
  693. feature->params = devm_kmemdup(binfo->dev,
  694. finfo->params, finfo->param_size,
  695. GFP_KERNEL);
  696. if (!feature->params)
  697. return -ENOMEM;
  698. feature->param_size = finfo->param_size;
  699. }
  700. /*
  701. * the FIU header feature has some fundamental functions (sriov
  702. * set, port enable/disable) needed for the dfl bus device and
  703. * other sub features. So its mmio resource should be mapped by
  704. * DFL bus device. And we should not assign it to feature
  705. * devices (dfl-fme/afu) again.
  706. */
  707. if (is_header_feature(feature)) {
  708. feature->resource_index = -1;
  709. feature->ioaddr =
  710. devm_ioremap_resource(binfo->dev,
  711. &finfo->mmio_res);
  712. if (IS_ERR(feature->ioaddr))
  713. return PTR_ERR(feature->ioaddr);
  714. } else {
  715. feature->resource_index = res_idx;
  716. fdev->resource[res_idx++] = finfo->mmio_res;
  717. }
  718. if (finfo->nr_irqs) {
  719. ctx = devm_kcalloc(binfo->dev, finfo->nr_irqs,
  720. sizeof(*ctx), GFP_KERNEL);
  721. if (!ctx)
  722. return -ENOMEM;
  723. for (i = 0; i < finfo->nr_irqs; i++)
  724. ctx[i].irq =
  725. binfo->irq_table[finfo->irq_base + i];
  726. feature->irq_ctx = ctx;
  727. feature->nr_irqs = finfo->nr_irqs;
  728. }
  729. list_del(&finfo->node);
  730. kfree(finfo);
  731. }
  732. ret = platform_device_add(binfo->feature_dev);
  733. if (!ret) {
  734. if (type == PORT_ID)
  735. dfl_fpga_cdev_add_port_dev(binfo->cdev,
  736. binfo->feature_dev);
  737. else
  738. binfo->cdev->fme_dev =
  739. get_device(&binfo->feature_dev->dev);
  740. /*
  741. * reset it to avoid build_info_free() freeing their resource.
  742. *
  743. * The resource of successfully registered feature devices
  744. * will be freed by platform_device_unregister(). See the
  745. * comments in build_info_create_dev().
  746. */
  747. binfo->feature_dev = NULL;
  748. }
  749. return ret;
  750. }
  751. static int
  752. build_info_create_dev(struct build_feature_devs_info *binfo,
  753. enum dfl_id_type type)
  754. {
  755. struct platform_device *fdev;
  756. if (type >= DFL_ID_MAX)
  757. return -EINVAL;
  758. /*
  759. * we use -ENODEV as the initialization indicator which indicates
  760. * whether the id need to be reclaimed
  761. */
  762. fdev = platform_device_alloc(dfl_devs[type].name, -ENODEV);
  763. if (!fdev)
  764. return -ENOMEM;
  765. binfo->feature_dev = fdev;
  766. binfo->feature_num = 0;
  767. INIT_LIST_HEAD(&binfo->sub_features);
  768. fdev->id = dfl_id_alloc(type, &fdev->dev);
  769. if (fdev->id < 0)
  770. return fdev->id;
  771. fdev->dev.parent = &binfo->cdev->region->dev;
  772. fdev->dev.devt = dfl_get_devt(dfl_devs[type].devt_type, fdev->id);
  773. return 0;
  774. }
  775. static void build_info_free(struct build_feature_devs_info *binfo)
  776. {
  777. struct dfl_feature_info *finfo, *p;
  778. /*
  779. * it is a valid id, free it. See comments in
  780. * build_info_create_dev()
  781. */
  782. if (binfo->feature_dev && binfo->feature_dev->id >= 0) {
  783. dfl_id_free(feature_dev_id_type(binfo->feature_dev),
  784. binfo->feature_dev->id);
  785. list_for_each_entry_safe(finfo, p, &binfo->sub_features, node) {
  786. list_del(&finfo->node);
  787. kfree(finfo);
  788. }
  789. }
  790. platform_device_put(binfo->feature_dev);
  791. devm_kfree(binfo->dev, binfo);
  792. }
  793. static inline u32 feature_size(u64 value)
  794. {
  795. u32 ofst = FIELD_GET(DFH_NEXT_HDR_OFST, value);
  796. /* workaround for private features with invalid size, use 4K instead */
  797. return ofst ? ofst : 4096;
  798. }
  799. static u16 feature_id(u64 value)
  800. {
  801. u16 id = FIELD_GET(DFH_ID, value);
  802. u8 type = FIELD_GET(DFH_TYPE, value);
  803. if (type == DFH_TYPE_FIU)
  804. return FEATURE_ID_FIU_HEADER;
  805. else if (type == DFH_TYPE_PRIVATE)
  806. return id;
  807. else if (type == DFH_TYPE_AFU)
  808. return FEATURE_ID_AFU;
  809. WARN_ON(1);
  810. return 0;
  811. }
  812. static u64 *find_param(u64 *params, resource_size_t max, int param_id)
  813. {
  814. u64 *end = params + max / sizeof(u64);
  815. u64 v, next;
  816. while (params < end) {
  817. v = *params;
  818. if (param_id == FIELD_GET(DFHv1_PARAM_HDR_ID, v))
  819. return params;
  820. if (FIELD_GET(DFHv1_PARAM_HDR_NEXT_EOP, v))
  821. break;
  822. next = FIELD_GET(DFHv1_PARAM_HDR_NEXT_OFFSET, v);
  823. params += next;
  824. }
  825. return NULL;
  826. }
  827. /**
  828. * dfh_find_param() - find parameter block for the given parameter id
  829. * @dfl_dev: dfl device
  830. * @param_id: id of dfl parameter
  831. * @psize: destination to store size of parameter data in bytes
  832. *
  833. * Return: pointer to start of parameter data, PTR_ERR otherwise.
  834. */
  835. void *dfh_find_param(struct dfl_device *dfl_dev, int param_id, size_t *psize)
  836. {
  837. u64 *phdr = find_param(dfl_dev->params, dfl_dev->param_size, param_id);
  838. if (!phdr)
  839. return ERR_PTR(-ENOENT);
  840. if (psize)
  841. *psize = (FIELD_GET(DFHv1_PARAM_HDR_NEXT_OFFSET, *phdr) - 1) * sizeof(u64);
  842. return phdr + 1;
  843. }
  844. EXPORT_SYMBOL_GPL(dfh_find_param);
  845. static int parse_feature_irqs(struct build_feature_devs_info *binfo,
  846. resource_size_t ofst, struct dfl_feature_info *finfo)
  847. {
  848. void __iomem *base = binfo->ioaddr + ofst;
  849. unsigned int i, ibase, inr = 0;
  850. void *params = finfo->params;
  851. enum dfl_id_type type;
  852. u16 fid = finfo->fid;
  853. int virq;
  854. u64 *p;
  855. u64 v;
  856. switch (finfo->dfh_version) {
  857. case 0:
  858. /*
  859. * DFHv0 only provides MMIO resource information for each feature
  860. * in the DFL header. There is no generic interrupt information.
  861. * Instead, features with interrupt functionality provide
  862. * the information in feature specific registers.
  863. */
  864. type = feature_dev_id_type(binfo->feature_dev);
  865. if (type == PORT_ID) {
  866. switch (fid) {
  867. case PORT_FEATURE_ID_UINT:
  868. v = readq(base + PORT_UINT_CAP);
  869. ibase = FIELD_GET(PORT_UINT_CAP_FST_VECT, v);
  870. inr = FIELD_GET(PORT_UINT_CAP_INT_NUM, v);
  871. break;
  872. case PORT_FEATURE_ID_ERROR:
  873. v = readq(base + PORT_ERROR_CAP);
  874. ibase = FIELD_GET(PORT_ERROR_CAP_INT_VECT, v);
  875. inr = FIELD_GET(PORT_ERROR_CAP_SUPP_INT, v);
  876. break;
  877. }
  878. } else if (type == FME_ID) {
  879. switch (fid) {
  880. case FME_FEATURE_ID_GLOBAL_ERR:
  881. v = readq(base + FME_ERROR_CAP);
  882. ibase = FIELD_GET(FME_ERROR_CAP_INT_VECT, v);
  883. inr = FIELD_GET(FME_ERROR_CAP_SUPP_INT, v);
  884. break;
  885. }
  886. }
  887. break;
  888. case 1:
  889. /*
  890. * DFHv1 provides interrupt resource information in DFHv1
  891. * parameter blocks.
  892. */
  893. p = find_param(params, finfo->param_size, DFHv1_PARAM_ID_MSI_X);
  894. if (!p)
  895. break;
  896. p++;
  897. ibase = FIELD_GET(DFHv1_PARAM_MSI_X_STARTV, *p);
  898. inr = FIELD_GET(DFHv1_PARAM_MSI_X_NUMV, *p);
  899. break;
  900. default:
  901. dev_warn(binfo->dev, "unexpected DFH version %d\n", finfo->dfh_version);
  902. break;
  903. }
  904. if (!inr) {
  905. finfo->irq_base = 0;
  906. finfo->nr_irqs = 0;
  907. return 0;
  908. }
  909. dev_dbg(binfo->dev, "feature: 0x%x, irq_base: %u, nr_irqs: %u\n",
  910. fid, ibase, inr);
  911. if (ibase + inr > binfo->nr_irqs) {
  912. dev_err(binfo->dev,
  913. "Invalid interrupt number in feature 0x%x\n", fid);
  914. return -EINVAL;
  915. }
  916. for (i = 0; i < inr; i++) {
  917. virq = binfo->irq_table[ibase + i];
  918. if (virq < 0 || virq > NR_IRQS) {
  919. dev_err(binfo->dev,
  920. "Invalid irq table entry for feature 0x%x\n",
  921. fid);
  922. return -EINVAL;
  923. }
  924. }
  925. finfo->irq_base = ibase;
  926. finfo->nr_irqs = inr;
  927. return 0;
  928. }
  929. static int dfh_get_param_size(void __iomem *dfh_base, resource_size_t max)
  930. {
  931. int size = 0;
  932. u64 v, next;
  933. if (!FIELD_GET(DFHv1_CSR_SIZE_GRP_HAS_PARAMS,
  934. readq(dfh_base + DFHv1_CSR_SIZE_GRP)))
  935. return 0;
  936. while (size + DFHv1_PARAM_HDR < max) {
  937. v = readq(dfh_base + DFHv1_PARAM_HDR + size);
  938. next = FIELD_GET(DFHv1_PARAM_HDR_NEXT_OFFSET, v);
  939. if (!next)
  940. return -EINVAL;
  941. size += next * sizeof(u64);
  942. if (FIELD_GET(DFHv1_PARAM_HDR_NEXT_EOP, v))
  943. return size;
  944. }
  945. return -ENOENT;
  946. }
  947. /*
  948. * when create sub feature instances, for private features, it doesn't need
  949. * to provide resource size and feature id as they could be read from DFH
  950. * register. For afu sub feature, its register region only contains user
  951. * defined registers, so never trust any information from it, just use the
  952. * resource size information provided by its parent FIU.
  953. */
  954. static int
  955. create_feature_instance(struct build_feature_devs_info *binfo,
  956. resource_size_t ofst, resource_size_t size, u16 fid)
  957. {
  958. struct dfl_feature_info *finfo;
  959. resource_size_t start, end;
  960. int dfh_psize = 0;
  961. u8 revision = 0;
  962. u64 v, addr_off;
  963. u8 dfh_ver = 0;
  964. int ret;
  965. if (fid != FEATURE_ID_AFU) {
  966. v = readq(binfo->ioaddr + ofst);
  967. revision = FIELD_GET(DFH_REVISION, v);
  968. dfh_ver = FIELD_GET(DFH_VERSION, v);
  969. /* read feature size and id if inputs are invalid */
  970. size = size ? size : feature_size(v);
  971. fid = fid ? fid : feature_id(v);
  972. if (dfh_ver == 1) {
  973. dfh_psize = dfh_get_param_size(binfo->ioaddr + ofst, size);
  974. if (dfh_psize < 0) {
  975. dev_err(binfo->dev,
  976. "failed to read size of DFHv1 parameters %d\n",
  977. dfh_psize);
  978. return dfh_psize;
  979. }
  980. dev_dbg(binfo->dev, "dfhv1_psize %d\n", dfh_psize);
  981. }
  982. }
  983. if (binfo->len - ofst < size)
  984. return -EINVAL;
  985. finfo = kzalloc(struct_size(finfo, params, dfh_psize / sizeof(u64)), GFP_KERNEL);
  986. if (!finfo)
  987. return -ENOMEM;
  988. memcpy_fromio(finfo->params, binfo->ioaddr + ofst + DFHv1_PARAM_HDR, dfh_psize);
  989. finfo->param_size = dfh_psize;
  990. finfo->fid = fid;
  991. finfo->revision = revision;
  992. finfo->dfh_version = dfh_ver;
  993. if (dfh_ver == 1) {
  994. v = readq(binfo->ioaddr + ofst + DFHv1_CSR_ADDR);
  995. addr_off = FIELD_GET(DFHv1_CSR_ADDR_MASK, v);
  996. if (FIELD_GET(DFHv1_CSR_ADDR_REL, v))
  997. start = addr_off << 1;
  998. else
  999. start = binfo->start + ofst + addr_off;
  1000. v = readq(binfo->ioaddr + ofst + DFHv1_CSR_SIZE_GRP);
  1001. end = start + FIELD_GET(DFHv1_CSR_SIZE_GRP_SIZE, v) - 1;
  1002. } else {
  1003. start = binfo->start + ofst;
  1004. end = start + size - 1;
  1005. }
  1006. finfo->mmio_res.flags = IORESOURCE_MEM;
  1007. finfo->mmio_res.start = start;
  1008. finfo->mmio_res.end = end;
  1009. ret = parse_feature_irqs(binfo, ofst, finfo);
  1010. if (ret) {
  1011. kfree(finfo);
  1012. return ret;
  1013. }
  1014. list_add_tail(&finfo->node, &binfo->sub_features);
  1015. binfo->feature_num++;
  1016. return 0;
  1017. }
  1018. static int parse_feature_port_afu(struct build_feature_devs_info *binfo,
  1019. resource_size_t ofst)
  1020. {
  1021. u64 v = readq(binfo->ioaddr + PORT_HDR_CAP);
  1022. u32 size = FIELD_GET(PORT_CAP_MMIO_SIZE, v) << 10;
  1023. WARN_ON(!size);
  1024. return create_feature_instance(binfo, ofst, size, FEATURE_ID_AFU);
  1025. }
  1026. #define is_feature_dev_detected(binfo) (!!(binfo)->feature_dev)
  1027. static int parse_feature_afu(struct build_feature_devs_info *binfo,
  1028. resource_size_t ofst)
  1029. {
  1030. if (!is_feature_dev_detected(binfo)) {
  1031. dev_err(binfo->dev, "this AFU does not belong to any FIU.\n");
  1032. return -EINVAL;
  1033. }
  1034. switch (feature_dev_id_type(binfo->feature_dev)) {
  1035. case PORT_ID:
  1036. return parse_feature_port_afu(binfo, ofst);
  1037. default:
  1038. dev_info(binfo->dev, "AFU belonging to FIU %s is not supported yet.\n",
  1039. binfo->feature_dev->name);
  1040. }
  1041. return 0;
  1042. }
  1043. static int build_info_prepare(struct build_feature_devs_info *binfo,
  1044. resource_size_t start, resource_size_t len)
  1045. {
  1046. struct device *dev = binfo->dev;
  1047. void __iomem *ioaddr;
  1048. if (!devm_request_mem_region(dev, start, len, dev_name(dev))) {
  1049. dev_err(dev, "request region fail, start:%pa, len:%pa\n",
  1050. &start, &len);
  1051. return -EBUSY;
  1052. }
  1053. ioaddr = devm_ioremap(dev, start, len);
  1054. if (!ioaddr) {
  1055. dev_err(dev, "ioremap region fail, start:%pa, len:%pa\n",
  1056. &start, &len);
  1057. return -ENOMEM;
  1058. }
  1059. binfo->start = start;
  1060. binfo->len = len;
  1061. binfo->ioaddr = ioaddr;
  1062. return 0;
  1063. }
  1064. static void build_info_complete(struct build_feature_devs_info *binfo)
  1065. {
  1066. devm_iounmap(binfo->dev, binfo->ioaddr);
  1067. devm_release_mem_region(binfo->dev, binfo->start, binfo->len);
  1068. }
  1069. static int parse_feature_fiu(struct build_feature_devs_info *binfo,
  1070. resource_size_t ofst)
  1071. {
  1072. int ret = 0;
  1073. u32 offset;
  1074. u16 id;
  1075. u64 v;
  1076. if (is_feature_dev_detected(binfo)) {
  1077. build_info_complete(binfo);
  1078. ret = build_info_commit_dev(binfo);
  1079. if (ret)
  1080. return ret;
  1081. ret = build_info_prepare(binfo, binfo->start + ofst,
  1082. binfo->len - ofst);
  1083. if (ret)
  1084. return ret;
  1085. }
  1086. v = readq(binfo->ioaddr + DFH);
  1087. id = FIELD_GET(DFH_ID, v);
  1088. /* create platform device for dfl feature dev */
  1089. ret = build_info_create_dev(binfo, dfh_id_to_type(id));
  1090. if (ret)
  1091. return ret;
  1092. ret = create_feature_instance(binfo, 0, 0, 0);
  1093. if (ret)
  1094. return ret;
  1095. /*
  1096. * find and parse FIU's child AFU via its NEXT_AFU register.
  1097. * please note that only Port has valid NEXT_AFU pointer per spec.
  1098. */
  1099. v = readq(binfo->ioaddr + NEXT_AFU);
  1100. offset = FIELD_GET(NEXT_AFU_NEXT_DFH_OFST, v);
  1101. if (offset)
  1102. return parse_feature_afu(binfo, offset);
  1103. dev_dbg(binfo->dev, "No AFUs detected on FIU %d\n", id);
  1104. return ret;
  1105. }
  1106. static int parse_feature_private(struct build_feature_devs_info *binfo,
  1107. resource_size_t ofst)
  1108. {
  1109. if (!is_feature_dev_detected(binfo)) {
  1110. dev_err(binfo->dev, "the private feature 0x%x does not belong to any AFU.\n",
  1111. feature_id(readq(binfo->ioaddr + ofst)));
  1112. return -EINVAL;
  1113. }
  1114. return create_feature_instance(binfo, ofst, 0, 0);
  1115. }
  1116. /**
  1117. * parse_feature - parse a feature on given device feature list
  1118. *
  1119. * @binfo: build feature devices information.
  1120. * @ofst: offset to current FIU header
  1121. */
  1122. static int parse_feature(struct build_feature_devs_info *binfo,
  1123. resource_size_t ofst)
  1124. {
  1125. u64 v;
  1126. u32 type;
  1127. v = readq(binfo->ioaddr + ofst + DFH);
  1128. type = FIELD_GET(DFH_TYPE, v);
  1129. switch (type) {
  1130. case DFH_TYPE_AFU:
  1131. return parse_feature_afu(binfo, ofst);
  1132. case DFH_TYPE_PRIVATE:
  1133. return parse_feature_private(binfo, ofst);
  1134. case DFH_TYPE_FIU:
  1135. return parse_feature_fiu(binfo, ofst);
  1136. default:
  1137. dev_info(binfo->dev,
  1138. "Feature Type %x is not supported.\n", type);
  1139. }
  1140. return 0;
  1141. }
  1142. static int parse_feature_list(struct build_feature_devs_info *binfo,
  1143. resource_size_t start, resource_size_t len)
  1144. {
  1145. resource_size_t end = start + len;
  1146. int ret = 0;
  1147. u32 ofst = 0;
  1148. u64 v;
  1149. ret = build_info_prepare(binfo, start, len);
  1150. if (ret)
  1151. return ret;
  1152. /* walk through the device feature list via DFH's next DFH pointer. */
  1153. for (; start < end; start += ofst) {
  1154. if (end - start < DFH_SIZE) {
  1155. dev_err(binfo->dev, "The region is too small to contain a feature.\n");
  1156. return -EINVAL;
  1157. }
  1158. ret = parse_feature(binfo, start - binfo->start);
  1159. if (ret)
  1160. return ret;
  1161. v = readq(binfo->ioaddr + start - binfo->start + DFH);
  1162. ofst = FIELD_GET(DFH_NEXT_HDR_OFST, v);
  1163. /* stop parsing if EOL(End of List) is set or offset is 0 */
  1164. if ((v & DFH_EOL) || !ofst)
  1165. break;
  1166. }
  1167. /* commit current feature device when reach the end of list */
  1168. build_info_complete(binfo);
  1169. if (is_feature_dev_detected(binfo))
  1170. ret = build_info_commit_dev(binfo);
  1171. return ret;
  1172. }
  1173. struct dfl_fpga_enum_info *dfl_fpga_enum_info_alloc(struct device *dev)
  1174. {
  1175. struct dfl_fpga_enum_info *info;
  1176. get_device(dev);
  1177. info = devm_kzalloc(dev, sizeof(*info), GFP_KERNEL);
  1178. if (!info) {
  1179. put_device(dev);
  1180. return NULL;
  1181. }
  1182. info->dev = dev;
  1183. INIT_LIST_HEAD(&info->dfls);
  1184. return info;
  1185. }
  1186. EXPORT_SYMBOL_GPL(dfl_fpga_enum_info_alloc);
  1187. void dfl_fpga_enum_info_free(struct dfl_fpga_enum_info *info)
  1188. {
  1189. struct dfl_fpga_enum_dfl *tmp, *dfl;
  1190. struct device *dev;
  1191. if (!info)
  1192. return;
  1193. dev = info->dev;
  1194. /* remove all device feature lists in the list. */
  1195. list_for_each_entry_safe(dfl, tmp, &info->dfls, node) {
  1196. list_del(&dfl->node);
  1197. devm_kfree(dev, dfl);
  1198. }
  1199. /* remove irq table */
  1200. if (info->irq_table)
  1201. devm_kfree(dev, info->irq_table);
  1202. devm_kfree(dev, info);
  1203. put_device(dev);
  1204. }
  1205. EXPORT_SYMBOL_GPL(dfl_fpga_enum_info_free);
  1206. /**
  1207. * dfl_fpga_enum_info_add_dfl - add info of a device feature list to enum info
  1208. *
  1209. * @info: ptr to dfl_fpga_enum_info
  1210. * @start: mmio resource address of the device feature list.
  1211. * @len: mmio resource length of the device feature list.
  1212. *
  1213. * One FPGA device may have one or more Device Feature Lists (DFLs), use this
  1214. * function to add information of each DFL to common data structure for next
  1215. * step enumeration.
  1216. *
  1217. * Return: 0 on success, negative error code otherwise.
  1218. */
  1219. int dfl_fpga_enum_info_add_dfl(struct dfl_fpga_enum_info *info,
  1220. resource_size_t start, resource_size_t len)
  1221. {
  1222. struct dfl_fpga_enum_dfl *dfl;
  1223. dfl = devm_kzalloc(info->dev, sizeof(*dfl), GFP_KERNEL);
  1224. if (!dfl)
  1225. return -ENOMEM;
  1226. dfl->start = start;
  1227. dfl->len = len;
  1228. list_add_tail(&dfl->node, &info->dfls);
  1229. return 0;
  1230. }
  1231. EXPORT_SYMBOL_GPL(dfl_fpga_enum_info_add_dfl);
  1232. /**
  1233. * dfl_fpga_enum_info_add_irq - add irq table to enum info
  1234. *
  1235. * @info: ptr to dfl_fpga_enum_info
  1236. * @nr_irqs: number of irqs of the DFL fpga device to be enumerated.
  1237. * @irq_table: Linux IRQ numbers for all irqs, indexed by local irq index of
  1238. * this device.
  1239. *
  1240. * One FPGA device may have several interrupts. This function adds irq
  1241. * information of the DFL fpga device to enum info for next step enumeration.
  1242. * This function should be called before dfl_fpga_feature_devs_enumerate().
  1243. * As we only support one irq domain for all DFLs in the same enum info, adding
  1244. * irq table a second time for the same enum info will return error.
  1245. *
  1246. * If we need to enumerate DFLs which belong to different irq domains, we
  1247. * should fill more enum info and enumerate them one by one.
  1248. *
  1249. * Return: 0 on success, negative error code otherwise.
  1250. */
  1251. int dfl_fpga_enum_info_add_irq(struct dfl_fpga_enum_info *info,
  1252. unsigned int nr_irqs, int *irq_table)
  1253. {
  1254. if (!nr_irqs || !irq_table)
  1255. return -EINVAL;
  1256. if (info->irq_table)
  1257. return -EEXIST;
  1258. info->irq_table = devm_kmemdup(info->dev, irq_table,
  1259. sizeof(int) * nr_irqs, GFP_KERNEL);
  1260. if (!info->irq_table)
  1261. return -ENOMEM;
  1262. info->nr_irqs = nr_irqs;
  1263. return 0;
  1264. }
  1265. EXPORT_SYMBOL_GPL(dfl_fpga_enum_info_add_irq);
  1266. static int remove_feature_dev(struct device *dev, void *data)
  1267. {
  1268. struct platform_device *pdev = to_platform_device(dev);
  1269. enum dfl_id_type type = feature_dev_id_type(pdev);
  1270. int id = pdev->id;
  1271. platform_device_unregister(pdev);
  1272. dfl_id_free(type, id);
  1273. return 0;
  1274. }
  1275. static void remove_feature_devs(struct dfl_fpga_cdev *cdev)
  1276. {
  1277. device_for_each_child(&cdev->region->dev, NULL, remove_feature_dev);
  1278. }
  1279. /**
  1280. * dfl_fpga_feature_devs_enumerate - enumerate feature devices
  1281. * @info: information for enumeration.
  1282. *
  1283. * This function creates a container device (base FPGA region), enumerates
  1284. * feature devices based on the enumeration info and creates platform devices
  1285. * under the container device.
  1286. *
  1287. * Return: dfl_fpga_cdev struct on success, -errno on failure
  1288. */
  1289. struct dfl_fpga_cdev *
  1290. dfl_fpga_feature_devs_enumerate(struct dfl_fpga_enum_info *info)
  1291. {
  1292. struct build_feature_devs_info *binfo;
  1293. struct dfl_fpga_enum_dfl *dfl;
  1294. struct dfl_fpga_cdev *cdev;
  1295. int ret = 0;
  1296. if (!info->dev)
  1297. return ERR_PTR(-ENODEV);
  1298. cdev = devm_kzalloc(info->dev, sizeof(*cdev), GFP_KERNEL);
  1299. if (!cdev)
  1300. return ERR_PTR(-ENOMEM);
  1301. cdev->parent = info->dev;
  1302. mutex_init(&cdev->lock);
  1303. INIT_LIST_HEAD(&cdev->port_dev_list);
  1304. cdev->region = fpga_region_register(info->dev, NULL, NULL);
  1305. if (IS_ERR(cdev->region)) {
  1306. ret = PTR_ERR(cdev->region);
  1307. goto free_cdev_exit;
  1308. }
  1309. /* create and init build info for enumeration */
  1310. binfo = devm_kzalloc(info->dev, sizeof(*binfo), GFP_KERNEL);
  1311. if (!binfo) {
  1312. ret = -ENOMEM;
  1313. goto unregister_region_exit;
  1314. }
  1315. binfo->dev = info->dev;
  1316. binfo->cdev = cdev;
  1317. binfo->nr_irqs = info->nr_irqs;
  1318. if (info->nr_irqs)
  1319. binfo->irq_table = info->irq_table;
  1320. /*
  1321. * start enumeration for all feature devices based on Device Feature
  1322. * Lists.
  1323. */
  1324. list_for_each_entry(dfl, &info->dfls, node) {
  1325. ret = parse_feature_list(binfo, dfl->start, dfl->len);
  1326. if (ret) {
  1327. remove_feature_devs(cdev);
  1328. build_info_free(binfo);
  1329. goto unregister_region_exit;
  1330. }
  1331. }
  1332. build_info_free(binfo);
  1333. return cdev;
  1334. unregister_region_exit:
  1335. fpga_region_unregister(cdev->region);
  1336. free_cdev_exit:
  1337. devm_kfree(info->dev, cdev);
  1338. return ERR_PTR(ret);
  1339. }
  1340. EXPORT_SYMBOL_GPL(dfl_fpga_feature_devs_enumerate);
  1341. /**
  1342. * dfl_fpga_feature_devs_remove - remove all feature devices
  1343. * @cdev: fpga container device.
  1344. *
  1345. * Remove the container device and all feature devices under given container
  1346. * devices.
  1347. */
  1348. void dfl_fpga_feature_devs_remove(struct dfl_fpga_cdev *cdev)
  1349. {
  1350. struct dfl_feature_platform_data *pdata, *ptmp;
  1351. mutex_lock(&cdev->lock);
  1352. if (cdev->fme_dev)
  1353. put_device(cdev->fme_dev);
  1354. list_for_each_entry_safe(pdata, ptmp, &cdev->port_dev_list, node) {
  1355. struct platform_device *port_dev = pdata->dev;
  1356. /* remove released ports */
  1357. if (!device_is_registered(&port_dev->dev)) {
  1358. dfl_id_free(feature_dev_id_type(port_dev),
  1359. port_dev->id);
  1360. platform_device_put(port_dev);
  1361. }
  1362. list_del(&pdata->node);
  1363. put_device(&port_dev->dev);
  1364. }
  1365. mutex_unlock(&cdev->lock);
  1366. remove_feature_devs(cdev);
  1367. fpga_region_unregister(cdev->region);
  1368. devm_kfree(cdev->parent, cdev);
  1369. }
  1370. EXPORT_SYMBOL_GPL(dfl_fpga_feature_devs_remove);
  1371. /**
  1372. * __dfl_fpga_cdev_find_port - find a port under given container device
  1373. *
  1374. * @cdev: container device
  1375. * @data: data passed to match function
  1376. * @match: match function used to find specific port from the port device list
  1377. *
  1378. * Find a port device under container device. This function needs to be
  1379. * invoked with lock held.
  1380. *
  1381. * Return: pointer to port's platform device if successful, NULL otherwise.
  1382. *
  1383. * NOTE: you will need to drop the device reference with put_device() after use.
  1384. */
  1385. struct platform_device *
  1386. __dfl_fpga_cdev_find_port(struct dfl_fpga_cdev *cdev, void *data,
  1387. int (*match)(struct platform_device *, void *))
  1388. {
  1389. struct dfl_feature_platform_data *pdata;
  1390. struct platform_device *port_dev;
  1391. list_for_each_entry(pdata, &cdev->port_dev_list, node) {
  1392. port_dev = pdata->dev;
  1393. if (match(port_dev, data) && get_device(&port_dev->dev))
  1394. return port_dev;
  1395. }
  1396. return NULL;
  1397. }
  1398. EXPORT_SYMBOL_GPL(__dfl_fpga_cdev_find_port);
  1399. static int __init dfl_fpga_init(void)
  1400. {
  1401. int ret;
  1402. ret = bus_register(&dfl_bus_type);
  1403. if (ret)
  1404. return ret;
  1405. dfl_ids_init();
  1406. ret = dfl_chardev_init();
  1407. if (ret) {
  1408. dfl_ids_destroy();
  1409. bus_unregister(&dfl_bus_type);
  1410. }
  1411. return ret;
  1412. }
  1413. /**
  1414. * dfl_fpga_cdev_release_port - release a port platform device
  1415. *
  1416. * @cdev: parent container device.
  1417. * @port_id: id of the port platform device.
  1418. *
  1419. * This function allows user to release a port platform device. This is a
  1420. * mandatory step before turn a port from PF into VF for SRIOV support.
  1421. *
  1422. * Return: 0 on success, negative error code otherwise.
  1423. */
  1424. int dfl_fpga_cdev_release_port(struct dfl_fpga_cdev *cdev, int port_id)
  1425. {
  1426. struct dfl_feature_platform_data *pdata;
  1427. struct platform_device *port_pdev;
  1428. int ret = -ENODEV;
  1429. mutex_lock(&cdev->lock);
  1430. port_pdev = __dfl_fpga_cdev_find_port(cdev, &port_id,
  1431. dfl_fpga_check_port_id);
  1432. if (!port_pdev)
  1433. goto unlock_exit;
  1434. if (!device_is_registered(&port_pdev->dev)) {
  1435. ret = -EBUSY;
  1436. goto put_dev_exit;
  1437. }
  1438. pdata = dev_get_platdata(&port_pdev->dev);
  1439. mutex_lock(&pdata->lock);
  1440. ret = dfl_feature_dev_use_begin(pdata, true);
  1441. mutex_unlock(&pdata->lock);
  1442. if (ret)
  1443. goto put_dev_exit;
  1444. platform_device_del(port_pdev);
  1445. cdev->released_port_num++;
  1446. put_dev_exit:
  1447. put_device(&port_pdev->dev);
  1448. unlock_exit:
  1449. mutex_unlock(&cdev->lock);
  1450. return ret;
  1451. }
  1452. EXPORT_SYMBOL_GPL(dfl_fpga_cdev_release_port);
  1453. /**
  1454. * dfl_fpga_cdev_assign_port - assign a port platform device back
  1455. *
  1456. * @cdev: parent container device.
  1457. * @port_id: id of the port platform device.
  1458. *
  1459. * This function allows user to assign a port platform device back. This is
  1460. * a mandatory step after disable SRIOV support.
  1461. *
  1462. * Return: 0 on success, negative error code otherwise.
  1463. */
  1464. int dfl_fpga_cdev_assign_port(struct dfl_fpga_cdev *cdev, int port_id)
  1465. {
  1466. struct dfl_feature_platform_data *pdata;
  1467. struct platform_device *port_pdev;
  1468. int ret = -ENODEV;
  1469. mutex_lock(&cdev->lock);
  1470. port_pdev = __dfl_fpga_cdev_find_port(cdev, &port_id,
  1471. dfl_fpga_check_port_id);
  1472. if (!port_pdev)
  1473. goto unlock_exit;
  1474. if (device_is_registered(&port_pdev->dev)) {
  1475. ret = -EBUSY;
  1476. goto put_dev_exit;
  1477. }
  1478. ret = platform_device_add(port_pdev);
  1479. if (ret)
  1480. goto put_dev_exit;
  1481. pdata = dev_get_platdata(&port_pdev->dev);
  1482. mutex_lock(&pdata->lock);
  1483. dfl_feature_dev_use_end(pdata);
  1484. mutex_unlock(&pdata->lock);
  1485. cdev->released_port_num--;
  1486. put_dev_exit:
  1487. put_device(&port_pdev->dev);
  1488. unlock_exit:
  1489. mutex_unlock(&cdev->lock);
  1490. return ret;
  1491. }
  1492. EXPORT_SYMBOL_GPL(dfl_fpga_cdev_assign_port);
  1493. static void config_port_access_mode(struct device *fme_dev, int port_id,
  1494. bool is_vf)
  1495. {
  1496. void __iomem *base;
  1497. u64 v;
  1498. base = dfl_get_feature_ioaddr_by_id(fme_dev, FME_FEATURE_ID_HEADER);
  1499. v = readq(base + FME_HDR_PORT_OFST(port_id));
  1500. v &= ~FME_PORT_OFST_ACC_CTRL;
  1501. v |= FIELD_PREP(FME_PORT_OFST_ACC_CTRL,
  1502. is_vf ? FME_PORT_OFST_ACC_VF : FME_PORT_OFST_ACC_PF);
  1503. writeq(v, base + FME_HDR_PORT_OFST(port_id));
  1504. }
  1505. #define config_port_vf_mode(dev, id) config_port_access_mode(dev, id, true)
  1506. #define config_port_pf_mode(dev, id) config_port_access_mode(dev, id, false)
  1507. /**
  1508. * dfl_fpga_cdev_config_ports_pf - configure ports to PF access mode
  1509. *
  1510. * @cdev: parent container device.
  1511. *
  1512. * This function is needed in sriov configuration routine. It could be used to
  1513. * configure the all released ports from VF access mode to PF.
  1514. */
  1515. void dfl_fpga_cdev_config_ports_pf(struct dfl_fpga_cdev *cdev)
  1516. {
  1517. struct dfl_feature_platform_data *pdata;
  1518. mutex_lock(&cdev->lock);
  1519. list_for_each_entry(pdata, &cdev->port_dev_list, node) {
  1520. if (device_is_registered(&pdata->dev->dev))
  1521. continue;
  1522. config_port_pf_mode(cdev->fme_dev, pdata->id);
  1523. }
  1524. mutex_unlock(&cdev->lock);
  1525. }
  1526. EXPORT_SYMBOL_GPL(dfl_fpga_cdev_config_ports_pf);
  1527. /**
  1528. * dfl_fpga_cdev_config_ports_vf - configure ports to VF access mode
  1529. *
  1530. * @cdev: parent container device.
  1531. * @num_vfs: VF device number.
  1532. *
  1533. * This function is needed in sriov configuration routine. It could be used to
  1534. * configure the released ports from PF access mode to VF.
  1535. *
  1536. * Return: 0 on success, negative error code otherwise.
  1537. */
  1538. int dfl_fpga_cdev_config_ports_vf(struct dfl_fpga_cdev *cdev, int num_vfs)
  1539. {
  1540. struct dfl_feature_platform_data *pdata;
  1541. int ret = 0;
  1542. mutex_lock(&cdev->lock);
  1543. /*
  1544. * can't turn multiple ports into 1 VF device, only 1 port for 1 VF
  1545. * device, so if released port number doesn't match VF device number,
  1546. * then reject the request with -EINVAL error code.
  1547. */
  1548. if (cdev->released_port_num != num_vfs) {
  1549. ret = -EINVAL;
  1550. goto done;
  1551. }
  1552. list_for_each_entry(pdata, &cdev->port_dev_list, node) {
  1553. if (device_is_registered(&pdata->dev->dev))
  1554. continue;
  1555. config_port_vf_mode(cdev->fme_dev, pdata->id);
  1556. }
  1557. done:
  1558. mutex_unlock(&cdev->lock);
  1559. return ret;
  1560. }
  1561. EXPORT_SYMBOL_GPL(dfl_fpga_cdev_config_ports_vf);
  1562. static irqreturn_t dfl_irq_handler(int irq, void *arg)
  1563. {
  1564. struct eventfd_ctx *trigger = arg;
  1565. eventfd_signal(trigger);
  1566. return IRQ_HANDLED;
  1567. }
  1568. static int do_set_irq_trigger(struct dfl_feature *feature, unsigned int idx,
  1569. int fd)
  1570. {
  1571. struct platform_device *pdev = feature->dev;
  1572. struct eventfd_ctx *trigger;
  1573. int irq, ret;
  1574. irq = feature->irq_ctx[idx].irq;
  1575. if (feature->irq_ctx[idx].trigger) {
  1576. free_irq(irq, feature->irq_ctx[idx].trigger);
  1577. kfree(feature->irq_ctx[idx].name);
  1578. eventfd_ctx_put(feature->irq_ctx[idx].trigger);
  1579. feature->irq_ctx[idx].trigger = NULL;
  1580. }
  1581. if (fd < 0)
  1582. return 0;
  1583. feature->irq_ctx[idx].name =
  1584. kasprintf(GFP_KERNEL, "fpga-irq[%u](%s-%x)", idx,
  1585. dev_name(&pdev->dev), feature->id);
  1586. if (!feature->irq_ctx[idx].name)
  1587. return -ENOMEM;
  1588. trigger = eventfd_ctx_fdget(fd);
  1589. if (IS_ERR(trigger)) {
  1590. ret = PTR_ERR(trigger);
  1591. goto free_name;
  1592. }
  1593. ret = request_irq(irq, dfl_irq_handler, 0,
  1594. feature->irq_ctx[idx].name, trigger);
  1595. if (!ret) {
  1596. feature->irq_ctx[idx].trigger = trigger;
  1597. return ret;
  1598. }
  1599. eventfd_ctx_put(trigger);
  1600. free_name:
  1601. kfree(feature->irq_ctx[idx].name);
  1602. return ret;
  1603. }
  1604. /**
  1605. * dfl_fpga_set_irq_triggers - set eventfd triggers for dfl feature interrupts
  1606. *
  1607. * @feature: dfl sub feature.
  1608. * @start: start of irq index in this dfl sub feature.
  1609. * @count: number of irqs.
  1610. * @fds: eventfds to bind with irqs. unbind related irq if fds[n] is negative.
  1611. * unbind "count" specified number of irqs if fds ptr is NULL.
  1612. *
  1613. * Bind given eventfds with irqs in this dfl sub feature. Unbind related irq if
  1614. * fds[n] is negative. Unbind "count" specified number of irqs if fds ptr is
  1615. * NULL.
  1616. *
  1617. * Return: 0 on success, negative error code otherwise.
  1618. */
  1619. int dfl_fpga_set_irq_triggers(struct dfl_feature *feature, unsigned int start,
  1620. unsigned int count, int32_t *fds)
  1621. {
  1622. unsigned int i;
  1623. int ret = 0;
  1624. /* overflow */
  1625. if (unlikely(start + count < start))
  1626. return -EINVAL;
  1627. /* exceeds nr_irqs */
  1628. if (start + count > feature->nr_irqs)
  1629. return -EINVAL;
  1630. for (i = 0; i < count; i++) {
  1631. int fd = fds ? fds[i] : -1;
  1632. ret = do_set_irq_trigger(feature, start + i, fd);
  1633. if (ret) {
  1634. while (i--)
  1635. do_set_irq_trigger(feature, start + i, -1);
  1636. break;
  1637. }
  1638. }
  1639. return ret;
  1640. }
  1641. EXPORT_SYMBOL_GPL(dfl_fpga_set_irq_triggers);
  1642. /**
  1643. * dfl_feature_ioctl_get_num_irqs - dfl feature _GET_IRQ_NUM ioctl interface.
  1644. * @pdev: the feature device which has the sub feature
  1645. * @feature: the dfl sub feature
  1646. * @arg: ioctl argument
  1647. *
  1648. * Return: 0 on success, negative error code otherwise.
  1649. */
  1650. long dfl_feature_ioctl_get_num_irqs(struct platform_device *pdev,
  1651. struct dfl_feature *feature,
  1652. unsigned long arg)
  1653. {
  1654. return put_user(feature->nr_irqs, (__u32 __user *)arg);
  1655. }
  1656. EXPORT_SYMBOL_GPL(dfl_feature_ioctl_get_num_irqs);
  1657. /**
  1658. * dfl_feature_ioctl_set_irq - dfl feature _SET_IRQ ioctl interface.
  1659. * @pdev: the feature device which has the sub feature
  1660. * @feature: the dfl sub feature
  1661. * @arg: ioctl argument
  1662. *
  1663. * Return: 0 on success, negative error code otherwise.
  1664. */
  1665. long dfl_feature_ioctl_set_irq(struct platform_device *pdev,
  1666. struct dfl_feature *feature,
  1667. unsigned long arg)
  1668. {
  1669. struct dfl_feature_platform_data *pdata = dev_get_platdata(&pdev->dev);
  1670. struct dfl_fpga_irq_set hdr;
  1671. s32 *fds;
  1672. long ret;
  1673. if (!feature->nr_irqs)
  1674. return -ENOENT;
  1675. if (copy_from_user(&hdr, (void __user *)arg, sizeof(hdr)))
  1676. return -EFAULT;
  1677. if (!hdr.count || (hdr.start + hdr.count > feature->nr_irqs) ||
  1678. (hdr.start + hdr.count < hdr.start))
  1679. return -EINVAL;
  1680. fds = memdup_array_user((void __user *)(arg + sizeof(hdr)),
  1681. hdr.count, sizeof(s32));
  1682. if (IS_ERR(fds))
  1683. return PTR_ERR(fds);
  1684. mutex_lock(&pdata->lock);
  1685. ret = dfl_fpga_set_irq_triggers(feature, hdr.start, hdr.count, fds);
  1686. mutex_unlock(&pdata->lock);
  1687. kfree(fds);
  1688. return ret;
  1689. }
  1690. EXPORT_SYMBOL_GPL(dfl_feature_ioctl_set_irq);
  1691. static void __exit dfl_fpga_exit(void)
  1692. {
  1693. dfl_chardev_uinit();
  1694. dfl_ids_destroy();
  1695. bus_unregister(&dfl_bus_type);
  1696. }
  1697. module_init(dfl_fpga_init);
  1698. module_exit(dfl_fpga_exit);
  1699. MODULE_DESCRIPTION("FPGA Device Feature List (DFL) Support");
  1700. MODULE_AUTHOR("Intel Corporation");
  1701. MODULE_LICENSE("GPL v2");