core.c 14 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551
  1. // SPDX-License-Identifier: GPL-2.0-or-later
  2. /*
  3. * Copyright (c) 2016 Mellanox Technologies. All rights reserved.
  4. * Copyright (c) 2016 Jiri Pirko <jiri@mellanox.com>
  5. */
  6. #include <net/genetlink.h>
  7. #define CREATE_TRACE_POINTS
  8. #include <trace/events/devlink.h>
  9. #include "devl_internal.h"
  10. EXPORT_TRACEPOINT_SYMBOL_GPL(devlink_hwmsg);
  11. EXPORT_TRACEPOINT_SYMBOL_GPL(devlink_hwerr);
  12. EXPORT_TRACEPOINT_SYMBOL_GPL(devlink_trap_report);
  13. DEFINE_XARRAY_FLAGS(devlinks, XA_FLAGS_ALLOC);
  14. static struct devlink *devlinks_xa_get(unsigned long index)
  15. {
  16. struct devlink *devlink;
  17. rcu_read_lock();
  18. devlink = xa_find(&devlinks, &index, index, DEVLINK_REGISTERED);
  19. if (!devlink || !devlink_try_get(devlink))
  20. devlink = NULL;
  21. rcu_read_unlock();
  22. return devlink;
  23. }
  24. /* devlink_rels xarray contains 1:1 relationships between
  25. * devlink object and related nested devlink instance.
  26. * The xarray index is used to get the nested object from
  27. * the nested-in object code.
  28. */
  29. static DEFINE_XARRAY_FLAGS(devlink_rels, XA_FLAGS_ALLOC1);
  30. #define DEVLINK_REL_IN_USE XA_MARK_0
  31. struct devlink_rel {
  32. u32 index;
  33. refcount_t refcount;
  34. u32 devlink_index;
  35. struct {
  36. u32 devlink_index;
  37. u32 obj_index;
  38. devlink_rel_notify_cb_t *notify_cb;
  39. devlink_rel_cleanup_cb_t *cleanup_cb;
  40. struct delayed_work notify_work;
  41. } nested_in;
  42. };
  43. static void devlink_rel_free(struct devlink_rel *rel)
  44. {
  45. xa_erase(&devlink_rels, rel->index);
  46. kfree(rel);
  47. }
  48. static void __devlink_rel_get(struct devlink_rel *rel)
  49. {
  50. refcount_inc(&rel->refcount);
  51. }
  52. static void __devlink_rel_put(struct devlink_rel *rel)
  53. {
  54. if (refcount_dec_and_test(&rel->refcount))
  55. devlink_rel_free(rel);
  56. }
  57. static void devlink_rel_nested_in_notify_work(struct work_struct *work)
  58. {
  59. struct devlink_rel *rel = container_of(work, struct devlink_rel,
  60. nested_in.notify_work.work);
  61. struct devlink *devlink;
  62. devlink = devlinks_xa_get(rel->nested_in.devlink_index);
  63. if (!devlink)
  64. goto rel_put;
  65. if (!devl_trylock(devlink)) {
  66. devlink_put(devlink);
  67. goto reschedule_work;
  68. }
  69. if (!devl_is_registered(devlink)) {
  70. devl_unlock(devlink);
  71. devlink_put(devlink);
  72. goto rel_put;
  73. }
  74. if (!xa_get_mark(&devlink_rels, rel->index, DEVLINK_REL_IN_USE))
  75. rel->nested_in.cleanup_cb(devlink, rel->nested_in.obj_index, rel->index);
  76. rel->nested_in.notify_cb(devlink, rel->nested_in.obj_index);
  77. devl_unlock(devlink);
  78. devlink_put(devlink);
  79. rel_put:
  80. __devlink_rel_put(rel);
  81. return;
  82. reschedule_work:
  83. schedule_delayed_work(&rel->nested_in.notify_work, 1);
  84. }
  85. static void devlink_rel_nested_in_notify_work_schedule(struct devlink_rel *rel)
  86. {
  87. __devlink_rel_get(rel);
  88. schedule_delayed_work(&rel->nested_in.notify_work, 0);
  89. }
  90. static struct devlink_rel *devlink_rel_alloc(void)
  91. {
  92. struct devlink_rel *rel;
  93. static u32 next;
  94. int err;
  95. rel = kzalloc(sizeof(*rel), GFP_KERNEL);
  96. if (!rel)
  97. return ERR_PTR(-ENOMEM);
  98. err = xa_alloc_cyclic(&devlink_rels, &rel->index, rel,
  99. xa_limit_32b, &next, GFP_KERNEL);
  100. if (err < 0) {
  101. kfree(rel);
  102. return ERR_PTR(err);
  103. }
  104. refcount_set(&rel->refcount, 1);
  105. INIT_DELAYED_WORK(&rel->nested_in.notify_work,
  106. &devlink_rel_nested_in_notify_work);
  107. return rel;
  108. }
  109. static void devlink_rel_put(struct devlink *devlink)
  110. {
  111. struct devlink_rel *rel = devlink->rel;
  112. if (!rel)
  113. return;
  114. xa_clear_mark(&devlink_rels, rel->index, DEVLINK_REL_IN_USE);
  115. devlink_rel_nested_in_notify_work_schedule(rel);
  116. __devlink_rel_put(rel);
  117. devlink->rel = NULL;
  118. }
  119. void devlink_rel_nested_in_clear(u32 rel_index)
  120. {
  121. xa_clear_mark(&devlink_rels, rel_index, DEVLINK_REL_IN_USE);
  122. }
  123. int devlink_rel_nested_in_add(u32 *rel_index, u32 devlink_index,
  124. u32 obj_index, devlink_rel_notify_cb_t *notify_cb,
  125. devlink_rel_cleanup_cb_t *cleanup_cb,
  126. struct devlink *devlink)
  127. {
  128. struct devlink_rel *rel = devlink_rel_alloc();
  129. ASSERT_DEVLINK_NOT_REGISTERED(devlink);
  130. if (IS_ERR(rel))
  131. return PTR_ERR(rel);
  132. rel->devlink_index = devlink->index;
  133. rel->nested_in.devlink_index = devlink_index;
  134. rel->nested_in.obj_index = obj_index;
  135. rel->nested_in.notify_cb = notify_cb;
  136. rel->nested_in.cleanup_cb = cleanup_cb;
  137. *rel_index = rel->index;
  138. xa_set_mark(&devlink_rels, rel->index, DEVLINK_REL_IN_USE);
  139. devlink->rel = rel;
  140. return 0;
  141. }
  142. /**
  143. * devlink_rel_nested_in_notify - Notify the object this devlink
  144. * instance is nested in.
  145. * @devlink: devlink
  146. *
  147. * This is called upon network namespace change of devlink instance.
  148. * In case this devlink instance is nested in another devlink object,
  149. * a notification of a change of this object should be sent
  150. * over netlink. The parent devlink instance lock needs to be
  151. * taken during the notification preparation.
  152. * However, since the devlink lock of nested instance is held here,
  153. * we would end with wrong devlink instance lock ordering and
  154. * deadlock. Therefore the work is utilized to avoid that.
  155. */
  156. void devlink_rel_nested_in_notify(struct devlink *devlink)
  157. {
  158. struct devlink_rel *rel = devlink->rel;
  159. if (!rel)
  160. return;
  161. devlink_rel_nested_in_notify_work_schedule(rel);
  162. }
  163. static struct devlink_rel *devlink_rel_find(unsigned long rel_index)
  164. {
  165. return xa_find(&devlink_rels, &rel_index, rel_index,
  166. DEVLINK_REL_IN_USE);
  167. }
  168. static struct devlink *devlink_rel_devlink_get(u32 rel_index)
  169. {
  170. struct devlink_rel *rel;
  171. u32 devlink_index;
  172. if (!rel_index)
  173. return NULL;
  174. xa_lock(&devlink_rels);
  175. rel = devlink_rel_find(rel_index);
  176. if (rel)
  177. devlink_index = rel->devlink_index;
  178. xa_unlock(&devlink_rels);
  179. if (!rel)
  180. return NULL;
  181. return devlinks_xa_get(devlink_index);
  182. }
  183. int devlink_rel_devlink_handle_put(struct sk_buff *msg, struct devlink *devlink,
  184. u32 rel_index, int attrtype,
  185. bool *msg_updated)
  186. {
  187. struct net *net = devlink_net(devlink);
  188. struct devlink *rel_devlink;
  189. int err;
  190. rel_devlink = devlink_rel_devlink_get(rel_index);
  191. if (!rel_devlink)
  192. return 0;
  193. err = devlink_nl_put_nested_handle(msg, net, rel_devlink, attrtype);
  194. devlink_put(rel_devlink);
  195. if (!err && msg_updated)
  196. *msg_updated = true;
  197. return err;
  198. }
  199. void *devlink_priv(struct devlink *devlink)
  200. {
  201. return &devlink->priv;
  202. }
  203. EXPORT_SYMBOL_GPL(devlink_priv);
  204. struct devlink *priv_to_devlink(void *priv)
  205. {
  206. return container_of(priv, struct devlink, priv);
  207. }
  208. EXPORT_SYMBOL_GPL(priv_to_devlink);
  209. struct device *devlink_to_dev(const struct devlink *devlink)
  210. {
  211. return devlink->dev;
  212. }
  213. EXPORT_SYMBOL_GPL(devlink_to_dev);
  214. struct net *devlink_net(const struct devlink *devlink)
  215. {
  216. return read_pnet(&devlink->_net);
  217. }
  218. EXPORT_SYMBOL_GPL(devlink_net);
  219. void devl_assert_locked(struct devlink *devlink)
  220. {
  221. lockdep_assert_held(&devlink->lock);
  222. }
  223. EXPORT_SYMBOL_GPL(devl_assert_locked);
  224. #ifdef CONFIG_LOCKDEP
  225. /* For use in conjunction with LOCKDEP only e.g. rcu_dereference_protected() */
  226. bool devl_lock_is_held(struct devlink *devlink)
  227. {
  228. return lockdep_is_held(&devlink->lock);
  229. }
  230. EXPORT_SYMBOL_GPL(devl_lock_is_held);
  231. #endif
  232. void devl_lock(struct devlink *devlink)
  233. {
  234. mutex_lock(&devlink->lock);
  235. }
  236. EXPORT_SYMBOL_GPL(devl_lock);
  237. int devl_trylock(struct devlink *devlink)
  238. {
  239. return mutex_trylock(&devlink->lock);
  240. }
  241. EXPORT_SYMBOL_GPL(devl_trylock);
  242. void devl_unlock(struct devlink *devlink)
  243. {
  244. mutex_unlock(&devlink->lock);
  245. }
  246. EXPORT_SYMBOL_GPL(devl_unlock);
  247. /**
  248. * devlink_try_get() - try to obtain a reference on a devlink instance
  249. * @devlink: instance to reference
  250. *
  251. * Obtain a reference on a devlink instance. A reference on a devlink instance
  252. * only implies that it's safe to take the instance lock. It does not imply
  253. * that the instance is registered, use devl_is_registered() after taking
  254. * the instance lock to check registration status.
  255. */
  256. struct devlink *__must_check devlink_try_get(struct devlink *devlink)
  257. {
  258. if (refcount_inc_not_zero(&devlink->refcount))
  259. return devlink;
  260. return NULL;
  261. }
  262. static void devlink_release(struct work_struct *work)
  263. {
  264. struct devlink *devlink;
  265. devlink = container_of(to_rcu_work(work), struct devlink, rwork);
  266. mutex_destroy(&devlink->lock);
  267. lockdep_unregister_key(&devlink->lock_key);
  268. put_device(devlink->dev);
  269. kvfree(devlink);
  270. }
  271. void devlink_put(struct devlink *devlink)
  272. {
  273. if (refcount_dec_and_test(&devlink->refcount))
  274. queue_rcu_work(system_wq, &devlink->rwork);
  275. }
  276. struct devlink *devlinks_xa_find_get(struct net *net, unsigned long *indexp)
  277. {
  278. struct devlink *devlink = NULL;
  279. rcu_read_lock();
  280. retry:
  281. devlink = xa_find(&devlinks, indexp, ULONG_MAX, DEVLINK_REGISTERED);
  282. if (!devlink)
  283. goto unlock;
  284. if (!devlink_try_get(devlink))
  285. goto next;
  286. if (!net_eq(devlink_net(devlink), net)) {
  287. devlink_put(devlink);
  288. goto next;
  289. }
  290. unlock:
  291. rcu_read_unlock();
  292. return devlink;
  293. next:
  294. (*indexp)++;
  295. goto retry;
  296. }
  297. /**
  298. * devl_register - Register devlink instance
  299. * @devlink: devlink
  300. */
  301. int devl_register(struct devlink *devlink)
  302. {
  303. ASSERT_DEVLINK_NOT_REGISTERED(devlink);
  304. devl_assert_locked(devlink);
  305. xa_set_mark(&devlinks, devlink->index, DEVLINK_REGISTERED);
  306. devlink_notify_register(devlink);
  307. devlink_rel_nested_in_notify(devlink);
  308. return 0;
  309. }
  310. EXPORT_SYMBOL_GPL(devl_register);
  311. void devlink_register(struct devlink *devlink)
  312. {
  313. devl_lock(devlink);
  314. devl_register(devlink);
  315. devl_unlock(devlink);
  316. }
  317. EXPORT_SYMBOL_GPL(devlink_register);
  318. /**
  319. * devl_unregister - Unregister devlink instance
  320. * @devlink: devlink
  321. */
  322. void devl_unregister(struct devlink *devlink)
  323. {
  324. ASSERT_DEVLINK_REGISTERED(devlink);
  325. devl_assert_locked(devlink);
  326. devlink_notify_unregister(devlink);
  327. xa_clear_mark(&devlinks, devlink->index, DEVLINK_REGISTERED);
  328. devlink_rel_put(devlink);
  329. }
  330. EXPORT_SYMBOL_GPL(devl_unregister);
  331. void devlink_unregister(struct devlink *devlink)
  332. {
  333. devl_lock(devlink);
  334. devl_unregister(devlink);
  335. devl_unlock(devlink);
  336. }
  337. EXPORT_SYMBOL_GPL(devlink_unregister);
  338. /**
  339. * devlink_alloc_ns - Allocate new devlink instance resources
  340. * in specific namespace
  341. *
  342. * @ops: ops
  343. * @priv_size: size of user private data
  344. * @net: net namespace
  345. * @dev: parent device
  346. *
  347. * Allocate new devlink instance resources, including devlink index
  348. * and name.
  349. */
  350. struct devlink *devlink_alloc_ns(const struct devlink_ops *ops,
  351. size_t priv_size, struct net *net,
  352. struct device *dev)
  353. {
  354. struct devlink *devlink;
  355. static u32 last_id;
  356. int ret;
  357. WARN_ON(!ops || !dev);
  358. if (!devlink_reload_actions_valid(ops))
  359. return NULL;
  360. devlink = kvzalloc(struct_size(devlink, priv, priv_size), GFP_KERNEL);
  361. if (!devlink)
  362. return NULL;
  363. ret = xa_alloc_cyclic(&devlinks, &devlink->index, devlink, xa_limit_31b,
  364. &last_id, GFP_KERNEL);
  365. if (ret < 0)
  366. goto err_xa_alloc;
  367. devlink->dev = get_device(dev);
  368. devlink->ops = ops;
  369. xa_init_flags(&devlink->ports, XA_FLAGS_ALLOC);
  370. xa_init_flags(&devlink->params, XA_FLAGS_ALLOC);
  371. xa_init_flags(&devlink->snapshot_ids, XA_FLAGS_ALLOC);
  372. xa_init_flags(&devlink->nested_rels, XA_FLAGS_ALLOC);
  373. write_pnet(&devlink->_net, net);
  374. INIT_LIST_HEAD(&devlink->rate_list);
  375. INIT_LIST_HEAD(&devlink->linecard_list);
  376. INIT_LIST_HEAD(&devlink->sb_list);
  377. INIT_LIST_HEAD_RCU(&devlink->dpipe_table_list);
  378. INIT_LIST_HEAD(&devlink->resource_list);
  379. INIT_LIST_HEAD(&devlink->region_list);
  380. INIT_LIST_HEAD(&devlink->reporter_list);
  381. INIT_LIST_HEAD(&devlink->trap_list);
  382. INIT_LIST_HEAD(&devlink->trap_group_list);
  383. INIT_LIST_HEAD(&devlink->trap_policer_list);
  384. INIT_RCU_WORK(&devlink->rwork, devlink_release);
  385. lockdep_register_key(&devlink->lock_key);
  386. mutex_init(&devlink->lock);
  387. lockdep_set_class(&devlink->lock, &devlink->lock_key);
  388. refcount_set(&devlink->refcount, 1);
  389. return devlink;
  390. err_xa_alloc:
  391. kvfree(devlink);
  392. return NULL;
  393. }
  394. EXPORT_SYMBOL_GPL(devlink_alloc_ns);
  395. /**
  396. * devlink_free - Free devlink instance resources
  397. *
  398. * @devlink: devlink
  399. */
  400. void devlink_free(struct devlink *devlink)
  401. {
  402. ASSERT_DEVLINK_NOT_REGISTERED(devlink);
  403. WARN_ON(!list_empty(&devlink->trap_policer_list));
  404. WARN_ON(!list_empty(&devlink->trap_group_list));
  405. WARN_ON(!list_empty(&devlink->trap_list));
  406. WARN_ON(!list_empty(&devlink->reporter_list));
  407. WARN_ON(!list_empty(&devlink->region_list));
  408. WARN_ON(!list_empty(&devlink->resource_list));
  409. WARN_ON(!list_empty(&devlink->dpipe_table_list));
  410. WARN_ON(!list_empty(&devlink->sb_list));
  411. WARN_ON(!list_empty(&devlink->rate_list));
  412. WARN_ON(!list_empty(&devlink->linecard_list));
  413. WARN_ON(!xa_empty(&devlink->ports));
  414. xa_destroy(&devlink->nested_rels);
  415. xa_destroy(&devlink->snapshot_ids);
  416. xa_destroy(&devlink->params);
  417. xa_destroy(&devlink->ports);
  418. xa_erase(&devlinks, devlink->index);
  419. devlink_put(devlink);
  420. }
  421. EXPORT_SYMBOL_GPL(devlink_free);
  422. static void __net_exit devlink_pernet_pre_exit(struct net *net)
  423. {
  424. struct devlink *devlink;
  425. u32 actions_performed;
  426. unsigned long index;
  427. int err;
  428. /* In case network namespace is getting destroyed, reload
  429. * all devlink instances from this namespace into init_net.
  430. */
  431. devlinks_xa_for_each_registered_get(net, index, devlink) {
  432. devl_dev_lock(devlink, true);
  433. err = 0;
  434. if (devl_is_registered(devlink))
  435. err = devlink_reload(devlink, &init_net,
  436. DEVLINK_RELOAD_ACTION_DRIVER_REINIT,
  437. DEVLINK_RELOAD_LIMIT_UNSPEC,
  438. &actions_performed, NULL);
  439. devl_dev_unlock(devlink, true);
  440. devlink_put(devlink);
  441. if (err && err != -EOPNOTSUPP)
  442. pr_warn("Failed to reload devlink instance into init_net\n");
  443. }
  444. }
  445. static struct pernet_operations devlink_pernet_ops __net_initdata = {
  446. .pre_exit = devlink_pernet_pre_exit,
  447. };
  448. static struct notifier_block devlink_port_netdevice_nb = {
  449. .notifier_call = devlink_port_netdevice_event,
  450. };
  451. static int __init devlink_init(void)
  452. {
  453. int err;
  454. err = register_pernet_subsys(&devlink_pernet_ops);
  455. if (err)
  456. goto out;
  457. err = genl_register_family(&devlink_nl_family);
  458. if (err)
  459. goto out_unreg_pernet_subsys;
  460. err = register_netdevice_notifier(&devlink_port_netdevice_nb);
  461. if (!err)
  462. return 0;
  463. genl_unregister_family(&devlink_nl_family);
  464. out_unreg_pernet_subsys:
  465. unregister_pernet_subsys(&devlink_pernet_ops);
  466. out:
  467. WARN_ON(err);
  468. return err;
  469. }
  470. subsys_initcall(devlink_init);