page_pool_user.c 11 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439
  1. // SPDX-License-Identifier: GPL-2.0
  2. #include <linux/mutex.h>
  3. #include <linux/netdevice.h>
  4. #include <linux/xarray.h>
  5. #include <net/net_debug.h>
  6. #include <net/netdev_rx_queue.h>
  7. #include <net/page_pool/helpers.h>
  8. #include <net/page_pool/types.h>
  9. #include <net/sock.h>
  10. #include "devmem.h"
  11. #include "page_pool_priv.h"
  12. #include "netdev-genl-gen.h"
  13. static DEFINE_XARRAY_FLAGS(page_pools, XA_FLAGS_ALLOC1);
  14. /* Protects: page_pools, netdevice->page_pools, pool->slow.netdev, pool->user.
  15. * Ordering: inside rtnl_lock
  16. */
  17. static DEFINE_MUTEX(page_pools_lock);
  18. /* Page pools are only reachable from user space (via netlink) if they are
  19. * linked to a netdev at creation time. Following page pool "visibility"
  20. * states are possible:
  21. * - normal
  22. * - user.list: linked to real netdev, netdev: real netdev
  23. * - orphaned - real netdev has disappeared
  24. * - user.list: linked to lo, netdev: lo
  25. * - invisible - either (a) created without netdev linking, (b) unlisted due
  26. * to error, or (c) the entire namespace which owned this pool disappeared
  27. * - user.list: unhashed, netdev: unknown
  28. */
  29. typedef int (*pp_nl_fill_cb)(struct sk_buff *rsp, const struct page_pool *pool,
  30. const struct genl_info *info);
  31. static int
  32. netdev_nl_page_pool_get_do(struct genl_info *info, u32 id, pp_nl_fill_cb fill)
  33. {
  34. struct page_pool *pool;
  35. struct sk_buff *rsp;
  36. int err;
  37. mutex_lock(&page_pools_lock);
  38. pool = xa_load(&page_pools, id);
  39. if (!pool || hlist_unhashed(&pool->user.list) ||
  40. !net_eq(dev_net(pool->slow.netdev), genl_info_net(info))) {
  41. err = -ENOENT;
  42. goto err_unlock;
  43. }
  44. rsp = genlmsg_new(GENLMSG_DEFAULT_SIZE, GFP_KERNEL);
  45. if (!rsp) {
  46. err = -ENOMEM;
  47. goto err_unlock;
  48. }
  49. err = fill(rsp, pool, info);
  50. if (err)
  51. goto err_free_msg;
  52. mutex_unlock(&page_pools_lock);
  53. return genlmsg_reply(rsp, info);
  54. err_free_msg:
  55. nlmsg_free(rsp);
  56. err_unlock:
  57. mutex_unlock(&page_pools_lock);
  58. return err;
  59. }
  60. struct page_pool_dump_cb {
  61. unsigned long ifindex;
  62. u32 pp_id;
  63. };
  64. static int
  65. netdev_nl_page_pool_get_dump(struct sk_buff *skb, struct netlink_callback *cb,
  66. pp_nl_fill_cb fill)
  67. {
  68. struct page_pool_dump_cb *state = (void *)cb->ctx;
  69. const struct genl_info *info = genl_info_dump(cb);
  70. struct net *net = sock_net(skb->sk);
  71. struct net_device *netdev;
  72. struct page_pool *pool;
  73. int err = 0;
  74. rtnl_lock();
  75. mutex_lock(&page_pools_lock);
  76. for_each_netdev_dump(net, netdev, state->ifindex) {
  77. hlist_for_each_entry(pool, &netdev->page_pools, user.list) {
  78. if (state->pp_id && state->pp_id < pool->user.id)
  79. continue;
  80. state->pp_id = pool->user.id;
  81. err = fill(skb, pool, info);
  82. if (err)
  83. goto out;
  84. }
  85. state->pp_id = 0;
  86. }
  87. out:
  88. mutex_unlock(&page_pools_lock);
  89. rtnl_unlock();
  90. return err;
  91. }
  92. static int
  93. page_pool_nl_stats_fill(struct sk_buff *rsp, const struct page_pool *pool,
  94. const struct genl_info *info)
  95. {
  96. #ifdef CONFIG_PAGE_POOL_STATS
  97. struct page_pool_stats stats = {};
  98. struct nlattr *nest;
  99. void *hdr;
  100. if (!page_pool_get_stats(pool, &stats))
  101. return 0;
  102. hdr = genlmsg_iput(rsp, info);
  103. if (!hdr)
  104. return -EMSGSIZE;
  105. nest = nla_nest_start(rsp, NETDEV_A_PAGE_POOL_STATS_INFO);
  106. if (nla_put_uint(rsp, NETDEV_A_PAGE_POOL_ID, pool->user.id) ||
  107. (pool->slow.netdev->ifindex != LOOPBACK_IFINDEX &&
  108. nla_put_u32(rsp, NETDEV_A_PAGE_POOL_IFINDEX,
  109. pool->slow.netdev->ifindex)))
  110. goto err_cancel_nest;
  111. nla_nest_end(rsp, nest);
  112. if (nla_put_uint(rsp, NETDEV_A_PAGE_POOL_STATS_ALLOC_FAST,
  113. stats.alloc_stats.fast) ||
  114. nla_put_uint(rsp, NETDEV_A_PAGE_POOL_STATS_ALLOC_SLOW,
  115. stats.alloc_stats.slow) ||
  116. nla_put_uint(rsp, NETDEV_A_PAGE_POOL_STATS_ALLOC_SLOW_HIGH_ORDER,
  117. stats.alloc_stats.slow_high_order) ||
  118. nla_put_uint(rsp, NETDEV_A_PAGE_POOL_STATS_ALLOC_EMPTY,
  119. stats.alloc_stats.empty) ||
  120. nla_put_uint(rsp, NETDEV_A_PAGE_POOL_STATS_ALLOC_REFILL,
  121. stats.alloc_stats.refill) ||
  122. nla_put_uint(rsp, NETDEV_A_PAGE_POOL_STATS_ALLOC_WAIVE,
  123. stats.alloc_stats.waive) ||
  124. nla_put_uint(rsp, NETDEV_A_PAGE_POOL_STATS_RECYCLE_CACHED,
  125. stats.recycle_stats.cached) ||
  126. nla_put_uint(rsp, NETDEV_A_PAGE_POOL_STATS_RECYCLE_CACHE_FULL,
  127. stats.recycle_stats.cache_full) ||
  128. nla_put_uint(rsp, NETDEV_A_PAGE_POOL_STATS_RECYCLE_RING,
  129. stats.recycle_stats.ring) ||
  130. nla_put_uint(rsp, NETDEV_A_PAGE_POOL_STATS_RECYCLE_RING_FULL,
  131. stats.recycle_stats.ring_full) ||
  132. nla_put_uint(rsp, NETDEV_A_PAGE_POOL_STATS_RECYCLE_RELEASED_REFCNT,
  133. stats.recycle_stats.released_refcnt))
  134. goto err_cancel_msg;
  135. genlmsg_end(rsp, hdr);
  136. return 0;
  137. err_cancel_nest:
  138. nla_nest_cancel(rsp, nest);
  139. err_cancel_msg:
  140. genlmsg_cancel(rsp, hdr);
  141. return -EMSGSIZE;
  142. #else
  143. GENL_SET_ERR_MSG(info, "kernel built without CONFIG_PAGE_POOL_STATS");
  144. return -EOPNOTSUPP;
  145. #endif
  146. }
  147. int netdev_nl_page_pool_stats_get_doit(struct sk_buff *skb,
  148. struct genl_info *info)
  149. {
  150. struct nlattr *tb[ARRAY_SIZE(netdev_page_pool_info_nl_policy)];
  151. struct nlattr *nest;
  152. int err;
  153. u32 id;
  154. if (GENL_REQ_ATTR_CHECK(info, NETDEV_A_PAGE_POOL_STATS_INFO))
  155. return -EINVAL;
  156. nest = info->attrs[NETDEV_A_PAGE_POOL_STATS_INFO];
  157. err = nla_parse_nested(tb, ARRAY_SIZE(tb) - 1, nest,
  158. netdev_page_pool_info_nl_policy,
  159. info->extack);
  160. if (err)
  161. return err;
  162. if (NL_REQ_ATTR_CHECK(info->extack, nest, tb, NETDEV_A_PAGE_POOL_ID))
  163. return -EINVAL;
  164. if (tb[NETDEV_A_PAGE_POOL_IFINDEX]) {
  165. NL_SET_ERR_MSG_ATTR(info->extack,
  166. tb[NETDEV_A_PAGE_POOL_IFINDEX],
  167. "selecting by ifindex not supported");
  168. return -EINVAL;
  169. }
  170. id = nla_get_uint(tb[NETDEV_A_PAGE_POOL_ID]);
  171. return netdev_nl_page_pool_get_do(info, id, page_pool_nl_stats_fill);
  172. }
  173. int netdev_nl_page_pool_stats_get_dumpit(struct sk_buff *skb,
  174. struct netlink_callback *cb)
  175. {
  176. return netdev_nl_page_pool_get_dump(skb, cb, page_pool_nl_stats_fill);
  177. }
  178. static int
  179. page_pool_nl_fill(struct sk_buff *rsp, const struct page_pool *pool,
  180. const struct genl_info *info)
  181. {
  182. struct net_devmem_dmabuf_binding *binding = pool->mp_priv;
  183. size_t inflight, refsz;
  184. void *hdr;
  185. hdr = genlmsg_iput(rsp, info);
  186. if (!hdr)
  187. return -EMSGSIZE;
  188. if (nla_put_uint(rsp, NETDEV_A_PAGE_POOL_ID, pool->user.id))
  189. goto err_cancel;
  190. if (pool->slow.netdev->ifindex != LOOPBACK_IFINDEX &&
  191. nla_put_u32(rsp, NETDEV_A_PAGE_POOL_IFINDEX,
  192. pool->slow.netdev->ifindex))
  193. goto err_cancel;
  194. if (pool->user.napi_id &&
  195. nla_put_uint(rsp, NETDEV_A_PAGE_POOL_NAPI_ID, pool->user.napi_id))
  196. goto err_cancel;
  197. inflight = page_pool_inflight(pool, false);
  198. refsz = PAGE_SIZE << pool->p.order;
  199. if (nla_put_uint(rsp, NETDEV_A_PAGE_POOL_INFLIGHT, inflight) ||
  200. nla_put_uint(rsp, NETDEV_A_PAGE_POOL_INFLIGHT_MEM,
  201. inflight * refsz))
  202. goto err_cancel;
  203. if (pool->user.detach_time &&
  204. nla_put_uint(rsp, NETDEV_A_PAGE_POOL_DETACH_TIME,
  205. pool->user.detach_time))
  206. goto err_cancel;
  207. if (binding && nla_put_u32(rsp, NETDEV_A_PAGE_POOL_DMABUF, binding->id))
  208. goto err_cancel;
  209. genlmsg_end(rsp, hdr);
  210. return 0;
  211. err_cancel:
  212. genlmsg_cancel(rsp, hdr);
  213. return -EMSGSIZE;
  214. }
  215. static void netdev_nl_page_pool_event(const struct page_pool *pool, u32 cmd)
  216. {
  217. struct genl_info info;
  218. struct sk_buff *ntf;
  219. struct net *net;
  220. lockdep_assert_held(&page_pools_lock);
  221. /* 'invisible' page pools don't matter */
  222. if (hlist_unhashed(&pool->user.list))
  223. return;
  224. net = dev_net(pool->slow.netdev);
  225. if (!genl_has_listeners(&netdev_nl_family, net, NETDEV_NLGRP_PAGE_POOL))
  226. return;
  227. genl_info_init_ntf(&info, &netdev_nl_family, cmd);
  228. ntf = genlmsg_new(GENLMSG_DEFAULT_SIZE, GFP_KERNEL);
  229. if (!ntf)
  230. return;
  231. if (page_pool_nl_fill(ntf, pool, &info)) {
  232. nlmsg_free(ntf);
  233. return;
  234. }
  235. genlmsg_multicast_netns(&netdev_nl_family, net, ntf,
  236. 0, NETDEV_NLGRP_PAGE_POOL, GFP_KERNEL);
  237. }
  238. int netdev_nl_page_pool_get_doit(struct sk_buff *skb, struct genl_info *info)
  239. {
  240. u32 id;
  241. if (GENL_REQ_ATTR_CHECK(info, NETDEV_A_PAGE_POOL_ID))
  242. return -EINVAL;
  243. id = nla_get_uint(info->attrs[NETDEV_A_PAGE_POOL_ID]);
  244. return netdev_nl_page_pool_get_do(info, id, page_pool_nl_fill);
  245. }
  246. int netdev_nl_page_pool_get_dumpit(struct sk_buff *skb,
  247. struct netlink_callback *cb)
  248. {
  249. return netdev_nl_page_pool_get_dump(skb, cb, page_pool_nl_fill);
  250. }
  251. int page_pool_list(struct page_pool *pool)
  252. {
  253. static u32 id_alloc_next;
  254. int err;
  255. mutex_lock(&page_pools_lock);
  256. err = xa_alloc_cyclic(&page_pools, &pool->user.id, pool, xa_limit_32b,
  257. &id_alloc_next, GFP_KERNEL);
  258. if (err < 0)
  259. goto err_unlock;
  260. INIT_HLIST_NODE(&pool->user.list);
  261. if (pool->slow.netdev) {
  262. hlist_add_head(&pool->user.list,
  263. &pool->slow.netdev->page_pools);
  264. pool->user.napi_id = pool->p.napi ? pool->p.napi->napi_id : 0;
  265. netdev_nl_page_pool_event(pool, NETDEV_CMD_PAGE_POOL_ADD_NTF);
  266. }
  267. mutex_unlock(&page_pools_lock);
  268. return 0;
  269. err_unlock:
  270. mutex_unlock(&page_pools_lock);
  271. return err;
  272. }
  273. void page_pool_detached(struct page_pool *pool)
  274. {
  275. mutex_lock(&page_pools_lock);
  276. pool->user.detach_time = ktime_get_boottime_seconds();
  277. netdev_nl_page_pool_event(pool, NETDEV_CMD_PAGE_POOL_CHANGE_NTF);
  278. mutex_unlock(&page_pools_lock);
  279. }
  280. void page_pool_unlist(struct page_pool *pool)
  281. {
  282. mutex_lock(&page_pools_lock);
  283. netdev_nl_page_pool_event(pool, NETDEV_CMD_PAGE_POOL_DEL_NTF);
  284. xa_erase(&page_pools, pool->user.id);
  285. if (!hlist_unhashed(&pool->user.list))
  286. hlist_del(&pool->user.list);
  287. mutex_unlock(&page_pools_lock);
  288. }
  289. int page_pool_check_memory_provider(struct net_device *dev,
  290. struct netdev_rx_queue *rxq)
  291. {
  292. void *binding = rxq->mp_params.mp_priv;
  293. struct page_pool *pool;
  294. struct hlist_node *n;
  295. if (!binding)
  296. return 0;
  297. mutex_lock(&page_pools_lock);
  298. hlist_for_each_entry_safe(pool, n, &dev->page_pools, user.list) {
  299. if (pool->mp_priv != binding)
  300. continue;
  301. if (pool->slow.queue_idx == get_netdev_rx_queue_index(rxq)) {
  302. mutex_unlock(&page_pools_lock);
  303. return 0;
  304. }
  305. }
  306. mutex_unlock(&page_pools_lock);
  307. return -ENODATA;
  308. }
  309. static void page_pool_unreg_netdev_wipe(struct net_device *netdev)
  310. {
  311. struct page_pool *pool;
  312. struct hlist_node *n;
  313. mutex_lock(&page_pools_lock);
  314. hlist_for_each_entry_safe(pool, n, &netdev->page_pools, user.list) {
  315. hlist_del_init(&pool->user.list);
  316. pool->slow.netdev = NET_PTR_POISON;
  317. }
  318. mutex_unlock(&page_pools_lock);
  319. }
  320. static void page_pool_unreg_netdev(struct net_device *netdev)
  321. {
  322. struct page_pool *pool, *last;
  323. struct net_device *lo;
  324. lo = dev_net(netdev)->loopback_dev;
  325. mutex_lock(&page_pools_lock);
  326. last = NULL;
  327. hlist_for_each_entry(pool, &netdev->page_pools, user.list) {
  328. pool->slow.netdev = lo;
  329. netdev_nl_page_pool_event(pool,
  330. NETDEV_CMD_PAGE_POOL_CHANGE_NTF);
  331. last = pool;
  332. }
  333. if (last)
  334. hlist_splice_init(&netdev->page_pools, &last->user.list,
  335. &lo->page_pools);
  336. mutex_unlock(&page_pools_lock);
  337. }
  338. static int
  339. page_pool_netdevice_event(struct notifier_block *nb,
  340. unsigned long event, void *ptr)
  341. {
  342. struct net_device *netdev = netdev_notifier_info_to_dev(ptr);
  343. if (event != NETDEV_UNREGISTER)
  344. return NOTIFY_DONE;
  345. if (hlist_empty(&netdev->page_pools))
  346. return NOTIFY_OK;
  347. if (netdev->ifindex != LOOPBACK_IFINDEX)
  348. page_pool_unreg_netdev(netdev);
  349. else
  350. page_pool_unreg_netdev_wipe(netdev);
  351. return NOTIFY_OK;
  352. }
  353. static struct notifier_block page_pool_netdevice_nb = {
  354. .notifier_call = page_pool_netdevice_event,
  355. };
  356. static int __init page_pool_user_init(void)
  357. {
  358. return register_netdevice_notifier(&page_pool_netdevice_nb);
  359. }
  360. subsys_initcall(page_pool_user_init);