local_object.c 12 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491
  1. // SPDX-License-Identifier: GPL-2.0-or-later
  2. /* Local endpoint object management
  3. *
  4. * Copyright (C) 2016 Red Hat, Inc. All Rights Reserved.
  5. * Written by David Howells (dhowells@redhat.com)
  6. */
  7. #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
  8. #include <linux/module.h>
  9. #include <linux/net.h>
  10. #include <linux/skbuff.h>
  11. #include <linux/slab.h>
  12. #include <linux/udp.h>
  13. #include <linux/ip.h>
  14. #include <linux/hashtable.h>
  15. #include <net/sock.h>
  16. #include <net/udp.h>
  17. #include <net/udp_tunnel.h>
  18. #include <net/af_rxrpc.h>
  19. #include "ar-internal.h"
  20. static void rxrpc_local_rcu(struct rcu_head *);
  21. /*
  22. * Handle an ICMP/ICMP6 error turning up at the tunnel. Push it through the
  23. * usual mechanism so that it gets parsed and presented through the UDP
  24. * socket's error_report().
  25. */
  26. static void rxrpc_encap_err_rcv(struct sock *sk, struct sk_buff *skb, int err,
  27. __be16 port, u32 info, u8 *payload)
  28. {
  29. if (ip_hdr(skb)->version == IPVERSION)
  30. return ip_icmp_error(sk, skb, err, port, info, payload);
  31. if (IS_ENABLED(CONFIG_AF_RXRPC_IPV6))
  32. return ipv6_icmp_error(sk, skb, err, port, info, payload);
  33. }
  34. /*
  35. * Set or clear the Don't Fragment flag on a socket.
  36. */
  37. void rxrpc_local_dont_fragment(const struct rxrpc_local *local, bool set)
  38. {
  39. if (set)
  40. ip_sock_set_mtu_discover(local->socket->sk, IP_PMTUDISC_DO);
  41. else
  42. ip_sock_set_mtu_discover(local->socket->sk, IP_PMTUDISC_DONT);
  43. }
  44. /*
  45. * Compare a local to an address. Return -ve, 0 or +ve to indicate less than,
  46. * same or greater than.
  47. *
  48. * We explicitly don't compare the RxRPC service ID as we want to reject
  49. * conflicting uses by differing services. Further, we don't want to share
  50. * addresses with different options (IPv6), so we don't compare those bits
  51. * either.
  52. */
  53. static long rxrpc_local_cmp_key(const struct rxrpc_local *local,
  54. const struct sockaddr_rxrpc *srx)
  55. {
  56. long diff;
  57. diff = ((local->srx.transport_type - srx->transport_type) ?:
  58. (local->srx.transport_len - srx->transport_len) ?:
  59. (local->srx.transport.family - srx->transport.family));
  60. if (diff != 0)
  61. return diff;
  62. switch (srx->transport.family) {
  63. case AF_INET:
  64. /* If the choice of UDP port is left up to the transport, then
  65. * the endpoint record doesn't match.
  66. */
  67. return ((u16 __force)local->srx.transport.sin.sin_port -
  68. (u16 __force)srx->transport.sin.sin_port) ?:
  69. memcmp(&local->srx.transport.sin.sin_addr,
  70. &srx->transport.sin.sin_addr,
  71. sizeof(struct in_addr));
  72. #ifdef CONFIG_AF_RXRPC_IPV6
  73. case AF_INET6:
  74. /* If the choice of UDP6 port is left up to the transport, then
  75. * the endpoint record doesn't match.
  76. */
  77. return ((u16 __force)local->srx.transport.sin6.sin6_port -
  78. (u16 __force)srx->transport.sin6.sin6_port) ?:
  79. memcmp(&local->srx.transport.sin6.sin6_addr,
  80. &srx->transport.sin6.sin6_addr,
  81. sizeof(struct in6_addr));
  82. #endif
  83. default:
  84. BUG();
  85. }
  86. }
  87. static void rxrpc_client_conn_reap_timeout(struct timer_list *timer)
  88. {
  89. struct rxrpc_local *local =
  90. container_of(timer, struct rxrpc_local, client_conn_reap_timer);
  91. if (!local->kill_all_client_conns &&
  92. test_and_set_bit(RXRPC_CLIENT_CONN_REAP_TIMER, &local->client_conn_flags))
  93. rxrpc_wake_up_io_thread(local);
  94. }
  95. /*
  96. * Allocate a new local endpoint.
  97. */
  98. static struct rxrpc_local *rxrpc_alloc_local(struct net *net,
  99. const struct sockaddr_rxrpc *srx)
  100. {
  101. struct rxrpc_local *local;
  102. u32 tmp;
  103. local = kzalloc(sizeof(struct rxrpc_local), GFP_KERNEL);
  104. if (local) {
  105. refcount_set(&local->ref, 1);
  106. atomic_set(&local->active_users, 1);
  107. local->net = net;
  108. local->rxnet = rxrpc_net(net);
  109. INIT_HLIST_NODE(&local->link);
  110. init_completion(&local->io_thread_ready);
  111. #ifdef CONFIG_AF_RXRPC_INJECT_RX_DELAY
  112. skb_queue_head_init(&local->rx_delay_queue);
  113. #endif
  114. skb_queue_head_init(&local->rx_queue);
  115. INIT_LIST_HEAD(&local->conn_attend_q);
  116. INIT_LIST_HEAD(&local->call_attend_q);
  117. local->client_bundles = RB_ROOT;
  118. spin_lock_init(&local->client_bundles_lock);
  119. local->kill_all_client_conns = false;
  120. INIT_LIST_HEAD(&local->idle_client_conns);
  121. timer_setup(&local->client_conn_reap_timer,
  122. rxrpc_client_conn_reap_timeout, 0);
  123. spin_lock_init(&local->lock);
  124. rwlock_init(&local->services_lock);
  125. local->debug_id = atomic_inc_return(&rxrpc_debug_id);
  126. memcpy(&local->srx, srx, sizeof(*srx));
  127. local->srx.srx_service = 0;
  128. idr_init(&local->conn_ids);
  129. get_random_bytes(&tmp, sizeof(tmp));
  130. tmp &= 0x3fffffff;
  131. if (tmp == 0)
  132. tmp = 1;
  133. idr_set_cursor(&local->conn_ids, tmp);
  134. INIT_LIST_HEAD(&local->new_client_calls);
  135. spin_lock_init(&local->client_call_lock);
  136. trace_rxrpc_local(local->debug_id, rxrpc_local_new, 1, 1);
  137. }
  138. _leave(" = %p", local);
  139. return local;
  140. }
  141. /*
  142. * create the local socket
  143. * - must be called with rxrpc_local_mutex locked
  144. */
  145. static int rxrpc_open_socket(struct rxrpc_local *local, struct net *net)
  146. {
  147. struct udp_tunnel_sock_cfg tuncfg = {NULL};
  148. struct sockaddr_rxrpc *srx = &local->srx;
  149. struct udp_port_cfg udp_conf = {0};
  150. struct task_struct *io_thread;
  151. struct sock *usk;
  152. int ret;
  153. _enter("%p{%d,%d}",
  154. local, srx->transport_type, srx->transport.family);
  155. udp_conf.family = srx->transport.family;
  156. udp_conf.use_udp_checksums = true;
  157. if (udp_conf.family == AF_INET) {
  158. udp_conf.local_ip = srx->transport.sin.sin_addr;
  159. udp_conf.local_udp_port = srx->transport.sin.sin_port;
  160. #if IS_ENABLED(CONFIG_AF_RXRPC_IPV6)
  161. } else {
  162. udp_conf.local_ip6 = srx->transport.sin6.sin6_addr;
  163. udp_conf.local_udp_port = srx->transport.sin6.sin6_port;
  164. udp_conf.use_udp6_tx_checksums = true;
  165. udp_conf.use_udp6_rx_checksums = true;
  166. #endif
  167. }
  168. ret = udp_sock_create(net, &udp_conf, &local->socket);
  169. if (ret < 0) {
  170. _leave(" = %d [socket]", ret);
  171. return ret;
  172. }
  173. tuncfg.encap_type = UDP_ENCAP_RXRPC;
  174. tuncfg.encap_rcv = rxrpc_encap_rcv;
  175. tuncfg.encap_err_rcv = rxrpc_encap_err_rcv;
  176. tuncfg.sk_user_data = local;
  177. setup_udp_tunnel_sock(net, local->socket, &tuncfg);
  178. /* set the socket up */
  179. usk = local->socket->sk;
  180. usk->sk_error_report = rxrpc_error_report;
  181. switch (srx->transport.family) {
  182. case AF_INET6:
  183. /* we want to receive ICMPv6 errors */
  184. ip6_sock_set_recverr(usk);
  185. /* Fall through and set IPv4 options too otherwise we don't get
  186. * errors from IPv4 packets sent through the IPv6 socket.
  187. */
  188. fallthrough;
  189. case AF_INET:
  190. /* we want to receive ICMP errors */
  191. ip_sock_set_recverr(usk);
  192. /* we want to set the don't fragment bit */
  193. rxrpc_local_dont_fragment(local, true);
  194. /* We want receive timestamps. */
  195. sock_enable_timestamps(usk);
  196. break;
  197. default:
  198. BUG();
  199. }
  200. io_thread = kthread_run(rxrpc_io_thread, local,
  201. "krxrpcio/%u", ntohs(udp_conf.local_udp_port));
  202. if (IS_ERR(io_thread)) {
  203. ret = PTR_ERR(io_thread);
  204. goto error_sock;
  205. }
  206. wait_for_completion(&local->io_thread_ready);
  207. WRITE_ONCE(local->io_thread, io_thread);
  208. _leave(" = 0");
  209. return 0;
  210. error_sock:
  211. kernel_sock_shutdown(local->socket, SHUT_RDWR);
  212. local->socket->sk->sk_user_data = NULL;
  213. sock_release(local->socket);
  214. local->socket = NULL;
  215. return ret;
  216. }
  217. /*
  218. * Look up or create a new local endpoint using the specified local address.
  219. */
  220. struct rxrpc_local *rxrpc_lookup_local(struct net *net,
  221. const struct sockaddr_rxrpc *srx)
  222. {
  223. struct rxrpc_local *local;
  224. struct rxrpc_net *rxnet = rxrpc_net(net);
  225. struct hlist_node *cursor;
  226. long diff;
  227. int ret;
  228. _enter("{%d,%d,%pISp}",
  229. srx->transport_type, srx->transport.family, &srx->transport);
  230. mutex_lock(&rxnet->local_mutex);
  231. hlist_for_each(cursor, &rxnet->local_endpoints) {
  232. local = hlist_entry(cursor, struct rxrpc_local, link);
  233. diff = rxrpc_local_cmp_key(local, srx);
  234. if (diff != 0)
  235. continue;
  236. /* Services aren't allowed to share transport sockets, so
  237. * reject that here. It is possible that the object is dying -
  238. * but it may also still have the local transport address that
  239. * we want bound.
  240. */
  241. if (srx->srx_service) {
  242. local = NULL;
  243. goto addr_in_use;
  244. }
  245. /* Found a match. We want to replace a dying object.
  246. * Attempting to bind the transport socket may still fail if
  247. * we're attempting to use a local address that the dying
  248. * object is still using.
  249. */
  250. if (!rxrpc_use_local(local, rxrpc_local_use_lookup))
  251. break;
  252. goto found;
  253. }
  254. local = rxrpc_alloc_local(net, srx);
  255. if (!local)
  256. goto nomem;
  257. ret = rxrpc_open_socket(local, net);
  258. if (ret < 0)
  259. goto sock_error;
  260. if (cursor) {
  261. hlist_replace_rcu(cursor, &local->link);
  262. cursor->pprev = NULL;
  263. } else {
  264. hlist_add_head_rcu(&local->link, &rxnet->local_endpoints);
  265. }
  266. found:
  267. mutex_unlock(&rxnet->local_mutex);
  268. _leave(" = %p", local);
  269. return local;
  270. nomem:
  271. ret = -ENOMEM;
  272. sock_error:
  273. mutex_unlock(&rxnet->local_mutex);
  274. if (local)
  275. call_rcu(&local->rcu, rxrpc_local_rcu);
  276. _leave(" = %d", ret);
  277. return ERR_PTR(ret);
  278. addr_in_use:
  279. mutex_unlock(&rxnet->local_mutex);
  280. _leave(" = -EADDRINUSE");
  281. return ERR_PTR(-EADDRINUSE);
  282. }
  283. /*
  284. * Get a ref on a local endpoint.
  285. */
  286. struct rxrpc_local *rxrpc_get_local(struct rxrpc_local *local,
  287. enum rxrpc_local_trace why)
  288. {
  289. int r, u;
  290. u = atomic_read(&local->active_users);
  291. __refcount_inc(&local->ref, &r);
  292. trace_rxrpc_local(local->debug_id, why, r + 1, u);
  293. return local;
  294. }
  295. /*
  296. * Get a ref on a local endpoint unless its usage has already reached 0.
  297. */
  298. struct rxrpc_local *rxrpc_get_local_maybe(struct rxrpc_local *local,
  299. enum rxrpc_local_trace why)
  300. {
  301. int r, u;
  302. if (local && __refcount_inc_not_zero(&local->ref, &r)) {
  303. u = atomic_read(&local->active_users);
  304. trace_rxrpc_local(local->debug_id, why, r + 1, u);
  305. return local;
  306. }
  307. return NULL;
  308. }
  309. /*
  310. * Drop a ref on a local endpoint.
  311. */
  312. void rxrpc_put_local(struct rxrpc_local *local, enum rxrpc_local_trace why)
  313. {
  314. unsigned int debug_id;
  315. bool dead;
  316. int r, u;
  317. if (local) {
  318. debug_id = local->debug_id;
  319. u = atomic_read(&local->active_users);
  320. dead = __refcount_dec_and_test(&local->ref, &r);
  321. trace_rxrpc_local(debug_id, why, r, u);
  322. if (dead)
  323. call_rcu(&local->rcu, rxrpc_local_rcu);
  324. }
  325. }
  326. /*
  327. * Start using a local endpoint.
  328. */
  329. struct rxrpc_local *rxrpc_use_local(struct rxrpc_local *local,
  330. enum rxrpc_local_trace why)
  331. {
  332. local = rxrpc_get_local_maybe(local, rxrpc_local_get_for_use);
  333. if (!local)
  334. return NULL;
  335. if (!__rxrpc_use_local(local, why)) {
  336. rxrpc_put_local(local, rxrpc_local_put_for_use);
  337. return NULL;
  338. }
  339. return local;
  340. }
  341. /*
  342. * Cease using a local endpoint. Once the number of active users reaches 0, we
  343. * start the closure of the transport in the I/O thread..
  344. */
  345. void rxrpc_unuse_local(struct rxrpc_local *local, enum rxrpc_local_trace why)
  346. {
  347. unsigned int debug_id;
  348. int r, u;
  349. if (local) {
  350. debug_id = local->debug_id;
  351. r = refcount_read(&local->ref);
  352. u = atomic_dec_return(&local->active_users);
  353. trace_rxrpc_local(debug_id, why, r, u);
  354. if (u == 0)
  355. kthread_stop(local->io_thread);
  356. }
  357. }
  358. /*
  359. * Destroy a local endpoint's socket and then hand the record to RCU to dispose
  360. * of.
  361. *
  362. * Closing the socket cannot be done from bottom half context or RCU callback
  363. * context because it might sleep.
  364. */
  365. void rxrpc_destroy_local(struct rxrpc_local *local)
  366. {
  367. struct socket *socket = local->socket;
  368. struct rxrpc_net *rxnet = local->rxnet;
  369. _enter("%d", local->debug_id);
  370. local->dead = true;
  371. mutex_lock(&rxnet->local_mutex);
  372. hlist_del_init_rcu(&local->link);
  373. mutex_unlock(&rxnet->local_mutex);
  374. rxrpc_clean_up_local_conns(local);
  375. rxrpc_service_connection_reaper(&rxnet->service_conn_reaper);
  376. ASSERT(!local->service);
  377. if (socket) {
  378. local->socket = NULL;
  379. kernel_sock_shutdown(socket, SHUT_RDWR);
  380. socket->sk->sk_user_data = NULL;
  381. sock_release(socket);
  382. }
  383. /* At this point, there should be no more packets coming in to the
  384. * local endpoint.
  385. */
  386. #ifdef CONFIG_AF_RXRPC_INJECT_RX_DELAY
  387. rxrpc_purge_queue(&local->rx_delay_queue);
  388. #endif
  389. rxrpc_purge_queue(&local->rx_queue);
  390. rxrpc_purge_client_connections(local);
  391. if (local->tx_alloc.va)
  392. __page_frag_cache_drain(virt_to_page(local->tx_alloc.va),
  393. local->tx_alloc.pagecnt_bias);
  394. }
  395. /*
  396. * Destroy a local endpoint after the RCU grace period expires.
  397. */
  398. static void rxrpc_local_rcu(struct rcu_head *rcu)
  399. {
  400. struct rxrpc_local *local = container_of(rcu, struct rxrpc_local, rcu);
  401. rxrpc_see_local(local, rxrpc_local_free);
  402. kfree(local);
  403. }
  404. /*
  405. * Verify the local endpoint list is empty by this point.
  406. */
  407. void rxrpc_destroy_all_locals(struct rxrpc_net *rxnet)
  408. {
  409. struct rxrpc_local *local;
  410. _enter("");
  411. flush_workqueue(rxrpc_workqueue);
  412. if (!hlist_empty(&rxnet->local_endpoints)) {
  413. mutex_lock(&rxnet->local_mutex);
  414. hlist_for_each_entry(local, &rxnet->local_endpoints, link) {
  415. pr_err("AF_RXRPC: Leaked local %p {%d}\n",
  416. local, refcount_read(&local->ref));
  417. }
  418. mutex_unlock(&rxnet->local_mutex);
  419. BUG();
  420. }
  421. }