bareudp.c 21 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832
  1. // SPDX-License-Identifier: GPL-2.0
  2. /* Bareudp: UDP tunnel encasulation for different Payload types like
  3. * MPLS, NSH, IP, etc.
  4. * Copyright (c) 2019 Nokia, Inc.
  5. * Authors: Martin Varghese, <martin.varghese@nokia.com>
  6. */
  7. #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
  8. #include <linux/kernel.h>
  9. #include <linux/module.h>
  10. #include <linux/etherdevice.h>
  11. #include <linux/hash.h>
  12. #include <net/dst_metadata.h>
  13. #include <net/gro_cells.h>
  14. #include <net/rtnetlink.h>
  15. #include <net/protocol.h>
  16. #include <net/ip6_tunnel.h>
  17. #include <net/ip_tunnels.h>
  18. #include <net/udp_tunnel.h>
  19. #include <net/bareudp.h>
  20. #define BAREUDP_BASE_HLEN sizeof(struct udphdr)
  21. #define BAREUDP_IPV4_HLEN (sizeof(struct iphdr) + \
  22. sizeof(struct udphdr))
  23. #define BAREUDP_IPV6_HLEN (sizeof(struct ipv6hdr) + \
  24. sizeof(struct udphdr))
  25. static bool log_ecn_error = true;
  26. module_param(log_ecn_error, bool, 0644);
  27. MODULE_PARM_DESC(log_ecn_error, "Log packets received with corrupted ECN");
  28. /* per-network namespace private data for this module */
  29. static unsigned int bareudp_net_id;
  30. struct bareudp_net {
  31. struct list_head bareudp_list;
  32. };
  33. struct bareudp_conf {
  34. __be16 ethertype;
  35. __be16 port;
  36. u16 sport_min;
  37. bool multi_proto_mode;
  38. };
  39. /* Pseudo network device */
  40. struct bareudp_dev {
  41. struct net *net; /* netns for packet i/o */
  42. struct net_device *dev; /* netdev for bareudp tunnel */
  43. __be16 ethertype;
  44. __be16 port;
  45. u16 sport_min;
  46. bool multi_proto_mode;
  47. struct socket __rcu *sock;
  48. struct list_head next; /* bareudp node on namespace list */
  49. struct gro_cells gro_cells;
  50. };
  51. static int bareudp_udp_encap_recv(struct sock *sk, struct sk_buff *skb)
  52. {
  53. struct metadata_dst *tun_dst = NULL;
  54. IP_TUNNEL_DECLARE_FLAGS(key) = { };
  55. struct bareudp_dev *bareudp;
  56. unsigned short family;
  57. unsigned int len;
  58. __be16 proto;
  59. void *oiph;
  60. int err;
  61. int nh;
  62. bareudp = rcu_dereference_sk_user_data(sk);
  63. if (!bareudp)
  64. goto drop;
  65. if (skb->protocol == htons(ETH_P_IP))
  66. family = AF_INET;
  67. else
  68. family = AF_INET6;
  69. if (bareudp->ethertype == htons(ETH_P_IP)) {
  70. __u8 ipversion;
  71. if (skb_copy_bits(skb, BAREUDP_BASE_HLEN, &ipversion,
  72. sizeof(ipversion))) {
  73. DEV_STATS_INC(bareudp->dev, rx_dropped);
  74. goto drop;
  75. }
  76. ipversion >>= 4;
  77. if (ipversion == 4) {
  78. proto = htons(ETH_P_IP);
  79. } else if (ipversion == 6 && bareudp->multi_proto_mode) {
  80. proto = htons(ETH_P_IPV6);
  81. } else {
  82. DEV_STATS_INC(bareudp->dev, rx_dropped);
  83. goto drop;
  84. }
  85. } else if (bareudp->ethertype == htons(ETH_P_MPLS_UC)) {
  86. struct iphdr *tunnel_hdr;
  87. tunnel_hdr = (struct iphdr *)skb_network_header(skb);
  88. if (tunnel_hdr->version == 4) {
  89. if (!ipv4_is_multicast(tunnel_hdr->daddr)) {
  90. proto = bareudp->ethertype;
  91. } else if (bareudp->multi_proto_mode &&
  92. ipv4_is_multicast(tunnel_hdr->daddr)) {
  93. proto = htons(ETH_P_MPLS_MC);
  94. } else {
  95. DEV_STATS_INC(bareudp->dev, rx_dropped);
  96. goto drop;
  97. }
  98. } else {
  99. int addr_type;
  100. struct ipv6hdr *tunnel_hdr_v6;
  101. tunnel_hdr_v6 = (struct ipv6hdr *)skb_network_header(skb);
  102. addr_type =
  103. ipv6_addr_type((struct in6_addr *)&tunnel_hdr_v6->daddr);
  104. if (!(addr_type & IPV6_ADDR_MULTICAST)) {
  105. proto = bareudp->ethertype;
  106. } else if (bareudp->multi_proto_mode &&
  107. (addr_type & IPV6_ADDR_MULTICAST)) {
  108. proto = htons(ETH_P_MPLS_MC);
  109. } else {
  110. DEV_STATS_INC(bareudp->dev, rx_dropped);
  111. goto drop;
  112. }
  113. }
  114. } else {
  115. proto = bareudp->ethertype;
  116. }
  117. if (iptunnel_pull_header(skb, BAREUDP_BASE_HLEN,
  118. proto,
  119. !net_eq(bareudp->net,
  120. dev_net(bareudp->dev)))) {
  121. DEV_STATS_INC(bareudp->dev, rx_dropped);
  122. goto drop;
  123. }
  124. __set_bit(IP_TUNNEL_KEY_BIT, key);
  125. tun_dst = udp_tun_rx_dst(skb, family, key, 0, 0);
  126. if (!tun_dst) {
  127. DEV_STATS_INC(bareudp->dev, rx_dropped);
  128. goto drop;
  129. }
  130. skb_dst_set(skb, &tun_dst->dst);
  131. skb->dev = bareudp->dev;
  132. skb_reset_mac_header(skb);
  133. /* Save offset of outer header relative to skb->head,
  134. * because we are going to reset the network header to the inner header
  135. * and might change skb->head.
  136. */
  137. nh = skb_network_header(skb) - skb->head;
  138. skb_reset_network_header(skb);
  139. if (!pskb_inet_may_pull(skb)) {
  140. DEV_STATS_INC(bareudp->dev, rx_length_errors);
  141. DEV_STATS_INC(bareudp->dev, rx_errors);
  142. goto drop;
  143. }
  144. /* Get the outer header. */
  145. oiph = skb->head + nh;
  146. if (!ipv6_mod_enabled() || family == AF_INET)
  147. err = IP_ECN_decapsulate(oiph, skb);
  148. else
  149. err = IP6_ECN_decapsulate(oiph, skb);
  150. if (unlikely(err)) {
  151. if (log_ecn_error) {
  152. if (!ipv6_mod_enabled() || family == AF_INET)
  153. net_info_ratelimited("non-ECT from %pI4 "
  154. "with TOS=%#x\n",
  155. &((struct iphdr *)oiph)->saddr,
  156. ((struct iphdr *)oiph)->tos);
  157. else
  158. net_info_ratelimited("non-ECT from %pI6\n",
  159. &((struct ipv6hdr *)oiph)->saddr);
  160. }
  161. if (err > 1) {
  162. DEV_STATS_INC(bareudp->dev, rx_frame_errors);
  163. DEV_STATS_INC(bareudp->dev, rx_errors);
  164. goto drop;
  165. }
  166. }
  167. len = skb->len;
  168. err = gro_cells_receive(&bareudp->gro_cells, skb);
  169. if (likely(err == NET_RX_SUCCESS))
  170. dev_sw_netstats_rx_add(bareudp->dev, len);
  171. return 0;
  172. drop:
  173. /* Consume bad packet */
  174. kfree_skb(skb);
  175. return 0;
  176. }
  177. static int bareudp_err_lookup(struct sock *sk, struct sk_buff *skb)
  178. {
  179. return 0;
  180. }
  181. static int bareudp_init(struct net_device *dev)
  182. {
  183. struct bareudp_dev *bareudp = netdev_priv(dev);
  184. int err;
  185. err = gro_cells_init(&bareudp->gro_cells, dev);
  186. if (err)
  187. return err;
  188. return 0;
  189. }
  190. static void bareudp_uninit(struct net_device *dev)
  191. {
  192. struct bareudp_dev *bareudp = netdev_priv(dev);
  193. gro_cells_destroy(&bareudp->gro_cells);
  194. }
  195. static struct socket *bareudp_create_sock(struct net *net, __be16 port)
  196. {
  197. struct udp_port_cfg udp_conf;
  198. struct socket *sock;
  199. int err;
  200. memset(&udp_conf, 0, sizeof(udp_conf));
  201. if (ipv6_mod_enabled())
  202. udp_conf.family = AF_INET6;
  203. else
  204. udp_conf.family = AF_INET;
  205. udp_conf.local_udp_port = port;
  206. /* Open UDP socket */
  207. err = udp_sock_create(net, &udp_conf, &sock);
  208. if (err < 0)
  209. return ERR_PTR(err);
  210. udp_allow_gso(sock->sk);
  211. return sock;
  212. }
  213. /* Create new listen socket if needed */
  214. static int bareudp_socket_create(struct bareudp_dev *bareudp, __be16 port)
  215. {
  216. struct udp_tunnel_sock_cfg tunnel_cfg;
  217. struct socket *sock;
  218. sock = bareudp_create_sock(bareudp->net, port);
  219. if (IS_ERR(sock))
  220. return PTR_ERR(sock);
  221. /* Mark socket as an encapsulation socket */
  222. memset(&tunnel_cfg, 0, sizeof(tunnel_cfg));
  223. tunnel_cfg.sk_user_data = bareudp;
  224. tunnel_cfg.encap_type = 1;
  225. tunnel_cfg.encap_rcv = bareudp_udp_encap_recv;
  226. tunnel_cfg.encap_err_lookup = bareudp_err_lookup;
  227. tunnel_cfg.encap_destroy = NULL;
  228. setup_udp_tunnel_sock(bareudp->net, sock, &tunnel_cfg);
  229. rcu_assign_pointer(bareudp->sock, sock);
  230. return 0;
  231. }
  232. static int bareudp_open(struct net_device *dev)
  233. {
  234. struct bareudp_dev *bareudp = netdev_priv(dev);
  235. int ret = 0;
  236. ret = bareudp_socket_create(bareudp, bareudp->port);
  237. return ret;
  238. }
  239. static void bareudp_sock_release(struct bareudp_dev *bareudp)
  240. {
  241. struct socket *sock;
  242. sock = bareudp->sock;
  243. rcu_assign_pointer(bareudp->sock, NULL);
  244. synchronize_net();
  245. udp_tunnel_sock_release(sock);
  246. }
  247. static int bareudp_stop(struct net_device *dev)
  248. {
  249. struct bareudp_dev *bareudp = netdev_priv(dev);
  250. bareudp_sock_release(bareudp);
  251. return 0;
  252. }
  253. static int bareudp_xmit_skb(struct sk_buff *skb, struct net_device *dev,
  254. struct bareudp_dev *bareudp,
  255. const struct ip_tunnel_info *info)
  256. {
  257. bool udp_sum = test_bit(IP_TUNNEL_CSUM_BIT, info->key.tun_flags);
  258. bool xnet = !net_eq(bareudp->net, dev_net(bareudp->dev));
  259. bool use_cache = ip_tunnel_dst_cache_usable(skb, info);
  260. struct socket *sock = rcu_dereference(bareudp->sock);
  261. const struct ip_tunnel_key *key = &info->key;
  262. struct rtable *rt;
  263. __be16 sport, df;
  264. int min_headroom;
  265. __u8 tos, ttl;
  266. __be32 saddr;
  267. int err;
  268. if (!skb_vlan_inet_prepare(skb, skb->protocol != htons(ETH_P_TEB)))
  269. return -EINVAL;
  270. if (!sock)
  271. return -ESHUTDOWN;
  272. sport = udp_flow_src_port(bareudp->net, skb,
  273. bareudp->sport_min, USHRT_MAX,
  274. true);
  275. rt = udp_tunnel_dst_lookup(skb, dev, bareudp->net, 0, &saddr, &info->key,
  276. sport, bareudp->port, key->tos,
  277. use_cache ?
  278. (struct dst_cache *)&info->dst_cache : NULL);
  279. if (IS_ERR(rt))
  280. return PTR_ERR(rt);
  281. skb_tunnel_check_pmtu(skb, &rt->dst,
  282. BAREUDP_IPV4_HLEN + info->options_len, false);
  283. tos = ip_tunnel_ecn_encap(key->tos, ip_hdr(skb), skb);
  284. ttl = key->ttl;
  285. df = test_bit(IP_TUNNEL_DONT_FRAGMENT_BIT, key->tun_flags) ?
  286. htons(IP_DF) : 0;
  287. skb_scrub_packet(skb, xnet);
  288. err = -ENOSPC;
  289. if (!skb_pull(skb, skb_network_offset(skb)))
  290. goto free_dst;
  291. min_headroom = LL_RESERVED_SPACE(rt->dst.dev) + rt->dst.header_len +
  292. BAREUDP_BASE_HLEN + info->options_len + sizeof(struct iphdr);
  293. err = skb_cow_head(skb, min_headroom);
  294. if (unlikely(err))
  295. goto free_dst;
  296. err = udp_tunnel_handle_offloads(skb, udp_sum);
  297. if (err)
  298. goto free_dst;
  299. skb_set_inner_protocol(skb, bareudp->ethertype);
  300. udp_tunnel_xmit_skb(rt, sock->sk, skb, saddr, info->key.u.ipv4.dst,
  301. tos, ttl, df, sport, bareudp->port,
  302. !net_eq(bareudp->net, dev_net(bareudp->dev)),
  303. !test_bit(IP_TUNNEL_CSUM_BIT,
  304. info->key.tun_flags));
  305. return 0;
  306. free_dst:
  307. dst_release(&rt->dst);
  308. return err;
  309. }
  310. static int bareudp6_xmit_skb(struct sk_buff *skb, struct net_device *dev,
  311. struct bareudp_dev *bareudp,
  312. const struct ip_tunnel_info *info)
  313. {
  314. bool udp_sum = test_bit(IP_TUNNEL_CSUM_BIT, info->key.tun_flags);
  315. bool xnet = !net_eq(bareudp->net, dev_net(bareudp->dev));
  316. bool use_cache = ip_tunnel_dst_cache_usable(skb, info);
  317. struct socket *sock = rcu_dereference(bareudp->sock);
  318. const struct ip_tunnel_key *key = &info->key;
  319. struct dst_entry *dst = NULL;
  320. struct in6_addr saddr, daddr;
  321. int min_headroom;
  322. __u8 prio, ttl;
  323. __be16 sport;
  324. int err;
  325. if (!skb_vlan_inet_prepare(skb, skb->protocol != htons(ETH_P_TEB)))
  326. return -EINVAL;
  327. if (!sock)
  328. return -ESHUTDOWN;
  329. sport = udp_flow_src_port(bareudp->net, skb,
  330. bareudp->sport_min, USHRT_MAX,
  331. true);
  332. dst = udp_tunnel6_dst_lookup(skb, dev, bareudp->net, sock, 0, &saddr,
  333. key, sport, bareudp->port, key->tos,
  334. use_cache ?
  335. (struct dst_cache *) &info->dst_cache : NULL);
  336. if (IS_ERR(dst))
  337. return PTR_ERR(dst);
  338. skb_tunnel_check_pmtu(skb, dst, BAREUDP_IPV6_HLEN + info->options_len,
  339. false);
  340. prio = ip_tunnel_ecn_encap(key->tos, ip_hdr(skb), skb);
  341. ttl = key->ttl;
  342. skb_scrub_packet(skb, xnet);
  343. err = -ENOSPC;
  344. if (!skb_pull(skb, skb_network_offset(skb)))
  345. goto free_dst;
  346. min_headroom = LL_RESERVED_SPACE(dst->dev) + dst->header_len +
  347. BAREUDP_BASE_HLEN + info->options_len + sizeof(struct ipv6hdr);
  348. err = skb_cow_head(skb, min_headroom);
  349. if (unlikely(err))
  350. goto free_dst;
  351. err = udp_tunnel_handle_offloads(skb, udp_sum);
  352. if (err)
  353. goto free_dst;
  354. daddr = info->key.u.ipv6.dst;
  355. udp_tunnel6_xmit_skb(dst, sock->sk, skb, dev,
  356. &saddr, &daddr, prio, ttl,
  357. info->key.label, sport, bareudp->port,
  358. !test_bit(IP_TUNNEL_CSUM_BIT,
  359. info->key.tun_flags));
  360. return 0;
  361. free_dst:
  362. dst_release(dst);
  363. return err;
  364. }
  365. static bool bareudp_proto_valid(struct bareudp_dev *bareudp, __be16 proto)
  366. {
  367. if (bareudp->ethertype == proto)
  368. return true;
  369. if (!bareudp->multi_proto_mode)
  370. return false;
  371. if (bareudp->ethertype == htons(ETH_P_MPLS_UC) &&
  372. proto == htons(ETH_P_MPLS_MC))
  373. return true;
  374. if (bareudp->ethertype == htons(ETH_P_IP) &&
  375. proto == htons(ETH_P_IPV6))
  376. return true;
  377. return false;
  378. }
  379. static netdev_tx_t bareudp_xmit(struct sk_buff *skb, struct net_device *dev)
  380. {
  381. struct bareudp_dev *bareudp = netdev_priv(dev);
  382. struct ip_tunnel_info *info = NULL;
  383. int err;
  384. if (!bareudp_proto_valid(bareudp, skb->protocol)) {
  385. err = -EINVAL;
  386. goto tx_error;
  387. }
  388. info = skb_tunnel_info(skb);
  389. if (unlikely(!info || !(info->mode & IP_TUNNEL_INFO_TX))) {
  390. err = -EINVAL;
  391. goto tx_error;
  392. }
  393. rcu_read_lock();
  394. if (ipv6_mod_enabled() && info->mode & IP_TUNNEL_INFO_IPV6)
  395. err = bareudp6_xmit_skb(skb, dev, bareudp, info);
  396. else
  397. err = bareudp_xmit_skb(skb, dev, bareudp, info);
  398. rcu_read_unlock();
  399. if (likely(!err))
  400. return NETDEV_TX_OK;
  401. tx_error:
  402. dev_kfree_skb(skb);
  403. if (err == -ELOOP)
  404. DEV_STATS_INC(dev, collisions);
  405. else if (err == -ENETUNREACH)
  406. DEV_STATS_INC(dev, tx_carrier_errors);
  407. DEV_STATS_INC(dev, tx_errors);
  408. return NETDEV_TX_OK;
  409. }
  410. static int bareudp_fill_metadata_dst(struct net_device *dev,
  411. struct sk_buff *skb)
  412. {
  413. struct ip_tunnel_info *info = skb_tunnel_info(skb);
  414. struct bareudp_dev *bareudp = netdev_priv(dev);
  415. bool use_cache;
  416. __be16 sport;
  417. use_cache = ip_tunnel_dst_cache_usable(skb, info);
  418. sport = udp_flow_src_port(bareudp->net, skb,
  419. bareudp->sport_min, USHRT_MAX,
  420. true);
  421. if (!ipv6_mod_enabled() || ip_tunnel_info_af(info) == AF_INET) {
  422. struct rtable *rt;
  423. __be32 saddr;
  424. rt = udp_tunnel_dst_lookup(skb, dev, bareudp->net, 0, &saddr,
  425. &info->key, sport, bareudp->port,
  426. info->key.tos,
  427. use_cache ? &info->dst_cache : NULL);
  428. if (IS_ERR(rt))
  429. return PTR_ERR(rt);
  430. ip_rt_put(rt);
  431. info->key.u.ipv4.src = saddr;
  432. } else if (ip_tunnel_info_af(info) == AF_INET6) {
  433. struct dst_entry *dst;
  434. struct in6_addr saddr;
  435. struct socket *sock = rcu_dereference(bareudp->sock);
  436. dst = udp_tunnel6_dst_lookup(skb, dev, bareudp->net, sock,
  437. 0, &saddr, &info->key,
  438. sport, bareudp->port, info->key.tos,
  439. use_cache ? &info->dst_cache : NULL);
  440. if (IS_ERR(dst))
  441. return PTR_ERR(dst);
  442. dst_release(dst);
  443. info->key.u.ipv6.src = saddr;
  444. } else {
  445. return -EINVAL;
  446. }
  447. info->key.tp_src = sport;
  448. info->key.tp_dst = bareudp->port;
  449. return 0;
  450. }
  451. static const struct net_device_ops bareudp_netdev_ops = {
  452. .ndo_init = bareudp_init,
  453. .ndo_uninit = bareudp_uninit,
  454. .ndo_open = bareudp_open,
  455. .ndo_stop = bareudp_stop,
  456. .ndo_start_xmit = bareudp_xmit,
  457. .ndo_fill_metadata_dst = bareudp_fill_metadata_dst,
  458. };
  459. static const struct nla_policy bareudp_policy[IFLA_BAREUDP_MAX + 1] = {
  460. [IFLA_BAREUDP_PORT] = { .type = NLA_U16 },
  461. [IFLA_BAREUDP_ETHERTYPE] = { .type = NLA_U16 },
  462. [IFLA_BAREUDP_SRCPORT_MIN] = { .type = NLA_U16 },
  463. [IFLA_BAREUDP_MULTIPROTO_MODE] = { .type = NLA_FLAG },
  464. };
  465. /* Info for udev, that this is a virtual tunnel endpoint */
  466. static const struct device_type bareudp_type = {
  467. .name = "bareudp",
  468. };
  469. /* Initialize the device structure. */
  470. static void bareudp_setup(struct net_device *dev)
  471. {
  472. dev->netdev_ops = &bareudp_netdev_ops;
  473. dev->needs_free_netdev = true;
  474. SET_NETDEV_DEVTYPE(dev, &bareudp_type);
  475. dev->features |= NETIF_F_SG | NETIF_F_HW_CSUM | NETIF_F_FRAGLIST;
  476. dev->features |= NETIF_F_RXCSUM;
  477. dev->features |= NETIF_F_GSO_SOFTWARE;
  478. dev->hw_features |= NETIF_F_SG | NETIF_F_HW_CSUM | NETIF_F_FRAGLIST;
  479. dev->hw_features |= NETIF_F_RXCSUM;
  480. dev->hw_features |= NETIF_F_GSO_SOFTWARE;
  481. dev->hard_header_len = 0;
  482. dev->addr_len = 0;
  483. dev->mtu = ETH_DATA_LEN;
  484. dev->min_mtu = IPV4_MIN_MTU;
  485. dev->max_mtu = IP_MAX_MTU - BAREUDP_BASE_HLEN;
  486. dev->type = ARPHRD_NONE;
  487. netif_keep_dst(dev);
  488. dev->priv_flags |= IFF_NO_QUEUE;
  489. dev->lltx = true;
  490. dev->flags = IFF_POINTOPOINT | IFF_NOARP | IFF_MULTICAST;
  491. dev->pcpu_stat_type = NETDEV_PCPU_STAT_TSTATS;
  492. }
  493. static int bareudp_validate(struct nlattr *tb[], struct nlattr *data[],
  494. struct netlink_ext_ack *extack)
  495. {
  496. if (!data) {
  497. NL_SET_ERR_MSG(extack,
  498. "Not enough attributes provided to perform the operation");
  499. return -EINVAL;
  500. }
  501. return 0;
  502. }
  503. static int bareudp2info(struct nlattr *data[], struct bareudp_conf *conf,
  504. struct netlink_ext_ack *extack)
  505. {
  506. memset(conf, 0, sizeof(*conf));
  507. if (!data[IFLA_BAREUDP_PORT]) {
  508. NL_SET_ERR_MSG(extack, "port not specified");
  509. return -EINVAL;
  510. }
  511. if (!data[IFLA_BAREUDP_ETHERTYPE]) {
  512. NL_SET_ERR_MSG(extack, "ethertype not specified");
  513. return -EINVAL;
  514. }
  515. conf->port = nla_get_u16(data[IFLA_BAREUDP_PORT]);
  516. conf->ethertype = nla_get_u16(data[IFLA_BAREUDP_ETHERTYPE]);
  517. if (data[IFLA_BAREUDP_SRCPORT_MIN])
  518. conf->sport_min = nla_get_u16(data[IFLA_BAREUDP_SRCPORT_MIN]);
  519. if (data[IFLA_BAREUDP_MULTIPROTO_MODE])
  520. conf->multi_proto_mode = true;
  521. return 0;
  522. }
  523. static struct bareudp_dev *bareudp_find_dev(struct bareudp_net *bn,
  524. const struct bareudp_conf *conf)
  525. {
  526. struct bareudp_dev *bareudp, *t = NULL;
  527. list_for_each_entry(bareudp, &bn->bareudp_list, next) {
  528. if (conf->port == bareudp->port)
  529. t = bareudp;
  530. }
  531. return t;
  532. }
  533. static int bareudp_configure(struct net *net, struct net_device *dev,
  534. struct bareudp_conf *conf,
  535. struct netlink_ext_ack *extack)
  536. {
  537. struct bareudp_net *bn = net_generic(net, bareudp_net_id);
  538. struct bareudp_dev *t, *bareudp = netdev_priv(dev);
  539. int err;
  540. bareudp->net = net;
  541. bareudp->dev = dev;
  542. t = bareudp_find_dev(bn, conf);
  543. if (t) {
  544. NL_SET_ERR_MSG(extack, "Another bareudp device using the same port already exists");
  545. return -EBUSY;
  546. }
  547. if (conf->multi_proto_mode &&
  548. (conf->ethertype != htons(ETH_P_MPLS_UC) &&
  549. conf->ethertype != htons(ETH_P_IP))) {
  550. NL_SET_ERR_MSG(extack, "Cannot set multiproto mode for this ethertype (only IPv4 and unicast MPLS are supported)");
  551. return -EINVAL;
  552. }
  553. bareudp->port = conf->port;
  554. bareudp->ethertype = conf->ethertype;
  555. bareudp->sport_min = conf->sport_min;
  556. bareudp->multi_proto_mode = conf->multi_proto_mode;
  557. err = register_netdevice(dev);
  558. if (err)
  559. return err;
  560. list_add(&bareudp->next, &bn->bareudp_list);
  561. return 0;
  562. }
  563. static int bareudp_link_config(struct net_device *dev,
  564. struct nlattr *tb[])
  565. {
  566. int err;
  567. if (tb[IFLA_MTU]) {
  568. err = dev_set_mtu(dev, nla_get_u32(tb[IFLA_MTU]));
  569. if (err)
  570. return err;
  571. }
  572. return 0;
  573. }
  574. static void bareudp_dellink(struct net_device *dev, struct list_head *head)
  575. {
  576. struct bareudp_dev *bareudp = netdev_priv(dev);
  577. list_del(&bareudp->next);
  578. unregister_netdevice_queue(dev, head);
  579. }
  580. static int bareudp_newlink(struct net *net, struct net_device *dev,
  581. struct nlattr *tb[], struct nlattr *data[],
  582. struct netlink_ext_ack *extack)
  583. {
  584. struct bareudp_conf conf;
  585. int err;
  586. err = bareudp2info(data, &conf, extack);
  587. if (err)
  588. return err;
  589. err = bareudp_configure(net, dev, &conf, extack);
  590. if (err)
  591. return err;
  592. err = bareudp_link_config(dev, tb);
  593. if (err)
  594. goto err_unconfig;
  595. return 0;
  596. err_unconfig:
  597. bareudp_dellink(dev, NULL);
  598. return err;
  599. }
  600. static size_t bareudp_get_size(const struct net_device *dev)
  601. {
  602. return nla_total_size(sizeof(__be16)) + /* IFLA_BAREUDP_PORT */
  603. nla_total_size(sizeof(__be16)) + /* IFLA_BAREUDP_ETHERTYPE */
  604. nla_total_size(sizeof(__u16)) + /* IFLA_BAREUDP_SRCPORT_MIN */
  605. nla_total_size(0) + /* IFLA_BAREUDP_MULTIPROTO_MODE */
  606. 0;
  607. }
  608. static int bareudp_fill_info(struct sk_buff *skb, const struct net_device *dev)
  609. {
  610. struct bareudp_dev *bareudp = netdev_priv(dev);
  611. if (nla_put_be16(skb, IFLA_BAREUDP_PORT, bareudp->port))
  612. goto nla_put_failure;
  613. if (nla_put_be16(skb, IFLA_BAREUDP_ETHERTYPE, bareudp->ethertype))
  614. goto nla_put_failure;
  615. if (nla_put_u16(skb, IFLA_BAREUDP_SRCPORT_MIN, bareudp->sport_min))
  616. goto nla_put_failure;
  617. if (bareudp->multi_proto_mode &&
  618. nla_put_flag(skb, IFLA_BAREUDP_MULTIPROTO_MODE))
  619. goto nla_put_failure;
  620. return 0;
  621. nla_put_failure:
  622. return -EMSGSIZE;
  623. }
  624. static struct rtnl_link_ops bareudp_link_ops __read_mostly = {
  625. .kind = "bareudp",
  626. .maxtype = IFLA_BAREUDP_MAX,
  627. .policy = bareudp_policy,
  628. .priv_size = sizeof(struct bareudp_dev),
  629. .setup = bareudp_setup,
  630. .validate = bareudp_validate,
  631. .newlink = bareudp_newlink,
  632. .dellink = bareudp_dellink,
  633. .get_size = bareudp_get_size,
  634. .fill_info = bareudp_fill_info,
  635. };
  636. static __net_init int bareudp_init_net(struct net *net)
  637. {
  638. struct bareudp_net *bn = net_generic(net, bareudp_net_id);
  639. INIT_LIST_HEAD(&bn->bareudp_list);
  640. return 0;
  641. }
  642. static void bareudp_destroy_tunnels(struct net *net, struct list_head *head)
  643. {
  644. struct bareudp_net *bn = net_generic(net, bareudp_net_id);
  645. struct bareudp_dev *bareudp, *next;
  646. list_for_each_entry_safe(bareudp, next, &bn->bareudp_list, next)
  647. unregister_netdevice_queue(bareudp->dev, head);
  648. }
  649. static void __net_exit bareudp_exit_batch_rtnl(struct list_head *net_list,
  650. struct list_head *dev_kill_list)
  651. {
  652. struct net *net;
  653. list_for_each_entry(net, net_list, exit_list)
  654. bareudp_destroy_tunnels(net, dev_kill_list);
  655. }
  656. static struct pernet_operations bareudp_net_ops = {
  657. .init = bareudp_init_net,
  658. .exit_batch_rtnl = bareudp_exit_batch_rtnl,
  659. .id = &bareudp_net_id,
  660. .size = sizeof(struct bareudp_net),
  661. };
  662. static int __init bareudp_init_module(void)
  663. {
  664. int rc;
  665. rc = register_pernet_subsys(&bareudp_net_ops);
  666. if (rc)
  667. goto out1;
  668. rc = rtnl_link_register(&bareudp_link_ops);
  669. if (rc)
  670. goto out2;
  671. return 0;
  672. out2:
  673. unregister_pernet_subsys(&bareudp_net_ops);
  674. out1:
  675. return rc;
  676. }
  677. late_initcall(bareudp_init_module);
  678. static void __exit bareudp_cleanup_module(void)
  679. {
  680. rtnl_link_unregister(&bareudp_link_ops);
  681. unregister_pernet_subsys(&bareudp_net_ops);
  682. }
  683. module_exit(bareudp_cleanup_module);
  684. MODULE_ALIAS_RTNL_LINK("bareudp");
  685. MODULE_LICENSE("GPL");
  686. MODULE_AUTHOR("Martin Varghese <martin.varghese@nokia.com>");
  687. MODULE_DESCRIPTION("Interface driver for UDP encapsulated traffic");