ioam6.c 24 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043
  1. // SPDX-License-Identifier: GPL-2.0+
  2. /*
  3. * IPv6 IOAM implementation
  4. *
  5. * Author:
  6. * Justin Iurman <justin.iurman@uliege.be>
  7. */
  8. #include <linux/errno.h>
  9. #include <linux/types.h>
  10. #include <linux/kernel.h>
  11. #include <linux/net.h>
  12. #include <linux/ioam6.h>
  13. #include <linux/ioam6_genl.h>
  14. #include <linux/rhashtable.h>
  15. #include <linux/netdevice.h>
  16. #include <net/addrconf.h>
  17. #include <net/genetlink.h>
  18. #include <net/ioam6.h>
  19. #include <net/sch_generic.h>
  20. static void ioam6_ns_release(struct ioam6_namespace *ns)
  21. {
  22. kfree_rcu(ns, rcu);
  23. }
  24. static void ioam6_sc_release(struct ioam6_schema *sc)
  25. {
  26. kfree_rcu(sc, rcu);
  27. }
  28. static void ioam6_free_ns(void *ptr, void *arg)
  29. {
  30. struct ioam6_namespace *ns = (struct ioam6_namespace *)ptr;
  31. if (ns)
  32. ioam6_ns_release(ns);
  33. }
  34. static void ioam6_free_sc(void *ptr, void *arg)
  35. {
  36. struct ioam6_schema *sc = (struct ioam6_schema *)ptr;
  37. if (sc)
  38. ioam6_sc_release(sc);
  39. }
  40. static int ioam6_ns_cmpfn(struct rhashtable_compare_arg *arg, const void *obj)
  41. {
  42. const struct ioam6_namespace *ns = obj;
  43. return (ns->id != *(__be16 *)arg->key);
  44. }
  45. static int ioam6_sc_cmpfn(struct rhashtable_compare_arg *arg, const void *obj)
  46. {
  47. const struct ioam6_schema *sc = obj;
  48. return (sc->id != *(u32 *)arg->key);
  49. }
  50. static const struct rhashtable_params rht_ns_params = {
  51. .key_len = sizeof(__be16),
  52. .key_offset = offsetof(struct ioam6_namespace, id),
  53. .head_offset = offsetof(struct ioam6_namespace, head),
  54. .automatic_shrinking = true,
  55. .obj_cmpfn = ioam6_ns_cmpfn,
  56. };
  57. static const struct rhashtable_params rht_sc_params = {
  58. .key_len = sizeof(u32),
  59. .key_offset = offsetof(struct ioam6_schema, id),
  60. .head_offset = offsetof(struct ioam6_schema, head),
  61. .automatic_shrinking = true,
  62. .obj_cmpfn = ioam6_sc_cmpfn,
  63. };
  64. static struct genl_family ioam6_genl_family;
  65. static const struct nla_policy ioam6_genl_policy_addns[] = {
  66. [IOAM6_ATTR_NS_ID] = { .type = NLA_U16 },
  67. [IOAM6_ATTR_NS_DATA] = { .type = NLA_U32 },
  68. [IOAM6_ATTR_NS_DATA_WIDE] = { .type = NLA_U64 },
  69. };
  70. static const struct nla_policy ioam6_genl_policy_delns[] = {
  71. [IOAM6_ATTR_NS_ID] = { .type = NLA_U16 },
  72. };
  73. static const struct nla_policy ioam6_genl_policy_addsc[] = {
  74. [IOAM6_ATTR_SC_ID] = { .type = NLA_U32 },
  75. [IOAM6_ATTR_SC_DATA] = { .type = NLA_BINARY,
  76. .len = IOAM6_MAX_SCHEMA_DATA_LEN },
  77. };
  78. static const struct nla_policy ioam6_genl_policy_delsc[] = {
  79. [IOAM6_ATTR_SC_ID] = { .type = NLA_U32 },
  80. };
  81. static const struct nla_policy ioam6_genl_policy_ns_sc[] = {
  82. [IOAM6_ATTR_NS_ID] = { .type = NLA_U16 },
  83. [IOAM6_ATTR_SC_ID] = { .type = NLA_U32 },
  84. [IOAM6_ATTR_SC_NONE] = { .type = NLA_FLAG },
  85. };
  86. static int ioam6_genl_addns(struct sk_buff *skb, struct genl_info *info)
  87. {
  88. struct ioam6_pernet_data *nsdata;
  89. struct ioam6_namespace *ns;
  90. u64 data64;
  91. u32 data32;
  92. __be16 id;
  93. int err;
  94. if (!info->attrs[IOAM6_ATTR_NS_ID])
  95. return -EINVAL;
  96. id = cpu_to_be16(nla_get_u16(info->attrs[IOAM6_ATTR_NS_ID]));
  97. nsdata = ioam6_pernet(genl_info_net(info));
  98. mutex_lock(&nsdata->lock);
  99. ns = rhashtable_lookup_fast(&nsdata->namespaces, &id, rht_ns_params);
  100. if (ns) {
  101. err = -EEXIST;
  102. goto out_unlock;
  103. }
  104. ns = kzalloc(sizeof(*ns), GFP_KERNEL);
  105. if (!ns) {
  106. err = -ENOMEM;
  107. goto out_unlock;
  108. }
  109. ns->id = id;
  110. if (!info->attrs[IOAM6_ATTR_NS_DATA])
  111. data32 = IOAM6_U32_UNAVAILABLE;
  112. else
  113. data32 = nla_get_u32(info->attrs[IOAM6_ATTR_NS_DATA]);
  114. if (!info->attrs[IOAM6_ATTR_NS_DATA_WIDE])
  115. data64 = IOAM6_U64_UNAVAILABLE;
  116. else
  117. data64 = nla_get_u64(info->attrs[IOAM6_ATTR_NS_DATA_WIDE]);
  118. ns->data = cpu_to_be32(data32);
  119. ns->data_wide = cpu_to_be64(data64);
  120. err = rhashtable_lookup_insert_fast(&nsdata->namespaces, &ns->head,
  121. rht_ns_params);
  122. if (err)
  123. kfree(ns);
  124. out_unlock:
  125. mutex_unlock(&nsdata->lock);
  126. return err;
  127. }
  128. static int ioam6_genl_delns(struct sk_buff *skb, struct genl_info *info)
  129. {
  130. struct ioam6_pernet_data *nsdata;
  131. struct ioam6_namespace *ns;
  132. struct ioam6_schema *sc;
  133. __be16 id;
  134. int err;
  135. if (!info->attrs[IOAM6_ATTR_NS_ID])
  136. return -EINVAL;
  137. id = cpu_to_be16(nla_get_u16(info->attrs[IOAM6_ATTR_NS_ID]));
  138. nsdata = ioam6_pernet(genl_info_net(info));
  139. mutex_lock(&nsdata->lock);
  140. ns = rhashtable_lookup_fast(&nsdata->namespaces, &id, rht_ns_params);
  141. if (!ns) {
  142. err = -ENOENT;
  143. goto out_unlock;
  144. }
  145. sc = rcu_dereference_protected(ns->schema,
  146. lockdep_is_held(&nsdata->lock));
  147. err = rhashtable_remove_fast(&nsdata->namespaces, &ns->head,
  148. rht_ns_params);
  149. if (err)
  150. goto out_unlock;
  151. if (sc)
  152. rcu_assign_pointer(sc->ns, NULL);
  153. ioam6_ns_release(ns);
  154. out_unlock:
  155. mutex_unlock(&nsdata->lock);
  156. return err;
  157. }
  158. static int __ioam6_genl_dumpns_element(struct ioam6_namespace *ns,
  159. u32 portid,
  160. u32 seq,
  161. u32 flags,
  162. struct sk_buff *skb,
  163. u8 cmd)
  164. {
  165. struct ioam6_schema *sc;
  166. u64 data64;
  167. u32 data32;
  168. void *hdr;
  169. hdr = genlmsg_put(skb, portid, seq, &ioam6_genl_family, flags, cmd);
  170. if (!hdr)
  171. return -ENOMEM;
  172. data32 = be32_to_cpu(ns->data);
  173. data64 = be64_to_cpu(ns->data_wide);
  174. if (nla_put_u16(skb, IOAM6_ATTR_NS_ID, be16_to_cpu(ns->id)) ||
  175. (data32 != IOAM6_U32_UNAVAILABLE &&
  176. nla_put_u32(skb, IOAM6_ATTR_NS_DATA, data32)) ||
  177. (data64 != IOAM6_U64_UNAVAILABLE &&
  178. nla_put_u64_64bit(skb, IOAM6_ATTR_NS_DATA_WIDE,
  179. data64, IOAM6_ATTR_PAD)))
  180. goto nla_put_failure;
  181. rcu_read_lock();
  182. sc = rcu_dereference(ns->schema);
  183. if (sc && nla_put_u32(skb, IOAM6_ATTR_SC_ID, sc->id)) {
  184. rcu_read_unlock();
  185. goto nla_put_failure;
  186. }
  187. rcu_read_unlock();
  188. genlmsg_end(skb, hdr);
  189. return 0;
  190. nla_put_failure:
  191. genlmsg_cancel(skb, hdr);
  192. return -EMSGSIZE;
  193. }
  194. static int ioam6_genl_dumpns_start(struct netlink_callback *cb)
  195. {
  196. struct ioam6_pernet_data *nsdata = ioam6_pernet(sock_net(cb->skb->sk));
  197. struct rhashtable_iter *iter = (struct rhashtable_iter *)cb->args[0];
  198. if (!iter) {
  199. iter = kmalloc(sizeof(*iter), GFP_KERNEL);
  200. if (!iter)
  201. return -ENOMEM;
  202. cb->args[0] = (long)iter;
  203. }
  204. rhashtable_walk_enter(&nsdata->namespaces, iter);
  205. return 0;
  206. }
  207. static int ioam6_genl_dumpns_done(struct netlink_callback *cb)
  208. {
  209. struct rhashtable_iter *iter = (struct rhashtable_iter *)cb->args[0];
  210. rhashtable_walk_exit(iter);
  211. kfree(iter);
  212. return 0;
  213. }
  214. static int ioam6_genl_dumpns(struct sk_buff *skb, struct netlink_callback *cb)
  215. {
  216. struct rhashtable_iter *iter;
  217. struct ioam6_namespace *ns;
  218. int err;
  219. iter = (struct rhashtable_iter *)cb->args[0];
  220. rhashtable_walk_start(iter);
  221. for (;;) {
  222. ns = rhashtable_walk_next(iter);
  223. if (IS_ERR(ns)) {
  224. if (PTR_ERR(ns) == -EAGAIN)
  225. continue;
  226. err = PTR_ERR(ns);
  227. goto done;
  228. } else if (!ns) {
  229. break;
  230. }
  231. err = __ioam6_genl_dumpns_element(ns,
  232. NETLINK_CB(cb->skb).portid,
  233. cb->nlh->nlmsg_seq,
  234. NLM_F_MULTI,
  235. skb,
  236. IOAM6_CMD_DUMP_NAMESPACES);
  237. if (err)
  238. goto done;
  239. }
  240. err = skb->len;
  241. done:
  242. rhashtable_walk_stop(iter);
  243. return err;
  244. }
  245. static int ioam6_genl_addsc(struct sk_buff *skb, struct genl_info *info)
  246. {
  247. struct ioam6_pernet_data *nsdata;
  248. int len, len_aligned, err;
  249. struct ioam6_schema *sc;
  250. u32 id;
  251. if (!info->attrs[IOAM6_ATTR_SC_ID] || !info->attrs[IOAM6_ATTR_SC_DATA])
  252. return -EINVAL;
  253. id = nla_get_u32(info->attrs[IOAM6_ATTR_SC_ID]);
  254. nsdata = ioam6_pernet(genl_info_net(info));
  255. mutex_lock(&nsdata->lock);
  256. sc = rhashtable_lookup_fast(&nsdata->schemas, &id, rht_sc_params);
  257. if (sc) {
  258. err = -EEXIST;
  259. goto out_unlock;
  260. }
  261. len = nla_len(info->attrs[IOAM6_ATTR_SC_DATA]);
  262. len_aligned = ALIGN(len, 4);
  263. sc = kzalloc(sizeof(*sc) + len_aligned, GFP_KERNEL);
  264. if (!sc) {
  265. err = -ENOMEM;
  266. goto out_unlock;
  267. }
  268. sc->id = id;
  269. sc->len = len_aligned;
  270. sc->hdr = cpu_to_be32(sc->id | ((u8)(sc->len / 4) << 24));
  271. nla_memcpy(sc->data, info->attrs[IOAM6_ATTR_SC_DATA], len);
  272. err = rhashtable_lookup_insert_fast(&nsdata->schemas, &sc->head,
  273. rht_sc_params);
  274. if (err)
  275. goto free_sc;
  276. out_unlock:
  277. mutex_unlock(&nsdata->lock);
  278. return err;
  279. free_sc:
  280. kfree(sc);
  281. goto out_unlock;
  282. }
  283. static int ioam6_genl_delsc(struct sk_buff *skb, struct genl_info *info)
  284. {
  285. struct ioam6_pernet_data *nsdata;
  286. struct ioam6_namespace *ns;
  287. struct ioam6_schema *sc;
  288. int err;
  289. u32 id;
  290. if (!info->attrs[IOAM6_ATTR_SC_ID])
  291. return -EINVAL;
  292. id = nla_get_u32(info->attrs[IOAM6_ATTR_SC_ID]);
  293. nsdata = ioam6_pernet(genl_info_net(info));
  294. mutex_lock(&nsdata->lock);
  295. sc = rhashtable_lookup_fast(&nsdata->schemas, &id, rht_sc_params);
  296. if (!sc) {
  297. err = -ENOENT;
  298. goto out_unlock;
  299. }
  300. ns = rcu_dereference_protected(sc->ns, lockdep_is_held(&nsdata->lock));
  301. err = rhashtable_remove_fast(&nsdata->schemas, &sc->head,
  302. rht_sc_params);
  303. if (err)
  304. goto out_unlock;
  305. if (ns)
  306. rcu_assign_pointer(ns->schema, NULL);
  307. ioam6_sc_release(sc);
  308. out_unlock:
  309. mutex_unlock(&nsdata->lock);
  310. return err;
  311. }
  312. static int __ioam6_genl_dumpsc_element(struct ioam6_schema *sc,
  313. u32 portid, u32 seq, u32 flags,
  314. struct sk_buff *skb, u8 cmd)
  315. {
  316. struct ioam6_namespace *ns;
  317. void *hdr;
  318. hdr = genlmsg_put(skb, portid, seq, &ioam6_genl_family, flags, cmd);
  319. if (!hdr)
  320. return -ENOMEM;
  321. if (nla_put_u32(skb, IOAM6_ATTR_SC_ID, sc->id) ||
  322. nla_put(skb, IOAM6_ATTR_SC_DATA, sc->len, sc->data))
  323. goto nla_put_failure;
  324. rcu_read_lock();
  325. ns = rcu_dereference(sc->ns);
  326. if (ns && nla_put_u16(skb, IOAM6_ATTR_NS_ID, be16_to_cpu(ns->id))) {
  327. rcu_read_unlock();
  328. goto nla_put_failure;
  329. }
  330. rcu_read_unlock();
  331. genlmsg_end(skb, hdr);
  332. return 0;
  333. nla_put_failure:
  334. genlmsg_cancel(skb, hdr);
  335. return -EMSGSIZE;
  336. }
  337. static int ioam6_genl_dumpsc_start(struct netlink_callback *cb)
  338. {
  339. struct ioam6_pernet_data *nsdata = ioam6_pernet(sock_net(cb->skb->sk));
  340. struct rhashtable_iter *iter = (struct rhashtable_iter *)cb->args[0];
  341. if (!iter) {
  342. iter = kmalloc(sizeof(*iter), GFP_KERNEL);
  343. if (!iter)
  344. return -ENOMEM;
  345. cb->args[0] = (long)iter;
  346. }
  347. rhashtable_walk_enter(&nsdata->schemas, iter);
  348. return 0;
  349. }
  350. static int ioam6_genl_dumpsc_done(struct netlink_callback *cb)
  351. {
  352. struct rhashtable_iter *iter = (struct rhashtable_iter *)cb->args[0];
  353. rhashtable_walk_exit(iter);
  354. kfree(iter);
  355. return 0;
  356. }
  357. static int ioam6_genl_dumpsc(struct sk_buff *skb, struct netlink_callback *cb)
  358. {
  359. struct rhashtable_iter *iter;
  360. struct ioam6_schema *sc;
  361. int err;
  362. iter = (struct rhashtable_iter *)cb->args[0];
  363. rhashtable_walk_start(iter);
  364. for (;;) {
  365. sc = rhashtable_walk_next(iter);
  366. if (IS_ERR(sc)) {
  367. if (PTR_ERR(sc) == -EAGAIN)
  368. continue;
  369. err = PTR_ERR(sc);
  370. goto done;
  371. } else if (!sc) {
  372. break;
  373. }
  374. err = __ioam6_genl_dumpsc_element(sc,
  375. NETLINK_CB(cb->skb).portid,
  376. cb->nlh->nlmsg_seq,
  377. NLM_F_MULTI,
  378. skb,
  379. IOAM6_CMD_DUMP_SCHEMAS);
  380. if (err)
  381. goto done;
  382. }
  383. err = skb->len;
  384. done:
  385. rhashtable_walk_stop(iter);
  386. return err;
  387. }
  388. static int ioam6_genl_ns_set_schema(struct sk_buff *skb, struct genl_info *info)
  389. {
  390. struct ioam6_namespace *ns, *ns_ref;
  391. struct ioam6_schema *sc, *sc_ref;
  392. struct ioam6_pernet_data *nsdata;
  393. __be16 ns_id;
  394. u32 sc_id;
  395. int err;
  396. if (!info->attrs[IOAM6_ATTR_NS_ID] ||
  397. (!info->attrs[IOAM6_ATTR_SC_ID] &&
  398. !info->attrs[IOAM6_ATTR_SC_NONE]))
  399. return -EINVAL;
  400. ns_id = cpu_to_be16(nla_get_u16(info->attrs[IOAM6_ATTR_NS_ID]));
  401. nsdata = ioam6_pernet(genl_info_net(info));
  402. mutex_lock(&nsdata->lock);
  403. ns = rhashtable_lookup_fast(&nsdata->namespaces, &ns_id, rht_ns_params);
  404. if (!ns) {
  405. err = -ENOENT;
  406. goto out_unlock;
  407. }
  408. if (info->attrs[IOAM6_ATTR_SC_NONE]) {
  409. sc = NULL;
  410. } else {
  411. sc_id = nla_get_u32(info->attrs[IOAM6_ATTR_SC_ID]);
  412. sc = rhashtable_lookup_fast(&nsdata->schemas, &sc_id,
  413. rht_sc_params);
  414. if (!sc) {
  415. err = -ENOENT;
  416. goto out_unlock;
  417. }
  418. }
  419. sc_ref = rcu_dereference_protected(ns->schema,
  420. lockdep_is_held(&nsdata->lock));
  421. if (sc_ref)
  422. rcu_assign_pointer(sc_ref->ns, NULL);
  423. rcu_assign_pointer(ns->schema, sc);
  424. if (sc) {
  425. ns_ref = rcu_dereference_protected(sc->ns,
  426. lockdep_is_held(&nsdata->lock));
  427. if (ns_ref)
  428. rcu_assign_pointer(ns_ref->schema, NULL);
  429. rcu_assign_pointer(sc->ns, ns);
  430. }
  431. err = 0;
  432. out_unlock:
  433. mutex_unlock(&nsdata->lock);
  434. return err;
  435. }
  436. static const struct genl_ops ioam6_genl_ops[] = {
  437. {
  438. .cmd = IOAM6_CMD_ADD_NAMESPACE,
  439. .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP,
  440. .doit = ioam6_genl_addns,
  441. .flags = GENL_ADMIN_PERM,
  442. .policy = ioam6_genl_policy_addns,
  443. .maxattr = ARRAY_SIZE(ioam6_genl_policy_addns) - 1,
  444. },
  445. {
  446. .cmd = IOAM6_CMD_DEL_NAMESPACE,
  447. .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP,
  448. .doit = ioam6_genl_delns,
  449. .flags = GENL_ADMIN_PERM,
  450. .policy = ioam6_genl_policy_delns,
  451. .maxattr = ARRAY_SIZE(ioam6_genl_policy_delns) - 1,
  452. },
  453. {
  454. .cmd = IOAM6_CMD_DUMP_NAMESPACES,
  455. .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP,
  456. .start = ioam6_genl_dumpns_start,
  457. .dumpit = ioam6_genl_dumpns,
  458. .done = ioam6_genl_dumpns_done,
  459. .flags = GENL_ADMIN_PERM,
  460. },
  461. {
  462. .cmd = IOAM6_CMD_ADD_SCHEMA,
  463. .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP,
  464. .doit = ioam6_genl_addsc,
  465. .flags = GENL_ADMIN_PERM,
  466. .policy = ioam6_genl_policy_addsc,
  467. .maxattr = ARRAY_SIZE(ioam6_genl_policy_addsc) - 1,
  468. },
  469. {
  470. .cmd = IOAM6_CMD_DEL_SCHEMA,
  471. .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP,
  472. .doit = ioam6_genl_delsc,
  473. .flags = GENL_ADMIN_PERM,
  474. .policy = ioam6_genl_policy_delsc,
  475. .maxattr = ARRAY_SIZE(ioam6_genl_policy_delsc) - 1,
  476. },
  477. {
  478. .cmd = IOAM6_CMD_DUMP_SCHEMAS,
  479. .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP,
  480. .start = ioam6_genl_dumpsc_start,
  481. .dumpit = ioam6_genl_dumpsc,
  482. .done = ioam6_genl_dumpsc_done,
  483. .flags = GENL_ADMIN_PERM,
  484. },
  485. {
  486. .cmd = IOAM6_CMD_NS_SET_SCHEMA,
  487. .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP,
  488. .doit = ioam6_genl_ns_set_schema,
  489. .flags = GENL_ADMIN_PERM,
  490. .policy = ioam6_genl_policy_ns_sc,
  491. .maxattr = ARRAY_SIZE(ioam6_genl_policy_ns_sc) - 1,
  492. },
  493. };
  494. #define IOAM6_GENL_EV_GRP_OFFSET 0
  495. static const struct genl_multicast_group ioam6_mcgrps[] = {
  496. [IOAM6_GENL_EV_GRP_OFFSET] = { .name = IOAM6_GENL_EV_GRP_NAME,
  497. .flags = GENL_MCAST_CAP_NET_ADMIN },
  498. };
  499. static int ioam6_event_put_trace(struct sk_buff *skb,
  500. struct ioam6_trace_hdr *trace,
  501. unsigned int len)
  502. {
  503. if (nla_put_u16(skb, IOAM6_EVENT_ATTR_TRACE_NAMESPACE,
  504. be16_to_cpu(trace->namespace_id)) ||
  505. nla_put_u8(skb, IOAM6_EVENT_ATTR_TRACE_NODELEN, trace->nodelen) ||
  506. nla_put_u32(skb, IOAM6_EVENT_ATTR_TRACE_TYPE,
  507. be32_to_cpu(trace->type_be32)) ||
  508. nla_put(skb, IOAM6_EVENT_ATTR_TRACE_DATA,
  509. len - sizeof(struct ioam6_trace_hdr) - trace->remlen * 4,
  510. trace->data + trace->remlen * 4))
  511. return 1;
  512. return 0;
  513. }
  514. void ioam6_event(enum ioam6_event_type type, struct net *net, gfp_t gfp,
  515. void *opt, unsigned int opt_len)
  516. {
  517. struct nlmsghdr *nlh;
  518. struct sk_buff *skb;
  519. if (!genl_has_listeners(&ioam6_genl_family, net,
  520. IOAM6_GENL_EV_GRP_OFFSET))
  521. return;
  522. skb = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp);
  523. if (!skb)
  524. return;
  525. nlh = genlmsg_put(skb, 0, 0, &ioam6_genl_family, 0, type);
  526. if (!nlh)
  527. goto nla_put_failure;
  528. switch (type) {
  529. case IOAM6_EVENT_UNSPEC:
  530. WARN_ON_ONCE(1);
  531. break;
  532. case IOAM6_EVENT_TRACE:
  533. if (ioam6_event_put_trace(skb, (struct ioam6_trace_hdr *)opt,
  534. opt_len))
  535. goto nla_put_failure;
  536. break;
  537. }
  538. genlmsg_end(skb, nlh);
  539. genlmsg_multicast_netns(&ioam6_genl_family, net, skb, 0,
  540. IOAM6_GENL_EV_GRP_OFFSET, gfp);
  541. return;
  542. nla_put_failure:
  543. nlmsg_free(skb);
  544. }
  545. static struct genl_family ioam6_genl_family __ro_after_init = {
  546. .name = IOAM6_GENL_NAME,
  547. .version = IOAM6_GENL_VERSION,
  548. .netnsok = true,
  549. .parallel_ops = true,
  550. .ops = ioam6_genl_ops,
  551. .n_ops = ARRAY_SIZE(ioam6_genl_ops),
  552. .resv_start_op = IOAM6_CMD_NS_SET_SCHEMA + 1,
  553. .mcgrps = ioam6_mcgrps,
  554. .n_mcgrps = ARRAY_SIZE(ioam6_mcgrps),
  555. .module = THIS_MODULE,
  556. };
  557. struct ioam6_namespace *ioam6_namespace(struct net *net, __be16 id)
  558. {
  559. struct ioam6_pernet_data *nsdata = ioam6_pernet(net);
  560. return rhashtable_lookup_fast(&nsdata->namespaces, &id, rht_ns_params);
  561. }
  562. static void __ioam6_fill_trace_data(struct sk_buff *skb,
  563. struct ioam6_namespace *ns,
  564. struct ioam6_trace_hdr *trace,
  565. struct ioam6_schema *sc,
  566. u8 sclen, bool is_input)
  567. {
  568. struct timespec64 ts;
  569. ktime_t tstamp;
  570. u64 raw64;
  571. u32 raw32;
  572. u16 raw16;
  573. u8 *data;
  574. u8 byte;
  575. data = trace->data + trace->remlen * 4 - trace->nodelen * 4 - sclen * 4;
  576. /* hop_lim and node_id */
  577. if (trace->type.bit0) {
  578. byte = ipv6_hdr(skb)->hop_limit;
  579. if (is_input)
  580. byte--;
  581. raw32 = dev_net(skb_dst(skb)->dev)->ipv6.sysctl.ioam6_id;
  582. *(__be32 *)data = cpu_to_be32((byte << 24) | raw32);
  583. data += sizeof(__be32);
  584. }
  585. /* ingress_if_id and egress_if_id */
  586. if (trace->type.bit1) {
  587. if (!skb->dev)
  588. raw16 = IOAM6_U16_UNAVAILABLE;
  589. else
  590. raw16 = (__force u16)READ_ONCE(__in6_dev_get(skb->dev)->cnf.ioam6_id);
  591. *(__be16 *)data = cpu_to_be16(raw16);
  592. data += sizeof(__be16);
  593. if (skb_dst(skb)->dev->flags & IFF_LOOPBACK)
  594. raw16 = IOAM6_U16_UNAVAILABLE;
  595. else
  596. raw16 = (__force u16)READ_ONCE(__in6_dev_get(skb_dst(skb)->dev)->cnf.ioam6_id);
  597. *(__be16 *)data = cpu_to_be16(raw16);
  598. data += sizeof(__be16);
  599. }
  600. /* timestamp seconds */
  601. if (trace->type.bit2) {
  602. if (!skb->dev) {
  603. *(__be32 *)data = cpu_to_be32(IOAM6_U32_UNAVAILABLE);
  604. } else {
  605. tstamp = skb_tstamp_cond(skb, true);
  606. ts = ktime_to_timespec64(tstamp);
  607. *(__be32 *)data = cpu_to_be32((u32)ts.tv_sec);
  608. }
  609. data += sizeof(__be32);
  610. }
  611. /* timestamp subseconds */
  612. if (trace->type.bit3) {
  613. if (!skb->dev) {
  614. *(__be32 *)data = cpu_to_be32(IOAM6_U32_UNAVAILABLE);
  615. } else {
  616. if (!trace->type.bit2) {
  617. tstamp = skb_tstamp_cond(skb, true);
  618. ts = ktime_to_timespec64(tstamp);
  619. }
  620. *(__be32 *)data = cpu_to_be32((u32)(ts.tv_nsec / NSEC_PER_USEC));
  621. }
  622. data += sizeof(__be32);
  623. }
  624. /* transit delay */
  625. if (trace->type.bit4) {
  626. *(__be32 *)data = cpu_to_be32(IOAM6_U32_UNAVAILABLE);
  627. data += sizeof(__be32);
  628. }
  629. /* namespace data */
  630. if (trace->type.bit5) {
  631. *(__be32 *)data = ns->data;
  632. data += sizeof(__be32);
  633. }
  634. /* queue depth */
  635. if (trace->type.bit6) {
  636. struct netdev_queue *queue;
  637. struct Qdisc *qdisc;
  638. __u32 qlen, backlog;
  639. if (skb_dst(skb)->dev->flags & IFF_LOOPBACK) {
  640. *(__be32 *)data = cpu_to_be32(IOAM6_U32_UNAVAILABLE);
  641. } else {
  642. queue = skb_get_tx_queue(skb_dst(skb)->dev, skb);
  643. qdisc = rcu_dereference(queue->qdisc);
  644. qdisc_qstats_qlen_backlog(qdisc, &qlen, &backlog);
  645. *(__be32 *)data = cpu_to_be32(backlog);
  646. }
  647. data += sizeof(__be32);
  648. }
  649. /* checksum complement */
  650. if (trace->type.bit7) {
  651. *(__be32 *)data = cpu_to_be32(IOAM6_U32_UNAVAILABLE);
  652. data += sizeof(__be32);
  653. }
  654. /* hop_lim and node_id (wide) */
  655. if (trace->type.bit8) {
  656. byte = ipv6_hdr(skb)->hop_limit;
  657. if (is_input)
  658. byte--;
  659. raw64 = dev_net(skb_dst(skb)->dev)->ipv6.sysctl.ioam6_id_wide;
  660. *(__be64 *)data = cpu_to_be64(((u64)byte << 56) | raw64);
  661. data += sizeof(__be64);
  662. }
  663. /* ingress_if_id and egress_if_id (wide) */
  664. if (trace->type.bit9) {
  665. if (!skb->dev)
  666. raw32 = IOAM6_U32_UNAVAILABLE;
  667. else
  668. raw32 = READ_ONCE(__in6_dev_get(skb->dev)->cnf.ioam6_id_wide);
  669. *(__be32 *)data = cpu_to_be32(raw32);
  670. data += sizeof(__be32);
  671. if (skb_dst(skb)->dev->flags & IFF_LOOPBACK)
  672. raw32 = IOAM6_U32_UNAVAILABLE;
  673. else
  674. raw32 = READ_ONCE(__in6_dev_get(skb_dst(skb)->dev)->cnf.ioam6_id_wide);
  675. *(__be32 *)data = cpu_to_be32(raw32);
  676. data += sizeof(__be32);
  677. }
  678. /* namespace data (wide) */
  679. if (trace->type.bit10) {
  680. *(__be64 *)data = ns->data_wide;
  681. data += sizeof(__be64);
  682. }
  683. /* buffer occupancy */
  684. if (trace->type.bit11) {
  685. *(__be32 *)data = cpu_to_be32(IOAM6_U32_UNAVAILABLE);
  686. data += sizeof(__be32);
  687. }
  688. /* bit12 undefined: filled with empty value */
  689. if (trace->type.bit12) {
  690. *(__be32 *)data = cpu_to_be32(IOAM6_U32_UNAVAILABLE);
  691. data += sizeof(__be32);
  692. }
  693. /* bit13 undefined: filled with empty value */
  694. if (trace->type.bit13) {
  695. *(__be32 *)data = cpu_to_be32(IOAM6_U32_UNAVAILABLE);
  696. data += sizeof(__be32);
  697. }
  698. /* bit14 undefined: filled with empty value */
  699. if (trace->type.bit14) {
  700. *(__be32 *)data = cpu_to_be32(IOAM6_U32_UNAVAILABLE);
  701. data += sizeof(__be32);
  702. }
  703. /* bit15 undefined: filled with empty value */
  704. if (trace->type.bit15) {
  705. *(__be32 *)data = cpu_to_be32(IOAM6_U32_UNAVAILABLE);
  706. data += sizeof(__be32);
  707. }
  708. /* bit16 undefined: filled with empty value */
  709. if (trace->type.bit16) {
  710. *(__be32 *)data = cpu_to_be32(IOAM6_U32_UNAVAILABLE);
  711. data += sizeof(__be32);
  712. }
  713. /* bit17 undefined: filled with empty value */
  714. if (trace->type.bit17) {
  715. *(__be32 *)data = cpu_to_be32(IOAM6_U32_UNAVAILABLE);
  716. data += sizeof(__be32);
  717. }
  718. /* bit18 undefined: filled with empty value */
  719. if (trace->type.bit18) {
  720. *(__be32 *)data = cpu_to_be32(IOAM6_U32_UNAVAILABLE);
  721. data += sizeof(__be32);
  722. }
  723. /* bit19 undefined: filled with empty value */
  724. if (trace->type.bit19) {
  725. *(__be32 *)data = cpu_to_be32(IOAM6_U32_UNAVAILABLE);
  726. data += sizeof(__be32);
  727. }
  728. /* bit20 undefined: filled with empty value */
  729. if (trace->type.bit20) {
  730. *(__be32 *)data = cpu_to_be32(IOAM6_U32_UNAVAILABLE);
  731. data += sizeof(__be32);
  732. }
  733. /* bit21 undefined: filled with empty value */
  734. if (trace->type.bit21) {
  735. *(__be32 *)data = cpu_to_be32(IOAM6_U32_UNAVAILABLE);
  736. data += sizeof(__be32);
  737. }
  738. /* opaque state snapshot */
  739. if (trace->type.bit22) {
  740. if (!sc) {
  741. *(__be32 *)data = cpu_to_be32(IOAM6_U32_UNAVAILABLE >> 8);
  742. } else {
  743. *(__be32 *)data = sc->hdr;
  744. data += sizeof(__be32);
  745. memcpy(data, sc->data, sc->len);
  746. }
  747. }
  748. }
  749. /* called with rcu_read_lock() */
  750. void ioam6_fill_trace_data(struct sk_buff *skb,
  751. struct ioam6_namespace *ns,
  752. struct ioam6_trace_hdr *trace,
  753. bool is_input)
  754. {
  755. struct ioam6_schema *sc;
  756. u8 sclen = 0;
  757. /* Skip if Overflow flag is set
  758. */
  759. if (trace->overflow)
  760. return;
  761. /* NodeLen does not include Opaque State Snapshot length. We need to
  762. * take it into account if the corresponding bit is set (bit 22) and
  763. * if the current IOAM namespace has an active schema attached to it
  764. */
  765. sc = rcu_dereference(ns->schema);
  766. if (trace->type.bit22) {
  767. sclen = sizeof_field(struct ioam6_schema, hdr) / 4;
  768. if (sc)
  769. sclen += sc->len / 4;
  770. }
  771. /* If there is no space remaining, we set the Overflow flag and we
  772. * skip without filling the trace
  773. */
  774. if (!trace->remlen || trace->remlen < trace->nodelen + sclen) {
  775. trace->overflow = 1;
  776. return;
  777. }
  778. __ioam6_fill_trace_data(skb, ns, trace, sc, sclen, is_input);
  779. trace->remlen -= trace->nodelen + sclen;
  780. }
  781. static int __net_init ioam6_net_init(struct net *net)
  782. {
  783. struct ioam6_pernet_data *nsdata;
  784. int err = -ENOMEM;
  785. nsdata = kzalloc(sizeof(*nsdata), GFP_KERNEL);
  786. if (!nsdata)
  787. goto out;
  788. mutex_init(&nsdata->lock);
  789. net->ipv6.ioam6_data = nsdata;
  790. err = rhashtable_init(&nsdata->namespaces, &rht_ns_params);
  791. if (err)
  792. goto free_nsdata;
  793. err = rhashtable_init(&nsdata->schemas, &rht_sc_params);
  794. if (err)
  795. goto free_rht_ns;
  796. out:
  797. return err;
  798. free_rht_ns:
  799. rhashtable_destroy(&nsdata->namespaces);
  800. free_nsdata:
  801. kfree(nsdata);
  802. net->ipv6.ioam6_data = NULL;
  803. goto out;
  804. }
  805. static void __net_exit ioam6_net_exit(struct net *net)
  806. {
  807. struct ioam6_pernet_data *nsdata = ioam6_pernet(net);
  808. rhashtable_free_and_destroy(&nsdata->namespaces, ioam6_free_ns, NULL);
  809. rhashtable_free_and_destroy(&nsdata->schemas, ioam6_free_sc, NULL);
  810. kfree(nsdata);
  811. }
  812. static struct pernet_operations ioam6_net_ops = {
  813. .init = ioam6_net_init,
  814. .exit = ioam6_net_exit,
  815. };
  816. int __init ioam6_init(void)
  817. {
  818. int err = register_pernet_subsys(&ioam6_net_ops);
  819. if (err)
  820. goto out;
  821. err = genl_register_family(&ioam6_genl_family);
  822. if (err)
  823. goto out_unregister_pernet_subsys;
  824. #ifdef CONFIG_IPV6_IOAM6_LWTUNNEL
  825. err = ioam6_iptunnel_init();
  826. if (err)
  827. goto out_unregister_genl;
  828. #endif
  829. pr_info("In-situ OAM (IOAM) with IPv6\n");
  830. out:
  831. return err;
  832. #ifdef CONFIG_IPV6_IOAM6_LWTUNNEL
  833. out_unregister_genl:
  834. genl_unregister_family(&ioam6_genl_family);
  835. #endif
  836. out_unregister_pernet_subsys:
  837. unregister_pernet_subsys(&ioam6_net_ops);
  838. goto out;
  839. }
  840. void ioam6_exit(void)
  841. {
  842. #ifdef CONFIG_IPV6_IOAM6_LWTUNNEL
  843. ioam6_iptunnel_exit();
  844. #endif
  845. genl_unregister_family(&ioam6_genl_family);
  846. unregister_pernet_subsys(&ioam6_net_ops);
  847. }