smc_clc.c 37 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323132413251326132713281329133013311332133313341335133613371338133913401341134213431344134513461347134813491350135113521353135413551356135713581359136013611362136313641365136613671368
  1. // SPDX-License-Identifier: GPL-2.0
  2. /*
  3. * Shared Memory Communications over RDMA (SMC-R) and RoCE
  4. *
  5. * CLC (connection layer control) handshake over initial TCP socket to
  6. * prepare for RDMA traffic
  7. *
  8. * Copyright IBM Corp. 2016, 2018
  9. *
  10. * Author(s): Ursula Braun <ubraun@linux.vnet.ibm.com>
  11. */
  12. #include <linux/in.h>
  13. #include <linux/inetdevice.h>
  14. #include <linux/if_ether.h>
  15. #include <linux/sched/signal.h>
  16. #include <linux/utsname.h>
  17. #include <linux/ctype.h>
  18. #include <net/addrconf.h>
  19. #include <net/sock.h>
  20. #include <net/tcp.h>
  21. #include "smc.h"
  22. #include "smc_core.h"
  23. #include "smc_clc.h"
  24. #include "smc_ib.h"
  25. #include "smc_ism.h"
  26. #include "smc_netlink.h"
  27. #define SMCR_CLC_ACCEPT_CONFIRM_LEN 68
  28. #define SMCD_CLC_ACCEPT_CONFIRM_LEN 48
  29. #define SMCD_CLC_ACCEPT_CONFIRM_LEN_V2 78
  30. #define SMCR_CLC_ACCEPT_CONFIRM_LEN_V2 108
  31. #define SMC_CLC_RECV_BUF_LEN 100
  32. /* eye catcher "SMCR" EBCDIC for CLC messages */
  33. static const char SMC_EYECATCHER[4] = {'\xe2', '\xd4', '\xc3', '\xd9'};
  34. /* eye catcher "SMCD" EBCDIC for CLC messages */
  35. static const char SMCD_EYECATCHER[4] = {'\xe2', '\xd4', '\xc3', '\xc4'};
  36. static u8 smc_hostname[SMC_MAX_HOSTNAME_LEN];
  37. struct smc_clc_eid_table {
  38. rwlock_t lock;
  39. struct list_head list;
  40. u8 ueid_cnt;
  41. u8 seid_enabled;
  42. };
  43. static struct smc_clc_eid_table smc_clc_eid_table;
  44. struct smc_clc_eid_entry {
  45. struct list_head list;
  46. u8 eid[SMC_MAX_EID_LEN];
  47. };
  48. /* The size of a user EID is 32 characters.
  49. * Valid characters should be (single-byte character set) A-Z, 0-9, '.' and '-'.
  50. * Blanks should only be used to pad to the expected size.
  51. * First character must be alphanumeric.
  52. */
  53. static bool smc_clc_ueid_valid(char *ueid)
  54. {
  55. char *end = ueid + SMC_MAX_EID_LEN;
  56. while (--end >= ueid && isspace(*end))
  57. ;
  58. if (end < ueid)
  59. return false;
  60. if (!isalnum(*ueid) || islower(*ueid))
  61. return false;
  62. while (ueid <= end) {
  63. if ((!isalnum(*ueid) || islower(*ueid)) && *ueid != '.' &&
  64. *ueid != '-')
  65. return false;
  66. ueid++;
  67. }
  68. return true;
  69. }
  70. static int smc_clc_ueid_add(char *ueid)
  71. {
  72. struct smc_clc_eid_entry *new_ueid, *tmp_ueid;
  73. int rc;
  74. if (!smc_clc_ueid_valid(ueid))
  75. return -EINVAL;
  76. /* add a new ueid entry to the ueid table if there isn't one */
  77. new_ueid = kzalloc(sizeof(*new_ueid), GFP_KERNEL);
  78. if (!new_ueid)
  79. return -ENOMEM;
  80. memcpy(new_ueid->eid, ueid, SMC_MAX_EID_LEN);
  81. write_lock(&smc_clc_eid_table.lock);
  82. if (smc_clc_eid_table.ueid_cnt >= SMC_MAX_UEID) {
  83. rc = -ERANGE;
  84. goto err_out;
  85. }
  86. list_for_each_entry(tmp_ueid, &smc_clc_eid_table.list, list) {
  87. if (!memcmp(tmp_ueid->eid, ueid, SMC_MAX_EID_LEN)) {
  88. rc = -EEXIST;
  89. goto err_out;
  90. }
  91. }
  92. list_add_tail(&new_ueid->list, &smc_clc_eid_table.list);
  93. smc_clc_eid_table.ueid_cnt++;
  94. write_unlock(&smc_clc_eid_table.lock);
  95. return 0;
  96. err_out:
  97. write_unlock(&smc_clc_eid_table.lock);
  98. kfree(new_ueid);
  99. return rc;
  100. }
  101. int smc_clc_ueid_count(void)
  102. {
  103. int count;
  104. read_lock(&smc_clc_eid_table.lock);
  105. count = smc_clc_eid_table.ueid_cnt;
  106. read_unlock(&smc_clc_eid_table.lock);
  107. return count;
  108. }
  109. int smc_nl_add_ueid(struct sk_buff *skb, struct genl_info *info)
  110. {
  111. struct nlattr *nla_ueid = info->attrs[SMC_NLA_EID_TABLE_ENTRY];
  112. char *ueid;
  113. if (!nla_ueid || nla_len(nla_ueid) != SMC_MAX_EID_LEN + 1)
  114. return -EINVAL;
  115. ueid = (char *)nla_data(nla_ueid);
  116. return smc_clc_ueid_add(ueid);
  117. }
  118. /* remove one or all ueid entries from the table */
  119. static int smc_clc_ueid_remove(char *ueid)
  120. {
  121. struct smc_clc_eid_entry *lst_ueid, *tmp_ueid;
  122. int rc = -ENOENT;
  123. /* remove table entry */
  124. write_lock(&smc_clc_eid_table.lock);
  125. list_for_each_entry_safe(lst_ueid, tmp_ueid, &smc_clc_eid_table.list,
  126. list) {
  127. if (!ueid || !memcmp(lst_ueid->eid, ueid, SMC_MAX_EID_LEN)) {
  128. list_del(&lst_ueid->list);
  129. smc_clc_eid_table.ueid_cnt--;
  130. kfree(lst_ueid);
  131. rc = 0;
  132. }
  133. }
  134. #if IS_ENABLED(CONFIG_S390)
  135. if (!rc && !smc_clc_eid_table.ueid_cnt) {
  136. smc_clc_eid_table.seid_enabled = 1;
  137. rc = -EAGAIN; /* indicate success and enabling of seid */
  138. }
  139. #endif
  140. write_unlock(&smc_clc_eid_table.lock);
  141. return rc;
  142. }
  143. int smc_nl_remove_ueid(struct sk_buff *skb, struct genl_info *info)
  144. {
  145. struct nlattr *nla_ueid = info->attrs[SMC_NLA_EID_TABLE_ENTRY];
  146. char *ueid;
  147. if (!nla_ueid || nla_len(nla_ueid) != SMC_MAX_EID_LEN + 1)
  148. return -EINVAL;
  149. ueid = (char *)nla_data(nla_ueid);
  150. return smc_clc_ueid_remove(ueid);
  151. }
  152. int smc_nl_flush_ueid(struct sk_buff *skb, struct genl_info *info)
  153. {
  154. smc_clc_ueid_remove(NULL);
  155. return 0;
  156. }
  157. static int smc_nl_ueid_dumpinfo(struct sk_buff *skb, u32 portid, u32 seq,
  158. u32 flags, char *ueid)
  159. {
  160. char ueid_str[SMC_MAX_EID_LEN + 1];
  161. void *hdr;
  162. hdr = genlmsg_put(skb, portid, seq, &smc_gen_nl_family,
  163. flags, SMC_NETLINK_DUMP_UEID);
  164. if (!hdr)
  165. return -ENOMEM;
  166. memcpy(ueid_str, ueid, SMC_MAX_EID_LEN);
  167. ueid_str[SMC_MAX_EID_LEN] = 0;
  168. if (nla_put_string(skb, SMC_NLA_EID_TABLE_ENTRY, ueid_str)) {
  169. genlmsg_cancel(skb, hdr);
  170. return -EMSGSIZE;
  171. }
  172. genlmsg_end(skb, hdr);
  173. return 0;
  174. }
  175. static int _smc_nl_ueid_dump(struct sk_buff *skb, u32 portid, u32 seq,
  176. int start_idx)
  177. {
  178. struct smc_clc_eid_entry *lst_ueid;
  179. int idx = 0;
  180. read_lock(&smc_clc_eid_table.lock);
  181. list_for_each_entry(lst_ueid, &smc_clc_eid_table.list, list) {
  182. if (idx++ < start_idx)
  183. continue;
  184. if (smc_nl_ueid_dumpinfo(skb, portid, seq, NLM_F_MULTI,
  185. lst_ueid->eid)) {
  186. --idx;
  187. break;
  188. }
  189. }
  190. read_unlock(&smc_clc_eid_table.lock);
  191. return idx;
  192. }
  193. int smc_nl_dump_ueid(struct sk_buff *skb, struct netlink_callback *cb)
  194. {
  195. struct smc_nl_dmp_ctx *cb_ctx = smc_nl_dmp_ctx(cb);
  196. int idx;
  197. idx = _smc_nl_ueid_dump(skb, NETLINK_CB(cb->skb).portid,
  198. cb->nlh->nlmsg_seq, cb_ctx->pos[0]);
  199. cb_ctx->pos[0] = idx;
  200. return skb->len;
  201. }
  202. int smc_nl_dump_seid(struct sk_buff *skb, struct netlink_callback *cb)
  203. {
  204. struct smc_nl_dmp_ctx *cb_ctx = smc_nl_dmp_ctx(cb);
  205. char seid_str[SMC_MAX_EID_LEN + 1];
  206. u8 seid_enabled;
  207. void *hdr;
  208. u8 *seid;
  209. if (cb_ctx->pos[0])
  210. return skb->len;
  211. hdr = genlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq,
  212. &smc_gen_nl_family, NLM_F_MULTI,
  213. SMC_NETLINK_DUMP_SEID);
  214. if (!hdr)
  215. return -ENOMEM;
  216. if (!smc_ism_is_v2_capable())
  217. goto end;
  218. smc_ism_get_system_eid(&seid);
  219. memcpy(seid_str, seid, SMC_MAX_EID_LEN);
  220. seid_str[SMC_MAX_EID_LEN] = 0;
  221. if (nla_put_string(skb, SMC_NLA_SEID_ENTRY, seid_str))
  222. goto err;
  223. read_lock(&smc_clc_eid_table.lock);
  224. seid_enabled = smc_clc_eid_table.seid_enabled;
  225. read_unlock(&smc_clc_eid_table.lock);
  226. if (nla_put_u8(skb, SMC_NLA_SEID_ENABLED, seid_enabled))
  227. goto err;
  228. end:
  229. genlmsg_end(skb, hdr);
  230. cb_ctx->pos[0]++;
  231. return skb->len;
  232. err:
  233. genlmsg_cancel(skb, hdr);
  234. return -EMSGSIZE;
  235. }
  236. int smc_nl_enable_seid(struct sk_buff *skb, struct genl_info *info)
  237. {
  238. #if IS_ENABLED(CONFIG_S390)
  239. write_lock(&smc_clc_eid_table.lock);
  240. smc_clc_eid_table.seid_enabled = 1;
  241. write_unlock(&smc_clc_eid_table.lock);
  242. return 0;
  243. #else
  244. return -EOPNOTSUPP;
  245. #endif
  246. }
  247. int smc_nl_disable_seid(struct sk_buff *skb, struct genl_info *info)
  248. {
  249. int rc = 0;
  250. #if IS_ENABLED(CONFIG_S390)
  251. write_lock(&smc_clc_eid_table.lock);
  252. if (!smc_clc_eid_table.ueid_cnt)
  253. rc = -ENOENT;
  254. else
  255. smc_clc_eid_table.seid_enabled = 0;
  256. write_unlock(&smc_clc_eid_table.lock);
  257. #else
  258. rc = -EOPNOTSUPP;
  259. #endif
  260. return rc;
  261. }
  262. static bool _smc_clc_match_ueid(u8 *peer_ueid)
  263. {
  264. struct smc_clc_eid_entry *tmp_ueid;
  265. list_for_each_entry(tmp_ueid, &smc_clc_eid_table.list, list) {
  266. if (!memcmp(tmp_ueid->eid, peer_ueid, SMC_MAX_EID_LEN))
  267. return true;
  268. }
  269. return false;
  270. }
  271. bool smc_clc_match_eid(u8 *negotiated_eid,
  272. struct smc_clc_v2_extension *smc_v2_ext,
  273. u8 *peer_eid, u8 *local_eid)
  274. {
  275. bool match = false;
  276. int i;
  277. negotiated_eid[0] = 0;
  278. read_lock(&smc_clc_eid_table.lock);
  279. if (peer_eid && local_eid &&
  280. smc_clc_eid_table.seid_enabled &&
  281. smc_v2_ext->hdr.flag.seid &&
  282. !memcmp(peer_eid, local_eid, SMC_MAX_EID_LEN)) {
  283. memcpy(negotiated_eid, peer_eid, SMC_MAX_EID_LEN);
  284. match = true;
  285. goto out;
  286. }
  287. for (i = 0; i < smc_v2_ext->hdr.eid_cnt; i++) {
  288. if (_smc_clc_match_ueid(smc_v2_ext->user_eids[i])) {
  289. memcpy(negotiated_eid, smc_v2_ext->user_eids[i],
  290. SMC_MAX_EID_LEN);
  291. match = true;
  292. goto out;
  293. }
  294. }
  295. out:
  296. read_unlock(&smc_clc_eid_table.lock);
  297. return match;
  298. }
  299. /* check arriving CLC proposal */
  300. static bool smc_clc_msg_prop_valid(struct smc_clc_msg_proposal *pclc)
  301. {
  302. struct smc_clc_msg_proposal_prefix *pclc_prfx;
  303. struct smc_clc_smcd_v2_extension *smcd_v2_ext;
  304. struct smc_clc_msg_hdr *hdr = &pclc->hdr;
  305. struct smc_clc_v2_extension *v2_ext;
  306. pclc_prfx = smc_clc_proposal_get_prefix(pclc);
  307. if (!pclc_prfx ||
  308. pclc_prfx->ipv6_prefixes_cnt > SMC_CLC_MAX_V6_PREFIX)
  309. return false;
  310. if (hdr->version == SMC_V1) {
  311. if (hdr->typev1 == SMC_TYPE_N)
  312. return false;
  313. if (ntohs(hdr->length) !=
  314. sizeof(*pclc) + ntohs(pclc->iparea_offset) +
  315. sizeof(*pclc_prfx) +
  316. pclc_prfx->ipv6_prefixes_cnt *
  317. sizeof(struct smc_clc_ipv6_prefix) +
  318. sizeof(struct smc_clc_msg_trail))
  319. return false;
  320. } else {
  321. v2_ext = smc_get_clc_v2_ext(pclc);
  322. if ((hdr->typev2 != SMC_TYPE_N &&
  323. (!v2_ext || v2_ext->hdr.eid_cnt > SMC_CLC_MAX_UEID)) ||
  324. (smcd_indicated(hdr->typev2) &&
  325. v2_ext->hdr.ism_gid_cnt > SMCD_CLC_MAX_V2_GID_ENTRIES))
  326. return false;
  327. if (ntohs(hdr->length) !=
  328. sizeof(*pclc) +
  329. sizeof(struct smc_clc_msg_smcd) +
  330. (hdr->typev1 != SMC_TYPE_N ?
  331. sizeof(*pclc_prfx) +
  332. pclc_prfx->ipv6_prefixes_cnt *
  333. sizeof(struct smc_clc_ipv6_prefix) : 0) +
  334. (hdr->typev2 != SMC_TYPE_N ?
  335. sizeof(*v2_ext) +
  336. v2_ext->hdr.eid_cnt * SMC_MAX_EID_LEN : 0) +
  337. (smcd_indicated(hdr->typev2) ?
  338. sizeof(*smcd_v2_ext) + v2_ext->hdr.ism_gid_cnt *
  339. sizeof(struct smc_clc_smcd_gid_chid) :
  340. 0) +
  341. sizeof(struct smc_clc_msg_trail))
  342. return false;
  343. }
  344. return true;
  345. }
  346. /* check arriving CLC accept or confirm */
  347. static bool
  348. smc_clc_msg_acc_conf_valid(struct smc_clc_msg_accept_confirm *clc)
  349. {
  350. struct smc_clc_msg_hdr *hdr = &clc->hdr;
  351. if (hdr->typev1 != SMC_TYPE_R && hdr->typev1 != SMC_TYPE_D)
  352. return false;
  353. if (hdr->version == SMC_V1) {
  354. if ((hdr->typev1 == SMC_TYPE_R &&
  355. ntohs(hdr->length) != SMCR_CLC_ACCEPT_CONFIRM_LEN) ||
  356. (hdr->typev1 == SMC_TYPE_D &&
  357. ntohs(hdr->length) != SMCD_CLC_ACCEPT_CONFIRM_LEN))
  358. return false;
  359. } else {
  360. if (hdr->typev1 == SMC_TYPE_D &&
  361. ntohs(hdr->length) < SMCD_CLC_ACCEPT_CONFIRM_LEN_V2)
  362. return false;
  363. if (hdr->typev1 == SMC_TYPE_R &&
  364. ntohs(hdr->length) < SMCR_CLC_ACCEPT_CONFIRM_LEN_V2)
  365. return false;
  366. }
  367. return true;
  368. }
  369. /* check arriving CLC decline */
  370. static bool
  371. smc_clc_msg_decl_valid(struct smc_clc_msg_decline *dclc)
  372. {
  373. struct smc_clc_msg_hdr *hdr = &dclc->hdr;
  374. if (hdr->typev1 != SMC_TYPE_R && hdr->typev1 != SMC_TYPE_D)
  375. return false;
  376. if (hdr->version == SMC_V1) {
  377. if (ntohs(hdr->length) != sizeof(struct smc_clc_msg_decline))
  378. return false;
  379. } else {
  380. if (ntohs(hdr->length) != sizeof(struct smc_clc_msg_decline_v2))
  381. return false;
  382. }
  383. return true;
  384. }
  385. static int smc_clc_fill_fce_v2x(struct smc_clc_first_contact_ext_v2x *fce_v2x,
  386. struct smc_init_info *ini)
  387. {
  388. int ret = sizeof(*fce_v2x);
  389. memset(fce_v2x, 0, sizeof(*fce_v2x));
  390. fce_v2x->fce_v2_base.os_type = SMC_CLC_OS_LINUX;
  391. fce_v2x->fce_v2_base.release = ini->release_nr;
  392. memcpy(fce_v2x->fce_v2_base.hostname,
  393. smc_hostname, sizeof(smc_hostname));
  394. if (ini->is_smcd && ini->release_nr < SMC_RELEASE_1) {
  395. ret = sizeof(struct smc_clc_first_contact_ext);
  396. goto out;
  397. }
  398. if (ini->release_nr >= SMC_RELEASE_1) {
  399. if (!ini->is_smcd) {
  400. fce_v2x->max_conns = ini->max_conns;
  401. fce_v2x->max_links = ini->max_links;
  402. }
  403. fce_v2x->feature_mask = htons(ini->feature_mask);
  404. }
  405. out:
  406. return ret;
  407. }
  408. /* check if received message has a correct header length and contains valid
  409. * heading and trailing eyecatchers
  410. */
  411. static bool smc_clc_msg_hdr_valid(struct smc_clc_msg_hdr *clcm, bool check_trl)
  412. {
  413. struct smc_clc_msg_accept_confirm *clc;
  414. struct smc_clc_msg_proposal *pclc;
  415. struct smc_clc_msg_decline *dclc;
  416. struct smc_clc_msg_trail *trl;
  417. if (memcmp(clcm->eyecatcher, SMC_EYECATCHER, sizeof(SMC_EYECATCHER)) &&
  418. memcmp(clcm->eyecatcher, SMCD_EYECATCHER, sizeof(SMCD_EYECATCHER)))
  419. return false;
  420. switch (clcm->type) {
  421. case SMC_CLC_PROPOSAL:
  422. pclc = (struct smc_clc_msg_proposal *)clcm;
  423. if (!smc_clc_msg_prop_valid(pclc))
  424. return false;
  425. trl = (struct smc_clc_msg_trail *)
  426. ((u8 *)pclc + ntohs(pclc->hdr.length) - sizeof(*trl));
  427. break;
  428. case SMC_CLC_ACCEPT:
  429. case SMC_CLC_CONFIRM:
  430. clc = (struct smc_clc_msg_accept_confirm *)clcm;
  431. if (!smc_clc_msg_acc_conf_valid(clc))
  432. return false;
  433. trl = (struct smc_clc_msg_trail *)
  434. ((u8 *)clc + ntohs(clc->hdr.length) - sizeof(*trl));
  435. break;
  436. case SMC_CLC_DECLINE:
  437. dclc = (struct smc_clc_msg_decline *)clcm;
  438. if (!smc_clc_msg_decl_valid(dclc))
  439. return false;
  440. check_trl = false;
  441. break;
  442. default:
  443. return false;
  444. }
  445. if (check_trl &&
  446. memcmp(trl->eyecatcher, SMC_EYECATCHER, sizeof(SMC_EYECATCHER)) &&
  447. memcmp(trl->eyecatcher, SMCD_EYECATCHER, sizeof(SMCD_EYECATCHER)))
  448. return false;
  449. return true;
  450. }
  451. /* find ipv4 addr on device and get the prefix len, fill CLC proposal msg */
  452. static int smc_clc_prfx_set4_rcu(struct dst_entry *dst, __be32 ipv4,
  453. struct smc_clc_msg_proposal_prefix *prop)
  454. {
  455. struct in_device *in_dev = __in_dev_get_rcu(dst->dev);
  456. const struct in_ifaddr *ifa;
  457. if (!in_dev)
  458. return -ENODEV;
  459. in_dev_for_each_ifa_rcu(ifa, in_dev) {
  460. if (!inet_ifa_match(ipv4, ifa))
  461. continue;
  462. prop->prefix_len = inet_mask_len(ifa->ifa_mask);
  463. prop->outgoing_subnet = ifa->ifa_address & ifa->ifa_mask;
  464. /* prop->ipv6_prefixes_cnt = 0; already done by memset before */
  465. return 0;
  466. }
  467. return -ENOENT;
  468. }
  469. /* fill CLC proposal msg with ipv6 prefixes from device */
  470. static int smc_clc_prfx_set6_rcu(struct dst_entry *dst,
  471. struct smc_clc_msg_proposal_prefix *prop,
  472. struct smc_clc_ipv6_prefix *ipv6_prfx)
  473. {
  474. #if IS_ENABLED(CONFIG_IPV6)
  475. struct inet6_dev *in6_dev = __in6_dev_get(dst->dev);
  476. struct inet6_ifaddr *ifa;
  477. int cnt = 0;
  478. if (!in6_dev)
  479. return -ENODEV;
  480. /* use a maximum of 8 IPv6 prefixes from device */
  481. list_for_each_entry(ifa, &in6_dev->addr_list, if_list) {
  482. if (ipv6_addr_type(&ifa->addr) & IPV6_ADDR_LINKLOCAL)
  483. continue;
  484. ipv6_addr_prefix(&ipv6_prfx[cnt].prefix,
  485. &ifa->addr, ifa->prefix_len);
  486. ipv6_prfx[cnt].prefix_len = ifa->prefix_len;
  487. cnt++;
  488. if (cnt == SMC_CLC_MAX_V6_PREFIX)
  489. break;
  490. }
  491. prop->ipv6_prefixes_cnt = cnt;
  492. if (cnt)
  493. return 0;
  494. #endif
  495. return -ENOENT;
  496. }
  497. /* retrieve and set prefixes in CLC proposal msg */
  498. static int smc_clc_prfx_set(struct socket *clcsock,
  499. struct smc_clc_msg_proposal_prefix *prop,
  500. struct smc_clc_ipv6_prefix *ipv6_prfx)
  501. {
  502. struct dst_entry *dst = sk_dst_get(clcsock->sk);
  503. struct sockaddr_storage addrs;
  504. struct sockaddr_in6 *addr6;
  505. struct sockaddr_in *addr;
  506. int rc = -ENOENT;
  507. if (!dst) {
  508. rc = -ENOTCONN;
  509. goto out;
  510. }
  511. if (!dst->dev) {
  512. rc = -ENODEV;
  513. goto out_rel;
  514. }
  515. /* get address to which the internal TCP socket is bound */
  516. if (kernel_getsockname(clcsock, (struct sockaddr *)&addrs) < 0)
  517. goto out_rel;
  518. /* analyze IP specific data of net_device belonging to TCP socket */
  519. addr6 = (struct sockaddr_in6 *)&addrs;
  520. rcu_read_lock();
  521. if (addrs.ss_family == PF_INET) {
  522. /* IPv4 */
  523. addr = (struct sockaddr_in *)&addrs;
  524. rc = smc_clc_prfx_set4_rcu(dst, addr->sin_addr.s_addr, prop);
  525. } else if (ipv6_addr_v4mapped(&addr6->sin6_addr)) {
  526. /* mapped IPv4 address - peer is IPv4 only */
  527. rc = smc_clc_prfx_set4_rcu(dst, addr6->sin6_addr.s6_addr32[3],
  528. prop);
  529. } else {
  530. /* IPv6 */
  531. rc = smc_clc_prfx_set6_rcu(dst, prop, ipv6_prfx);
  532. }
  533. rcu_read_unlock();
  534. out_rel:
  535. dst_release(dst);
  536. out:
  537. return rc;
  538. }
  539. /* match ipv4 addrs of dev against addr in CLC proposal */
  540. static int smc_clc_prfx_match4_rcu(struct net_device *dev,
  541. struct smc_clc_msg_proposal_prefix *prop)
  542. {
  543. struct in_device *in_dev = __in_dev_get_rcu(dev);
  544. const struct in_ifaddr *ifa;
  545. if (!in_dev)
  546. return -ENODEV;
  547. in_dev_for_each_ifa_rcu(ifa, in_dev) {
  548. if (prop->prefix_len == inet_mask_len(ifa->ifa_mask) &&
  549. inet_ifa_match(prop->outgoing_subnet, ifa))
  550. return 0;
  551. }
  552. return -ENOENT;
  553. }
  554. /* match ipv6 addrs of dev against addrs in CLC proposal */
  555. static int smc_clc_prfx_match6_rcu(struct net_device *dev,
  556. struct smc_clc_msg_proposal_prefix *prop)
  557. {
  558. #if IS_ENABLED(CONFIG_IPV6)
  559. struct inet6_dev *in6_dev = __in6_dev_get(dev);
  560. struct smc_clc_ipv6_prefix *ipv6_prfx;
  561. struct inet6_ifaddr *ifa;
  562. int i, max;
  563. if (!in6_dev)
  564. return -ENODEV;
  565. /* ipv6 prefix list starts behind smc_clc_msg_proposal_prefix */
  566. ipv6_prfx = (struct smc_clc_ipv6_prefix *)((u8 *)prop + sizeof(*prop));
  567. max = min_t(u8, prop->ipv6_prefixes_cnt, SMC_CLC_MAX_V6_PREFIX);
  568. list_for_each_entry(ifa, &in6_dev->addr_list, if_list) {
  569. if (ipv6_addr_type(&ifa->addr) & IPV6_ADDR_LINKLOCAL)
  570. continue;
  571. for (i = 0; i < max; i++) {
  572. if (ifa->prefix_len == ipv6_prfx[i].prefix_len &&
  573. ipv6_prefix_equal(&ifa->addr, &ipv6_prfx[i].prefix,
  574. ifa->prefix_len))
  575. return 0;
  576. }
  577. }
  578. #endif
  579. return -ENOENT;
  580. }
  581. /* check if proposed prefixes match one of our device prefixes */
  582. int smc_clc_prfx_match(struct socket *clcsock,
  583. struct smc_clc_msg_proposal_prefix *prop)
  584. {
  585. struct dst_entry *dst = sk_dst_get(clcsock->sk);
  586. int rc;
  587. if (!dst) {
  588. rc = -ENOTCONN;
  589. goto out;
  590. }
  591. if (!dst->dev) {
  592. rc = -ENODEV;
  593. goto out_rel;
  594. }
  595. rcu_read_lock();
  596. if (!prop->ipv6_prefixes_cnt)
  597. rc = smc_clc_prfx_match4_rcu(dst->dev, prop);
  598. else
  599. rc = smc_clc_prfx_match6_rcu(dst->dev, prop);
  600. rcu_read_unlock();
  601. out_rel:
  602. dst_release(dst);
  603. out:
  604. return rc;
  605. }
  606. /* Wait for data on the tcp-socket, analyze received data
  607. * Returns:
  608. * 0 if success and it was not a decline that we received.
  609. * SMC_CLC_DECL_REPLY if decline received for fallback w/o another decl send.
  610. * clcsock error, -EINTR, -ECONNRESET, -EPROTO otherwise.
  611. */
  612. int smc_clc_wait_msg(struct smc_sock *smc, void *buf, int buflen,
  613. u8 expected_type, unsigned long timeout)
  614. {
  615. long rcvtimeo = smc->clcsock->sk->sk_rcvtimeo;
  616. struct sock *clc_sk = smc->clcsock->sk;
  617. struct smc_clc_msg_hdr *clcm = buf;
  618. struct msghdr msg = {NULL, 0};
  619. int reason_code = 0;
  620. struct kvec vec = {buf, buflen};
  621. int len, datlen, recvlen;
  622. bool check_trl = true;
  623. int krflags;
  624. /* peek the first few bytes to determine length of data to receive
  625. * so we don't consume any subsequent CLC message or payload data
  626. * in the TCP byte stream
  627. */
  628. /*
  629. * Caller must make sure that buflen is no less than
  630. * sizeof(struct smc_clc_msg_hdr)
  631. */
  632. krflags = MSG_PEEK | MSG_WAITALL;
  633. clc_sk->sk_rcvtimeo = timeout;
  634. iov_iter_kvec(&msg.msg_iter, ITER_DEST, &vec, 1,
  635. sizeof(struct smc_clc_msg_hdr));
  636. len = sock_recvmsg(smc->clcsock, &msg, krflags);
  637. if (signal_pending(current)) {
  638. reason_code = -EINTR;
  639. clc_sk->sk_err = EINTR;
  640. smc->sk.sk_err = EINTR;
  641. goto out;
  642. }
  643. if (clc_sk->sk_err) {
  644. reason_code = -clc_sk->sk_err;
  645. if (clc_sk->sk_err == EAGAIN &&
  646. expected_type == SMC_CLC_DECLINE)
  647. clc_sk->sk_err = 0; /* reset for fallback usage */
  648. else
  649. smc->sk.sk_err = clc_sk->sk_err;
  650. goto out;
  651. }
  652. if (!len) { /* peer has performed orderly shutdown */
  653. smc->sk.sk_err = ECONNRESET;
  654. reason_code = -ECONNRESET;
  655. goto out;
  656. }
  657. if (len < 0) {
  658. if (len != -EAGAIN || expected_type != SMC_CLC_DECLINE)
  659. smc->sk.sk_err = -len;
  660. reason_code = len;
  661. goto out;
  662. }
  663. datlen = ntohs(clcm->length);
  664. if ((len < sizeof(struct smc_clc_msg_hdr)) ||
  665. (clcm->version < SMC_V1) ||
  666. ((clcm->type != SMC_CLC_DECLINE) &&
  667. (clcm->type != expected_type))) {
  668. smc->sk.sk_err = EPROTO;
  669. reason_code = -EPROTO;
  670. goto out;
  671. }
  672. /* receive the complete CLC message */
  673. memset(&msg, 0, sizeof(struct msghdr));
  674. if (datlen > buflen) {
  675. check_trl = false;
  676. recvlen = buflen;
  677. } else {
  678. recvlen = datlen;
  679. }
  680. iov_iter_kvec(&msg.msg_iter, ITER_DEST, &vec, 1, recvlen);
  681. krflags = MSG_WAITALL;
  682. len = sock_recvmsg(smc->clcsock, &msg, krflags);
  683. if (len < recvlen || !smc_clc_msg_hdr_valid(clcm, check_trl)) {
  684. smc->sk.sk_err = EPROTO;
  685. reason_code = -EPROTO;
  686. goto out;
  687. }
  688. datlen -= len;
  689. while (datlen) {
  690. u8 tmp[SMC_CLC_RECV_BUF_LEN];
  691. vec.iov_base = &tmp;
  692. vec.iov_len = SMC_CLC_RECV_BUF_LEN;
  693. /* receive remaining proposal message */
  694. recvlen = datlen > SMC_CLC_RECV_BUF_LEN ?
  695. SMC_CLC_RECV_BUF_LEN : datlen;
  696. iov_iter_kvec(&msg.msg_iter, ITER_DEST, &vec, 1, recvlen);
  697. len = sock_recvmsg(smc->clcsock, &msg, krflags);
  698. if (len < recvlen) {
  699. smc->sk.sk_err = EPROTO;
  700. reason_code = -EPROTO;
  701. goto out;
  702. }
  703. datlen -= len;
  704. }
  705. if (clcm->type == SMC_CLC_DECLINE) {
  706. struct smc_clc_msg_decline *dclc;
  707. dclc = (struct smc_clc_msg_decline *)clcm;
  708. reason_code = SMC_CLC_DECL_PEERDECL;
  709. smc->peer_diagnosis = ntohl(dclc->peer_diagnosis);
  710. if (((struct smc_clc_msg_decline *)buf)->hdr.typev2 &
  711. SMC_FIRST_CONTACT_MASK) {
  712. smc->conn.lgr->sync_err = 1;
  713. smc_lgr_terminate_sched(smc->conn.lgr);
  714. }
  715. }
  716. out:
  717. clc_sk->sk_rcvtimeo = rcvtimeo;
  718. return reason_code;
  719. }
  720. /* send CLC DECLINE message across internal TCP socket */
  721. int smc_clc_send_decline(struct smc_sock *smc, u32 peer_diag_info, u8 version)
  722. {
  723. struct smc_clc_msg_decline *dclc_v1;
  724. struct smc_clc_msg_decline_v2 dclc;
  725. struct msghdr msg;
  726. int len, send_len;
  727. struct kvec vec;
  728. dclc_v1 = (struct smc_clc_msg_decline *)&dclc;
  729. memset(&dclc, 0, sizeof(dclc));
  730. memcpy(dclc.hdr.eyecatcher, SMC_EYECATCHER, sizeof(SMC_EYECATCHER));
  731. dclc.hdr.type = SMC_CLC_DECLINE;
  732. dclc.hdr.version = version;
  733. dclc.os_type = version == SMC_V1 ? 0 : SMC_CLC_OS_LINUX;
  734. dclc.hdr.typev2 = (peer_diag_info == SMC_CLC_DECL_SYNCERR) ?
  735. SMC_FIRST_CONTACT_MASK : 0;
  736. if ((!smc_conn_lgr_valid(&smc->conn) || !smc->conn.lgr->is_smcd) &&
  737. smc_ib_is_valid_local_systemid())
  738. memcpy(dclc.id_for_peer, local_systemid,
  739. sizeof(local_systemid));
  740. dclc.peer_diagnosis = htonl(peer_diag_info);
  741. if (version == SMC_V1) {
  742. memcpy(dclc_v1->trl.eyecatcher, SMC_EYECATCHER,
  743. sizeof(SMC_EYECATCHER));
  744. send_len = sizeof(*dclc_v1);
  745. } else {
  746. memcpy(dclc.trl.eyecatcher, SMC_EYECATCHER,
  747. sizeof(SMC_EYECATCHER));
  748. send_len = sizeof(dclc);
  749. }
  750. dclc.hdr.length = htons(send_len);
  751. memset(&msg, 0, sizeof(msg));
  752. vec.iov_base = &dclc;
  753. vec.iov_len = send_len;
  754. len = kernel_sendmsg(smc->clcsock, &msg, &vec, 1, send_len);
  755. if (len < 0 || len < send_len)
  756. len = -EPROTO;
  757. return len > 0 ? 0 : len;
  758. }
  759. /* send CLC PROPOSAL message across internal TCP socket */
  760. int smc_clc_send_proposal(struct smc_sock *smc, struct smc_init_info *ini)
  761. {
  762. struct smc_clc_smcd_v2_extension *smcd_v2_ext;
  763. struct smc_clc_msg_proposal_prefix *pclc_prfx;
  764. struct smc_clc_msg_proposal *pclc_base;
  765. struct smc_clc_smcd_gid_chid *gidchids;
  766. struct smc_clc_msg_proposal_area *pclc;
  767. struct smc_clc_ipv6_prefix *ipv6_prfx;
  768. struct net *net = sock_net(&smc->sk);
  769. struct smc_clc_v2_extension *v2_ext;
  770. struct smc_clc_msg_smcd *pclc_smcd;
  771. struct smc_clc_msg_trail *trl;
  772. struct smcd_dev *smcd;
  773. int len, i, plen, rc;
  774. int reason_code = 0;
  775. struct kvec vec[8];
  776. struct msghdr msg;
  777. pclc = kzalloc(sizeof(*pclc), GFP_KERNEL);
  778. if (!pclc)
  779. return -ENOMEM;
  780. pclc_base = &pclc->pclc_base;
  781. pclc_smcd = &pclc->pclc_smcd;
  782. pclc_prfx = &pclc->pclc_prfx;
  783. ipv6_prfx = pclc->pclc_prfx_ipv6;
  784. v2_ext = container_of(&pclc->pclc_v2_ext,
  785. struct smc_clc_v2_extension, fixed);
  786. smcd_v2_ext = container_of(&pclc->pclc_smcd_v2_ext,
  787. struct smc_clc_smcd_v2_extension, fixed);
  788. gidchids = pclc->pclc_gidchids;
  789. trl = &pclc->pclc_trl;
  790. pclc_base->hdr.version = SMC_V2;
  791. pclc_base->hdr.typev1 = ini->smc_type_v1;
  792. pclc_base->hdr.typev2 = ini->smc_type_v2;
  793. plen = sizeof(*pclc_base) + sizeof(*pclc_smcd) + sizeof(*trl);
  794. /* retrieve ip prefixes for CLC proposal msg */
  795. if (ini->smc_type_v1 != SMC_TYPE_N) {
  796. rc = smc_clc_prfx_set(smc->clcsock, pclc_prfx, ipv6_prfx);
  797. if (rc) {
  798. if (ini->smc_type_v2 == SMC_TYPE_N) {
  799. kfree(pclc);
  800. return SMC_CLC_DECL_CNFERR;
  801. }
  802. pclc_base->hdr.typev1 = SMC_TYPE_N;
  803. } else {
  804. pclc_base->iparea_offset = htons(sizeof(*pclc_smcd));
  805. plen += sizeof(*pclc_prfx) +
  806. pclc_prfx->ipv6_prefixes_cnt *
  807. sizeof(ipv6_prfx[0]);
  808. }
  809. }
  810. /* build SMC Proposal CLC message */
  811. memcpy(pclc_base->hdr.eyecatcher, SMC_EYECATCHER,
  812. sizeof(SMC_EYECATCHER));
  813. pclc_base->hdr.type = SMC_CLC_PROPOSAL;
  814. if (smcr_indicated(ini->smc_type_v1)) {
  815. /* add SMC-R specifics */
  816. memcpy(pclc_base->lcl.id_for_peer, local_systemid,
  817. sizeof(local_systemid));
  818. memcpy(pclc_base->lcl.gid, ini->ib_gid, SMC_GID_SIZE);
  819. memcpy(pclc_base->lcl.mac, &ini->ib_dev->mac[ini->ib_port - 1],
  820. ETH_ALEN);
  821. }
  822. if (smcd_indicated(ini->smc_type_v1)) {
  823. struct smcd_gid smcd_gid;
  824. /* add SMC-D specifics */
  825. if (ini->ism_dev[0]) {
  826. smcd = ini->ism_dev[0];
  827. smcd->ops->get_local_gid(smcd, &smcd_gid);
  828. pclc_smcd->ism.gid = htonll(smcd_gid.gid);
  829. pclc_smcd->ism.chid =
  830. htons(smc_ism_get_chid(ini->ism_dev[0]));
  831. }
  832. }
  833. if (ini->smc_type_v2 == SMC_TYPE_N) {
  834. pclc_smcd->v2_ext_offset = 0;
  835. } else {
  836. struct smc_clc_eid_entry *ueident;
  837. u16 v2_ext_offset;
  838. v2_ext->hdr.flag.release = SMC_RELEASE;
  839. v2_ext_offset = sizeof(*pclc_smcd) -
  840. offsetofend(struct smc_clc_msg_smcd, v2_ext_offset);
  841. if (ini->smc_type_v1 != SMC_TYPE_N)
  842. v2_ext_offset += sizeof(*pclc_prfx) +
  843. pclc_prfx->ipv6_prefixes_cnt *
  844. sizeof(ipv6_prfx[0]);
  845. pclc_smcd->v2_ext_offset = htons(v2_ext_offset);
  846. plen += sizeof(*v2_ext);
  847. v2_ext->feature_mask = htons(SMC_FEATURE_MASK);
  848. read_lock(&smc_clc_eid_table.lock);
  849. v2_ext->hdr.eid_cnt = smc_clc_eid_table.ueid_cnt;
  850. plen += smc_clc_eid_table.ueid_cnt * SMC_MAX_EID_LEN;
  851. i = 0;
  852. list_for_each_entry(ueident, &smc_clc_eid_table.list, list) {
  853. memcpy(v2_ext->user_eids[i++], ueident->eid,
  854. sizeof(ueident->eid));
  855. }
  856. read_unlock(&smc_clc_eid_table.lock);
  857. }
  858. if (smcd_indicated(ini->smc_type_v2)) {
  859. struct smcd_gid smcd_gid;
  860. u8 *eid = NULL;
  861. int entry = 0;
  862. v2_ext->hdr.flag.seid = smc_clc_eid_table.seid_enabled;
  863. v2_ext->hdr.smcd_v2_ext_offset = htons(sizeof(*v2_ext) -
  864. offsetofend(struct smc_clnt_opts_area_hdr,
  865. smcd_v2_ext_offset) +
  866. v2_ext->hdr.eid_cnt * SMC_MAX_EID_LEN);
  867. smc_ism_get_system_eid(&eid);
  868. if (eid && v2_ext->hdr.flag.seid)
  869. memcpy(smcd_v2_ext->system_eid, eid, SMC_MAX_EID_LEN);
  870. plen += sizeof(*smcd_v2_ext);
  871. if (ini->ism_offered_cnt) {
  872. for (i = 1; i <= ini->ism_offered_cnt; i++) {
  873. smcd = ini->ism_dev[i];
  874. smcd->ops->get_local_gid(smcd, &smcd_gid);
  875. gidchids[entry].chid =
  876. htons(smc_ism_get_chid(ini->ism_dev[i]));
  877. gidchids[entry].gid = htonll(smcd_gid.gid);
  878. if (smc_ism_is_emulated(smcd)) {
  879. /* an Emulated-ISM device takes two
  880. * entries. CHID of the second entry
  881. * repeats that of the first entry.
  882. */
  883. gidchids[entry + 1].chid =
  884. gidchids[entry].chid;
  885. gidchids[entry + 1].gid =
  886. htonll(smcd_gid.gid_ext);
  887. entry++;
  888. }
  889. entry++;
  890. }
  891. plen += entry * sizeof(struct smc_clc_smcd_gid_chid);
  892. }
  893. v2_ext->hdr.ism_gid_cnt = entry;
  894. }
  895. if (smcr_indicated(ini->smc_type_v2)) {
  896. memcpy(v2_ext->roce, ini->smcrv2.ib_gid_v2, SMC_GID_SIZE);
  897. v2_ext->max_conns = net->smc.sysctl_max_conns_per_lgr;
  898. v2_ext->max_links = net->smc.sysctl_max_links_per_lgr;
  899. }
  900. pclc_base->hdr.length = htons(plen);
  901. memcpy(trl->eyecatcher, SMC_EYECATCHER, sizeof(SMC_EYECATCHER));
  902. /* send SMC Proposal CLC message */
  903. memset(&msg, 0, sizeof(msg));
  904. i = 0;
  905. vec[i].iov_base = pclc_base;
  906. vec[i++].iov_len = sizeof(*pclc_base);
  907. vec[i].iov_base = pclc_smcd;
  908. vec[i++].iov_len = sizeof(*pclc_smcd);
  909. if (ini->smc_type_v1 != SMC_TYPE_N) {
  910. vec[i].iov_base = pclc_prfx;
  911. vec[i++].iov_len = sizeof(*pclc_prfx);
  912. if (pclc_prfx->ipv6_prefixes_cnt > 0) {
  913. vec[i].iov_base = ipv6_prfx;
  914. vec[i++].iov_len = pclc_prfx->ipv6_prefixes_cnt *
  915. sizeof(ipv6_prfx[0]);
  916. }
  917. }
  918. if (ini->smc_type_v2 != SMC_TYPE_N) {
  919. vec[i].iov_base = v2_ext;
  920. vec[i++].iov_len = sizeof(*v2_ext) +
  921. (v2_ext->hdr.eid_cnt * SMC_MAX_EID_LEN);
  922. if (smcd_indicated(ini->smc_type_v2)) {
  923. vec[i].iov_base = smcd_v2_ext;
  924. vec[i++].iov_len = sizeof(*smcd_v2_ext);
  925. if (ini->ism_offered_cnt) {
  926. vec[i].iov_base = gidchids;
  927. vec[i++].iov_len = v2_ext->hdr.ism_gid_cnt *
  928. sizeof(struct smc_clc_smcd_gid_chid);
  929. }
  930. }
  931. }
  932. vec[i].iov_base = trl;
  933. vec[i++].iov_len = sizeof(*trl);
  934. /* due to the few bytes needed for clc-handshake this cannot block */
  935. len = kernel_sendmsg(smc->clcsock, &msg, vec, i, plen);
  936. if (len < 0) {
  937. smc->sk.sk_err = smc->clcsock->sk->sk_err;
  938. reason_code = -smc->sk.sk_err;
  939. } else if (len < ntohs(pclc_base->hdr.length)) {
  940. reason_code = -ENETUNREACH;
  941. smc->sk.sk_err = -reason_code;
  942. }
  943. kfree(pclc);
  944. return reason_code;
  945. }
  946. static void
  947. smcd_clc_prep_confirm_accept(struct smc_connection *conn,
  948. struct smc_clc_msg_accept_confirm *clc,
  949. int first_contact, u8 version,
  950. u8 *eid, struct smc_init_info *ini,
  951. int *fce_len,
  952. struct smc_clc_first_contact_ext_v2x *fce_v2x,
  953. struct smc_clc_msg_trail *trl)
  954. {
  955. struct smcd_dev *smcd = conn->lgr->smcd;
  956. struct smcd_gid smcd_gid;
  957. u16 chid;
  958. int len;
  959. /* SMC-D specific settings */
  960. memcpy(clc->hdr.eyecatcher, SMCD_EYECATCHER,
  961. sizeof(SMCD_EYECATCHER));
  962. smcd->ops->get_local_gid(smcd, &smcd_gid);
  963. clc->hdr.typev1 = SMC_TYPE_D;
  964. clc->d0.gid = htonll(smcd_gid.gid);
  965. clc->d0.token = htonll(conn->rmb_desc->token);
  966. clc->d0.dmbe_size = conn->rmbe_size_comp;
  967. clc->d0.dmbe_idx = 0;
  968. memcpy(&clc->d0.linkid, conn->lgr->id, SMC_LGR_ID_SIZE);
  969. if (version == SMC_V1) {
  970. clc->hdr.length = htons(SMCD_CLC_ACCEPT_CONFIRM_LEN);
  971. } else {
  972. chid = smc_ism_get_chid(smcd);
  973. clc->d1.chid = htons(chid);
  974. if (eid && eid[0])
  975. memcpy(clc->d1.eid, eid, SMC_MAX_EID_LEN);
  976. if (__smc_ism_is_emulated(chid))
  977. clc->d1.gid_ext = htonll(smcd_gid.gid_ext);
  978. len = SMCD_CLC_ACCEPT_CONFIRM_LEN_V2;
  979. if (first_contact) {
  980. *fce_len = smc_clc_fill_fce_v2x(fce_v2x, ini);
  981. len += *fce_len;
  982. }
  983. clc->hdr.length = htons(len);
  984. }
  985. memcpy(trl->eyecatcher, SMCD_EYECATCHER,
  986. sizeof(SMCD_EYECATCHER));
  987. }
  988. static void
  989. smcr_clc_prep_confirm_accept(struct smc_connection *conn,
  990. struct smc_clc_msg_accept_confirm *clc,
  991. int first_contact, u8 version,
  992. u8 *eid, struct smc_init_info *ini,
  993. int *fce_len,
  994. struct smc_clc_first_contact_ext_v2x *fce_v2x,
  995. struct smc_clc_fce_gid_ext *gle,
  996. struct smc_clc_msg_trail *trl)
  997. {
  998. struct smc_link *link = conn->lnk;
  999. int len;
  1000. /* SMC-R specific settings */
  1001. memcpy(clc->hdr.eyecatcher, SMC_EYECATCHER,
  1002. sizeof(SMC_EYECATCHER));
  1003. clc->hdr.typev1 = SMC_TYPE_R;
  1004. memcpy(clc->r0.lcl.id_for_peer, local_systemid,
  1005. sizeof(local_systemid));
  1006. memcpy(&clc->r0.lcl.gid, link->gid, SMC_GID_SIZE);
  1007. memcpy(&clc->r0.lcl.mac, &link->smcibdev->mac[link->ibport - 1],
  1008. ETH_ALEN);
  1009. hton24(clc->r0.qpn, link->roce_qp->qp_num);
  1010. clc->r0.rmb_rkey =
  1011. htonl(conn->rmb_desc->mr[link->link_idx]->rkey);
  1012. clc->r0.rmbe_idx = 1; /* for now: 1 RMB = 1 RMBE */
  1013. clc->r0.rmbe_alert_token = htonl(conn->alert_token_local);
  1014. switch (clc->hdr.type) {
  1015. case SMC_CLC_ACCEPT:
  1016. clc->r0.qp_mtu = link->path_mtu;
  1017. break;
  1018. case SMC_CLC_CONFIRM:
  1019. clc->r0.qp_mtu = min(link->path_mtu, link->peer_mtu);
  1020. break;
  1021. }
  1022. clc->r0.rmbe_size = conn->rmbe_size_comp;
  1023. clc->r0.rmb_dma_addr = conn->rmb_desc->is_vm ?
  1024. cpu_to_be64((uintptr_t)conn->rmb_desc->cpu_addr) :
  1025. cpu_to_be64((u64)sg_dma_address
  1026. (conn->rmb_desc->sgt[link->link_idx].sgl));
  1027. hton24(clc->r0.psn, link->psn_initial);
  1028. if (version == SMC_V1) {
  1029. clc->hdr.length = htons(SMCR_CLC_ACCEPT_CONFIRM_LEN);
  1030. } else {
  1031. if (eid && eid[0])
  1032. memcpy(clc->r1.eid, eid, SMC_MAX_EID_LEN);
  1033. len = SMCR_CLC_ACCEPT_CONFIRM_LEN_V2;
  1034. if (first_contact) {
  1035. *fce_len = smc_clc_fill_fce_v2x(fce_v2x, ini);
  1036. len += *fce_len;
  1037. fce_v2x->fce_v2_base.v2_direct =
  1038. !link->lgr->uses_gateway;
  1039. if (clc->hdr.type == SMC_CLC_CONFIRM) {
  1040. memset(gle, 0, sizeof(*gle));
  1041. gle->gid_cnt = ini->smcrv2.gidlist.len;
  1042. len += sizeof(*gle);
  1043. len += gle->gid_cnt * sizeof(gle->gid[0]);
  1044. }
  1045. }
  1046. clc->hdr.length = htons(len);
  1047. }
  1048. memcpy(trl->eyecatcher, SMC_EYECATCHER, sizeof(SMC_EYECATCHER));
  1049. }
  1050. /* build and send CLC CONFIRM / ACCEPT message */
  1051. static int smc_clc_send_confirm_accept(struct smc_sock *smc,
  1052. struct smc_clc_msg_accept_confirm *clc,
  1053. int first_contact, u8 version,
  1054. u8 *eid, struct smc_init_info *ini)
  1055. {
  1056. struct smc_clc_first_contact_ext_v2x fce_v2x;
  1057. struct smc_connection *conn = &smc->conn;
  1058. struct smc_clc_fce_gid_ext gle;
  1059. struct smc_clc_msg_trail trl;
  1060. int i, fce_len;
  1061. struct kvec vec[5];
  1062. struct msghdr msg;
  1063. /* send SMC Confirm CLC msg */
  1064. clc->hdr.version = version; /* SMC version */
  1065. if (first_contact)
  1066. clc->hdr.typev2 |= SMC_FIRST_CONTACT_MASK;
  1067. if (conn->lgr->is_smcd)
  1068. smcd_clc_prep_confirm_accept(conn, clc, first_contact,
  1069. version, eid, ini, &fce_len,
  1070. &fce_v2x, &trl);
  1071. else
  1072. smcr_clc_prep_confirm_accept(conn, clc, first_contact,
  1073. version, eid, ini, &fce_len,
  1074. &fce_v2x, &gle, &trl);
  1075. memset(&msg, 0, sizeof(msg));
  1076. i = 0;
  1077. vec[i].iov_base = clc;
  1078. if (version > SMC_V1)
  1079. vec[i++].iov_len = (clc->hdr.typev1 == SMC_TYPE_D ?
  1080. SMCD_CLC_ACCEPT_CONFIRM_LEN_V2 :
  1081. SMCR_CLC_ACCEPT_CONFIRM_LEN_V2) -
  1082. sizeof(trl);
  1083. else
  1084. vec[i++].iov_len = (clc->hdr.typev1 == SMC_TYPE_D ?
  1085. SMCD_CLC_ACCEPT_CONFIRM_LEN :
  1086. SMCR_CLC_ACCEPT_CONFIRM_LEN) -
  1087. sizeof(trl);
  1088. if (version > SMC_V1 && first_contact) {
  1089. vec[i].iov_base = &fce_v2x;
  1090. vec[i++].iov_len = fce_len;
  1091. if (!conn->lgr->is_smcd) {
  1092. if (clc->hdr.type == SMC_CLC_CONFIRM) {
  1093. vec[i].iov_base = &gle;
  1094. vec[i++].iov_len = sizeof(gle);
  1095. vec[i].iov_base = &ini->smcrv2.gidlist.list;
  1096. vec[i++].iov_len = gle.gid_cnt *
  1097. sizeof(gle.gid[0]);
  1098. }
  1099. }
  1100. }
  1101. vec[i].iov_base = &trl;
  1102. vec[i++].iov_len = sizeof(trl);
  1103. return kernel_sendmsg(smc->clcsock, &msg, vec, 1,
  1104. ntohs(clc->hdr.length));
  1105. }
  1106. /* send CLC CONFIRM message across internal TCP socket */
  1107. int smc_clc_send_confirm(struct smc_sock *smc, bool clnt_first_contact,
  1108. u8 version, u8 *eid, struct smc_init_info *ini)
  1109. {
  1110. struct smc_clc_msg_accept_confirm cclc;
  1111. int reason_code = 0;
  1112. int len;
  1113. /* send SMC Confirm CLC msg */
  1114. memset(&cclc, 0, sizeof(cclc));
  1115. cclc.hdr.type = SMC_CLC_CONFIRM;
  1116. len = smc_clc_send_confirm_accept(smc, &cclc, clnt_first_contact,
  1117. version, eid, ini);
  1118. if (len < ntohs(cclc.hdr.length)) {
  1119. if (len >= 0) {
  1120. reason_code = -ENETUNREACH;
  1121. smc->sk.sk_err = -reason_code;
  1122. } else {
  1123. smc->sk.sk_err = smc->clcsock->sk->sk_err;
  1124. reason_code = -smc->sk.sk_err;
  1125. }
  1126. }
  1127. return reason_code;
  1128. }
  1129. /* send CLC ACCEPT message across internal TCP socket */
  1130. int smc_clc_send_accept(struct smc_sock *new_smc, bool srv_first_contact,
  1131. u8 version, u8 *negotiated_eid, struct smc_init_info *ini)
  1132. {
  1133. struct smc_clc_msg_accept_confirm aclc;
  1134. int len;
  1135. memset(&aclc, 0, sizeof(aclc));
  1136. aclc.hdr.type = SMC_CLC_ACCEPT;
  1137. len = smc_clc_send_confirm_accept(new_smc, &aclc, srv_first_contact,
  1138. version, negotiated_eid, ini);
  1139. if (len < ntohs(aclc.hdr.length))
  1140. len = len >= 0 ? -EPROTO : -new_smc->clcsock->sk->sk_err;
  1141. return len > 0 ? 0 : len;
  1142. }
  1143. int smc_clc_srv_v2x_features_validate(struct smc_sock *smc,
  1144. struct smc_clc_msg_proposal *pclc,
  1145. struct smc_init_info *ini)
  1146. {
  1147. struct smc_clc_v2_extension *pclc_v2_ext;
  1148. struct net *net = sock_net(&smc->sk);
  1149. ini->max_conns = SMC_CONN_PER_LGR_MAX;
  1150. ini->max_links = SMC_LINKS_ADD_LNK_MAX;
  1151. ini->feature_mask = SMC_FEATURE_MASK;
  1152. if ((!(ini->smcd_version & SMC_V2) && !(ini->smcr_version & SMC_V2)) ||
  1153. ini->release_nr < SMC_RELEASE_1)
  1154. return 0;
  1155. pclc_v2_ext = smc_get_clc_v2_ext(pclc);
  1156. if (!pclc_v2_ext)
  1157. return SMC_CLC_DECL_NOV2EXT;
  1158. if (ini->smcr_version & SMC_V2) {
  1159. ini->max_conns = min_t(u8, pclc_v2_ext->max_conns,
  1160. net->smc.sysctl_max_conns_per_lgr);
  1161. if (ini->max_conns < SMC_CONN_PER_LGR_MIN)
  1162. return SMC_CLC_DECL_MAXCONNERR;
  1163. ini->max_links = min_t(u8, pclc_v2_ext->max_links,
  1164. net->smc.sysctl_max_links_per_lgr);
  1165. if (ini->max_links < SMC_LINKS_ADD_LNK_MIN)
  1166. return SMC_CLC_DECL_MAXLINKERR;
  1167. }
  1168. return 0;
  1169. }
  1170. int smc_clc_clnt_v2x_features_validate(struct smc_clc_first_contact_ext *fce,
  1171. struct smc_init_info *ini)
  1172. {
  1173. struct smc_clc_first_contact_ext_v2x *fce_v2x =
  1174. (struct smc_clc_first_contact_ext_v2x *)fce;
  1175. if (ini->release_nr < SMC_RELEASE_1)
  1176. return 0;
  1177. if (!ini->is_smcd) {
  1178. if (fce_v2x->max_conns < SMC_CONN_PER_LGR_MIN)
  1179. return SMC_CLC_DECL_MAXCONNERR;
  1180. ini->max_conns = fce_v2x->max_conns;
  1181. if (fce_v2x->max_links > SMC_LINKS_ADD_LNK_MAX ||
  1182. fce_v2x->max_links < SMC_LINKS_ADD_LNK_MIN)
  1183. return SMC_CLC_DECL_MAXLINKERR;
  1184. ini->max_links = fce_v2x->max_links;
  1185. }
  1186. /* common supplemental features of server and client */
  1187. ini->feature_mask = ntohs(fce_v2x->feature_mask) & SMC_FEATURE_MASK;
  1188. return 0;
  1189. }
  1190. int smc_clc_v2x_features_confirm_check(struct smc_clc_msg_accept_confirm *cclc,
  1191. struct smc_init_info *ini)
  1192. {
  1193. struct smc_clc_first_contact_ext *fce =
  1194. smc_get_clc_first_contact_ext(cclc, ini->is_smcd);
  1195. struct smc_clc_first_contact_ext_v2x *fce_v2x =
  1196. (struct smc_clc_first_contact_ext_v2x *)fce;
  1197. if (cclc->hdr.version == SMC_V1 ||
  1198. !(cclc->hdr.typev2 & SMC_FIRST_CONTACT_MASK))
  1199. return 0;
  1200. if (ini->release_nr != fce->release)
  1201. return SMC_CLC_DECL_RELEASEERR;
  1202. if (fce->release < SMC_RELEASE_1)
  1203. return 0;
  1204. if (!ini->is_smcd) {
  1205. if (fce_v2x->max_conns != ini->max_conns)
  1206. return SMC_CLC_DECL_MAXCONNERR;
  1207. if (fce_v2x->max_links != ini->max_links)
  1208. return SMC_CLC_DECL_MAXLINKERR;
  1209. }
  1210. /* common supplemental features returned by client */
  1211. ini->feature_mask = ntohs(fce_v2x->feature_mask);
  1212. return 0;
  1213. }
  1214. void smc_clc_get_hostname(u8 **host)
  1215. {
  1216. *host = &smc_hostname[0];
  1217. }
  1218. void __init smc_clc_init(void)
  1219. {
  1220. struct new_utsname *u;
  1221. memset(smc_hostname, _S, sizeof(smc_hostname)); /* ASCII blanks */
  1222. u = utsname();
  1223. memcpy(smc_hostname, u->nodename,
  1224. min_t(size_t, strlen(u->nodename), sizeof(smc_hostname)));
  1225. INIT_LIST_HEAD(&smc_clc_eid_table.list);
  1226. rwlock_init(&smc_clc_eid_table.lock);
  1227. smc_clc_eid_table.ueid_cnt = 0;
  1228. #if IS_ENABLED(CONFIG_S390)
  1229. smc_clc_eid_table.seid_enabled = 1;
  1230. #else
  1231. smc_clc_eid_table.seid_enabled = 0;
  1232. #endif
  1233. }
  1234. void smc_clc_exit(void)
  1235. {
  1236. smc_clc_ueid_remove(NULL);
  1237. }