smc_clc.c 18 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603
  1. // SPDX-License-Identifier: GPL-2.0
  2. /*
  3. * Shared Memory Communications over RDMA (SMC-R) and RoCE
  4. *
  5. * CLC (connection layer control) handshake over initial TCP socket to
  6. * prepare for RDMA traffic
  7. *
  8. * Copyright IBM Corp. 2016, 2018
  9. *
  10. * Author(s): Ursula Braun <ubraun@linux.vnet.ibm.com>
  11. */
  12. #include <linux/in.h>
  13. #include <linux/inetdevice.h>
  14. #include <linux/if_ether.h>
  15. #include <linux/sched/signal.h>
  16. #include <net/addrconf.h>
  17. #include <net/sock.h>
  18. #include <net/tcp.h>
  19. #include "smc.h"
  20. #include "smc_core.h"
  21. #include "smc_clc.h"
  22. #include "smc_ib.h"
  23. #include "smc_ism.h"
  24. #define SMCR_CLC_ACCEPT_CONFIRM_LEN 68
  25. #define SMCD_CLC_ACCEPT_CONFIRM_LEN 48
  26. /* eye catcher "SMCR" EBCDIC for CLC messages */
  27. static const char SMC_EYECATCHER[4] = {'\xe2', '\xd4', '\xc3', '\xd9'};
  28. /* eye catcher "SMCD" EBCDIC for CLC messages */
  29. static const char SMCD_EYECATCHER[4] = {'\xe2', '\xd4', '\xc3', '\xc4'};
  30. /* check if received message has a correct header length and contains valid
  31. * heading and trailing eyecatchers
  32. */
  33. static bool smc_clc_msg_hdr_valid(struct smc_clc_msg_hdr *clcm)
  34. {
  35. struct smc_clc_msg_proposal_prefix *pclc_prfx;
  36. struct smc_clc_msg_accept_confirm *clc;
  37. struct smc_clc_msg_proposal *pclc;
  38. struct smc_clc_msg_decline *dclc;
  39. struct smc_clc_msg_trail *trl;
  40. if (memcmp(clcm->eyecatcher, SMC_EYECATCHER, sizeof(SMC_EYECATCHER)) &&
  41. memcmp(clcm->eyecatcher, SMCD_EYECATCHER, sizeof(SMCD_EYECATCHER)))
  42. return false;
  43. switch (clcm->type) {
  44. case SMC_CLC_PROPOSAL:
  45. if (clcm->path != SMC_TYPE_R && clcm->path != SMC_TYPE_D &&
  46. clcm->path != SMC_TYPE_B)
  47. return false;
  48. pclc = (struct smc_clc_msg_proposal *)clcm;
  49. pclc_prfx = smc_clc_proposal_get_prefix(pclc);
  50. if (ntohs(pclc->hdr.length) !=
  51. sizeof(*pclc) + ntohs(pclc->iparea_offset) +
  52. sizeof(*pclc_prfx) +
  53. pclc_prfx->ipv6_prefixes_cnt *
  54. sizeof(struct smc_clc_ipv6_prefix) +
  55. sizeof(*trl))
  56. return false;
  57. trl = (struct smc_clc_msg_trail *)
  58. ((u8 *)pclc + ntohs(pclc->hdr.length) - sizeof(*trl));
  59. break;
  60. case SMC_CLC_ACCEPT:
  61. case SMC_CLC_CONFIRM:
  62. if (clcm->path != SMC_TYPE_R && clcm->path != SMC_TYPE_D)
  63. return false;
  64. clc = (struct smc_clc_msg_accept_confirm *)clcm;
  65. if ((clcm->path == SMC_TYPE_R &&
  66. ntohs(clc->hdr.length) != SMCR_CLC_ACCEPT_CONFIRM_LEN) ||
  67. (clcm->path == SMC_TYPE_D &&
  68. ntohs(clc->hdr.length) != SMCD_CLC_ACCEPT_CONFIRM_LEN))
  69. return false;
  70. trl = (struct smc_clc_msg_trail *)
  71. ((u8 *)clc + ntohs(clc->hdr.length) - sizeof(*trl));
  72. break;
  73. case SMC_CLC_DECLINE:
  74. dclc = (struct smc_clc_msg_decline *)clcm;
  75. if (ntohs(dclc->hdr.length) != sizeof(*dclc))
  76. return false;
  77. trl = &dclc->trl;
  78. break;
  79. default:
  80. return false;
  81. }
  82. if (memcmp(trl->eyecatcher, SMC_EYECATCHER, sizeof(SMC_EYECATCHER)) &&
  83. memcmp(trl->eyecatcher, SMCD_EYECATCHER, sizeof(SMCD_EYECATCHER)))
  84. return false;
  85. return true;
  86. }
  87. /* find ipv4 addr on device and get the prefix len, fill CLC proposal msg */
  88. static int smc_clc_prfx_set4_rcu(struct dst_entry *dst, __be32 ipv4,
  89. struct smc_clc_msg_proposal_prefix *prop)
  90. {
  91. struct in_device *in_dev = __in_dev_get_rcu(dst->dev);
  92. if (!in_dev)
  93. return -ENODEV;
  94. for_ifa(in_dev) {
  95. if (!inet_ifa_match(ipv4, ifa))
  96. continue;
  97. prop->prefix_len = inet_mask_len(ifa->ifa_mask);
  98. prop->outgoing_subnet = ifa->ifa_address & ifa->ifa_mask;
  99. /* prop->ipv6_prefixes_cnt = 0; already done by memset before */
  100. return 0;
  101. } endfor_ifa(in_dev);
  102. return -ENOENT;
  103. }
  104. /* fill CLC proposal msg with ipv6 prefixes from device */
  105. static int smc_clc_prfx_set6_rcu(struct dst_entry *dst,
  106. struct smc_clc_msg_proposal_prefix *prop,
  107. struct smc_clc_ipv6_prefix *ipv6_prfx)
  108. {
  109. #if IS_ENABLED(CONFIG_IPV6)
  110. struct inet6_dev *in6_dev = __in6_dev_get(dst->dev);
  111. struct inet6_ifaddr *ifa;
  112. int cnt = 0;
  113. if (!in6_dev)
  114. return -ENODEV;
  115. /* use a maximum of 8 IPv6 prefixes from device */
  116. list_for_each_entry(ifa, &in6_dev->addr_list, if_list) {
  117. if (ipv6_addr_type(&ifa->addr) & IPV6_ADDR_LINKLOCAL)
  118. continue;
  119. ipv6_addr_prefix(&ipv6_prfx[cnt].prefix,
  120. &ifa->addr, ifa->prefix_len);
  121. ipv6_prfx[cnt].prefix_len = ifa->prefix_len;
  122. cnt++;
  123. if (cnt == SMC_CLC_MAX_V6_PREFIX)
  124. break;
  125. }
  126. prop->ipv6_prefixes_cnt = cnt;
  127. if (cnt)
  128. return 0;
  129. #endif
  130. return -ENOENT;
  131. }
  132. /* retrieve and set prefixes in CLC proposal msg */
  133. static int smc_clc_prfx_set(struct socket *clcsock,
  134. struct smc_clc_msg_proposal_prefix *prop,
  135. struct smc_clc_ipv6_prefix *ipv6_prfx)
  136. {
  137. struct dst_entry *dst = sk_dst_get(clcsock->sk);
  138. struct sockaddr_storage addrs;
  139. struct sockaddr_in6 *addr6;
  140. struct sockaddr_in *addr;
  141. int rc = -ENOENT;
  142. memset(prop, 0, sizeof(*prop));
  143. if (!dst) {
  144. rc = -ENOTCONN;
  145. goto out;
  146. }
  147. if (!dst->dev) {
  148. rc = -ENODEV;
  149. goto out_rel;
  150. }
  151. /* get address to which the internal TCP socket is bound */
  152. kernel_getsockname(clcsock, (struct sockaddr *)&addrs);
  153. /* analyze IP specific data of net_device belonging to TCP socket */
  154. addr6 = (struct sockaddr_in6 *)&addrs;
  155. rcu_read_lock();
  156. if (addrs.ss_family == PF_INET) {
  157. /* IPv4 */
  158. addr = (struct sockaddr_in *)&addrs;
  159. rc = smc_clc_prfx_set4_rcu(dst, addr->sin_addr.s_addr, prop);
  160. } else if (ipv6_addr_v4mapped(&addr6->sin6_addr)) {
  161. /* mapped IPv4 address - peer is IPv4 only */
  162. rc = smc_clc_prfx_set4_rcu(dst, addr6->sin6_addr.s6_addr32[3],
  163. prop);
  164. } else {
  165. /* IPv6 */
  166. rc = smc_clc_prfx_set6_rcu(dst, prop, ipv6_prfx);
  167. }
  168. rcu_read_unlock();
  169. out_rel:
  170. dst_release(dst);
  171. out:
  172. return rc;
  173. }
  174. /* match ipv4 addrs of dev against addr in CLC proposal */
  175. static int smc_clc_prfx_match4_rcu(struct net_device *dev,
  176. struct smc_clc_msg_proposal_prefix *prop)
  177. {
  178. struct in_device *in_dev = __in_dev_get_rcu(dev);
  179. if (!in_dev)
  180. return -ENODEV;
  181. for_ifa(in_dev) {
  182. if (prop->prefix_len == inet_mask_len(ifa->ifa_mask) &&
  183. inet_ifa_match(prop->outgoing_subnet, ifa))
  184. return 0;
  185. } endfor_ifa(in_dev);
  186. return -ENOENT;
  187. }
  188. /* match ipv6 addrs of dev against addrs in CLC proposal */
  189. static int smc_clc_prfx_match6_rcu(struct net_device *dev,
  190. struct smc_clc_msg_proposal_prefix *prop)
  191. {
  192. #if IS_ENABLED(CONFIG_IPV6)
  193. struct inet6_dev *in6_dev = __in6_dev_get(dev);
  194. struct smc_clc_ipv6_prefix *ipv6_prfx;
  195. struct inet6_ifaddr *ifa;
  196. int i, max;
  197. if (!in6_dev)
  198. return -ENODEV;
  199. /* ipv6 prefix list starts behind smc_clc_msg_proposal_prefix */
  200. ipv6_prfx = (struct smc_clc_ipv6_prefix *)((u8 *)prop + sizeof(*prop));
  201. max = min_t(u8, prop->ipv6_prefixes_cnt, SMC_CLC_MAX_V6_PREFIX);
  202. list_for_each_entry(ifa, &in6_dev->addr_list, if_list) {
  203. if (ipv6_addr_type(&ifa->addr) & IPV6_ADDR_LINKLOCAL)
  204. continue;
  205. for (i = 0; i < max; i++) {
  206. if (ifa->prefix_len == ipv6_prfx[i].prefix_len &&
  207. ipv6_prefix_equal(&ifa->addr, &ipv6_prfx[i].prefix,
  208. ifa->prefix_len))
  209. return 0;
  210. }
  211. }
  212. #endif
  213. return -ENOENT;
  214. }
  215. /* check if proposed prefixes match one of our device prefixes */
  216. int smc_clc_prfx_match(struct socket *clcsock,
  217. struct smc_clc_msg_proposal_prefix *prop)
  218. {
  219. struct dst_entry *dst = sk_dst_get(clcsock->sk);
  220. int rc;
  221. if (!dst) {
  222. rc = -ENOTCONN;
  223. goto out;
  224. }
  225. if (!dst->dev) {
  226. rc = -ENODEV;
  227. goto out_rel;
  228. }
  229. rcu_read_lock();
  230. if (!prop->ipv6_prefixes_cnt)
  231. rc = smc_clc_prfx_match4_rcu(dst->dev, prop);
  232. else
  233. rc = smc_clc_prfx_match6_rcu(dst->dev, prop);
  234. rcu_read_unlock();
  235. out_rel:
  236. dst_release(dst);
  237. out:
  238. return rc;
  239. }
  240. /* Wait for data on the tcp-socket, analyze received data
  241. * Returns:
  242. * 0 if success and it was not a decline that we received.
  243. * SMC_CLC_DECL_REPLY if decline received for fallback w/o another decl send.
  244. * clcsock error, -EINTR, -ECONNRESET, -EPROTO otherwise.
  245. */
  246. int smc_clc_wait_msg(struct smc_sock *smc, void *buf, int buflen,
  247. u8 expected_type)
  248. {
  249. long rcvtimeo = smc->clcsock->sk->sk_rcvtimeo;
  250. struct sock *clc_sk = smc->clcsock->sk;
  251. struct smc_clc_msg_hdr *clcm = buf;
  252. struct msghdr msg = {NULL, 0};
  253. int reason_code = 0;
  254. struct kvec vec = {buf, buflen};
  255. int len, datlen;
  256. int krflags;
  257. /* peek the first few bytes to determine length of data to receive
  258. * so we don't consume any subsequent CLC message or payload data
  259. * in the TCP byte stream
  260. */
  261. /*
  262. * Caller must make sure that buflen is no less than
  263. * sizeof(struct smc_clc_msg_hdr)
  264. */
  265. krflags = MSG_PEEK | MSG_WAITALL;
  266. smc->clcsock->sk->sk_rcvtimeo = CLC_WAIT_TIME;
  267. iov_iter_kvec(&msg.msg_iter, READ | ITER_KVEC, &vec, 1,
  268. sizeof(struct smc_clc_msg_hdr));
  269. len = sock_recvmsg(smc->clcsock, &msg, krflags);
  270. if (signal_pending(current)) {
  271. reason_code = -EINTR;
  272. clc_sk->sk_err = EINTR;
  273. smc->sk.sk_err = EINTR;
  274. goto out;
  275. }
  276. if (clc_sk->sk_err) {
  277. reason_code = -clc_sk->sk_err;
  278. smc->sk.sk_err = clc_sk->sk_err;
  279. goto out;
  280. }
  281. if (!len) { /* peer has performed orderly shutdown */
  282. smc->sk.sk_err = ECONNRESET;
  283. reason_code = -ECONNRESET;
  284. goto out;
  285. }
  286. if (len < 0) {
  287. smc->sk.sk_err = -len;
  288. reason_code = len;
  289. goto out;
  290. }
  291. datlen = ntohs(clcm->length);
  292. if ((len < sizeof(struct smc_clc_msg_hdr)) ||
  293. (datlen > buflen) ||
  294. (clcm->version != SMC_CLC_V1) ||
  295. (clcm->path != SMC_TYPE_R && clcm->path != SMC_TYPE_D &&
  296. clcm->path != SMC_TYPE_B) ||
  297. ((clcm->type != SMC_CLC_DECLINE) &&
  298. (clcm->type != expected_type))) {
  299. smc->sk.sk_err = EPROTO;
  300. reason_code = -EPROTO;
  301. goto out;
  302. }
  303. /* receive the complete CLC message */
  304. memset(&msg, 0, sizeof(struct msghdr));
  305. iov_iter_kvec(&msg.msg_iter, READ | ITER_KVEC, &vec, 1, datlen);
  306. krflags = MSG_WAITALL;
  307. len = sock_recvmsg(smc->clcsock, &msg, krflags);
  308. if (len < datlen || !smc_clc_msg_hdr_valid(clcm)) {
  309. smc->sk.sk_err = EPROTO;
  310. reason_code = -EPROTO;
  311. goto out;
  312. }
  313. if (clcm->type == SMC_CLC_DECLINE) {
  314. struct smc_clc_msg_decline *dclc;
  315. dclc = (struct smc_clc_msg_decline *)clcm;
  316. reason_code = SMC_CLC_DECL_PEERDECL;
  317. smc->peer_diagnosis = ntohl(dclc->peer_diagnosis);
  318. if (((struct smc_clc_msg_decline *)buf)->hdr.flag) {
  319. smc->conn.lgr->sync_err = 1;
  320. smc_lgr_terminate(smc->conn.lgr);
  321. }
  322. }
  323. out:
  324. smc->clcsock->sk->sk_rcvtimeo = rcvtimeo;
  325. return reason_code;
  326. }
  327. /* send CLC DECLINE message across internal TCP socket */
  328. int smc_clc_send_decline(struct smc_sock *smc, u32 peer_diag_info)
  329. {
  330. struct smc_clc_msg_decline dclc;
  331. struct msghdr msg;
  332. struct kvec vec;
  333. int len;
  334. memset(&dclc, 0, sizeof(dclc));
  335. memcpy(dclc.hdr.eyecatcher, SMC_EYECATCHER, sizeof(SMC_EYECATCHER));
  336. dclc.hdr.type = SMC_CLC_DECLINE;
  337. dclc.hdr.length = htons(sizeof(struct smc_clc_msg_decline));
  338. dclc.hdr.version = SMC_CLC_V1;
  339. dclc.hdr.flag = (peer_diag_info == SMC_CLC_DECL_SYNCERR) ? 1 : 0;
  340. if (smc->conn.lgr && !smc->conn.lgr->is_smcd)
  341. memcpy(dclc.id_for_peer, local_systemid,
  342. sizeof(local_systemid));
  343. dclc.peer_diagnosis = htonl(peer_diag_info);
  344. memcpy(dclc.trl.eyecatcher, SMC_EYECATCHER, sizeof(SMC_EYECATCHER));
  345. memset(&msg, 0, sizeof(msg));
  346. vec.iov_base = &dclc;
  347. vec.iov_len = sizeof(struct smc_clc_msg_decline);
  348. len = kernel_sendmsg(smc->clcsock, &msg, &vec, 1,
  349. sizeof(struct smc_clc_msg_decline));
  350. if (len < sizeof(struct smc_clc_msg_decline))
  351. smc->sk.sk_err = EPROTO;
  352. if (len < 0)
  353. smc->sk.sk_err = -len;
  354. return sock_error(&smc->sk);
  355. }
  356. /* send CLC PROPOSAL message across internal TCP socket */
  357. int smc_clc_send_proposal(struct smc_sock *smc, int smc_type,
  358. struct smc_ib_device *ibdev, u8 ibport, u8 gid[],
  359. struct smcd_dev *ismdev)
  360. {
  361. struct smc_clc_ipv6_prefix ipv6_prfx[SMC_CLC_MAX_V6_PREFIX];
  362. struct smc_clc_msg_proposal_prefix pclc_prfx;
  363. struct smc_clc_msg_smcd pclc_smcd;
  364. struct smc_clc_msg_proposal pclc;
  365. struct smc_clc_msg_trail trl;
  366. int len, i, plen, rc;
  367. int reason_code = 0;
  368. struct kvec vec[5];
  369. struct msghdr msg;
  370. /* retrieve ip prefixes for CLC proposal msg */
  371. rc = smc_clc_prfx_set(smc->clcsock, &pclc_prfx, ipv6_prfx);
  372. if (rc)
  373. return SMC_CLC_DECL_CNFERR; /* configuration error */
  374. /* send SMC Proposal CLC message */
  375. plen = sizeof(pclc) + sizeof(pclc_prfx) +
  376. (pclc_prfx.ipv6_prefixes_cnt * sizeof(ipv6_prfx[0])) +
  377. sizeof(trl);
  378. memset(&pclc, 0, sizeof(pclc));
  379. memcpy(pclc.hdr.eyecatcher, SMC_EYECATCHER, sizeof(SMC_EYECATCHER));
  380. pclc.hdr.type = SMC_CLC_PROPOSAL;
  381. pclc.hdr.version = SMC_CLC_V1; /* SMC version */
  382. pclc.hdr.path = smc_type;
  383. if (smc_type == SMC_TYPE_R || smc_type == SMC_TYPE_B) {
  384. /* add SMC-R specifics */
  385. memcpy(pclc.lcl.id_for_peer, local_systemid,
  386. sizeof(local_systemid));
  387. memcpy(&pclc.lcl.gid, gid, SMC_GID_SIZE);
  388. memcpy(&pclc.lcl.mac, &ibdev->mac[ibport - 1], ETH_ALEN);
  389. pclc.iparea_offset = htons(0);
  390. }
  391. if (smc_type == SMC_TYPE_D || smc_type == SMC_TYPE_B) {
  392. /* add SMC-D specifics */
  393. memset(&pclc_smcd, 0, sizeof(pclc_smcd));
  394. plen += sizeof(pclc_smcd);
  395. pclc.iparea_offset = htons(SMC_CLC_PROPOSAL_MAX_OFFSET);
  396. pclc_smcd.gid = ismdev->local_gid;
  397. }
  398. pclc.hdr.length = htons(plen);
  399. memcpy(trl.eyecatcher, SMC_EYECATCHER, sizeof(SMC_EYECATCHER));
  400. memset(&msg, 0, sizeof(msg));
  401. i = 0;
  402. vec[i].iov_base = &pclc;
  403. vec[i++].iov_len = sizeof(pclc);
  404. if (smc_type == SMC_TYPE_D || smc_type == SMC_TYPE_B) {
  405. vec[i].iov_base = &pclc_smcd;
  406. vec[i++].iov_len = sizeof(pclc_smcd);
  407. }
  408. vec[i].iov_base = &pclc_prfx;
  409. vec[i++].iov_len = sizeof(pclc_prfx);
  410. if (pclc_prfx.ipv6_prefixes_cnt > 0) {
  411. vec[i].iov_base = &ipv6_prfx[0];
  412. vec[i++].iov_len = pclc_prfx.ipv6_prefixes_cnt *
  413. sizeof(ipv6_prfx[0]);
  414. }
  415. vec[i].iov_base = &trl;
  416. vec[i++].iov_len = sizeof(trl);
  417. /* due to the few bytes needed for clc-handshake this cannot block */
  418. len = kernel_sendmsg(smc->clcsock, &msg, vec, i, plen);
  419. if (len < 0) {
  420. smc->sk.sk_err = smc->clcsock->sk->sk_err;
  421. reason_code = -smc->sk.sk_err;
  422. } else if (len < (int)sizeof(pclc)) {
  423. reason_code = -ENETUNREACH;
  424. smc->sk.sk_err = -reason_code;
  425. }
  426. return reason_code;
  427. }
  428. /* send CLC CONFIRM message across internal TCP socket */
  429. int smc_clc_send_confirm(struct smc_sock *smc)
  430. {
  431. struct smc_connection *conn = &smc->conn;
  432. struct smc_clc_msg_accept_confirm cclc;
  433. struct smc_link *link;
  434. int reason_code = 0;
  435. struct msghdr msg;
  436. struct kvec vec;
  437. int len;
  438. /* send SMC Confirm CLC msg */
  439. memset(&cclc, 0, sizeof(cclc));
  440. cclc.hdr.type = SMC_CLC_CONFIRM;
  441. cclc.hdr.version = SMC_CLC_V1; /* SMC version */
  442. if (smc->conn.lgr->is_smcd) {
  443. /* SMC-D specific settings */
  444. memcpy(cclc.hdr.eyecatcher, SMCD_EYECATCHER,
  445. sizeof(SMCD_EYECATCHER));
  446. cclc.hdr.path = SMC_TYPE_D;
  447. cclc.hdr.length = htons(SMCD_CLC_ACCEPT_CONFIRM_LEN);
  448. cclc.gid = conn->lgr->smcd->local_gid;
  449. cclc.token = conn->rmb_desc->token;
  450. cclc.dmbe_size = conn->rmbe_size_short;
  451. cclc.dmbe_idx = 0;
  452. memcpy(&cclc.linkid, conn->lgr->id, SMC_LGR_ID_SIZE);
  453. memcpy(cclc.smcd_trl.eyecatcher, SMCD_EYECATCHER,
  454. sizeof(SMCD_EYECATCHER));
  455. } else {
  456. /* SMC-R specific settings */
  457. link = &conn->lgr->lnk[SMC_SINGLE_LINK];
  458. memcpy(cclc.hdr.eyecatcher, SMC_EYECATCHER,
  459. sizeof(SMC_EYECATCHER));
  460. cclc.hdr.path = SMC_TYPE_R;
  461. cclc.hdr.length = htons(SMCR_CLC_ACCEPT_CONFIRM_LEN);
  462. memcpy(cclc.lcl.id_for_peer, local_systemid,
  463. sizeof(local_systemid));
  464. memcpy(&cclc.lcl.gid, link->gid, SMC_GID_SIZE);
  465. memcpy(&cclc.lcl.mac, &link->smcibdev->mac[link->ibport - 1],
  466. ETH_ALEN);
  467. hton24(cclc.qpn, link->roce_qp->qp_num);
  468. cclc.rmb_rkey =
  469. htonl(conn->rmb_desc->mr_rx[SMC_SINGLE_LINK]->rkey);
  470. cclc.rmbe_idx = 1; /* for now: 1 RMB = 1 RMBE */
  471. cclc.rmbe_alert_token = htonl(conn->alert_token_local);
  472. cclc.qp_mtu = min(link->path_mtu, link->peer_mtu);
  473. cclc.rmbe_size = conn->rmbe_size_short;
  474. cclc.rmb_dma_addr = cpu_to_be64((u64)sg_dma_address
  475. (conn->rmb_desc->sgt[SMC_SINGLE_LINK].sgl));
  476. hton24(cclc.psn, link->psn_initial);
  477. memcpy(cclc.smcr_trl.eyecatcher, SMC_EYECATCHER,
  478. sizeof(SMC_EYECATCHER));
  479. }
  480. memset(&msg, 0, sizeof(msg));
  481. vec.iov_base = &cclc;
  482. vec.iov_len = ntohs(cclc.hdr.length);
  483. len = kernel_sendmsg(smc->clcsock, &msg, &vec, 1,
  484. ntohs(cclc.hdr.length));
  485. if (len < ntohs(cclc.hdr.length)) {
  486. if (len >= 0) {
  487. reason_code = -ENETUNREACH;
  488. smc->sk.sk_err = -reason_code;
  489. } else {
  490. smc->sk.sk_err = smc->clcsock->sk->sk_err;
  491. reason_code = -smc->sk.sk_err;
  492. }
  493. }
  494. return reason_code;
  495. }
  496. /* send CLC ACCEPT message across internal TCP socket */
  497. int smc_clc_send_accept(struct smc_sock *new_smc, int srv_first_contact)
  498. {
  499. struct smc_connection *conn = &new_smc->conn;
  500. struct smc_clc_msg_accept_confirm aclc;
  501. struct smc_link *link;
  502. struct msghdr msg;
  503. struct kvec vec;
  504. int rc = 0;
  505. int len;
  506. memset(&aclc, 0, sizeof(aclc));
  507. aclc.hdr.type = SMC_CLC_ACCEPT;
  508. aclc.hdr.version = SMC_CLC_V1; /* SMC version */
  509. if (srv_first_contact)
  510. aclc.hdr.flag = 1;
  511. if (new_smc->conn.lgr->is_smcd) {
  512. /* SMC-D specific settings */
  513. aclc.hdr.length = htons(SMCD_CLC_ACCEPT_CONFIRM_LEN);
  514. memcpy(aclc.hdr.eyecatcher, SMCD_EYECATCHER,
  515. sizeof(SMCD_EYECATCHER));
  516. aclc.hdr.path = SMC_TYPE_D;
  517. aclc.gid = conn->lgr->smcd->local_gid;
  518. aclc.token = conn->rmb_desc->token;
  519. aclc.dmbe_size = conn->rmbe_size_short;
  520. aclc.dmbe_idx = 0;
  521. memcpy(&aclc.linkid, conn->lgr->id, SMC_LGR_ID_SIZE);
  522. memcpy(aclc.smcd_trl.eyecatcher, SMCD_EYECATCHER,
  523. sizeof(SMCD_EYECATCHER));
  524. } else {
  525. /* SMC-R specific settings */
  526. aclc.hdr.length = htons(SMCR_CLC_ACCEPT_CONFIRM_LEN);
  527. memcpy(aclc.hdr.eyecatcher, SMC_EYECATCHER,
  528. sizeof(SMC_EYECATCHER));
  529. aclc.hdr.path = SMC_TYPE_R;
  530. link = &conn->lgr->lnk[SMC_SINGLE_LINK];
  531. memcpy(aclc.lcl.id_for_peer, local_systemid,
  532. sizeof(local_systemid));
  533. memcpy(&aclc.lcl.gid, link->gid, SMC_GID_SIZE);
  534. memcpy(&aclc.lcl.mac, link->smcibdev->mac[link->ibport - 1],
  535. ETH_ALEN);
  536. hton24(aclc.qpn, link->roce_qp->qp_num);
  537. aclc.rmb_rkey =
  538. htonl(conn->rmb_desc->mr_rx[SMC_SINGLE_LINK]->rkey);
  539. aclc.rmbe_idx = 1; /* as long as 1 RMB = 1 RMBE */
  540. aclc.rmbe_alert_token = htonl(conn->alert_token_local);
  541. aclc.qp_mtu = link->path_mtu;
  542. aclc.rmbe_size = conn->rmbe_size_short,
  543. aclc.rmb_dma_addr = cpu_to_be64((u64)sg_dma_address
  544. (conn->rmb_desc->sgt[SMC_SINGLE_LINK].sgl));
  545. hton24(aclc.psn, link->psn_initial);
  546. memcpy(aclc.smcr_trl.eyecatcher, SMC_EYECATCHER,
  547. sizeof(SMC_EYECATCHER));
  548. }
  549. memset(&msg, 0, sizeof(msg));
  550. vec.iov_base = &aclc;
  551. vec.iov_len = ntohs(aclc.hdr.length);
  552. len = kernel_sendmsg(new_smc->clcsock, &msg, &vec, 1,
  553. ntohs(aclc.hdr.length));
  554. if (len < ntohs(aclc.hdr.length)) {
  555. if (len >= 0)
  556. new_smc->sk.sk_err = EPROTO;
  557. else
  558. new_smc->sk.sk_err = new_smc->clcsock->sk->sk_err;
  559. rc = sock_error(&new_smc->sk);
  560. }
  561. return rc;
  562. }