• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * Shared Memory Communications over RDMA (SMC-R) and RoCE
4  *
5  * Monitoring SMC transport protocol sockets
6  *
7  * Copyright IBM Corp. 2016
8  *
9  * Author(s):  Ursula Braun <ubraun@linux.vnet.ibm.com>
10  */
11 
12 #include <linux/kernel.h>
13 #include <linux/module.h>
14 #include <linux/types.h>
15 #include <linux/init.h>
16 #include <linux/sock_diag.h>
17 #include <linux/inet_diag.h>
18 #include <linux/smc_diag.h>
19 #include <net/netlink.h>
20 #include <net/smc.h>
21 
22 #include "smc.h"
23 #include "smc_core.h"
24 
25 struct smc_diag_dump_ctx {
26 	int pos[2];
27 };
28 
smc_dump_context(struct netlink_callback * cb)29 static struct smc_diag_dump_ctx *smc_dump_context(struct netlink_callback *cb)
30 {
31 	return (struct smc_diag_dump_ctx *)cb->ctx;
32 }
33 
smc_gid_be16_convert(__u8 * buf,u8 * gid_raw)34 static void smc_gid_be16_convert(__u8 *buf, u8 *gid_raw)
35 {
36 	sprintf(buf, "%04x:%04x:%04x:%04x:%04x:%04x:%04x:%04x",
37 		be16_to_cpu(((__be16 *)gid_raw)[0]),
38 		be16_to_cpu(((__be16 *)gid_raw)[1]),
39 		be16_to_cpu(((__be16 *)gid_raw)[2]),
40 		be16_to_cpu(((__be16 *)gid_raw)[3]),
41 		be16_to_cpu(((__be16 *)gid_raw)[4]),
42 		be16_to_cpu(((__be16 *)gid_raw)[5]),
43 		be16_to_cpu(((__be16 *)gid_raw)[6]),
44 		be16_to_cpu(((__be16 *)gid_raw)[7]));
45 }
46 
smc_diag_msg_common_fill(struct smc_diag_msg * r,struct sock * sk)47 static void smc_diag_msg_common_fill(struct smc_diag_msg *r, struct sock *sk)
48 {
49 	struct smc_sock *smc = smc_sk(sk);
50 
51 	memset(r, 0, sizeof(*r));
52 	r->diag_family = sk->sk_family;
53 	sock_diag_save_cookie(sk, r->id.idiag_cookie);
54 	if (!smc->clcsock)
55 		return;
56 	r->id.idiag_sport = htons(smc->clcsock->sk->sk_num);
57 	r->id.idiag_dport = smc->clcsock->sk->sk_dport;
58 	r->id.idiag_if = smc->clcsock->sk->sk_bound_dev_if;
59 	if (sk->sk_protocol == SMCPROTO_SMC) {
60 		r->id.idiag_src[0] = smc->clcsock->sk->sk_rcv_saddr;
61 		r->id.idiag_dst[0] = smc->clcsock->sk->sk_daddr;
62 #if IS_ENABLED(CONFIG_IPV6)
63 	} else if (sk->sk_protocol == SMCPROTO_SMC6) {
64 		memcpy(&r->id.idiag_src, &smc->clcsock->sk->sk_v6_rcv_saddr,
65 		       sizeof(smc->clcsock->sk->sk_v6_rcv_saddr));
66 		memcpy(&r->id.idiag_dst, &smc->clcsock->sk->sk_v6_daddr,
67 		       sizeof(smc->clcsock->sk->sk_v6_daddr));
68 #endif
69 	}
70 }
71 
smc_diag_msg_attrs_fill(struct sock * sk,struct sk_buff * skb,struct smc_diag_msg * r,struct user_namespace * user_ns)72 static int smc_diag_msg_attrs_fill(struct sock *sk, struct sk_buff *skb,
73 				   struct smc_diag_msg *r,
74 				   struct user_namespace *user_ns)
75 {
76 	if (nla_put_u8(skb, SMC_DIAG_SHUTDOWN, sk->sk_shutdown))
77 		return 1;
78 
79 	r->diag_uid = from_kuid_munged(user_ns, sock_i_uid(sk));
80 	r->diag_inode = sock_i_ino(sk);
81 	return 0;
82 }
83 
__smc_diag_dump(struct sock * sk,struct sk_buff * skb,struct netlink_callback * cb,const struct smc_diag_req * req,struct nlattr * bc)84 static int __smc_diag_dump(struct sock *sk, struct sk_buff *skb,
85 			   struct netlink_callback *cb,
86 			   const struct smc_diag_req *req,
87 			   struct nlattr *bc)
88 {
89 	struct smc_sock *smc = smc_sk(sk);
90 	struct smc_diag_fallback fallback;
91 	struct user_namespace *user_ns;
92 	struct smc_diag_msg *r;
93 	struct nlmsghdr *nlh;
94 
95 	nlh = nlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq,
96 			cb->nlh->nlmsg_type, sizeof(*r), NLM_F_MULTI);
97 	if (!nlh)
98 		return -EMSGSIZE;
99 
100 	r = nlmsg_data(nlh);
101 	smc_diag_msg_common_fill(r, sk);
102 	r->diag_state = sk->sk_state;
103 	if (smc->use_fallback)
104 		r->diag_mode = SMC_DIAG_MODE_FALLBACK_TCP;
105 	else if (smc->conn.lgr && smc->conn.lgr->is_smcd)
106 		r->diag_mode = SMC_DIAG_MODE_SMCD;
107 	else
108 		r->diag_mode = SMC_DIAG_MODE_SMCR;
109 	user_ns = sk_user_ns(NETLINK_CB(cb->skb).sk);
110 	if (smc_diag_msg_attrs_fill(sk, skb, r, user_ns))
111 		goto errout;
112 
113 	fallback.reason = smc->fallback_rsn;
114 	fallback.peer_diagnosis = smc->peer_diagnosis;
115 	if (nla_put(skb, SMC_DIAG_FALLBACK, sizeof(fallback), &fallback) < 0)
116 		goto errout;
117 
118 	if ((req->diag_ext & (1 << (SMC_DIAG_CONNINFO - 1))) &&
119 	    smc->conn.alert_token_local) {
120 		struct smc_connection *conn = &smc->conn;
121 		struct smc_diag_conninfo cinfo = {
122 			.token = conn->alert_token_local,
123 			.sndbuf_size = conn->sndbuf_desc ?
124 				conn->sndbuf_desc->len : 0,
125 			.rmbe_size = conn->rmb_desc ? conn->rmb_desc->len : 0,
126 			.peer_rmbe_size = conn->peer_rmbe_size,
127 
128 			.rx_prod.wrap = conn->local_rx_ctrl.prod.wrap,
129 			.rx_prod.count = conn->local_rx_ctrl.prod.count,
130 			.rx_cons.wrap = conn->local_rx_ctrl.cons.wrap,
131 			.rx_cons.count = conn->local_rx_ctrl.cons.count,
132 
133 			.tx_prod.wrap = conn->local_tx_ctrl.prod.wrap,
134 			.tx_prod.count = conn->local_tx_ctrl.prod.count,
135 			.tx_cons.wrap = conn->local_tx_ctrl.cons.wrap,
136 			.tx_cons.count = conn->local_tx_ctrl.cons.count,
137 
138 			.tx_prod_flags =
139 				*(u8 *)&conn->local_tx_ctrl.prod_flags,
140 			.tx_conn_state_flags =
141 				*(u8 *)&conn->local_tx_ctrl.conn_state_flags,
142 			.rx_prod_flags = *(u8 *)&conn->local_rx_ctrl.prod_flags,
143 			.rx_conn_state_flags =
144 				*(u8 *)&conn->local_rx_ctrl.conn_state_flags,
145 
146 			.tx_prep.wrap = conn->tx_curs_prep.wrap,
147 			.tx_prep.count = conn->tx_curs_prep.count,
148 			.tx_sent.wrap = conn->tx_curs_sent.wrap,
149 			.tx_sent.count = conn->tx_curs_sent.count,
150 			.tx_fin.wrap = conn->tx_curs_fin.wrap,
151 			.tx_fin.count = conn->tx_curs_fin.count,
152 		};
153 
154 		if (nla_put(skb, SMC_DIAG_CONNINFO, sizeof(cinfo), &cinfo) < 0)
155 			goto errout;
156 	}
157 
158 	if (smc->conn.lgr && !smc->conn.lgr->is_smcd &&
159 	    (req->diag_ext & (1 << (SMC_DIAG_LGRINFO - 1))) &&
160 	    !list_empty(&smc->conn.lgr->list)) {
161 		struct smc_diag_lgrinfo linfo = {
162 			.role = smc->conn.lgr->role,
163 			.lnk[0].ibport = smc->conn.lgr->lnk[0].ibport,
164 			.lnk[0].link_id = smc->conn.lgr->lnk[0].link_id,
165 		};
166 
167 		memcpy(linfo.lnk[0].ibname,
168 		       smc->conn.lgr->lnk[0].smcibdev->ibdev->name,
169 		       sizeof(smc->conn.lgr->lnk[0].smcibdev->ibdev->name));
170 		smc_gid_be16_convert(linfo.lnk[0].gid,
171 				     smc->conn.lgr->lnk[0].gid);
172 		smc_gid_be16_convert(linfo.lnk[0].peer_gid,
173 				     smc->conn.lgr->lnk[0].peer_gid);
174 
175 		if (nla_put(skb, SMC_DIAG_LGRINFO, sizeof(linfo), &linfo) < 0)
176 			goto errout;
177 	}
178 	if (smc->conn.lgr && smc->conn.lgr->is_smcd &&
179 	    (req->diag_ext & (1 << (SMC_DIAG_DMBINFO - 1))) &&
180 	    !list_empty(&smc->conn.lgr->list)) {
181 		struct smc_connection *conn = &smc->conn;
182 		struct smcd_diag_dmbinfo dinfo;
183 
184 		memset(&dinfo, 0, sizeof(dinfo));
185 
186 		dinfo.linkid = *((u32 *)conn->lgr->id);
187 		dinfo.peer_gid = conn->lgr->peer_gid;
188 		dinfo.my_gid = conn->lgr->smcd->local_gid;
189 		dinfo.token = conn->rmb_desc->token;
190 		dinfo.peer_token = conn->peer_token;
191 
192 		if (nla_put(skb, SMC_DIAG_DMBINFO, sizeof(dinfo), &dinfo) < 0)
193 			goto errout;
194 	}
195 
196 	nlmsg_end(skb, nlh);
197 	return 0;
198 
199 errout:
200 	nlmsg_cancel(skb, nlh);
201 	return -EMSGSIZE;
202 }
203 
smc_diag_dump_proto(struct proto * prot,struct sk_buff * skb,struct netlink_callback * cb,int p_type)204 static int smc_diag_dump_proto(struct proto *prot, struct sk_buff *skb,
205 			       struct netlink_callback *cb, int p_type)
206 {
207 	struct smc_diag_dump_ctx *cb_ctx = smc_dump_context(cb);
208 	struct net *net = sock_net(skb->sk);
209 	int snum = cb_ctx->pos[p_type];
210 	struct nlattr *bc = NULL;
211 	struct hlist_head *head;
212 	int rc = 0, num = 0;
213 	struct sock *sk;
214 
215 	read_lock(&prot->h.smc_hash->lock);
216 	head = &prot->h.smc_hash->ht;
217 	if (hlist_empty(head))
218 		goto out;
219 
220 	sk_for_each(sk, head) {
221 		if (!net_eq(sock_net(sk), net))
222 			continue;
223 		if (num < snum)
224 			goto next;
225 		rc = __smc_diag_dump(sk, skb, cb, nlmsg_data(cb->nlh), bc);
226 		if (rc < 0)
227 			goto out;
228 next:
229 		num++;
230 	}
231 
232 out:
233 	read_unlock(&prot->h.smc_hash->lock);
234 	cb_ctx->pos[p_type] = num;
235 	return rc;
236 }
237 
smc_diag_dump(struct sk_buff * skb,struct netlink_callback * cb)238 static int smc_diag_dump(struct sk_buff *skb, struct netlink_callback *cb)
239 {
240 	int rc = 0;
241 
242 	rc = smc_diag_dump_proto(&smc_proto, skb, cb, SMCPROTO_SMC);
243 	if (!rc)
244 		smc_diag_dump_proto(&smc_proto6, skb, cb, SMCPROTO_SMC6);
245 	return skb->len;
246 }
247 
smc_diag_handler_dump(struct sk_buff * skb,struct nlmsghdr * h)248 static int smc_diag_handler_dump(struct sk_buff *skb, struct nlmsghdr *h)
249 {
250 	struct net *net = sock_net(skb->sk);
251 
252 	if (h->nlmsg_type == SOCK_DIAG_BY_FAMILY &&
253 	    h->nlmsg_flags & NLM_F_DUMP) {
254 		{
255 			struct netlink_dump_control c = {
256 				.dump = smc_diag_dump,
257 				.min_dump_alloc = SKB_WITH_OVERHEAD(32768),
258 			};
259 			return netlink_dump_start(net->diag_nlsk, skb, h, &c);
260 		}
261 	}
262 	return 0;
263 }
264 
265 static const struct sock_diag_handler smc_diag_handler = {
266 	.family = AF_SMC,
267 	.dump = smc_diag_handler_dump,
268 };
269 
smc_diag_init(void)270 static int __init smc_diag_init(void)
271 {
272 	return sock_diag_register(&smc_diag_handler);
273 }
274 
smc_diag_exit(void)275 static void __exit smc_diag_exit(void)
276 {
277 	sock_diag_unregister(&smc_diag_handler);
278 }
279 
280 module_init(smc_diag_init);
281 module_exit(smc_diag_exit);
282 MODULE_LICENSE("GPL");
283 MODULE_ALIAS_NET_PF_PROTO_TYPE(PF_NETLINK, NETLINK_SOCK_DIAG, 43 /* AF_SMC */);
284