1 /*
2 * GPL HEADER START
3 *
4 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
5 *
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License version 2 only,
8 * as published by the Free Software Foundation.
9 *
10 * This program is distributed in the hope that it will be useful, but
11 * WITHOUT ANY WARRANTY; without even the implied warranty of
12 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
13 * General Public License version 2 for more details (a copy is included
14 * in the LICENSE file that accompanied this code).
15 *
16 * You should have received a copy of the GNU General Public License
17 * version 2 along with this program; If not, see
18 * http://www.sun.com/software/products/lustre/docs/GPLv2.pdf
19 *
20 * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa Clara,
21 * CA 95054 USA or visit www.sun.com if you need additional information or
22 * have any questions.
23 *
24 * GPL HEADER END
25 */
26 /*
27 * Copyright (c) 2007, 2010, Oracle and/or its affiliates. All rights reserved.
28 * Use is subject to license terms.
29 *
30 * Copyright (c) 2011, 2012, Intel Corporation.
31 */
32 /*
33 * This file is part of Lustre, http://www.lustre.org/
34 * Lustre is a trademark of Sun Microsystems, Inc.
35 *
36 * lustre/ptlrpc/sec_null.c
37 *
38 * Author: Eric Mei <ericm@clusterfs.com>
39 */
40
41 #define DEBUG_SUBSYSTEM S_SEC
42
43 #include "../include/obd_support.h"
44 #include "../include/obd_cksum.h"
45 #include "../include/obd_class.h"
46 #include "../include/lustre_net.h"
47 #include "../include/lustre_sec.h"
48
49 #include "ptlrpc_internal.h"
50
51 static struct ptlrpc_sec_policy null_policy;
52 static struct ptlrpc_sec null_sec;
53 static struct ptlrpc_cli_ctx null_cli_ctx;
54 static struct ptlrpc_svc_ctx null_svc_ctx;
55
56 /*
57 * we can temporarily use the topmost 8-bits of lm_secflvr to identify
58 * the source sec part.
59 */
60 static inline
null_encode_sec_part(struct lustre_msg * msg,enum lustre_sec_part sp)61 void null_encode_sec_part(struct lustre_msg *msg, enum lustre_sec_part sp)
62 {
63 msg->lm_secflvr |= (((__u32) sp) & 0xFF) << 24;
64 }
65
66 static inline
null_decode_sec_part(struct lustre_msg * msg)67 enum lustre_sec_part null_decode_sec_part(struct lustre_msg *msg)
68 {
69 return (msg->lm_secflvr >> 24) & 0xFF;
70 }
71
null_ctx_refresh(struct ptlrpc_cli_ctx * ctx)72 static int null_ctx_refresh(struct ptlrpc_cli_ctx *ctx)
73 {
74 /* should never reach here */
75 LBUG();
76 return 0;
77 }
78
79 static
null_ctx_sign(struct ptlrpc_cli_ctx * ctx,struct ptlrpc_request * req)80 int null_ctx_sign(struct ptlrpc_cli_ctx *ctx, struct ptlrpc_request *req)
81 {
82 req->rq_reqbuf->lm_secflvr = SPTLRPC_FLVR_NULL;
83
84 if (!req->rq_import->imp_dlm_fake) {
85 struct obd_device *obd = req->rq_import->imp_obd;
86
87 null_encode_sec_part(req->rq_reqbuf,
88 obd->u.cli.cl_sp_me);
89 }
90 req->rq_reqdata_len = req->rq_reqlen;
91 return 0;
92 }
93
94 static
null_ctx_verify(struct ptlrpc_cli_ctx * ctx,struct ptlrpc_request * req)95 int null_ctx_verify(struct ptlrpc_cli_ctx *ctx, struct ptlrpc_request *req)
96 {
97 __u32 cksums, cksumc;
98
99 LASSERT(req->rq_repdata);
100
101 req->rq_repmsg = req->rq_repdata;
102 req->rq_replen = req->rq_repdata_len;
103
104 if (req->rq_early) {
105 cksums = lustre_msg_get_cksum(req->rq_repdata);
106 cksumc = lustre_msg_calc_cksum(req->rq_repmsg);
107 if (cksumc != cksums) {
108 CDEBUG(D_SEC,
109 "early reply checksum mismatch: %08x != %08x\n",
110 cksumc, cksums);
111 return -EINVAL;
112 }
113 }
114
115 return 0;
116 }
117
118 static
null_create_sec(struct obd_import * imp,struct ptlrpc_svc_ctx * svc_ctx,struct sptlrpc_flavor * sf)119 struct ptlrpc_sec *null_create_sec(struct obd_import *imp,
120 struct ptlrpc_svc_ctx *svc_ctx,
121 struct sptlrpc_flavor *sf)
122 {
123 LASSERT(SPTLRPC_FLVR_POLICY(sf->sf_rpc) == SPTLRPC_POLICY_NULL);
124
125 /* general layer has take a module reference for us, because we never
126 * really destroy the sec, simply release the reference here.
127 */
128 sptlrpc_policy_put(&null_policy);
129 return &null_sec;
130 }
131
132 static
null_destroy_sec(struct ptlrpc_sec * sec)133 void null_destroy_sec(struct ptlrpc_sec *sec)
134 {
135 LASSERT(sec == &null_sec);
136 }
137
138 static
null_lookup_ctx(struct ptlrpc_sec * sec,struct vfs_cred * vcred,int create,int remove_dead)139 struct ptlrpc_cli_ctx *null_lookup_ctx(struct ptlrpc_sec *sec,
140 struct vfs_cred *vcred,
141 int create, int remove_dead)
142 {
143 atomic_inc(&null_cli_ctx.cc_refcount);
144 return &null_cli_ctx;
145 }
146
147 static
null_flush_ctx_cache(struct ptlrpc_sec * sec,uid_t uid,int grace,int force)148 int null_flush_ctx_cache(struct ptlrpc_sec *sec,
149 uid_t uid,
150 int grace, int force)
151 {
152 return 0;
153 }
154
155 static
null_alloc_reqbuf(struct ptlrpc_sec * sec,struct ptlrpc_request * req,int msgsize)156 int null_alloc_reqbuf(struct ptlrpc_sec *sec,
157 struct ptlrpc_request *req,
158 int msgsize)
159 {
160 if (!req->rq_reqbuf) {
161 int alloc_size = size_roundup_power2(msgsize);
162
163 LASSERT(!req->rq_pool);
164 req->rq_reqbuf = libcfs_kvzalloc(alloc_size, GFP_NOFS);
165 if (!req->rq_reqbuf)
166 return -ENOMEM;
167
168 req->rq_reqbuf_len = alloc_size;
169 } else {
170 LASSERT(req->rq_pool);
171 LASSERT(req->rq_reqbuf_len >= msgsize);
172 memset(req->rq_reqbuf, 0, msgsize);
173 }
174
175 req->rq_reqmsg = req->rq_reqbuf;
176 return 0;
177 }
178
179 static
null_free_reqbuf(struct ptlrpc_sec * sec,struct ptlrpc_request * req)180 void null_free_reqbuf(struct ptlrpc_sec *sec,
181 struct ptlrpc_request *req)
182 {
183 if (!req->rq_pool) {
184 LASSERTF(req->rq_reqmsg == req->rq_reqbuf,
185 "req %p: reqmsg %p is not reqbuf %p in null sec\n",
186 req, req->rq_reqmsg, req->rq_reqbuf);
187 LASSERTF(req->rq_reqbuf_len >= req->rq_reqlen,
188 "req %p: reqlen %d should smaller than buflen %d\n",
189 req, req->rq_reqlen, req->rq_reqbuf_len);
190
191 kvfree(req->rq_reqbuf);
192 req->rq_reqbuf = NULL;
193 req->rq_reqbuf_len = 0;
194 }
195 }
196
197 static
null_alloc_repbuf(struct ptlrpc_sec * sec,struct ptlrpc_request * req,int msgsize)198 int null_alloc_repbuf(struct ptlrpc_sec *sec,
199 struct ptlrpc_request *req,
200 int msgsize)
201 {
202 /* add space for early replied */
203 msgsize += lustre_msg_early_size();
204
205 msgsize = size_roundup_power2(msgsize);
206
207 req->rq_repbuf = libcfs_kvzalloc(msgsize, GFP_NOFS);
208 if (!req->rq_repbuf)
209 return -ENOMEM;
210
211 req->rq_repbuf_len = msgsize;
212 return 0;
213 }
214
215 static
null_free_repbuf(struct ptlrpc_sec * sec,struct ptlrpc_request * req)216 void null_free_repbuf(struct ptlrpc_sec *sec,
217 struct ptlrpc_request *req)
218 {
219 LASSERT(req->rq_repbuf);
220
221 kvfree(req->rq_repbuf);
222 req->rq_repbuf = NULL;
223 req->rq_repbuf_len = 0;
224 }
225
226 static
null_enlarge_reqbuf(struct ptlrpc_sec * sec,struct ptlrpc_request * req,int segment,int newsize)227 int null_enlarge_reqbuf(struct ptlrpc_sec *sec,
228 struct ptlrpc_request *req,
229 int segment, int newsize)
230 {
231 struct lustre_msg *newbuf;
232 struct lustre_msg *oldbuf = req->rq_reqmsg;
233 int oldsize, newmsg_size, alloc_size;
234
235 LASSERT(req->rq_reqbuf);
236 LASSERT(req->rq_reqbuf == req->rq_reqmsg);
237 LASSERT(req->rq_reqbuf_len >= req->rq_reqlen);
238 LASSERT(req->rq_reqlen == lustre_packed_msg_size(oldbuf));
239
240 /* compute new message size */
241 oldsize = req->rq_reqbuf->lm_buflens[segment];
242 req->rq_reqbuf->lm_buflens[segment] = newsize;
243 newmsg_size = lustre_packed_msg_size(oldbuf);
244 req->rq_reqbuf->lm_buflens[segment] = oldsize;
245
246 /* request from pool should always have enough buffer */
247 LASSERT(!req->rq_pool || req->rq_reqbuf_len >= newmsg_size);
248
249 if (req->rq_reqbuf_len < newmsg_size) {
250 alloc_size = size_roundup_power2(newmsg_size);
251
252 newbuf = libcfs_kvzalloc(alloc_size, GFP_NOFS);
253 if (newbuf == NULL)
254 return -ENOMEM;
255
256 /* Must lock this, so that otherwise unprotected change of
257 * rq_reqmsg is not racing with parallel processing of
258 * imp_replay_list traversing threads. See LU-3333
259 * This is a bandaid at best, we really need to deal with this
260 * in request enlarging code before unpacking that's already
261 * there */
262 if (req->rq_import)
263 spin_lock(&req->rq_import->imp_lock);
264 memcpy(newbuf, req->rq_reqbuf, req->rq_reqlen);
265
266 kvfree(req->rq_reqbuf);
267 req->rq_reqbuf = req->rq_reqmsg = newbuf;
268 req->rq_reqbuf_len = alloc_size;
269
270 if (req->rq_import)
271 spin_unlock(&req->rq_import->imp_lock);
272 }
273
274 _sptlrpc_enlarge_msg_inplace(req->rq_reqmsg, segment, newsize);
275 req->rq_reqlen = newmsg_size;
276
277 return 0;
278 }
279
280 static struct ptlrpc_svc_ctx null_svc_ctx = {
281 .sc_refcount = ATOMIC_INIT(1),
282 .sc_policy = &null_policy,
283 };
284
285 static
null_accept(struct ptlrpc_request * req)286 int null_accept(struct ptlrpc_request *req)
287 {
288 LASSERT(SPTLRPC_FLVR_POLICY(req->rq_flvr.sf_rpc) ==
289 SPTLRPC_POLICY_NULL);
290
291 if (req->rq_flvr.sf_rpc != SPTLRPC_FLVR_NULL) {
292 CERROR("Invalid rpc flavor 0x%x\n", req->rq_flvr.sf_rpc);
293 return SECSVC_DROP;
294 }
295
296 req->rq_sp_from = null_decode_sec_part(req->rq_reqbuf);
297
298 req->rq_reqmsg = req->rq_reqbuf;
299 req->rq_reqlen = req->rq_reqdata_len;
300
301 req->rq_svc_ctx = &null_svc_ctx;
302 atomic_inc(&req->rq_svc_ctx->sc_refcount);
303
304 return SECSVC_OK;
305 }
306
307 static
null_alloc_rs(struct ptlrpc_request * req,int msgsize)308 int null_alloc_rs(struct ptlrpc_request *req, int msgsize)
309 {
310 struct ptlrpc_reply_state *rs;
311 int rs_size = sizeof(*rs) + msgsize;
312
313 LASSERT(msgsize % 8 == 0);
314
315 rs = req->rq_reply_state;
316
317 if (rs) {
318 /* pre-allocated */
319 LASSERT(rs->rs_size >= rs_size);
320 } else {
321 rs = libcfs_kvzalloc(rs_size, GFP_NOFS);
322 if (rs == NULL)
323 return -ENOMEM;
324
325 rs->rs_size = rs_size;
326 }
327
328 rs->rs_svc_ctx = req->rq_svc_ctx;
329 atomic_inc(&req->rq_svc_ctx->sc_refcount);
330
331 rs->rs_repbuf = (struct lustre_msg *) (rs + 1);
332 rs->rs_repbuf_len = rs_size - sizeof(*rs);
333 rs->rs_msg = rs->rs_repbuf;
334
335 req->rq_reply_state = rs;
336 return 0;
337 }
338
339 static
null_free_rs(struct ptlrpc_reply_state * rs)340 void null_free_rs(struct ptlrpc_reply_state *rs)
341 {
342 LASSERT_ATOMIC_GT(&rs->rs_svc_ctx->sc_refcount, 1);
343 atomic_dec(&rs->rs_svc_ctx->sc_refcount);
344
345 if (!rs->rs_prealloc)
346 kvfree(rs);
347 }
348
349 static
null_authorize(struct ptlrpc_request * req)350 int null_authorize(struct ptlrpc_request *req)
351 {
352 struct ptlrpc_reply_state *rs = req->rq_reply_state;
353
354 LASSERT(rs);
355
356 rs->rs_repbuf->lm_secflvr = SPTLRPC_FLVR_NULL;
357 rs->rs_repdata_len = req->rq_replen;
358
359 if (likely(req->rq_packed_final)) {
360 if (lustre_msghdr_get_flags(req->rq_reqmsg) & MSGHDR_AT_SUPPORT)
361 req->rq_reply_off = lustre_msg_early_size();
362 else
363 req->rq_reply_off = 0;
364 } else {
365 __u32 cksum;
366
367 cksum = lustre_msg_calc_cksum(rs->rs_repbuf);
368 lustre_msg_set_cksum(rs->rs_repbuf, cksum);
369 req->rq_reply_off = 0;
370 }
371
372 return 0;
373 }
374
375 static struct ptlrpc_ctx_ops null_ctx_ops = {
376 .refresh = null_ctx_refresh,
377 .sign = null_ctx_sign,
378 .verify = null_ctx_verify,
379 };
380
381 static struct ptlrpc_sec_cops null_sec_cops = {
382 .create_sec = null_create_sec,
383 .destroy_sec = null_destroy_sec,
384 .lookup_ctx = null_lookup_ctx,
385 .flush_ctx_cache = null_flush_ctx_cache,
386 .alloc_reqbuf = null_alloc_reqbuf,
387 .alloc_repbuf = null_alloc_repbuf,
388 .free_reqbuf = null_free_reqbuf,
389 .free_repbuf = null_free_repbuf,
390 .enlarge_reqbuf = null_enlarge_reqbuf,
391 };
392
393 static struct ptlrpc_sec_sops null_sec_sops = {
394 .accept = null_accept,
395 .alloc_rs = null_alloc_rs,
396 .authorize = null_authorize,
397 .free_rs = null_free_rs,
398 };
399
400 static struct ptlrpc_sec_policy null_policy = {
401 .sp_owner = THIS_MODULE,
402 .sp_name = "sec.null",
403 .sp_policy = SPTLRPC_POLICY_NULL,
404 .sp_cops = &null_sec_cops,
405 .sp_sops = &null_sec_sops,
406 };
407
null_init_internal(void)408 static void null_init_internal(void)
409 {
410 static HLIST_HEAD(__list);
411
412 null_sec.ps_policy = &null_policy;
413 atomic_set(&null_sec.ps_refcount, 1); /* always busy */
414 null_sec.ps_id = -1;
415 null_sec.ps_import = NULL;
416 null_sec.ps_flvr.sf_rpc = SPTLRPC_FLVR_NULL;
417 null_sec.ps_flvr.sf_flags = 0;
418 null_sec.ps_part = LUSTRE_SP_ANY;
419 null_sec.ps_dying = 0;
420 spin_lock_init(&null_sec.ps_lock);
421 atomic_set(&null_sec.ps_nctx, 1); /* for "null_cli_ctx" */
422 INIT_LIST_HEAD(&null_sec.ps_gc_list);
423 null_sec.ps_gc_interval = 0;
424 null_sec.ps_gc_next = 0;
425
426 hlist_add_head(&null_cli_ctx.cc_cache, &__list);
427 atomic_set(&null_cli_ctx.cc_refcount, 1); /* for hash */
428 null_cli_ctx.cc_sec = &null_sec;
429 null_cli_ctx.cc_ops = &null_ctx_ops;
430 null_cli_ctx.cc_expire = 0;
431 null_cli_ctx.cc_flags = PTLRPC_CTX_CACHED | PTLRPC_CTX_ETERNAL |
432 PTLRPC_CTX_UPTODATE;
433 null_cli_ctx.cc_vcred.vc_uid = 0;
434 spin_lock_init(&null_cli_ctx.cc_lock);
435 INIT_LIST_HEAD(&null_cli_ctx.cc_req_list);
436 INIT_LIST_HEAD(&null_cli_ctx.cc_gc_chain);
437 }
438
sptlrpc_null_init(void)439 int sptlrpc_null_init(void)
440 {
441 int rc;
442
443 null_init_internal();
444
445 rc = sptlrpc_register_policy(&null_policy);
446 if (rc)
447 CERROR("failed to register %s: %d\n", null_policy.sp_name, rc);
448
449 return rc;
450 }
451
sptlrpc_null_fini(void)452 void sptlrpc_null_fini(void)
453 {
454 int rc;
455
456 rc = sptlrpc_unregister_policy(&null_policy);
457 if (rc)
458 CERROR("failed to unregister %s: %d\n",
459 null_policy.sp_name, rc);
460 }
461