1 /*
2 FUSE: Filesystem in Userspace
3 Copyright (C) 2001-2008 Miklos Szeredi <miklos@szeredi.hu>
4
5 This program can be distributed under the terms of the GNU GPL.
6 See the file COPYING.
7 */
8
9 #include "fuse_i.h"
10
11 #include <linux/pagemap.h>
12 #include <linux/file.h>
13 #include <linux/sched.h>
14 #include <linux/namei.h>
15 #include <linux/slab.h>
16
17 #if BITS_PER_LONG >= 64
fuse_dentry_settime(struct dentry * entry,u64 time)18 static inline void fuse_dentry_settime(struct dentry *entry, u64 time)
19 {
20 entry->d_time = time;
21 }
22
fuse_dentry_time(struct dentry * entry)23 static inline u64 fuse_dentry_time(struct dentry *entry)
24 {
25 return entry->d_time;
26 }
27 #else
28 /*
29 * On 32 bit archs store the high 32 bits of time in d_fsdata
30 */
fuse_dentry_settime(struct dentry * entry,u64 time)31 static void fuse_dentry_settime(struct dentry *entry, u64 time)
32 {
33 entry->d_time = time;
34 entry->d_fsdata = (void *) (unsigned long) (time >> 32);
35 }
36
fuse_dentry_time(struct dentry * entry)37 static u64 fuse_dentry_time(struct dentry *entry)
38 {
39 return (u64) entry->d_time +
40 ((u64) (unsigned long) entry->d_fsdata << 32);
41 }
42 #endif
43
44 /*
45 * FUSE caches dentries and attributes with separate timeout. The
46 * time in jiffies until the dentry/attributes are valid is stored in
47 * dentry->d_time and fuse_inode->i_time respectively.
48 */
49
50 /*
51 * Calculate the time in jiffies until a dentry/attributes are valid
52 */
time_to_jiffies(unsigned long sec,unsigned long nsec)53 static u64 time_to_jiffies(unsigned long sec, unsigned long nsec)
54 {
55 if (sec || nsec) {
56 struct timespec ts = {sec, nsec};
57 return get_jiffies_64() + timespec_to_jiffies(&ts);
58 } else
59 return 0;
60 }
61
62 /*
63 * Set dentry and possibly attribute timeouts from the lookup/mk*
64 * replies
65 */
fuse_change_entry_timeout(struct dentry * entry,struct fuse_entry_out * o)66 static void fuse_change_entry_timeout(struct dentry *entry,
67 struct fuse_entry_out *o)
68 {
69 fuse_dentry_settime(entry,
70 time_to_jiffies(o->entry_valid, o->entry_valid_nsec));
71 }
72
attr_timeout(struct fuse_attr_out * o)73 static u64 attr_timeout(struct fuse_attr_out *o)
74 {
75 return time_to_jiffies(o->attr_valid, o->attr_valid_nsec);
76 }
77
entry_attr_timeout(struct fuse_entry_out * o)78 static u64 entry_attr_timeout(struct fuse_entry_out *o)
79 {
80 return time_to_jiffies(o->attr_valid, o->attr_valid_nsec);
81 }
82
83 /*
84 * Mark the attributes as stale, so that at the next call to
85 * ->getattr() they will be fetched from userspace
86 */
fuse_invalidate_attr(struct inode * inode)87 void fuse_invalidate_attr(struct inode *inode)
88 {
89 get_fuse_inode(inode)->i_time = 0;
90 }
91
92 /*
93 * Just mark the entry as stale, so that a next attempt to look it up
94 * will result in a new lookup call to userspace
95 *
96 * This is called when a dentry is about to become negative and the
97 * timeout is unknown (unlink, rmdir, rename and in some cases
98 * lookup)
99 */
fuse_invalidate_entry_cache(struct dentry * entry)100 void fuse_invalidate_entry_cache(struct dentry *entry)
101 {
102 fuse_dentry_settime(entry, 0);
103 }
104
105 /*
106 * Same as fuse_invalidate_entry_cache(), but also try to remove the
107 * dentry from the hash
108 */
fuse_invalidate_entry(struct dentry * entry)109 static void fuse_invalidate_entry(struct dentry *entry)
110 {
111 d_invalidate(entry);
112 fuse_invalidate_entry_cache(entry);
113 }
114
fuse_lookup_init(struct fuse_conn * fc,struct fuse_req * req,u64 nodeid,struct qstr * name,struct fuse_entry_out * outarg)115 static void fuse_lookup_init(struct fuse_conn *fc, struct fuse_req *req,
116 u64 nodeid, struct qstr *name,
117 struct fuse_entry_out *outarg)
118 {
119 memset(outarg, 0, sizeof(struct fuse_entry_out));
120 req->in.h.opcode = FUSE_LOOKUP;
121 req->in.h.nodeid = nodeid;
122 req->in.numargs = 1;
123 req->in.args[0].size = name->len + 1;
124 req->in.args[0].value = name->name;
125 req->out.numargs = 1;
126 if (fc->minor < 9)
127 req->out.args[0].size = FUSE_COMPAT_ENTRY_OUT_SIZE;
128 else
129 req->out.args[0].size = sizeof(struct fuse_entry_out);
130 req->out.args[0].value = outarg;
131 }
132
fuse_get_attr_version(struct fuse_conn * fc)133 u64 fuse_get_attr_version(struct fuse_conn *fc)
134 {
135 u64 curr_version;
136
137 /*
138 * The spin lock isn't actually needed on 64bit archs, but we
139 * don't yet care too much about such optimizations.
140 */
141 spin_lock(&fc->lock);
142 curr_version = fc->attr_version;
143 spin_unlock(&fc->lock);
144
145 return curr_version;
146 }
147
148 /*
149 * Check whether the dentry is still valid
150 *
151 * If the entry validity timeout has expired and the dentry is
152 * positive, try to redo the lookup. If the lookup results in a
153 * different inode, then let the VFS invalidate the dentry and redo
154 * the lookup once more. If the lookup results in the same inode,
155 * then refresh the attributes, timeouts and mark the dentry valid.
156 */
fuse_dentry_revalidate(struct dentry * entry,struct nameidata * nd)157 static int fuse_dentry_revalidate(struct dentry *entry, struct nameidata *nd)
158 {
159 struct inode *inode;
160
161 inode = ACCESS_ONCE(entry->d_inode);
162 if (inode && is_bad_inode(inode))
163 return 0;
164 else if (fuse_dentry_time(entry) < get_jiffies_64()) {
165 int err;
166 struct fuse_entry_out outarg;
167 struct fuse_conn *fc;
168 struct fuse_req *req;
169 struct fuse_forget_link *forget;
170 struct dentry *parent;
171 u64 attr_version;
172
173 /* For negative dentries, always do a fresh lookup */
174 if (!inode)
175 return 0;
176
177 if (nd && (nd->flags & LOOKUP_RCU))
178 return -ECHILD;
179
180 fc = get_fuse_conn(inode);
181 req = fuse_get_req(fc);
182 if (IS_ERR(req))
183 return 0;
184
185 forget = fuse_alloc_forget();
186 if (!forget) {
187 fuse_put_request(fc, req);
188 return 0;
189 }
190
191 attr_version = fuse_get_attr_version(fc);
192
193 parent = dget_parent(entry);
194 fuse_lookup_init(fc, req, get_node_id(parent->d_inode),
195 &entry->d_name, &outarg);
196 fuse_request_send(fc, req);
197 dput(parent);
198 err = req->out.h.error;
199 fuse_put_request(fc, req);
200 /* Zero nodeid is same as -ENOENT */
201 if (!err && !outarg.nodeid)
202 err = -ENOENT;
203 if (!err) {
204 struct fuse_inode *fi = get_fuse_inode(inode);
205 if (outarg.nodeid != get_node_id(inode)) {
206 fuse_queue_forget(fc, forget, outarg.nodeid, 1);
207 return 0;
208 }
209 spin_lock(&fc->lock);
210 fi->nlookup++;
211 spin_unlock(&fc->lock);
212 }
213 kfree(forget);
214 if (err || (outarg.attr.mode ^ inode->i_mode) & S_IFMT)
215 return 0;
216
217 fuse_change_attributes(inode, &outarg.attr,
218 entry_attr_timeout(&outarg),
219 attr_version);
220 fuse_change_entry_timeout(entry, &outarg);
221 }
222 return 1;
223 }
224
invalid_nodeid(u64 nodeid)225 static int invalid_nodeid(u64 nodeid)
226 {
227 return !nodeid || nodeid == FUSE_ROOT_ID;
228 }
229
230 const struct dentry_operations fuse_dentry_operations = {
231 .d_revalidate = fuse_dentry_revalidate,
232 };
233
fuse_valid_type(int m)234 int fuse_valid_type(int m)
235 {
236 return S_ISREG(m) || S_ISDIR(m) || S_ISLNK(m) || S_ISCHR(m) ||
237 S_ISBLK(m) || S_ISFIFO(m) || S_ISSOCK(m);
238 }
239
240 /*
241 * Add a directory inode to a dentry, ensuring that no other dentry
242 * refers to this inode. Called with fc->inst_mutex.
243 */
fuse_d_add_directory(struct dentry * entry,struct inode * inode)244 static struct dentry *fuse_d_add_directory(struct dentry *entry,
245 struct inode *inode)
246 {
247 struct dentry *alias = d_find_alias(inode);
248 if (alias && !(alias->d_flags & DCACHE_DISCONNECTED)) {
249 /* This tries to shrink the subtree below alias */
250 fuse_invalidate_entry(alias);
251 dput(alias);
252 if (!list_empty(&inode->i_dentry))
253 return ERR_PTR(-EBUSY);
254 } else {
255 dput(alias);
256 }
257 return d_splice_alias(inode, entry);
258 }
259
fuse_lookup_name(struct super_block * sb,u64 nodeid,struct qstr * name,struct fuse_entry_out * outarg,struct inode ** inode)260 int fuse_lookup_name(struct super_block *sb, u64 nodeid, struct qstr *name,
261 struct fuse_entry_out *outarg, struct inode **inode)
262 {
263 struct fuse_conn *fc = get_fuse_conn_super(sb);
264 struct fuse_req *req;
265 struct fuse_forget_link *forget;
266 u64 attr_version;
267 int err;
268
269 *inode = NULL;
270 err = -ENAMETOOLONG;
271 if (name->len > FUSE_NAME_MAX)
272 goto out;
273
274 req = fuse_get_req(fc);
275 err = PTR_ERR(req);
276 if (IS_ERR(req))
277 goto out;
278
279 forget = fuse_alloc_forget();
280 err = -ENOMEM;
281 if (!forget) {
282 fuse_put_request(fc, req);
283 goto out;
284 }
285
286 attr_version = fuse_get_attr_version(fc);
287
288 fuse_lookup_init(fc, req, nodeid, name, outarg);
289 fuse_request_send(fc, req);
290 err = req->out.h.error;
291 fuse_put_request(fc, req);
292 /* Zero nodeid is same as -ENOENT, but with valid timeout */
293 if (err || !outarg->nodeid)
294 goto out_put_forget;
295
296 err = -EIO;
297 if (!outarg->nodeid)
298 goto out_put_forget;
299 if (!fuse_valid_type(outarg->attr.mode))
300 goto out_put_forget;
301
302 *inode = fuse_iget(sb, outarg->nodeid, outarg->generation,
303 &outarg->attr, entry_attr_timeout(outarg),
304 attr_version);
305 err = -ENOMEM;
306 if (!*inode) {
307 fuse_queue_forget(fc, forget, outarg->nodeid, 1);
308 goto out;
309 }
310 err = 0;
311
312 out_put_forget:
313 kfree(forget);
314 out:
315 return err;
316 }
317
fuse_lookup(struct inode * dir,struct dentry * entry,struct nameidata * nd)318 static struct dentry *fuse_lookup(struct inode *dir, struct dentry *entry,
319 struct nameidata *nd)
320 {
321 int err;
322 struct fuse_entry_out outarg;
323 struct inode *inode;
324 struct dentry *newent;
325 struct fuse_conn *fc = get_fuse_conn(dir);
326 bool outarg_valid = true;
327
328 err = fuse_lookup_name(dir->i_sb, get_node_id(dir), &entry->d_name,
329 &outarg, &inode);
330 if (err == -ENOENT) {
331 outarg_valid = false;
332 err = 0;
333 }
334 if (err)
335 goto out_err;
336
337 err = -EIO;
338 if (inode && get_node_id(inode) == FUSE_ROOT_ID)
339 goto out_iput;
340
341 if (inode && S_ISDIR(inode->i_mode)) {
342 mutex_lock(&fc->inst_mutex);
343 newent = fuse_d_add_directory(entry, inode);
344 mutex_unlock(&fc->inst_mutex);
345 err = PTR_ERR(newent);
346 if (IS_ERR(newent))
347 goto out_iput;
348 } else {
349 newent = d_splice_alias(inode, entry);
350 }
351
352 entry = newent ? newent : entry;
353 if (outarg_valid)
354 fuse_change_entry_timeout(entry, &outarg);
355 else
356 fuse_invalidate_entry_cache(entry);
357
358 return newent;
359
360 out_iput:
361 iput(inode);
362 out_err:
363 return ERR_PTR(err);
364 }
365
366 /*
367 * Atomic create+open operation
368 *
369 * If the filesystem doesn't support this, then fall back to separate
370 * 'mknod' + 'open' requests.
371 */
fuse_create_open(struct inode * dir,struct dentry * entry,umode_t mode,struct nameidata * nd)372 static int fuse_create_open(struct inode *dir, struct dentry *entry,
373 umode_t mode, struct nameidata *nd)
374 {
375 int err;
376 struct inode *inode;
377 struct fuse_conn *fc = get_fuse_conn(dir);
378 struct fuse_req *req;
379 struct fuse_forget_link *forget;
380 struct fuse_create_in inarg;
381 struct fuse_open_out outopen;
382 struct fuse_entry_out outentry;
383 struct fuse_file *ff;
384 struct file *file;
385 int flags = nd->intent.open.flags;
386
387 if (fc->no_create)
388 return -ENOSYS;
389
390 forget = fuse_alloc_forget();
391 if (!forget)
392 return -ENOMEM;
393
394 req = fuse_get_req(fc);
395 err = PTR_ERR(req);
396 if (IS_ERR(req))
397 goto out_put_forget_req;
398
399 err = -ENOMEM;
400 ff = fuse_file_alloc(fc);
401 if (!ff)
402 goto out_put_request;
403
404 if (!fc->dont_mask)
405 mode &= ~current_umask();
406
407 flags &= ~O_NOCTTY;
408 memset(&inarg, 0, sizeof(inarg));
409 memset(&outentry, 0, sizeof(outentry));
410 inarg.flags = flags;
411 inarg.mode = mode;
412 inarg.umask = current_umask();
413 req->in.h.opcode = FUSE_CREATE;
414 req->in.h.nodeid = get_node_id(dir);
415 req->in.numargs = 2;
416 req->in.args[0].size = fc->minor < 12 ? sizeof(struct fuse_open_in) :
417 sizeof(inarg);
418 req->in.args[0].value = &inarg;
419 req->in.args[1].size = entry->d_name.len + 1;
420 req->in.args[1].value = entry->d_name.name;
421 req->out.numargs = 2;
422 if (fc->minor < 9)
423 req->out.args[0].size = FUSE_COMPAT_ENTRY_OUT_SIZE;
424 else
425 req->out.args[0].size = sizeof(outentry);
426 req->out.args[0].value = &outentry;
427 req->out.args[1].size = sizeof(outopen);
428 req->out.args[1].value = &outopen;
429 fuse_request_send(fc, req);
430 err = req->out.h.error;
431 if (err) {
432 if (err == -ENOSYS)
433 fc->no_create = 1;
434 goto out_free_ff;
435 }
436
437 err = -EIO;
438 if (!S_ISREG(outentry.attr.mode) || invalid_nodeid(outentry.nodeid))
439 goto out_free_ff;
440
441 fuse_put_request(fc, req);
442 ff->fh = outopen.fh;
443 ff->nodeid = outentry.nodeid;
444 ff->open_flags = outopen.open_flags;
445 inode = fuse_iget(dir->i_sb, outentry.nodeid, outentry.generation,
446 &outentry.attr, entry_attr_timeout(&outentry), 0);
447 if (!inode) {
448 flags &= ~(O_CREAT | O_EXCL | O_TRUNC);
449 fuse_sync_release(ff, flags);
450 fuse_queue_forget(fc, forget, outentry.nodeid, 1);
451 return -ENOMEM;
452 }
453 kfree(forget);
454 d_instantiate(entry, inode);
455 fuse_change_entry_timeout(entry, &outentry);
456 fuse_invalidate_attr(dir);
457 file = lookup_instantiate_filp(nd, entry, generic_file_open);
458 if (IS_ERR(file)) {
459 fuse_sync_release(ff, flags);
460 return PTR_ERR(file);
461 }
462 file->private_data = fuse_file_get(ff);
463 fuse_finish_open(inode, file);
464 return 0;
465
466 out_free_ff:
467 fuse_file_free(ff);
468 out_put_request:
469 fuse_put_request(fc, req);
470 out_put_forget_req:
471 kfree(forget);
472 return err;
473 }
474
475 /*
476 * Code shared between mknod, mkdir, symlink and link
477 */
create_new_entry(struct fuse_conn * fc,struct fuse_req * req,struct inode * dir,struct dentry * entry,umode_t mode)478 static int create_new_entry(struct fuse_conn *fc, struct fuse_req *req,
479 struct inode *dir, struct dentry *entry,
480 umode_t mode)
481 {
482 struct fuse_entry_out outarg;
483 struct inode *inode;
484 int err;
485 struct fuse_forget_link *forget;
486
487 forget = fuse_alloc_forget();
488 if (!forget) {
489 fuse_put_request(fc, req);
490 return -ENOMEM;
491 }
492
493 memset(&outarg, 0, sizeof(outarg));
494 req->in.h.nodeid = get_node_id(dir);
495 req->out.numargs = 1;
496 if (fc->minor < 9)
497 req->out.args[0].size = FUSE_COMPAT_ENTRY_OUT_SIZE;
498 else
499 req->out.args[0].size = sizeof(outarg);
500 req->out.args[0].value = &outarg;
501 fuse_request_send(fc, req);
502 err = req->out.h.error;
503 fuse_put_request(fc, req);
504 if (err)
505 goto out_put_forget_req;
506
507 err = -EIO;
508 if (invalid_nodeid(outarg.nodeid))
509 goto out_put_forget_req;
510
511 if ((outarg.attr.mode ^ mode) & S_IFMT)
512 goto out_put_forget_req;
513
514 inode = fuse_iget(dir->i_sb, outarg.nodeid, outarg.generation,
515 &outarg.attr, entry_attr_timeout(&outarg), 0);
516 if (!inode) {
517 fuse_queue_forget(fc, forget, outarg.nodeid, 1);
518 return -ENOMEM;
519 }
520 kfree(forget);
521
522 if (S_ISDIR(inode->i_mode)) {
523 struct dentry *alias;
524 mutex_lock(&fc->inst_mutex);
525 alias = d_find_alias(inode);
526 if (alias) {
527 /* New directory must have moved since mkdir */
528 mutex_unlock(&fc->inst_mutex);
529 dput(alias);
530 iput(inode);
531 return -EBUSY;
532 }
533 d_instantiate(entry, inode);
534 mutex_unlock(&fc->inst_mutex);
535 } else
536 d_instantiate(entry, inode);
537
538 fuse_change_entry_timeout(entry, &outarg);
539 fuse_invalidate_attr(dir);
540 return 0;
541
542 out_put_forget_req:
543 kfree(forget);
544 return err;
545 }
546
fuse_mknod(struct inode * dir,struct dentry * entry,umode_t mode,dev_t rdev)547 static int fuse_mknod(struct inode *dir, struct dentry *entry, umode_t mode,
548 dev_t rdev)
549 {
550 struct fuse_mknod_in inarg;
551 struct fuse_conn *fc = get_fuse_conn(dir);
552 struct fuse_req *req = fuse_get_req(fc);
553 if (IS_ERR(req))
554 return PTR_ERR(req);
555
556 if (!fc->dont_mask)
557 mode &= ~current_umask();
558
559 memset(&inarg, 0, sizeof(inarg));
560 inarg.mode = mode;
561 inarg.rdev = new_encode_dev(rdev);
562 inarg.umask = current_umask();
563 req->in.h.opcode = FUSE_MKNOD;
564 req->in.numargs = 2;
565 req->in.args[0].size = fc->minor < 12 ? FUSE_COMPAT_MKNOD_IN_SIZE :
566 sizeof(inarg);
567 req->in.args[0].value = &inarg;
568 req->in.args[1].size = entry->d_name.len + 1;
569 req->in.args[1].value = entry->d_name.name;
570 return create_new_entry(fc, req, dir, entry, mode);
571 }
572
fuse_create(struct inode * dir,struct dentry * entry,umode_t mode,struct nameidata * nd)573 static int fuse_create(struct inode *dir, struct dentry *entry, umode_t mode,
574 struct nameidata *nd)
575 {
576 if (nd) {
577 int err = fuse_create_open(dir, entry, mode, nd);
578 if (err != -ENOSYS)
579 return err;
580 /* Fall back on mknod */
581 }
582 return fuse_mknod(dir, entry, mode, 0);
583 }
584
fuse_mkdir(struct inode * dir,struct dentry * entry,umode_t mode)585 static int fuse_mkdir(struct inode *dir, struct dentry *entry, umode_t mode)
586 {
587 struct fuse_mkdir_in inarg;
588 struct fuse_conn *fc = get_fuse_conn(dir);
589 struct fuse_req *req = fuse_get_req(fc);
590 if (IS_ERR(req))
591 return PTR_ERR(req);
592
593 if (!fc->dont_mask)
594 mode &= ~current_umask();
595
596 memset(&inarg, 0, sizeof(inarg));
597 inarg.mode = mode;
598 inarg.umask = current_umask();
599 req->in.h.opcode = FUSE_MKDIR;
600 req->in.numargs = 2;
601 req->in.args[0].size = sizeof(inarg);
602 req->in.args[0].value = &inarg;
603 req->in.args[1].size = entry->d_name.len + 1;
604 req->in.args[1].value = entry->d_name.name;
605 return create_new_entry(fc, req, dir, entry, S_IFDIR);
606 }
607
fuse_symlink(struct inode * dir,struct dentry * entry,const char * link)608 static int fuse_symlink(struct inode *dir, struct dentry *entry,
609 const char *link)
610 {
611 struct fuse_conn *fc = get_fuse_conn(dir);
612 unsigned len = strlen(link) + 1;
613 struct fuse_req *req = fuse_get_req(fc);
614 if (IS_ERR(req))
615 return PTR_ERR(req);
616
617 req->in.h.opcode = FUSE_SYMLINK;
618 req->in.numargs = 2;
619 req->in.args[0].size = entry->d_name.len + 1;
620 req->in.args[0].value = entry->d_name.name;
621 req->in.args[1].size = len;
622 req->in.args[1].value = link;
623 return create_new_entry(fc, req, dir, entry, S_IFLNK);
624 }
625
fuse_unlink(struct inode * dir,struct dentry * entry)626 static int fuse_unlink(struct inode *dir, struct dentry *entry)
627 {
628 int err;
629 struct fuse_conn *fc = get_fuse_conn(dir);
630 struct fuse_req *req = fuse_get_req(fc);
631 if (IS_ERR(req))
632 return PTR_ERR(req);
633
634 req->in.h.opcode = FUSE_UNLINK;
635 req->in.h.nodeid = get_node_id(dir);
636 req->in.numargs = 1;
637 req->in.args[0].size = entry->d_name.len + 1;
638 req->in.args[0].value = entry->d_name.name;
639 fuse_request_send(fc, req);
640 err = req->out.h.error;
641 fuse_put_request(fc, req);
642 if (!err) {
643 struct inode *inode = entry->d_inode;
644 struct fuse_inode *fi = get_fuse_inode(inode);
645
646 spin_lock(&fc->lock);
647 fi->attr_version = ++fc->attr_version;
648 /*
649 * If i_nlink == 0 then unlink doesn't make sense, yet this can
650 * happen if userspace filesystem is careless. It would be
651 * difficult to enforce correct nlink usage so just ignore this
652 * condition here
653 */
654 if (inode->i_nlink > 0)
655 drop_nlink(inode);
656 spin_unlock(&fc->lock);
657 fuse_invalidate_attr(inode);
658 fuse_invalidate_attr(dir);
659 fuse_invalidate_entry_cache(entry);
660 } else if (err == -EINTR)
661 fuse_invalidate_entry(entry);
662 return err;
663 }
664
fuse_rmdir(struct inode * dir,struct dentry * entry)665 static int fuse_rmdir(struct inode *dir, struct dentry *entry)
666 {
667 int err;
668 struct fuse_conn *fc = get_fuse_conn(dir);
669 struct fuse_req *req = fuse_get_req(fc);
670 if (IS_ERR(req))
671 return PTR_ERR(req);
672
673 req->in.h.opcode = FUSE_RMDIR;
674 req->in.h.nodeid = get_node_id(dir);
675 req->in.numargs = 1;
676 req->in.args[0].size = entry->d_name.len + 1;
677 req->in.args[0].value = entry->d_name.name;
678 fuse_request_send(fc, req);
679 err = req->out.h.error;
680 fuse_put_request(fc, req);
681 if (!err) {
682 clear_nlink(entry->d_inode);
683 fuse_invalidate_attr(dir);
684 fuse_invalidate_entry_cache(entry);
685 } else if (err == -EINTR)
686 fuse_invalidate_entry(entry);
687 return err;
688 }
689
fuse_rename(struct inode * olddir,struct dentry * oldent,struct inode * newdir,struct dentry * newent)690 static int fuse_rename(struct inode *olddir, struct dentry *oldent,
691 struct inode *newdir, struct dentry *newent)
692 {
693 int err;
694 struct fuse_rename_in inarg;
695 struct fuse_conn *fc = get_fuse_conn(olddir);
696 struct fuse_req *req = fuse_get_req(fc);
697
698 if (IS_ERR(req))
699 return PTR_ERR(req);
700
701 memset(&inarg, 0, sizeof(inarg));
702 inarg.newdir = get_node_id(newdir);
703 req->in.h.opcode = FUSE_RENAME;
704 req->in.h.nodeid = get_node_id(olddir);
705 req->in.numargs = 3;
706 req->in.args[0].size = sizeof(inarg);
707 req->in.args[0].value = &inarg;
708 req->in.args[1].size = oldent->d_name.len + 1;
709 req->in.args[1].value = oldent->d_name.name;
710 req->in.args[2].size = newent->d_name.len + 1;
711 req->in.args[2].value = newent->d_name.name;
712 fuse_request_send(fc, req);
713 err = req->out.h.error;
714 fuse_put_request(fc, req);
715 if (!err) {
716 /* ctime changes */
717 fuse_invalidate_attr(oldent->d_inode);
718
719 fuse_invalidate_attr(olddir);
720 if (olddir != newdir)
721 fuse_invalidate_attr(newdir);
722
723 /* newent will end up negative */
724 if (newent->d_inode) {
725 fuse_invalidate_attr(newent->d_inode);
726 fuse_invalidate_entry_cache(newent);
727 }
728 } else if (err == -EINTR) {
729 /* If request was interrupted, DEITY only knows if the
730 rename actually took place. If the invalidation
731 fails (e.g. some process has CWD under the renamed
732 directory), then there can be inconsistency between
733 the dcache and the real filesystem. Tough luck. */
734 fuse_invalidate_entry(oldent);
735 if (newent->d_inode)
736 fuse_invalidate_entry(newent);
737 }
738
739 return err;
740 }
741
fuse_link(struct dentry * entry,struct inode * newdir,struct dentry * newent)742 static int fuse_link(struct dentry *entry, struct inode *newdir,
743 struct dentry *newent)
744 {
745 int err;
746 struct fuse_link_in inarg;
747 struct inode *inode = entry->d_inode;
748 struct fuse_conn *fc = get_fuse_conn(inode);
749 struct fuse_req *req = fuse_get_req(fc);
750 if (IS_ERR(req))
751 return PTR_ERR(req);
752
753 memset(&inarg, 0, sizeof(inarg));
754 inarg.oldnodeid = get_node_id(inode);
755 req->in.h.opcode = FUSE_LINK;
756 req->in.numargs = 2;
757 req->in.args[0].size = sizeof(inarg);
758 req->in.args[0].value = &inarg;
759 req->in.args[1].size = newent->d_name.len + 1;
760 req->in.args[1].value = newent->d_name.name;
761 err = create_new_entry(fc, req, newdir, newent, inode->i_mode);
762 /* Contrary to "normal" filesystems it can happen that link
763 makes two "logical" inodes point to the same "physical"
764 inode. We invalidate the attributes of the old one, so it
765 will reflect changes in the backing inode (link count,
766 etc.)
767 */
768 if (!err) {
769 struct fuse_inode *fi = get_fuse_inode(inode);
770
771 spin_lock(&fc->lock);
772 fi->attr_version = ++fc->attr_version;
773 inc_nlink(inode);
774 spin_unlock(&fc->lock);
775 fuse_invalidate_attr(inode);
776 } else if (err == -EINTR) {
777 fuse_invalidate_attr(inode);
778 }
779 return err;
780 }
781
fuse_fillattr(struct inode * inode,struct fuse_attr * attr,struct kstat * stat)782 static void fuse_fillattr(struct inode *inode, struct fuse_attr *attr,
783 struct kstat *stat)
784 {
785 stat->dev = inode->i_sb->s_dev;
786 stat->ino = attr->ino;
787 stat->mode = (inode->i_mode & S_IFMT) | (attr->mode & 07777);
788 stat->nlink = attr->nlink;
789 stat->uid = attr->uid;
790 stat->gid = attr->gid;
791 stat->rdev = inode->i_rdev;
792 stat->atime.tv_sec = attr->atime;
793 stat->atime.tv_nsec = attr->atimensec;
794 stat->mtime.tv_sec = attr->mtime;
795 stat->mtime.tv_nsec = attr->mtimensec;
796 stat->ctime.tv_sec = attr->ctime;
797 stat->ctime.tv_nsec = attr->ctimensec;
798 stat->size = attr->size;
799 stat->blocks = attr->blocks;
800 stat->blksize = (1 << inode->i_blkbits);
801 }
802
fuse_do_getattr(struct inode * inode,struct kstat * stat,struct file * file)803 static int fuse_do_getattr(struct inode *inode, struct kstat *stat,
804 struct file *file)
805 {
806 int err;
807 struct fuse_getattr_in inarg;
808 struct fuse_attr_out outarg;
809 struct fuse_conn *fc = get_fuse_conn(inode);
810 struct fuse_req *req;
811 u64 attr_version;
812
813 req = fuse_get_req(fc);
814 if (IS_ERR(req))
815 return PTR_ERR(req);
816
817 attr_version = fuse_get_attr_version(fc);
818
819 memset(&inarg, 0, sizeof(inarg));
820 memset(&outarg, 0, sizeof(outarg));
821 /* Directories have separate file-handle space */
822 if (file && S_ISREG(inode->i_mode)) {
823 struct fuse_file *ff = file->private_data;
824
825 inarg.getattr_flags |= FUSE_GETATTR_FH;
826 inarg.fh = ff->fh;
827 }
828 req->in.h.opcode = FUSE_GETATTR;
829 req->in.h.nodeid = get_node_id(inode);
830 req->in.numargs = 1;
831 req->in.args[0].size = sizeof(inarg);
832 req->in.args[0].value = &inarg;
833 req->out.numargs = 1;
834 if (fc->minor < 9)
835 req->out.args[0].size = FUSE_COMPAT_ATTR_OUT_SIZE;
836 else
837 req->out.args[0].size = sizeof(outarg);
838 req->out.args[0].value = &outarg;
839 fuse_request_send(fc, req);
840 err = req->out.h.error;
841 fuse_put_request(fc, req);
842 if (!err) {
843 if ((inode->i_mode ^ outarg.attr.mode) & S_IFMT) {
844 make_bad_inode(inode);
845 err = -EIO;
846 } else {
847 fuse_change_attributes(inode, &outarg.attr,
848 attr_timeout(&outarg),
849 attr_version);
850 if (stat)
851 fuse_fillattr(inode, &outarg.attr, stat);
852 }
853 }
854 return err;
855 }
856
fuse_update_attributes(struct inode * inode,struct kstat * stat,struct file * file,bool * refreshed)857 int fuse_update_attributes(struct inode *inode, struct kstat *stat,
858 struct file *file, bool *refreshed)
859 {
860 struct fuse_inode *fi = get_fuse_inode(inode);
861 int err;
862 bool r;
863
864 if (fi->i_time < get_jiffies_64()) {
865 r = true;
866 err = fuse_do_getattr(inode, stat, file);
867 } else {
868 r = false;
869 err = 0;
870 if (stat) {
871 generic_fillattr(inode, stat);
872 stat->mode = fi->orig_i_mode;
873 stat->ino = fi->orig_ino;
874 }
875 }
876
877 if (refreshed != NULL)
878 *refreshed = r;
879
880 return err;
881 }
882
fuse_reverse_inval_entry(struct super_block * sb,u64 parent_nodeid,u64 child_nodeid,struct qstr * name)883 int fuse_reverse_inval_entry(struct super_block *sb, u64 parent_nodeid,
884 u64 child_nodeid, struct qstr *name)
885 {
886 int err = -ENOTDIR;
887 struct inode *parent;
888 struct dentry *dir;
889 struct dentry *entry;
890
891 parent = ilookup5(sb, parent_nodeid, fuse_inode_eq, &parent_nodeid);
892 if (!parent)
893 return -ENOENT;
894
895 mutex_lock(&parent->i_mutex);
896 if (!S_ISDIR(parent->i_mode))
897 goto unlock;
898
899 err = -ENOENT;
900 dir = d_find_alias(parent);
901 if (!dir)
902 goto unlock;
903
904 entry = d_lookup(dir, name);
905 dput(dir);
906 if (!entry)
907 goto unlock;
908
909 fuse_invalidate_attr(parent);
910 fuse_invalidate_entry(entry);
911
912 if (child_nodeid != 0 && entry->d_inode) {
913 mutex_lock(&entry->d_inode->i_mutex);
914 if (get_node_id(entry->d_inode) != child_nodeid) {
915 err = -ENOENT;
916 goto badentry;
917 }
918 if (d_mountpoint(entry)) {
919 err = -EBUSY;
920 goto badentry;
921 }
922 if (S_ISDIR(entry->d_inode->i_mode)) {
923 shrink_dcache_parent(entry);
924 if (!simple_empty(entry)) {
925 err = -ENOTEMPTY;
926 goto badentry;
927 }
928 entry->d_inode->i_flags |= S_DEAD;
929 }
930 dont_mount(entry);
931 clear_nlink(entry->d_inode);
932 err = 0;
933 badentry:
934 mutex_unlock(&entry->d_inode->i_mutex);
935 if (!err)
936 d_delete(entry);
937 } else {
938 err = 0;
939 }
940 dput(entry);
941
942 unlock:
943 mutex_unlock(&parent->i_mutex);
944 iput(parent);
945 return err;
946 }
947
948 /*
949 * Calling into a user-controlled filesystem gives the filesystem
950 * daemon ptrace-like capabilities over the requester process. This
951 * means, that the filesystem daemon is able to record the exact
952 * filesystem operations performed, and can also control the behavior
953 * of the requester process in otherwise impossible ways. For example
954 * it can delay the operation for arbitrary length of time allowing
955 * DoS against the requester.
956 *
957 * For this reason only those processes can call into the filesystem,
958 * for which the owner of the mount has ptrace privilege. This
959 * excludes processes started by other users, suid or sgid processes.
960 */
fuse_allow_task(struct fuse_conn * fc,struct task_struct * task)961 int fuse_allow_task(struct fuse_conn *fc, struct task_struct *task)
962 {
963 const struct cred *cred;
964 int ret;
965
966 if (fc->flags & FUSE_ALLOW_OTHER)
967 return 1;
968
969 rcu_read_lock();
970 ret = 0;
971 cred = __task_cred(task);
972 if (cred->euid == fc->user_id &&
973 cred->suid == fc->user_id &&
974 cred->uid == fc->user_id &&
975 cred->egid == fc->group_id &&
976 cred->sgid == fc->group_id &&
977 cred->gid == fc->group_id)
978 ret = 1;
979 rcu_read_unlock();
980
981 return ret;
982 }
983
fuse_access(struct inode * inode,int mask)984 static int fuse_access(struct inode *inode, int mask)
985 {
986 struct fuse_conn *fc = get_fuse_conn(inode);
987 struct fuse_req *req;
988 struct fuse_access_in inarg;
989 int err;
990
991 if (fc->no_access)
992 return 0;
993
994 req = fuse_get_req(fc);
995 if (IS_ERR(req))
996 return PTR_ERR(req);
997
998 memset(&inarg, 0, sizeof(inarg));
999 inarg.mask = mask & (MAY_READ | MAY_WRITE | MAY_EXEC);
1000 req->in.h.opcode = FUSE_ACCESS;
1001 req->in.h.nodeid = get_node_id(inode);
1002 req->in.numargs = 1;
1003 req->in.args[0].size = sizeof(inarg);
1004 req->in.args[0].value = &inarg;
1005 fuse_request_send(fc, req);
1006 err = req->out.h.error;
1007 fuse_put_request(fc, req);
1008 if (err == -ENOSYS) {
1009 fc->no_access = 1;
1010 err = 0;
1011 }
1012 return err;
1013 }
1014
fuse_perm_getattr(struct inode * inode,int mask)1015 static int fuse_perm_getattr(struct inode *inode, int mask)
1016 {
1017 if (mask & MAY_NOT_BLOCK)
1018 return -ECHILD;
1019
1020 return fuse_do_getattr(inode, NULL, NULL);
1021 }
1022
1023 /*
1024 * Check permission. The two basic access models of FUSE are:
1025 *
1026 * 1) Local access checking ('default_permissions' mount option) based
1027 * on file mode. This is the plain old disk filesystem permission
1028 * modell.
1029 *
1030 * 2) "Remote" access checking, where server is responsible for
1031 * checking permission in each inode operation. An exception to this
1032 * is if ->permission() was invoked from sys_access() in which case an
1033 * access request is sent. Execute permission is still checked
1034 * locally based on file mode.
1035 */
fuse_permission(struct inode * inode,int mask)1036 static int fuse_permission(struct inode *inode, int mask)
1037 {
1038 struct fuse_conn *fc = get_fuse_conn(inode);
1039 bool refreshed = false;
1040 int err = 0;
1041
1042 if (!fuse_allow_task(fc, current))
1043 return -EACCES;
1044
1045 /*
1046 * If attributes are needed, refresh them before proceeding
1047 */
1048 if ((fc->flags & FUSE_DEFAULT_PERMISSIONS) ||
1049 ((mask & MAY_EXEC) && S_ISREG(inode->i_mode))) {
1050 struct fuse_inode *fi = get_fuse_inode(inode);
1051
1052 if (fi->i_time < get_jiffies_64()) {
1053 refreshed = true;
1054
1055 err = fuse_perm_getattr(inode, mask);
1056 if (err)
1057 return err;
1058 }
1059 }
1060
1061 if (fc->flags & FUSE_DEFAULT_PERMISSIONS) {
1062 err = generic_permission(inode, mask);
1063
1064 /* If permission is denied, try to refresh file
1065 attributes. This is also needed, because the root
1066 node will at first have no permissions */
1067 if (err == -EACCES && !refreshed) {
1068 err = fuse_perm_getattr(inode, mask);
1069 if (!err)
1070 err = generic_permission(inode, mask);
1071 }
1072
1073 /* Note: the opposite of the above test does not
1074 exist. So if permissions are revoked this won't be
1075 noticed immediately, only after the attribute
1076 timeout has expired */
1077 } else if (mask & (MAY_ACCESS | MAY_CHDIR)) {
1078 if (mask & MAY_NOT_BLOCK)
1079 return -ECHILD;
1080
1081 err = fuse_access(inode, mask);
1082 } else if ((mask & MAY_EXEC) && S_ISREG(inode->i_mode)) {
1083 if (!(inode->i_mode & S_IXUGO)) {
1084 if (refreshed)
1085 return -EACCES;
1086
1087 err = fuse_perm_getattr(inode, mask);
1088 if (!err && !(inode->i_mode & S_IXUGO))
1089 return -EACCES;
1090 }
1091 }
1092 return err;
1093 }
1094
parse_dirfile(char * buf,size_t nbytes,struct file * file,void * dstbuf,filldir_t filldir)1095 static int parse_dirfile(char *buf, size_t nbytes, struct file *file,
1096 void *dstbuf, filldir_t filldir)
1097 {
1098 while (nbytes >= FUSE_NAME_OFFSET) {
1099 struct fuse_dirent *dirent = (struct fuse_dirent *) buf;
1100 size_t reclen = FUSE_DIRENT_SIZE(dirent);
1101 int over;
1102 if (!dirent->namelen || dirent->namelen > FUSE_NAME_MAX)
1103 return -EIO;
1104 if (reclen > nbytes)
1105 break;
1106
1107 over = filldir(dstbuf, dirent->name, dirent->namelen,
1108 file->f_pos, dirent->ino, dirent->type);
1109 if (over)
1110 break;
1111
1112 buf += reclen;
1113 nbytes -= reclen;
1114 file->f_pos = dirent->off;
1115 }
1116
1117 return 0;
1118 }
1119
fuse_readdir(struct file * file,void * dstbuf,filldir_t filldir)1120 static int fuse_readdir(struct file *file, void *dstbuf, filldir_t filldir)
1121 {
1122 int err;
1123 size_t nbytes;
1124 struct page *page;
1125 struct inode *inode = file->f_path.dentry->d_inode;
1126 struct fuse_conn *fc = get_fuse_conn(inode);
1127 struct fuse_req *req;
1128
1129 if (is_bad_inode(inode))
1130 return -EIO;
1131
1132 req = fuse_get_req(fc);
1133 if (IS_ERR(req))
1134 return PTR_ERR(req);
1135
1136 page = alloc_page(GFP_KERNEL);
1137 if (!page) {
1138 fuse_put_request(fc, req);
1139 return -ENOMEM;
1140 }
1141 req->out.argpages = 1;
1142 req->num_pages = 1;
1143 req->pages[0] = page;
1144 fuse_read_fill(req, file, file->f_pos, PAGE_SIZE, FUSE_READDIR);
1145 fuse_request_send(fc, req);
1146 nbytes = req->out.args[0].size;
1147 err = req->out.h.error;
1148 fuse_put_request(fc, req);
1149 if (!err)
1150 err = parse_dirfile(page_address(page), nbytes, file, dstbuf,
1151 filldir);
1152
1153 __free_page(page);
1154 fuse_invalidate_attr(inode); /* atime changed */
1155 return err;
1156 }
1157
read_link(struct dentry * dentry)1158 static char *read_link(struct dentry *dentry)
1159 {
1160 struct inode *inode = dentry->d_inode;
1161 struct fuse_conn *fc = get_fuse_conn(inode);
1162 struct fuse_req *req = fuse_get_req(fc);
1163 char *link;
1164
1165 if (IS_ERR(req))
1166 return ERR_CAST(req);
1167
1168 link = (char *) __get_free_page(GFP_KERNEL);
1169 if (!link) {
1170 link = ERR_PTR(-ENOMEM);
1171 goto out;
1172 }
1173 req->in.h.opcode = FUSE_READLINK;
1174 req->in.h.nodeid = get_node_id(inode);
1175 req->out.argvar = 1;
1176 req->out.numargs = 1;
1177 req->out.args[0].size = PAGE_SIZE - 1;
1178 req->out.args[0].value = link;
1179 fuse_request_send(fc, req);
1180 if (req->out.h.error) {
1181 free_page((unsigned long) link);
1182 link = ERR_PTR(req->out.h.error);
1183 } else
1184 link[req->out.args[0].size] = '\0';
1185 out:
1186 fuse_put_request(fc, req);
1187 fuse_invalidate_attr(inode); /* atime changed */
1188 return link;
1189 }
1190
free_link(char * link)1191 static void free_link(char *link)
1192 {
1193 if (!IS_ERR(link))
1194 free_page((unsigned long) link);
1195 }
1196
fuse_follow_link(struct dentry * dentry,struct nameidata * nd)1197 static void *fuse_follow_link(struct dentry *dentry, struct nameidata *nd)
1198 {
1199 nd_set_link(nd, read_link(dentry));
1200 return NULL;
1201 }
1202
fuse_put_link(struct dentry * dentry,struct nameidata * nd,void * c)1203 static void fuse_put_link(struct dentry *dentry, struct nameidata *nd, void *c)
1204 {
1205 free_link(nd_get_link(nd));
1206 }
1207
fuse_dir_open(struct inode * inode,struct file * file)1208 static int fuse_dir_open(struct inode *inode, struct file *file)
1209 {
1210 return fuse_open_common(inode, file, true);
1211 }
1212
fuse_dir_release(struct inode * inode,struct file * file)1213 static int fuse_dir_release(struct inode *inode, struct file *file)
1214 {
1215 fuse_release_common(file, FUSE_RELEASEDIR);
1216
1217 return 0;
1218 }
1219
fuse_dir_fsync(struct file * file,loff_t start,loff_t end,int datasync)1220 static int fuse_dir_fsync(struct file *file, loff_t start, loff_t end,
1221 int datasync)
1222 {
1223 return fuse_fsync_common(file, start, end, datasync, 1);
1224 }
1225
fuse_dir_ioctl(struct file * file,unsigned int cmd,unsigned long arg)1226 static long fuse_dir_ioctl(struct file *file, unsigned int cmd,
1227 unsigned long arg)
1228 {
1229 struct fuse_conn *fc = get_fuse_conn(file->f_mapping->host);
1230
1231 /* FUSE_IOCTL_DIR only supported for API version >= 7.18 */
1232 if (fc->minor < 18)
1233 return -ENOTTY;
1234
1235 return fuse_ioctl_common(file, cmd, arg, FUSE_IOCTL_DIR);
1236 }
1237
fuse_dir_compat_ioctl(struct file * file,unsigned int cmd,unsigned long arg)1238 static long fuse_dir_compat_ioctl(struct file *file, unsigned int cmd,
1239 unsigned long arg)
1240 {
1241 struct fuse_conn *fc = get_fuse_conn(file->f_mapping->host);
1242
1243 if (fc->minor < 18)
1244 return -ENOTTY;
1245
1246 return fuse_ioctl_common(file, cmd, arg,
1247 FUSE_IOCTL_COMPAT | FUSE_IOCTL_DIR);
1248 }
1249
update_mtime(unsigned ivalid)1250 static bool update_mtime(unsigned ivalid)
1251 {
1252 /* Always update if mtime is explicitly set */
1253 if (ivalid & ATTR_MTIME_SET)
1254 return true;
1255
1256 /* If it's an open(O_TRUNC) or an ftruncate(), don't update */
1257 if ((ivalid & ATTR_SIZE) && (ivalid & (ATTR_OPEN | ATTR_FILE)))
1258 return false;
1259
1260 /* In all other cases update */
1261 return true;
1262 }
1263
iattr_to_fattr(struct iattr * iattr,struct fuse_setattr_in * arg)1264 static void iattr_to_fattr(struct iattr *iattr, struct fuse_setattr_in *arg)
1265 {
1266 unsigned ivalid = iattr->ia_valid;
1267
1268 if (ivalid & ATTR_MODE)
1269 arg->valid |= FATTR_MODE, arg->mode = iattr->ia_mode;
1270 if (ivalid & ATTR_UID)
1271 arg->valid |= FATTR_UID, arg->uid = iattr->ia_uid;
1272 if (ivalid & ATTR_GID)
1273 arg->valid |= FATTR_GID, arg->gid = iattr->ia_gid;
1274 if (ivalid & ATTR_SIZE)
1275 arg->valid |= FATTR_SIZE, arg->size = iattr->ia_size;
1276 if (ivalid & ATTR_ATIME) {
1277 arg->valid |= FATTR_ATIME;
1278 arg->atime = iattr->ia_atime.tv_sec;
1279 arg->atimensec = iattr->ia_atime.tv_nsec;
1280 if (!(ivalid & ATTR_ATIME_SET))
1281 arg->valid |= FATTR_ATIME_NOW;
1282 }
1283 if ((ivalid & ATTR_MTIME) && update_mtime(ivalid)) {
1284 arg->valid |= FATTR_MTIME;
1285 arg->mtime = iattr->ia_mtime.tv_sec;
1286 arg->mtimensec = iattr->ia_mtime.tv_nsec;
1287 if (!(ivalid & ATTR_MTIME_SET))
1288 arg->valid |= FATTR_MTIME_NOW;
1289 }
1290 }
1291
1292 /*
1293 * Prevent concurrent writepages on inode
1294 *
1295 * This is done by adding a negative bias to the inode write counter
1296 * and waiting for all pending writes to finish.
1297 */
fuse_set_nowrite(struct inode * inode)1298 void fuse_set_nowrite(struct inode *inode)
1299 {
1300 struct fuse_conn *fc = get_fuse_conn(inode);
1301 struct fuse_inode *fi = get_fuse_inode(inode);
1302
1303 BUG_ON(!mutex_is_locked(&inode->i_mutex));
1304
1305 spin_lock(&fc->lock);
1306 BUG_ON(fi->writectr < 0);
1307 fi->writectr += FUSE_NOWRITE;
1308 spin_unlock(&fc->lock);
1309 wait_event(fi->page_waitq, fi->writectr == FUSE_NOWRITE);
1310 }
1311
1312 /*
1313 * Allow writepages on inode
1314 *
1315 * Remove the bias from the writecounter and send any queued
1316 * writepages.
1317 */
__fuse_release_nowrite(struct inode * inode)1318 static void __fuse_release_nowrite(struct inode *inode)
1319 {
1320 struct fuse_inode *fi = get_fuse_inode(inode);
1321
1322 BUG_ON(fi->writectr != FUSE_NOWRITE);
1323 fi->writectr = 0;
1324 fuse_flush_writepages(inode);
1325 }
1326
fuse_release_nowrite(struct inode * inode)1327 void fuse_release_nowrite(struct inode *inode)
1328 {
1329 struct fuse_conn *fc = get_fuse_conn(inode);
1330
1331 spin_lock(&fc->lock);
1332 __fuse_release_nowrite(inode);
1333 spin_unlock(&fc->lock);
1334 }
1335
1336 /*
1337 * Set attributes, and at the same time refresh them.
1338 *
1339 * Truncation is slightly complicated, because the 'truncate' request
1340 * may fail, in which case we don't want to touch the mapping.
1341 * vmtruncate() doesn't allow for this case, so do the rlimit checking
1342 * and the actual truncation by hand.
1343 */
fuse_do_setattr(struct dentry * entry,struct iattr * attr,struct file * file)1344 static int fuse_do_setattr(struct dentry *entry, struct iattr *attr,
1345 struct file *file)
1346 {
1347 struct inode *inode = entry->d_inode;
1348 struct fuse_conn *fc = get_fuse_conn(inode);
1349 struct fuse_req *req;
1350 struct fuse_setattr_in inarg;
1351 struct fuse_attr_out outarg;
1352 bool is_truncate = false;
1353 loff_t oldsize;
1354 int err;
1355
1356 if (!fuse_allow_task(fc, current))
1357 return -EACCES;
1358
1359 if (!(fc->flags & FUSE_DEFAULT_PERMISSIONS))
1360 attr->ia_valid |= ATTR_FORCE;
1361
1362 err = inode_change_ok(inode, attr);
1363 if (err)
1364 return err;
1365
1366 if (attr->ia_valid & ATTR_OPEN) {
1367 if (fc->atomic_o_trunc)
1368 return 0;
1369 file = NULL;
1370 }
1371
1372 if (attr->ia_valid & ATTR_SIZE)
1373 is_truncate = true;
1374
1375 req = fuse_get_req(fc);
1376 if (IS_ERR(req))
1377 return PTR_ERR(req);
1378
1379 if (is_truncate)
1380 fuse_set_nowrite(inode);
1381
1382 memset(&inarg, 0, sizeof(inarg));
1383 memset(&outarg, 0, sizeof(outarg));
1384 iattr_to_fattr(attr, &inarg);
1385 if (file) {
1386 struct fuse_file *ff = file->private_data;
1387 inarg.valid |= FATTR_FH;
1388 inarg.fh = ff->fh;
1389 }
1390 if (attr->ia_valid & ATTR_SIZE) {
1391 /* For mandatory locking in truncate */
1392 inarg.valid |= FATTR_LOCKOWNER;
1393 inarg.lock_owner = fuse_lock_owner_id(fc, current->files);
1394 }
1395 req->in.h.opcode = FUSE_SETATTR;
1396 req->in.h.nodeid = get_node_id(inode);
1397 req->in.numargs = 1;
1398 req->in.args[0].size = sizeof(inarg);
1399 req->in.args[0].value = &inarg;
1400 req->out.numargs = 1;
1401 if (fc->minor < 9)
1402 req->out.args[0].size = FUSE_COMPAT_ATTR_OUT_SIZE;
1403 else
1404 req->out.args[0].size = sizeof(outarg);
1405 req->out.args[0].value = &outarg;
1406 fuse_request_send(fc, req);
1407 err = req->out.h.error;
1408 fuse_put_request(fc, req);
1409 if (err) {
1410 if (err == -EINTR)
1411 fuse_invalidate_attr(inode);
1412 goto error;
1413 }
1414
1415 if ((inode->i_mode ^ outarg.attr.mode) & S_IFMT) {
1416 make_bad_inode(inode);
1417 err = -EIO;
1418 goto error;
1419 }
1420
1421 spin_lock(&fc->lock);
1422 fuse_change_attributes_common(inode, &outarg.attr,
1423 attr_timeout(&outarg));
1424 oldsize = inode->i_size;
1425 i_size_write(inode, outarg.attr.size);
1426
1427 if (is_truncate) {
1428 /* NOTE: this may release/reacquire fc->lock */
1429 __fuse_release_nowrite(inode);
1430 }
1431 spin_unlock(&fc->lock);
1432
1433 /*
1434 * Only call invalidate_inode_pages2() after removing
1435 * FUSE_NOWRITE, otherwise fuse_launder_page() would deadlock.
1436 */
1437 if (S_ISREG(inode->i_mode) && oldsize != outarg.attr.size) {
1438 truncate_pagecache(inode, oldsize, outarg.attr.size);
1439 invalidate_inode_pages2(inode->i_mapping);
1440 }
1441
1442 return 0;
1443
1444 error:
1445 if (is_truncate)
1446 fuse_release_nowrite(inode);
1447
1448 return err;
1449 }
1450
fuse_setattr(struct dentry * entry,struct iattr * attr)1451 static int fuse_setattr(struct dentry *entry, struct iattr *attr)
1452 {
1453 if (attr->ia_valid & ATTR_FILE)
1454 return fuse_do_setattr(entry, attr, attr->ia_file);
1455 else
1456 return fuse_do_setattr(entry, attr, NULL);
1457 }
1458
fuse_getattr(struct vfsmount * mnt,struct dentry * entry,struct kstat * stat)1459 static int fuse_getattr(struct vfsmount *mnt, struct dentry *entry,
1460 struct kstat *stat)
1461 {
1462 struct inode *inode = entry->d_inode;
1463 struct fuse_conn *fc = get_fuse_conn(inode);
1464
1465 if (!fuse_allow_task(fc, current))
1466 return -EACCES;
1467
1468 return fuse_update_attributes(inode, stat, NULL, NULL);
1469 }
1470
fuse_setxattr(struct dentry * entry,const char * name,const void * value,size_t size,int flags)1471 static int fuse_setxattr(struct dentry *entry, const char *name,
1472 const void *value, size_t size, int flags)
1473 {
1474 struct inode *inode = entry->d_inode;
1475 struct fuse_conn *fc = get_fuse_conn(inode);
1476 struct fuse_req *req;
1477 struct fuse_setxattr_in inarg;
1478 int err;
1479
1480 if (fc->no_setxattr)
1481 return -EOPNOTSUPP;
1482
1483 req = fuse_get_req(fc);
1484 if (IS_ERR(req))
1485 return PTR_ERR(req);
1486
1487 memset(&inarg, 0, sizeof(inarg));
1488 inarg.size = size;
1489 inarg.flags = flags;
1490 req->in.h.opcode = FUSE_SETXATTR;
1491 req->in.h.nodeid = get_node_id(inode);
1492 req->in.numargs = 3;
1493 req->in.args[0].size = sizeof(inarg);
1494 req->in.args[0].value = &inarg;
1495 req->in.args[1].size = strlen(name) + 1;
1496 req->in.args[1].value = name;
1497 req->in.args[2].size = size;
1498 req->in.args[2].value = value;
1499 fuse_request_send(fc, req);
1500 err = req->out.h.error;
1501 fuse_put_request(fc, req);
1502 if (err == -ENOSYS) {
1503 fc->no_setxattr = 1;
1504 err = -EOPNOTSUPP;
1505 }
1506 if (!err)
1507 fuse_invalidate_attr(inode);
1508 return err;
1509 }
1510
fuse_getxattr(struct dentry * entry,const char * name,void * value,size_t size)1511 static ssize_t fuse_getxattr(struct dentry *entry, const char *name,
1512 void *value, size_t size)
1513 {
1514 struct inode *inode = entry->d_inode;
1515 struct fuse_conn *fc = get_fuse_conn(inode);
1516 struct fuse_req *req;
1517 struct fuse_getxattr_in inarg;
1518 struct fuse_getxattr_out outarg;
1519 ssize_t ret;
1520
1521 if (fc->no_getxattr)
1522 return -EOPNOTSUPP;
1523
1524 req = fuse_get_req(fc);
1525 if (IS_ERR(req))
1526 return PTR_ERR(req);
1527
1528 memset(&inarg, 0, sizeof(inarg));
1529 inarg.size = size;
1530 req->in.h.opcode = FUSE_GETXATTR;
1531 req->in.h.nodeid = get_node_id(inode);
1532 req->in.numargs = 2;
1533 req->in.args[0].size = sizeof(inarg);
1534 req->in.args[0].value = &inarg;
1535 req->in.args[1].size = strlen(name) + 1;
1536 req->in.args[1].value = name;
1537 /* This is really two different operations rolled into one */
1538 req->out.numargs = 1;
1539 if (size) {
1540 req->out.argvar = 1;
1541 req->out.args[0].size = size;
1542 req->out.args[0].value = value;
1543 } else {
1544 req->out.args[0].size = sizeof(outarg);
1545 req->out.args[0].value = &outarg;
1546 }
1547 fuse_request_send(fc, req);
1548 ret = req->out.h.error;
1549 if (!ret)
1550 ret = size ? req->out.args[0].size : outarg.size;
1551 else {
1552 if (ret == -ENOSYS) {
1553 fc->no_getxattr = 1;
1554 ret = -EOPNOTSUPP;
1555 }
1556 }
1557 fuse_put_request(fc, req);
1558 return ret;
1559 }
1560
fuse_listxattr(struct dentry * entry,char * list,size_t size)1561 static ssize_t fuse_listxattr(struct dentry *entry, char *list, size_t size)
1562 {
1563 struct inode *inode = entry->d_inode;
1564 struct fuse_conn *fc = get_fuse_conn(inode);
1565 struct fuse_req *req;
1566 struct fuse_getxattr_in inarg;
1567 struct fuse_getxattr_out outarg;
1568 ssize_t ret;
1569
1570 if (!fuse_allow_task(fc, current))
1571 return -EACCES;
1572
1573 if (fc->no_listxattr)
1574 return -EOPNOTSUPP;
1575
1576 req = fuse_get_req(fc);
1577 if (IS_ERR(req))
1578 return PTR_ERR(req);
1579
1580 memset(&inarg, 0, sizeof(inarg));
1581 inarg.size = size;
1582 req->in.h.opcode = FUSE_LISTXATTR;
1583 req->in.h.nodeid = get_node_id(inode);
1584 req->in.numargs = 1;
1585 req->in.args[0].size = sizeof(inarg);
1586 req->in.args[0].value = &inarg;
1587 /* This is really two different operations rolled into one */
1588 req->out.numargs = 1;
1589 if (size) {
1590 req->out.argvar = 1;
1591 req->out.args[0].size = size;
1592 req->out.args[0].value = list;
1593 } else {
1594 req->out.args[0].size = sizeof(outarg);
1595 req->out.args[0].value = &outarg;
1596 }
1597 fuse_request_send(fc, req);
1598 ret = req->out.h.error;
1599 if (!ret)
1600 ret = size ? req->out.args[0].size : outarg.size;
1601 else {
1602 if (ret == -ENOSYS) {
1603 fc->no_listxattr = 1;
1604 ret = -EOPNOTSUPP;
1605 }
1606 }
1607 fuse_put_request(fc, req);
1608 return ret;
1609 }
1610
fuse_removexattr(struct dentry * entry,const char * name)1611 static int fuse_removexattr(struct dentry *entry, const char *name)
1612 {
1613 struct inode *inode = entry->d_inode;
1614 struct fuse_conn *fc = get_fuse_conn(inode);
1615 struct fuse_req *req;
1616 int err;
1617
1618 if (fc->no_removexattr)
1619 return -EOPNOTSUPP;
1620
1621 req = fuse_get_req(fc);
1622 if (IS_ERR(req))
1623 return PTR_ERR(req);
1624
1625 req->in.h.opcode = FUSE_REMOVEXATTR;
1626 req->in.h.nodeid = get_node_id(inode);
1627 req->in.numargs = 1;
1628 req->in.args[0].size = strlen(name) + 1;
1629 req->in.args[0].value = name;
1630 fuse_request_send(fc, req);
1631 err = req->out.h.error;
1632 fuse_put_request(fc, req);
1633 if (err == -ENOSYS) {
1634 fc->no_removexattr = 1;
1635 err = -EOPNOTSUPP;
1636 }
1637 if (!err)
1638 fuse_invalidate_attr(inode);
1639 return err;
1640 }
1641
1642 static const struct inode_operations fuse_dir_inode_operations = {
1643 .lookup = fuse_lookup,
1644 .mkdir = fuse_mkdir,
1645 .symlink = fuse_symlink,
1646 .unlink = fuse_unlink,
1647 .rmdir = fuse_rmdir,
1648 .rename = fuse_rename,
1649 .link = fuse_link,
1650 .setattr = fuse_setattr,
1651 .create = fuse_create,
1652 .mknod = fuse_mknod,
1653 .permission = fuse_permission,
1654 .getattr = fuse_getattr,
1655 .setxattr = fuse_setxattr,
1656 .getxattr = fuse_getxattr,
1657 .listxattr = fuse_listxattr,
1658 .removexattr = fuse_removexattr,
1659 };
1660
1661 static const struct file_operations fuse_dir_operations = {
1662 .llseek = generic_file_llseek,
1663 .read = generic_read_dir,
1664 .readdir = fuse_readdir,
1665 .open = fuse_dir_open,
1666 .release = fuse_dir_release,
1667 .fsync = fuse_dir_fsync,
1668 .unlocked_ioctl = fuse_dir_ioctl,
1669 .compat_ioctl = fuse_dir_compat_ioctl,
1670 };
1671
1672 static const struct inode_operations fuse_common_inode_operations = {
1673 .setattr = fuse_setattr,
1674 .permission = fuse_permission,
1675 .getattr = fuse_getattr,
1676 .setxattr = fuse_setxattr,
1677 .getxattr = fuse_getxattr,
1678 .listxattr = fuse_listxattr,
1679 .removexattr = fuse_removexattr,
1680 };
1681
1682 static const struct inode_operations fuse_symlink_inode_operations = {
1683 .setattr = fuse_setattr,
1684 .follow_link = fuse_follow_link,
1685 .put_link = fuse_put_link,
1686 .readlink = generic_readlink,
1687 .getattr = fuse_getattr,
1688 .setxattr = fuse_setxattr,
1689 .getxattr = fuse_getxattr,
1690 .listxattr = fuse_listxattr,
1691 .removexattr = fuse_removexattr,
1692 };
1693
fuse_init_common(struct inode * inode)1694 void fuse_init_common(struct inode *inode)
1695 {
1696 inode->i_op = &fuse_common_inode_operations;
1697 }
1698
fuse_init_dir(struct inode * inode)1699 void fuse_init_dir(struct inode *inode)
1700 {
1701 inode->i_op = &fuse_dir_inode_operations;
1702 inode->i_fop = &fuse_dir_operations;
1703 }
1704
fuse_init_symlink(struct inode * inode)1705 void fuse_init_symlink(struct inode *inode)
1706 {
1707 inode->i_op = &fuse_symlink_inode_operations;
1708 }
1709