1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3 * Copyright (C) International Business Machines Corp., 2000-2004
4 * Portions Copyright (C) Christoph Hellwig, 2001-2002
5 */
6
7 #include <linux/fs.h>
8 #include <linux/module.h>
9 #include <linux/parser.h>
10 #include <linux/completion.h>
11 #include <linux/vfs.h>
12 #include <linux/quotaops.h>
13 #include <linux/mount.h>
14 #include <linux/moduleparam.h>
15 #include <linux/kthread.h>
16 #include <linux/posix_acl.h>
17 #include <linux/buffer_head.h>
18 #include <linux/exportfs.h>
19 #include <linux/crc32.h>
20 #include <linux/slab.h>
21 #include <linux/uaccess.h>
22 #include <linux/seq_file.h>
23 #include <linux/blkdev.h>
24
25 #include "jfs_incore.h"
26 #include "jfs_filsys.h"
27 #include "jfs_inode.h"
28 #include "jfs_metapage.h"
29 #include "jfs_superblock.h"
30 #include "jfs_dmap.h"
31 #include "jfs_imap.h"
32 #include "jfs_acl.h"
33 #include "jfs_debug.h"
34 #include "jfs_xattr.h"
35 #include "jfs_dinode.h"
36
37 MODULE_DESCRIPTION("The Journaled Filesystem (JFS)");
38 MODULE_AUTHOR("Steve Best/Dave Kleikamp/Barry Arndt, IBM");
39 MODULE_LICENSE("GPL");
40 MODULE_IMPORT_NS(ANDROID_GKI_VFS_EXPORT_ONLY);
41
42 static struct kmem_cache *jfs_inode_cachep;
43
44 static const struct super_operations jfs_super_operations;
45 static const struct export_operations jfs_export_operations;
46 static struct file_system_type jfs_fs_type;
47
48 #define MAX_COMMIT_THREADS 64
49 static int commit_threads;
50 module_param(commit_threads, int, 0);
51 MODULE_PARM_DESC(commit_threads, "Number of commit threads");
52
53 static struct task_struct *jfsCommitThread[MAX_COMMIT_THREADS];
54 struct task_struct *jfsIOthread;
55 struct task_struct *jfsSyncThread;
56
57 #ifdef CONFIG_JFS_DEBUG
58 int jfsloglevel = JFS_LOGLEVEL_WARN;
59 module_param(jfsloglevel, int, 0644);
60 MODULE_PARM_DESC(jfsloglevel, "Specify JFS loglevel (0, 1 or 2)");
61 #endif
62
jfs_handle_error(struct super_block * sb)63 static void jfs_handle_error(struct super_block *sb)
64 {
65 struct jfs_sb_info *sbi = JFS_SBI(sb);
66
67 if (sb_rdonly(sb))
68 return;
69
70 updateSuper(sb, FM_DIRTY);
71
72 if (sbi->flag & JFS_ERR_PANIC)
73 panic("JFS (device %s): panic forced after error\n",
74 sb->s_id);
75 else if (sbi->flag & JFS_ERR_REMOUNT_RO) {
76 jfs_err("ERROR: (device %s): remounting filesystem as read-only",
77 sb->s_id);
78 sb->s_flags |= SB_RDONLY;
79 }
80
81 /* nothing is done for continue beyond marking the superblock dirty */
82 }
83
jfs_error(struct super_block * sb,const char * fmt,...)84 void jfs_error(struct super_block *sb, const char *fmt, ...)
85 {
86 struct va_format vaf;
87 va_list args;
88
89 va_start(args, fmt);
90
91 vaf.fmt = fmt;
92 vaf.va = &args;
93
94 pr_err("ERROR: (device %s): %ps: %pV\n",
95 sb->s_id, __builtin_return_address(0), &vaf);
96
97 va_end(args);
98
99 jfs_handle_error(sb);
100 }
101
jfs_alloc_inode(struct super_block * sb)102 static struct inode *jfs_alloc_inode(struct super_block *sb)
103 {
104 struct jfs_inode_info *jfs_inode;
105
106 jfs_inode = alloc_inode_sb(sb, jfs_inode_cachep, GFP_NOFS);
107 if (!jfs_inode)
108 return NULL;
109 #ifdef CONFIG_QUOTA
110 memset(&jfs_inode->i_dquot, 0, sizeof(jfs_inode->i_dquot));
111 #endif
112 return &jfs_inode->vfs_inode;
113 }
114
jfs_free_inode(struct inode * inode)115 static void jfs_free_inode(struct inode *inode)
116 {
117 kmem_cache_free(jfs_inode_cachep, JFS_IP(inode));
118 }
119
jfs_statfs(struct dentry * dentry,struct kstatfs * buf)120 static int jfs_statfs(struct dentry *dentry, struct kstatfs *buf)
121 {
122 struct jfs_sb_info *sbi = JFS_SBI(dentry->d_sb);
123 s64 maxinodes;
124 struct inomap *imap = JFS_IP(sbi->ipimap)->i_imap;
125
126 jfs_info("In jfs_statfs");
127 buf->f_type = JFS_SUPER_MAGIC;
128 buf->f_bsize = sbi->bsize;
129 buf->f_blocks = sbi->bmap->db_mapsize;
130 buf->f_bfree = sbi->bmap->db_nfree;
131 buf->f_bavail = sbi->bmap->db_nfree;
132 /*
133 * If we really return the number of allocated & free inodes, some
134 * applications will fail because they won't see enough free inodes.
135 * We'll try to calculate some guess as to how many inodes we can
136 * really allocate
137 *
138 * buf->f_files = atomic_read(&imap->im_numinos);
139 * buf->f_ffree = atomic_read(&imap->im_numfree);
140 */
141 maxinodes = min((s64) atomic_read(&imap->im_numinos) +
142 ((sbi->bmap->db_nfree >> imap->im_l2nbperiext)
143 << L2INOSPEREXT), (s64) 0xffffffffLL);
144 buf->f_files = maxinodes;
145 buf->f_ffree = maxinodes - (atomic_read(&imap->im_numinos) -
146 atomic_read(&imap->im_numfree));
147 buf->f_fsid.val[0] = crc32_le(0, (char *)&sbi->uuid,
148 sizeof(sbi->uuid)/2);
149 buf->f_fsid.val[1] = crc32_le(0,
150 (char *)&sbi->uuid + sizeof(sbi->uuid)/2,
151 sizeof(sbi->uuid)/2);
152
153 buf->f_namelen = JFS_NAME_MAX;
154 return 0;
155 }
156
157 #ifdef CONFIG_QUOTA
158 static int jfs_quota_off(struct super_block *sb, int type);
159 static int jfs_quota_on(struct super_block *sb, int type, int format_id,
160 const struct path *path);
161
jfs_quota_off_umount(struct super_block * sb)162 static void jfs_quota_off_umount(struct super_block *sb)
163 {
164 int type;
165
166 for (type = 0; type < MAXQUOTAS; type++)
167 jfs_quota_off(sb, type);
168 }
169
170 static const struct quotactl_ops jfs_quotactl_ops = {
171 .quota_on = jfs_quota_on,
172 .quota_off = jfs_quota_off,
173 .quota_sync = dquot_quota_sync,
174 .get_state = dquot_get_state,
175 .set_info = dquot_set_dqinfo,
176 .get_dqblk = dquot_get_dqblk,
177 .set_dqblk = dquot_set_dqblk,
178 .get_nextdqblk = dquot_get_next_dqblk,
179 };
180 #else
jfs_quota_off_umount(struct super_block * sb)181 static inline void jfs_quota_off_umount(struct super_block *sb)
182 {
183 }
184 #endif
185
jfs_put_super(struct super_block * sb)186 static void jfs_put_super(struct super_block *sb)
187 {
188 struct jfs_sb_info *sbi = JFS_SBI(sb);
189 int rc;
190
191 jfs_info("In jfs_put_super");
192
193 jfs_quota_off_umount(sb);
194
195 rc = jfs_umount(sb);
196 if (rc)
197 jfs_err("jfs_umount failed with return code %d", rc);
198
199 unload_nls(sbi->nls_tab);
200
201 truncate_inode_pages(sbi->direct_inode->i_mapping, 0);
202 iput(sbi->direct_inode);
203
204 kfree(sbi);
205 }
206
207 enum {
208 Opt_integrity, Opt_nointegrity, Opt_iocharset, Opt_resize,
209 Opt_resize_nosize, Opt_errors, Opt_ignore, Opt_err, Opt_quota,
210 Opt_usrquota, Opt_grpquota, Opt_uid, Opt_gid, Opt_umask,
211 Opt_discard, Opt_nodiscard, Opt_discard_minblk
212 };
213
214 static const match_table_t tokens = {
215 {Opt_integrity, "integrity"},
216 {Opt_nointegrity, "nointegrity"},
217 {Opt_iocharset, "iocharset=%s"},
218 {Opt_resize, "resize=%u"},
219 {Opt_resize_nosize, "resize"},
220 {Opt_errors, "errors=%s"},
221 {Opt_ignore, "noquota"},
222 {Opt_quota, "quota"},
223 {Opt_usrquota, "usrquota"},
224 {Opt_grpquota, "grpquota"},
225 {Opt_uid, "uid=%u"},
226 {Opt_gid, "gid=%u"},
227 {Opt_umask, "umask=%u"},
228 {Opt_discard, "discard"},
229 {Opt_nodiscard, "nodiscard"},
230 {Opt_discard_minblk, "discard=%u"},
231 {Opt_err, NULL}
232 };
233
parse_options(char * options,struct super_block * sb,s64 * newLVSize,int * flag)234 static int parse_options(char *options, struct super_block *sb, s64 *newLVSize,
235 int *flag)
236 {
237 void *nls_map = (void *)-1; /* -1: no change; NULL: none */
238 char *p;
239 struct jfs_sb_info *sbi = JFS_SBI(sb);
240
241 *newLVSize = 0;
242
243 if (!options)
244 return 1;
245
246 while ((p = strsep(&options, ",")) != NULL) {
247 substring_t args[MAX_OPT_ARGS];
248 int token;
249 if (!*p)
250 continue;
251
252 token = match_token(p, tokens, args);
253 switch (token) {
254 case Opt_integrity:
255 *flag &= ~JFS_NOINTEGRITY;
256 break;
257 case Opt_nointegrity:
258 *flag |= JFS_NOINTEGRITY;
259 break;
260 case Opt_ignore:
261 /* Silently ignore the quota options */
262 /* Don't do anything ;-) */
263 break;
264 case Opt_iocharset:
265 if (nls_map && nls_map != (void *) -1)
266 unload_nls(nls_map);
267 if (!strcmp(args[0].from, "none"))
268 nls_map = NULL;
269 else {
270 nls_map = load_nls(args[0].from);
271 if (!nls_map) {
272 pr_err("JFS: charset not found\n");
273 goto cleanup;
274 }
275 }
276 break;
277 case Opt_resize:
278 {
279 char *resize = args[0].from;
280 int rc = kstrtoll(resize, 0, newLVSize);
281
282 if (rc)
283 goto cleanup;
284 break;
285 }
286 case Opt_resize_nosize:
287 {
288 *newLVSize = sb_bdev_nr_blocks(sb);
289 if (*newLVSize == 0)
290 pr_err("JFS: Cannot determine volume size\n");
291 break;
292 }
293 case Opt_errors:
294 {
295 char *errors = args[0].from;
296 if (!errors || !*errors)
297 goto cleanup;
298 if (!strcmp(errors, "continue")) {
299 *flag &= ~JFS_ERR_REMOUNT_RO;
300 *flag &= ~JFS_ERR_PANIC;
301 *flag |= JFS_ERR_CONTINUE;
302 } else if (!strcmp(errors, "remount-ro")) {
303 *flag &= ~JFS_ERR_CONTINUE;
304 *flag &= ~JFS_ERR_PANIC;
305 *flag |= JFS_ERR_REMOUNT_RO;
306 } else if (!strcmp(errors, "panic")) {
307 *flag &= ~JFS_ERR_CONTINUE;
308 *flag &= ~JFS_ERR_REMOUNT_RO;
309 *flag |= JFS_ERR_PANIC;
310 } else {
311 pr_err("JFS: %s is an invalid error handler\n",
312 errors);
313 goto cleanup;
314 }
315 break;
316 }
317
318 #ifdef CONFIG_QUOTA
319 case Opt_quota:
320 case Opt_usrquota:
321 *flag |= JFS_USRQUOTA;
322 break;
323 case Opt_grpquota:
324 *flag |= JFS_GRPQUOTA;
325 break;
326 #else
327 case Opt_usrquota:
328 case Opt_grpquota:
329 case Opt_quota:
330 pr_err("JFS: quota operations not supported\n");
331 break;
332 #endif
333 case Opt_uid:
334 {
335 char *uid = args[0].from;
336 uid_t val;
337 int rc = kstrtouint(uid, 0, &val);
338
339 if (rc)
340 goto cleanup;
341 sbi->uid = make_kuid(current_user_ns(), val);
342 if (!uid_valid(sbi->uid))
343 goto cleanup;
344 break;
345 }
346
347 case Opt_gid:
348 {
349 char *gid = args[0].from;
350 gid_t val;
351 int rc = kstrtouint(gid, 0, &val);
352
353 if (rc)
354 goto cleanup;
355 sbi->gid = make_kgid(current_user_ns(), val);
356 if (!gid_valid(sbi->gid))
357 goto cleanup;
358 break;
359 }
360
361 case Opt_umask:
362 {
363 char *umask = args[0].from;
364 int rc = kstrtouint(umask, 8, &sbi->umask);
365
366 if (rc)
367 goto cleanup;
368 if (sbi->umask & ~0777) {
369 pr_err("JFS: Invalid value of umask\n");
370 goto cleanup;
371 }
372 break;
373 }
374
375 case Opt_discard:
376 /* if set to 1, even copying files will cause
377 * trimming :O
378 * -> user has more control over the online trimming
379 */
380 sbi->minblks_trim = 64;
381 if (bdev_max_discard_sectors(sb->s_bdev))
382 *flag |= JFS_DISCARD;
383 else
384 pr_err("JFS: discard option not supported on device\n");
385 break;
386
387 case Opt_nodiscard:
388 *flag &= ~JFS_DISCARD;
389 break;
390
391 case Opt_discard_minblk:
392 {
393 char *minblks_trim = args[0].from;
394 int rc;
395 if (bdev_max_discard_sectors(sb->s_bdev)) {
396 *flag |= JFS_DISCARD;
397 rc = kstrtouint(minblks_trim, 0,
398 &sbi->minblks_trim);
399 if (rc)
400 goto cleanup;
401 } else
402 pr_err("JFS: discard option not supported on device\n");
403 break;
404 }
405
406 default:
407 printk("jfs: Unrecognized mount option \"%s\" or missing value\n",
408 p);
409 goto cleanup;
410 }
411 }
412
413 if (nls_map != (void *) -1) {
414 /* Discard old (if remount) */
415 unload_nls(sbi->nls_tab);
416 sbi->nls_tab = nls_map;
417 }
418 return 1;
419
420 cleanup:
421 if (nls_map && nls_map != (void *) -1)
422 unload_nls(nls_map);
423 return 0;
424 }
425
jfs_remount(struct super_block * sb,int * flags,char * data)426 static int jfs_remount(struct super_block *sb, int *flags, char *data)
427 {
428 s64 newLVSize = 0;
429 int rc = 0;
430 int flag = JFS_SBI(sb)->flag;
431 int ret;
432
433 sync_filesystem(sb);
434 if (!parse_options(data, sb, &newLVSize, &flag))
435 return -EINVAL;
436
437 if (newLVSize) {
438 if (sb_rdonly(sb)) {
439 pr_err("JFS: resize requires volume to be mounted read-write\n");
440 return -EROFS;
441 }
442 rc = jfs_extendfs(sb, newLVSize, 0);
443 if (rc)
444 return rc;
445 }
446
447 if (sb_rdonly(sb) && !(*flags & SB_RDONLY)) {
448 /*
449 * Invalidate any previously read metadata. fsck may have
450 * changed the on-disk data since we mounted r/o
451 */
452 truncate_inode_pages(JFS_SBI(sb)->direct_inode->i_mapping, 0);
453
454 JFS_SBI(sb)->flag = flag;
455 ret = jfs_mount_rw(sb, 1);
456
457 /* mark the fs r/w for quota activity */
458 sb->s_flags &= ~SB_RDONLY;
459
460 dquot_resume(sb, -1);
461 return ret;
462 }
463 if (!sb_rdonly(sb) && (*flags & SB_RDONLY)) {
464 rc = dquot_suspend(sb, -1);
465 if (rc < 0)
466 return rc;
467 rc = jfs_umount_rw(sb);
468 JFS_SBI(sb)->flag = flag;
469 return rc;
470 }
471 if ((JFS_SBI(sb)->flag & JFS_NOINTEGRITY) != (flag & JFS_NOINTEGRITY))
472 if (!sb_rdonly(sb)) {
473 rc = jfs_umount_rw(sb);
474 if (rc)
475 return rc;
476
477 JFS_SBI(sb)->flag = flag;
478 ret = jfs_mount_rw(sb, 1);
479 return ret;
480 }
481 JFS_SBI(sb)->flag = flag;
482
483 return 0;
484 }
485
jfs_fill_super(struct super_block * sb,void * data,int silent)486 static int jfs_fill_super(struct super_block *sb, void *data, int silent)
487 {
488 struct jfs_sb_info *sbi;
489 struct inode *inode;
490 int rc;
491 s64 newLVSize = 0;
492 int flag, ret = -EINVAL;
493
494 jfs_info("In jfs_read_super: s_flags=0x%lx", sb->s_flags);
495
496 sbi = kzalloc(sizeof(struct jfs_sb_info), GFP_KERNEL);
497 if (!sbi)
498 return -ENOMEM;
499
500 sb->s_fs_info = sbi;
501 sb->s_max_links = JFS_LINK_MAX;
502 sb->s_time_min = 0;
503 sb->s_time_max = U32_MAX;
504 sbi->sb = sb;
505 sbi->uid = INVALID_UID;
506 sbi->gid = INVALID_GID;
507 sbi->umask = -1;
508
509 /* initialize the mount flag and determine the default error handler */
510 flag = JFS_ERR_REMOUNT_RO;
511
512 if (!parse_options((char *) data, sb, &newLVSize, &flag))
513 goto out_kfree;
514 sbi->flag = flag;
515
516 #ifdef CONFIG_JFS_POSIX_ACL
517 sb->s_flags |= SB_POSIXACL;
518 #endif
519
520 if (newLVSize) {
521 pr_err("resize option for remount only\n");
522 goto out_kfree;
523 }
524
525 /*
526 * Initialize blocksize to 4K.
527 */
528 sb_set_blocksize(sb, PSIZE);
529
530 /*
531 * Set method vectors.
532 */
533 sb->s_op = &jfs_super_operations;
534 sb->s_export_op = &jfs_export_operations;
535 sb->s_xattr = jfs_xattr_handlers;
536 #ifdef CONFIG_QUOTA
537 sb->dq_op = &dquot_operations;
538 sb->s_qcop = &jfs_quotactl_ops;
539 sb->s_quota_types = QTYPE_MASK_USR | QTYPE_MASK_GRP;
540 #endif
541
542 /*
543 * Initialize direct-mapping inode/address-space
544 */
545 inode = new_inode(sb);
546 if (inode == NULL) {
547 ret = -ENOMEM;
548 goto out_unload;
549 }
550 inode->i_size = bdev_nr_bytes(sb->s_bdev);
551 inode->i_mapping->a_ops = &jfs_metapage_aops;
552 inode_fake_hash(inode);
553 mapping_set_gfp_mask(inode->i_mapping, GFP_NOFS);
554
555 sbi->direct_inode = inode;
556
557 rc = jfs_mount(sb);
558 if (rc) {
559 if (!silent)
560 jfs_err("jfs_mount failed w/return code = %d", rc);
561 goto out_mount_failed;
562 }
563 if (sb_rdonly(sb))
564 sbi->log = NULL;
565 else {
566 rc = jfs_mount_rw(sb, 0);
567 if (rc) {
568 if (!silent) {
569 jfs_err("jfs_mount_rw failed, return code = %d",
570 rc);
571 }
572 goto out_no_rw;
573 }
574 }
575
576 sb->s_magic = JFS_SUPER_MAGIC;
577
578 if (sbi->mntflag & JFS_OS2)
579 sb->s_d_op = &jfs_ci_dentry_operations;
580
581 inode = jfs_iget(sb, ROOT_I);
582 if (IS_ERR(inode)) {
583 ret = PTR_ERR(inode);
584 goto out_no_rw;
585 }
586 sb->s_root = d_make_root(inode);
587 if (!sb->s_root)
588 goto out_no_root;
589
590 /* logical blocks are represented by 40 bits in pxd_t, etc.
591 * and page cache is indexed by long
592 */
593 sb->s_maxbytes = min(((loff_t)sb->s_blocksize) << 40, MAX_LFS_FILESIZE);
594 sb->s_time_gran = 1;
595 return 0;
596
597 out_no_root:
598 jfs_err("jfs_read_super: get root dentry failed");
599
600 out_no_rw:
601 rc = jfs_umount(sb);
602 if (rc)
603 jfs_err("jfs_umount failed with return code %d", rc);
604 out_mount_failed:
605 filemap_write_and_wait(sbi->direct_inode->i_mapping);
606 truncate_inode_pages(sbi->direct_inode->i_mapping, 0);
607 make_bad_inode(sbi->direct_inode);
608 iput(sbi->direct_inode);
609 sbi->direct_inode = NULL;
610 out_unload:
611 unload_nls(sbi->nls_tab);
612 out_kfree:
613 kfree(sbi);
614 return ret;
615 }
616
jfs_freeze(struct super_block * sb)617 static int jfs_freeze(struct super_block *sb)
618 {
619 struct jfs_sb_info *sbi = JFS_SBI(sb);
620 struct jfs_log *log = sbi->log;
621 int rc = 0;
622
623 if (!sb_rdonly(sb)) {
624 txQuiesce(sb);
625 rc = lmLogShutdown(log);
626 if (rc) {
627 jfs_error(sb, "lmLogShutdown failed\n");
628
629 /* let operations fail rather than hang */
630 txResume(sb);
631
632 return rc;
633 }
634 rc = updateSuper(sb, FM_CLEAN);
635 if (rc) {
636 jfs_err("jfs_freeze: updateSuper failed");
637 /*
638 * Don't fail here. Everything succeeded except
639 * marking the superblock clean, so there's really
640 * no harm in leaving it frozen for now.
641 */
642 }
643 }
644 return 0;
645 }
646
jfs_unfreeze(struct super_block * sb)647 static int jfs_unfreeze(struct super_block *sb)
648 {
649 struct jfs_sb_info *sbi = JFS_SBI(sb);
650 struct jfs_log *log = sbi->log;
651 int rc = 0;
652
653 if (!sb_rdonly(sb)) {
654 rc = updateSuper(sb, FM_MOUNT);
655 if (rc) {
656 jfs_error(sb, "updateSuper failed\n");
657 goto out;
658 }
659 rc = lmLogInit(log);
660 if (rc)
661 jfs_error(sb, "lmLogInit failed\n");
662 out:
663 txResume(sb);
664 }
665 return rc;
666 }
667
jfs_do_mount(struct file_system_type * fs_type,int flags,const char * dev_name,void * data)668 static struct dentry *jfs_do_mount(struct file_system_type *fs_type,
669 int flags, const char *dev_name, void *data)
670 {
671 return mount_bdev(fs_type, flags, dev_name, data, jfs_fill_super);
672 }
673
jfs_sync_fs(struct super_block * sb,int wait)674 static int jfs_sync_fs(struct super_block *sb, int wait)
675 {
676 struct jfs_log *log = JFS_SBI(sb)->log;
677
678 /* log == NULL indicates read-only mount */
679 if (log) {
680 /*
681 * Write quota structures to quota file, sync_blockdev() will
682 * write them to disk later
683 */
684 dquot_writeback_dquots(sb, -1);
685 jfs_flush_journal(log, wait);
686 jfs_syncpt(log, 0);
687 }
688
689 return 0;
690 }
691
jfs_show_options(struct seq_file * seq,struct dentry * root)692 static int jfs_show_options(struct seq_file *seq, struct dentry *root)
693 {
694 struct jfs_sb_info *sbi = JFS_SBI(root->d_sb);
695
696 if (uid_valid(sbi->uid))
697 seq_printf(seq, ",uid=%d", from_kuid(&init_user_ns, sbi->uid));
698 if (gid_valid(sbi->gid))
699 seq_printf(seq, ",gid=%d", from_kgid(&init_user_ns, sbi->gid));
700 if (sbi->umask != -1)
701 seq_printf(seq, ",umask=%03o", sbi->umask);
702 if (sbi->flag & JFS_NOINTEGRITY)
703 seq_puts(seq, ",nointegrity");
704 if (sbi->flag & JFS_DISCARD)
705 seq_printf(seq, ",discard=%u", sbi->minblks_trim);
706 if (sbi->nls_tab)
707 seq_printf(seq, ",iocharset=%s", sbi->nls_tab->charset);
708 if (sbi->flag & JFS_ERR_CONTINUE)
709 seq_printf(seq, ",errors=continue");
710 if (sbi->flag & JFS_ERR_PANIC)
711 seq_printf(seq, ",errors=panic");
712
713 #ifdef CONFIG_QUOTA
714 if (sbi->flag & JFS_USRQUOTA)
715 seq_puts(seq, ",usrquota");
716
717 if (sbi->flag & JFS_GRPQUOTA)
718 seq_puts(seq, ",grpquota");
719 #endif
720
721 return 0;
722 }
723
724 #ifdef CONFIG_QUOTA
725
726 /* Read data from quotafile - avoid pagecache and such because we cannot afford
727 * acquiring the locks... As quota files are never truncated and quota code
728 * itself serializes the operations (and no one else should touch the files)
729 * we don't have to be afraid of races */
jfs_quota_read(struct super_block * sb,int type,char * data,size_t len,loff_t off)730 static ssize_t jfs_quota_read(struct super_block *sb, int type, char *data,
731 size_t len, loff_t off)
732 {
733 struct inode *inode = sb_dqopt(sb)->files[type];
734 sector_t blk = off >> sb->s_blocksize_bits;
735 int err = 0;
736 int offset = off & (sb->s_blocksize - 1);
737 int tocopy;
738 size_t toread;
739 struct buffer_head tmp_bh;
740 struct buffer_head *bh;
741 loff_t i_size = i_size_read(inode);
742
743 if (off > i_size)
744 return 0;
745 if (off+len > i_size)
746 len = i_size-off;
747 toread = len;
748 while (toread > 0) {
749 tocopy = min_t(size_t, sb->s_blocksize - offset, toread);
750
751 tmp_bh.b_state = 0;
752 tmp_bh.b_size = i_blocksize(inode);
753 err = jfs_get_block(inode, blk, &tmp_bh, 0);
754 if (err)
755 return err;
756 if (!buffer_mapped(&tmp_bh)) /* A hole? */
757 memset(data, 0, tocopy);
758 else {
759 bh = sb_bread(sb, tmp_bh.b_blocknr);
760 if (!bh)
761 return -EIO;
762 memcpy(data, bh->b_data+offset, tocopy);
763 brelse(bh);
764 }
765 offset = 0;
766 toread -= tocopy;
767 data += tocopy;
768 blk++;
769 }
770 return len;
771 }
772
773 /* Write to quotafile */
jfs_quota_write(struct super_block * sb,int type,const char * data,size_t len,loff_t off)774 static ssize_t jfs_quota_write(struct super_block *sb, int type,
775 const char *data, size_t len, loff_t off)
776 {
777 struct inode *inode = sb_dqopt(sb)->files[type];
778 sector_t blk = off >> sb->s_blocksize_bits;
779 int err = 0;
780 int offset = off & (sb->s_blocksize - 1);
781 int tocopy;
782 size_t towrite = len;
783 struct buffer_head tmp_bh;
784 struct buffer_head *bh;
785
786 inode_lock(inode);
787 while (towrite > 0) {
788 tocopy = min_t(size_t, sb->s_blocksize - offset, towrite);
789
790 tmp_bh.b_state = 0;
791 tmp_bh.b_size = i_blocksize(inode);
792 err = jfs_get_block(inode, blk, &tmp_bh, 1);
793 if (err)
794 goto out;
795 if (offset || tocopy != sb->s_blocksize)
796 bh = sb_bread(sb, tmp_bh.b_blocknr);
797 else
798 bh = sb_getblk(sb, tmp_bh.b_blocknr);
799 if (!bh) {
800 err = -EIO;
801 goto out;
802 }
803 lock_buffer(bh);
804 memcpy(bh->b_data+offset, data, tocopy);
805 flush_dcache_page(bh->b_page);
806 set_buffer_uptodate(bh);
807 mark_buffer_dirty(bh);
808 unlock_buffer(bh);
809 brelse(bh);
810 offset = 0;
811 towrite -= tocopy;
812 data += tocopy;
813 blk++;
814 }
815 out:
816 if (len == towrite) {
817 inode_unlock(inode);
818 return err;
819 }
820 if (inode->i_size < off+len-towrite)
821 i_size_write(inode, off+len-towrite);
822 inode->i_mtime = inode_set_ctime_current(inode);
823 mark_inode_dirty(inode);
824 inode_unlock(inode);
825 return len - towrite;
826 }
827
jfs_get_dquots(struct inode * inode)828 static struct dquot __rcu **jfs_get_dquots(struct inode *inode)
829 {
830 return JFS_IP(inode)->i_dquot;
831 }
832
jfs_quota_on(struct super_block * sb,int type,int format_id,const struct path * path)833 static int jfs_quota_on(struct super_block *sb, int type, int format_id,
834 const struct path *path)
835 {
836 int err;
837 struct inode *inode;
838
839 err = dquot_quota_on(sb, type, format_id, path);
840 if (err)
841 return err;
842
843 inode = d_inode(path->dentry);
844 inode_lock(inode);
845 JFS_IP(inode)->mode2 |= JFS_NOATIME_FL | JFS_IMMUTABLE_FL;
846 inode_set_flags(inode, S_NOATIME | S_IMMUTABLE,
847 S_NOATIME | S_IMMUTABLE);
848 inode_unlock(inode);
849 mark_inode_dirty(inode);
850
851 return 0;
852 }
853
jfs_quota_off(struct super_block * sb,int type)854 static int jfs_quota_off(struct super_block *sb, int type)
855 {
856 struct inode *inode = sb_dqopt(sb)->files[type];
857 int err;
858
859 if (!inode || !igrab(inode))
860 goto out;
861
862 err = dquot_quota_off(sb, type);
863 if (err)
864 goto out_put;
865
866 inode_lock(inode);
867 JFS_IP(inode)->mode2 &= ~(JFS_NOATIME_FL | JFS_IMMUTABLE_FL);
868 inode_set_flags(inode, 0, S_NOATIME | S_IMMUTABLE);
869 inode_unlock(inode);
870 mark_inode_dirty(inode);
871 out_put:
872 iput(inode);
873 return err;
874 out:
875 return dquot_quota_off(sb, type);
876 }
877 #endif
878
879 static const struct super_operations jfs_super_operations = {
880 .alloc_inode = jfs_alloc_inode,
881 .free_inode = jfs_free_inode,
882 .dirty_inode = jfs_dirty_inode,
883 .write_inode = jfs_write_inode,
884 .evict_inode = jfs_evict_inode,
885 .put_super = jfs_put_super,
886 .sync_fs = jfs_sync_fs,
887 .freeze_fs = jfs_freeze,
888 .unfreeze_fs = jfs_unfreeze,
889 .statfs = jfs_statfs,
890 .remount_fs = jfs_remount,
891 .show_options = jfs_show_options,
892 #ifdef CONFIG_QUOTA
893 .quota_read = jfs_quota_read,
894 .quota_write = jfs_quota_write,
895 .get_dquots = jfs_get_dquots,
896 #endif
897 };
898
899 static const struct export_operations jfs_export_operations = {
900 .fh_to_dentry = jfs_fh_to_dentry,
901 .fh_to_parent = jfs_fh_to_parent,
902 .get_parent = jfs_get_parent,
903 };
904
905 static struct file_system_type jfs_fs_type = {
906 .owner = THIS_MODULE,
907 .name = "jfs",
908 .mount = jfs_do_mount,
909 .kill_sb = kill_block_super,
910 .fs_flags = FS_REQUIRES_DEV,
911 };
912 MODULE_ALIAS_FS("jfs");
913
init_once(void * foo)914 static void init_once(void *foo)
915 {
916 struct jfs_inode_info *jfs_ip = (struct jfs_inode_info *) foo;
917
918 memset(jfs_ip, 0, sizeof(struct jfs_inode_info));
919 INIT_LIST_HEAD(&jfs_ip->anon_inode_list);
920 init_rwsem(&jfs_ip->rdwrlock);
921 mutex_init(&jfs_ip->commit_mutex);
922 init_rwsem(&jfs_ip->xattr_sem);
923 spin_lock_init(&jfs_ip->ag_lock);
924 jfs_ip->active_ag = -1;
925 inode_init_once(&jfs_ip->vfs_inode);
926 }
927
init_jfs_fs(void)928 static int __init init_jfs_fs(void)
929 {
930 int i;
931 int rc;
932
933 jfs_inode_cachep =
934 kmem_cache_create_usercopy("jfs_ip", sizeof(struct jfs_inode_info),
935 0, SLAB_RECLAIM_ACCOUNT|SLAB_MEM_SPREAD|SLAB_ACCOUNT,
936 offsetof(struct jfs_inode_info, i_inline_all),
937 sizeof_field(struct jfs_inode_info, i_inline_all),
938 init_once);
939 if (jfs_inode_cachep == NULL)
940 return -ENOMEM;
941
942 /*
943 * Metapage initialization
944 */
945 rc = metapage_init();
946 if (rc) {
947 jfs_err("metapage_init failed w/rc = %d", rc);
948 goto free_slab;
949 }
950
951 /*
952 * Transaction Manager initialization
953 */
954 rc = txInit();
955 if (rc) {
956 jfs_err("txInit failed w/rc = %d", rc);
957 goto free_metapage;
958 }
959
960 /*
961 * I/O completion thread (endio)
962 */
963 jfsIOthread = kthread_run(jfsIOWait, NULL, "jfsIO");
964 if (IS_ERR(jfsIOthread)) {
965 rc = PTR_ERR(jfsIOthread);
966 jfs_err("init_jfs_fs: fork failed w/rc = %d", rc);
967 goto end_txmngr;
968 }
969
970 if (commit_threads < 1)
971 commit_threads = num_online_cpus();
972 if (commit_threads > MAX_COMMIT_THREADS)
973 commit_threads = MAX_COMMIT_THREADS;
974
975 for (i = 0; i < commit_threads; i++) {
976 jfsCommitThread[i] = kthread_run(jfs_lazycommit, NULL,
977 "jfsCommit");
978 if (IS_ERR(jfsCommitThread[i])) {
979 rc = PTR_ERR(jfsCommitThread[i]);
980 jfs_err("init_jfs_fs: fork failed w/rc = %d", rc);
981 commit_threads = i;
982 goto kill_committask;
983 }
984 }
985
986 jfsSyncThread = kthread_run(jfs_sync, NULL, "jfsSync");
987 if (IS_ERR(jfsSyncThread)) {
988 rc = PTR_ERR(jfsSyncThread);
989 jfs_err("init_jfs_fs: fork failed w/rc = %d", rc);
990 goto kill_committask;
991 }
992
993 #ifdef PROC_FS_JFS
994 jfs_proc_init();
995 #endif
996
997 rc = register_filesystem(&jfs_fs_type);
998 if (!rc)
999 return 0;
1000
1001 #ifdef PROC_FS_JFS
1002 jfs_proc_clean();
1003 #endif
1004 kthread_stop(jfsSyncThread);
1005 kill_committask:
1006 for (i = 0; i < commit_threads; i++)
1007 kthread_stop(jfsCommitThread[i]);
1008 kthread_stop(jfsIOthread);
1009 end_txmngr:
1010 txExit();
1011 free_metapage:
1012 metapage_exit();
1013 free_slab:
1014 kmem_cache_destroy(jfs_inode_cachep);
1015 return rc;
1016 }
1017
exit_jfs_fs(void)1018 static void __exit exit_jfs_fs(void)
1019 {
1020 int i;
1021
1022 jfs_info("exit_jfs_fs called");
1023
1024 txExit();
1025 metapage_exit();
1026
1027 kthread_stop(jfsIOthread);
1028 for (i = 0; i < commit_threads; i++)
1029 kthread_stop(jfsCommitThread[i]);
1030 kthread_stop(jfsSyncThread);
1031 #ifdef PROC_FS_JFS
1032 jfs_proc_clean();
1033 #endif
1034 unregister_filesystem(&jfs_fs_type);
1035
1036 /*
1037 * Make sure all delayed rcu free inodes are flushed before we
1038 * destroy cache.
1039 */
1040 rcu_barrier();
1041 kmem_cache_destroy(jfs_inode_cachep);
1042 }
1043
1044 module_init(init_jfs_fs)
1045 module_exit(exit_jfs_fs)
1046