/fs/xfs/ |
D | kmem.c | 12 kmem_alloc(size_t size, xfs_km_flags_t flags) in kmem_alloc() argument 15 gfp_t lflags = kmem_flags_convert(flags); in kmem_alloc() 18 trace_kmem_alloc(size, flags, _RET_IP_); in kmem_alloc() 22 if (ptr || (flags & KM_MAYFAIL)) in kmem_alloc() 42 __kmem_vmalloc(size_t size, xfs_km_flags_t flags) in __kmem_vmalloc() argument 46 gfp_t lflags = kmem_flags_convert(flags); in __kmem_vmalloc() 48 if (flags & KM_NOFS) in __kmem_vmalloc() 53 if (flags & KM_NOFS) in __kmem_vmalloc() 66 kmem_alloc_io(size_t size, int align_mask, xfs_km_flags_t flags) in kmem_alloc_io() argument 70 trace_kmem_alloc_io(size, flags, _RET_IP_); in kmem_alloc_io() [all …]
|
D | xfs_qm_syscalls.c | 37 uint flags) in xfs_qm_scall_quotaoff() argument 50 if ((mp->m_qflags & flags) == 0) in xfs_qm_scall_quotaoff() 54 flags &= (XFS_ALL_QUOTA_ACCT | XFS_ALL_QUOTA_ENFD); in xfs_qm_scall_quotaoff() 68 if ((flags & XFS_ALL_QUOTA_ACCT) == 0) { in xfs_qm_scall_quotaoff() 69 mp->m_qflags &= ~(flags); in xfs_qm_scall_quotaoff() 87 if (flags & XFS_UQUOTA_ACCT) { in xfs_qm_scall_quotaoff() 89 flags |= (XFS_UQUOTA_CHKD | XFS_UQUOTA_ENFD); in xfs_qm_scall_quotaoff() 92 if (flags & XFS_GQUOTA_ACCT) { in xfs_qm_scall_quotaoff() 94 flags |= (XFS_GQUOTA_CHKD | XFS_GQUOTA_ENFD); in xfs_qm_scall_quotaoff() 97 if (flags & XFS_PQUOTA_ACCT) { in xfs_qm_scall_quotaoff() [all …]
|
D | kmem.h | 29 kmem_flags_convert(xfs_km_flags_t flags) in kmem_flags_convert() argument 33 BUG_ON(flags & ~(KM_NOFS|KM_MAYFAIL|KM_ZERO)); in kmem_flags_convert() 36 if (flags & KM_NOFS) in kmem_flags_convert() 46 if (flags & KM_MAYFAIL) in kmem_flags_convert() 49 if (flags & KM_ZERO) in kmem_flags_convert() 56 extern void *kmem_alloc_io(size_t size, int align_mask, xfs_km_flags_t flags); 66 kmem_zalloc(size_t size, xfs_km_flags_t flags) in kmem_zalloc() argument 68 return kmem_alloc(size, flags | KM_ZERO); in kmem_zalloc() 72 kmem_zalloc_large(size_t size, xfs_km_flags_t flags) in kmem_zalloc_large() argument 74 return kmem_alloc_large(size, flags | KM_ZERO); in kmem_zalloc_large() [all …]
|
D | xfs_inode.h | 117 __xfs_iflags_set(xfs_inode_t *ip, unsigned short flags) in __xfs_iflags_set() argument 119 ip->i_flags |= flags; in __xfs_iflags_set() 123 xfs_iflags_set(xfs_inode_t *ip, unsigned short flags) in xfs_iflags_set() argument 126 __xfs_iflags_set(ip, flags); in xfs_iflags_set() 131 xfs_iflags_clear(xfs_inode_t *ip, unsigned short flags) in xfs_iflags_clear() argument 134 ip->i_flags &= ~flags; in xfs_iflags_clear() 139 __xfs_iflags_test(xfs_inode_t *ip, unsigned short flags) in __xfs_iflags_test() argument 141 return (ip->i_flags & flags); in __xfs_iflags_test() 145 xfs_iflags_test(xfs_inode_t *ip, unsigned short flags) in xfs_iflags_test() argument 149 ret = __xfs_iflags_test(ip, flags); in xfs_iflags_test() [all …]
|
D | xfs_quotaops.c | 37 tstate->flags |= QCI_SYSFILE; in xfs_qm_fill_state() 67 state->s_state[USRQUOTA].flags |= QCI_ACCT_ENABLED; in xfs_fs_get_quota_state() 69 state->s_state[USRQUOTA].flags |= QCI_LIMITS_ENFORCED; in xfs_fs_get_quota_state() 71 state->s_state[GRPQUOTA].flags |= QCI_ACCT_ENABLED; in xfs_fs_get_quota_state() 73 state->s_state[GRPQUOTA].flags |= QCI_LIMITS_ENFORCED; in xfs_fs_get_quota_state() 75 state->s_state[PRJQUOTA].flags |= QCI_ACCT_ENABLED; in xfs_fs_get_quota_state() 77 state->s_state[PRJQUOTA].flags |= QCI_LIMITS_ENFORCED; in xfs_fs_get_quota_state() 140 unsigned int flags = 0; in xfs_quota_flags() local 143 flags |= XFS_UQUOTA_ACCT; in xfs_quota_flags() 145 flags |= XFS_PQUOTA_ACCT; in xfs_quota_flags() [all …]
|
/fs/ubifs/ |
D | ioctl.c | 32 unsigned int flags = ubifs_inode(inode)->flags; in ubifs_set_inode_flags() local 36 if (flags & UBIFS_SYNC_FL) in ubifs_set_inode_flags() 38 if (flags & UBIFS_APPEND_FL) in ubifs_set_inode_flags() 40 if (flags & UBIFS_IMMUTABLE_FL) in ubifs_set_inode_flags() 42 if (flags & UBIFS_DIRSYNC_FL) in ubifs_set_inode_flags() 44 if (flags & UBIFS_CRYPT_FL) in ubifs_set_inode_flags() 98 static int setflags(struct inode *inode, int flags) in setflags() argument 111 oldflags = ubifs2ioctl(ui->flags); in setflags() 112 err = vfs_ioc_setflags_prepare(inode, oldflags, flags); in setflags() 116 ui->flags = ioctl2ubifs(flags); in setflags() [all …]
|
D | lprops.c | 188 lp->flags &= ~LPROPS_CAT_MASK; in add_to_lpt_heap() 189 lp->flags |= LPROPS_UNCAT; in add_to_lpt_heap() 290 lprops->flags &= ~LPROPS_CAT_MASK; in ubifs_add_to_cat() 291 lprops->flags |= cat; in ubifs_add_to_cat() 346 cat = new_lprops->flags & LPROPS_CAT_MASK; in ubifs_replace_cat() 375 int cat = lprops->flags & LPROPS_CAT_MASK; in ubifs_ensure_cat() 399 if (lprops->flags & LPROPS_TAKEN) in ubifs_categorize_lprops() 403 ubifs_assert(c, !(lprops->flags & LPROPS_INDEX)); in ubifs_categorize_lprops() 408 if (lprops->flags & LPROPS_INDEX) in ubifs_categorize_lprops() 414 if (lprops->flags & LPROPS_INDEX) { in ubifs_categorize_lprops() [all …]
|
D | find.c | 44 int n, cat = lprops->flags & LPROPS_CAT_MASK; in valuable() 90 if (lprops->flags & LPROPS_TAKEN) in scan_for_dirty_cb() 99 if (data->exclude_index && lprops->flags & LPROPS_INDEX) in scan_for_dirty_cb() 151 if (lprops->flags & LPROPS_TAKEN) in scan_for_dirty() 155 if (exclude_index && (lprops->flags & LPROPS_INDEX)) in scan_for_dirty() 184 ubifs_assert(c, !(lprops->flags & LPROPS_TAKEN)); in scan_for_dirty() 185 ubifs_assert(c, !exclude_index || !(lprops->flags & LPROPS_INDEX)); in scan_for_dirty() 322 lp->lnum, lp->free, lp->dirty, lp->flags); in ubifs_find_dirty_leb() 325 lp->flags | LPROPS_TAKEN, 0); in ubifs_find_dirty_leb() 357 if (lprops->flags & LPROPS_TAKEN) in scan_for_free_cb() [all …]
|
/fs/ocfs2/dlm/ |
D | dlmunlock.c | 58 int flags, 87 int flags, int *call_ast, in dlmunlock_common() argument 97 flags & LKM_VALBLK); in dlmunlock_common() 109 if (in_use && !(flags & LKM_CANCEL)) { in dlmunlock_common() 118 if (master_node && !(flags & LKM_CANCEL)) { in dlmunlock_common() 141 if (flags & LKM_CANCEL) in dlmunlock_common() 150 if (flags & LKM_VALBLK) { in dlmunlock_common() 155 flags |= LKM_PUT_LVB; /* let the send function in dlmunlock_common() 162 if (flags & LKM_CANCEL) in dlmunlock_common() 169 flags, owner); in dlmunlock_common() [all …]
|
D | dlmconvert.c | 45 struct dlm_lock *lock, int flags, 50 struct dlm_lock *lock, int flags, int type); 63 struct dlm_lock *lock, int flags, int type) in dlmconvert_master() argument 74 status = __dlmconvert_master(dlm, res, lock, flags, type, in dlmconvert_master() 106 struct dlm_lock *lock, int flags, in __dlmconvert_master() argument 136 if (flags & LKM_VALBLK) { in __dlmconvert_master() 143 lock->lksb->flags |= DLM_LKSB_PUT_LVB; in __dlmconvert_master() 153 lock->lksb->flags |= DLM_LKSB_GET_LVB; in __dlmconvert_master() 159 flags &= ~(LKM_VALBLK); in __dlmconvert_master() 197 if (lock->lksb->flags & DLM_LKSB_PUT_LVB) in __dlmconvert_master() [all …]
|
D | dlmlock.c | 47 struct dlm_lock *lock, int flags); 105 struct dlm_lock *lock, int flags) in dlmlock_master() argument 151 if (flags & LKM_NOQUEUE) { in dlmlock_master() 188 lock->lksb->flags &= ~DLM_LKSB_GET_LVB; in dlm_revert_pending_lock() 201 struct dlm_lock *lock, int flags) in dlmlock_remote() argument 208 res->lockname.name, flags); in dlmlock_remote() 230 status = dlm_send_remote_lock_request(dlm, res, lock, flags); in dlmlock_remote() 287 struct dlm_lock *lock, int flags) in dlm_send_remote_lock_request() argument 298 create.flags = cpu_to_be32(flags); in dlm_send_remote_lock_request() 391 newlock->ml.flags = 0; in dlm_init_lock() [all …]
|
/fs/ |
D | splice.c | 78 buf->flags |= PIPE_BUF_FLAG_LRU; in page_cache_pipe_buf_steal() 96 buf->flags &= ~PIPE_BUF_FLAG_LRU; in page_cache_pipe_buf_release() 151 if (!(buf->flags & PIPE_BUF_FLAG_GIFT)) in user_page_pipe_buf_steal() 154 buf->flags |= PIPE_BUF_FLAG_LRU; in user_page_pipe_buf_steal() 208 buf->flags = 0; in splice_to_pipe() 297 unsigned int flags) in generic_file_splice_read() argument 367 unsigned int flags) in default_file_splice_read() argument 444 more = (sd->flags & SPLICE_F_MORE) ? MSG_MORE : 0; in pipe_to_sendpage() 553 if (sd->flags & SPLICE_F_NONBLOCK) in splice_from_pipe_next() 647 loff_t *ppos, size_t len, unsigned int flags, in splice_from_pipe() argument [all …]
|
D | namei.c | 131 getname_flags(const char __user *filename, int flags, int *empty) in getname_flags() argument 197 if (!(flags & LOOKUP_EMPTY)) { in getname_flags() 506 unsigned int flags; member 550 if (nd->flags & LOOKUP_RCU) { in __nd_alloc_stack() 607 if (!(nd->flags & LOOKUP_RCU)) { in terminate_walk() 612 if (nd->flags & LOOKUP_ROOT_GRABBED) { in terminate_walk() 614 nd->flags &= ~LOOKUP_ROOT_GRABBED; in terminate_walk() 617 nd->flags &= ~LOOKUP_RCU; in terminate_walk() 657 if (!nd->root.mnt || (nd->flags & LOOKUP_ROOT)) in legitimize_root() 659 nd->flags |= LOOKUP_ROOT_GRABBED; in legitimize_root() [all …]
|
D | timerfd.c | 64 unsigned long flags; in timerfd_triggered() local 66 spin_lock_irqsave(&ctx->wqh.lock, flags); in timerfd_triggered() 70 spin_unlock_irqrestore(&ctx->wqh.lock, flags); in timerfd_triggered() 100 unsigned long flags; in timerfd_clock_was_set() local 106 spin_lock_irqsave(&ctx->wqh.lock, flags); in timerfd_clock_was_set() 112 spin_unlock_irqrestore(&ctx->wqh.lock, flags); in timerfd_clock_was_set() 142 static void timerfd_setup_cancel(struct timerfd_ctx *ctx, int flags) in timerfd_setup_cancel() argument 147 (flags & TFD_TIMER_ABSTIME) && (flags & TFD_TIMER_CANCEL_ON_SET)) { in timerfd_setup_cancel() 172 static int timerfd_setup(struct timerfd_ctx *ctx, int flags, in timerfd_setup() argument 179 htmode = (flags & TFD_TIMER_ABSTIME) ? in timerfd_setup() [all …]
|
/fs/jfs/ |
D | ioctl.c | 41 static long jfs_map_ext2(unsigned long flags, int from) in jfs_map_ext2() argument 48 if (jfs_map[index].ext2_flag & flags) in jfs_map_ext2() 51 if (jfs_map[index].jfs_flag & flags) in jfs_map_ext2() 64 unsigned int flags; in jfs_ioctl() local 68 flags = jfs_inode->mode2 & JFS_FL_USER_VISIBLE; in jfs_ioctl() 69 flags = jfs_map_ext2(flags, 0); in jfs_ioctl() 70 return put_user(flags, (int __user *) arg); in jfs_ioctl() 83 if (get_user(flags, (int __user *) arg)) { in jfs_ioctl() 88 flags = jfs_map_ext2(flags, 1); in jfs_ioctl() 90 flags &= ~JFS_DIRSYNC_FL; in jfs_ioctl() [all …]
|
/fs/notify/fanotify/ |
D | fanotify_user.c | 531 struct path *path, unsigned int flags, __u64 mask, in fanotify_find_path() argument 537 dfd, filename, flags); in fanotify_find_path() 547 if ((flags & FAN_MARK_ONLYDIR) && in fanotify_find_path() 559 if (!(flags & FAN_MARK_DONT_FOLLOW)) in fanotify_find_path() 561 if (flags & FAN_MARK_ONLYDIR) in fanotify_find_path() 586 unsigned int flags, in fanotify_mark_remove_from_mask() argument 592 if (!(flags & FAN_MARK_IGNORED_MASK)) { in fanotify_mark_remove_from_mask() 606 unsigned int flags) in fanotify_remove_mark() argument 619 removed = fanotify_mark_remove_from_mask(fsn_mark, mask, flags, in fanotify_remove_mark() 636 unsigned int flags) in fanotify_remove_vfsmount_mark() argument [all …]
|
/fs/btrfs/ |
D | async-thread.c | 84 unsigned int flags, int limit_active, int thresh) in __btrfs_alloc_workqueue() argument 110 if (flags & WQ_HIGHPRI) in __btrfs_alloc_workqueue() 111 ret->normal_wq = alloc_workqueue("btrfs-%s-high", flags, in __btrfs_alloc_workqueue() 114 ret->normal_wq = alloc_workqueue("btrfs-%s", flags, in __btrfs_alloc_workqueue() 124 trace_btrfs_workqueue_alloc(ret, name, flags & WQ_HIGHPRI); in __btrfs_alloc_workqueue() 133 unsigned int flags, in btrfs_alloc_workqueue() argument 143 flags & ~WQ_HIGHPRI, in btrfs_alloc_workqueue() 150 if (flags & WQ_HIGHPRI) { in btrfs_alloc_workqueue() 151 ret->high = __btrfs_alloc_workqueue(fs_info, name, flags, in btrfs_alloc_workqueue() 228 unsigned long flags; in run_ordered_work() local [all …]
|
/fs/afs/ |
D | rotate.c | 33 fc->flags |= AFS_FS_CURSOR_INTR; in afs_begin_vnode_operation() 36 fc->flags |= AFS_FS_CURSOR_STOP; in afs_begin_vnode_operation() 44 fc->flags |= AFS_FS_CURSOR_CUR_ONLY; in afs_begin_vnode_operation() 80 if (fc->flags & AFS_FS_CURSOR_CUR_ONLY) { in afs_start_fs_iteration() 89 if (test_and_clear_bit(AFS_VNODE_CB_PROMISED, &vnode->flags)) in afs_start_fs_iteration() 123 if (fc->flags & AFS_FS_CURSOR_INTR) { in afs_sleep_and_retry() 154 if (fc->flags & AFS_FS_CURSOR_STOP) { in afs_select_fileserver() 170 fc->flags |= AFS_FS_CURSOR_STOP; in afs_select_fileserver() 185 if (fc->flags & AFS_FS_CURSOR_VNOVOL) { in afs_select_fileserver() 194 set_bit(AFS_VOLUME_NEEDS_UPDATE, &vnode->volume->flags); in afs_select_fileserver() [all …]
|
/fs/iomap/ |
D | direct-io.c | 32 unsigned flags; member 83 ret = dops->end_io(iocb, dio->size, ret, dio->flags); in iomap_dio_complete() 89 !(dio->flags & IOMAP_DIO_WRITE)) in iomap_dio_complete() 107 (dio->flags & IOMAP_DIO_WRITE) && inode->i_mapping->nrpages) { in iomap_dio_complete() 120 if (ret > 0 && (dio->flags & IOMAP_DIO_NEED_SYNC)) in iomap_dio_complete() 150 bool should_dirty = (dio->flags & IOMAP_DIO_DIRTY); in iomap_dio_bio_end_io() 160 } else if (dio->flags & IOMAP_DIO_WRITE) { in iomap_dio_bio_end_io() 184 int flags = REQ_SYNC | REQ_IDLE; in iomap_dio_zero() local 197 bio_set_op_attrs(bio, REQ_OP_WRITE, flags); in iomap_dio_zero() 219 dio->flags |= IOMAP_DIO_UNWRITTEN; in iomap_dio_bio_actor() [all …]
|
/fs/hfsplus/ |
D | ioctl.c | 63 unsigned int flags = 0; in hfsplus_getflags() local 66 flags |= FS_IMMUTABLE_FL; in hfsplus_getflags() 68 flags |= FS_APPEND_FL; in hfsplus_getflags() 70 flags |= FS_NODUMP_FL; in hfsplus_getflags() 71 return flags; in hfsplus_getflags() 77 unsigned int flags = hfsplus_getflags(inode); in hfsplus_ioctl_getflags() local 79 return put_user(flags, user_flags); in hfsplus_ioctl_getflags() 86 unsigned int flags, new_fl = 0; in hfsplus_ioctl_setflags() local 99 if (get_user(flags, user_flags)) { in hfsplus_ioctl_setflags() 106 err = vfs_ioc_setflags_prepare(inode, oldflags, flags); in hfsplus_ioctl_setflags() [all …]
|
/fs/fscache/ |
D | operation.c | 73 switch (op->flags & FSCACHE_OP_TYPE) { in fscache_enqueue_operation() 86 pr_err("Unexpected op type %lx", op->flags); in fscache_enqueue_operation() 103 if (test_and_clear_bit(FSCACHE_OP_WAITING, &op->flags)) in fscache_run_op() 104 wake_up_bit(&op->flags, FSCACHE_OP_WAITING); in fscache_run_op() 130 kdebug("objflags=%lx", object->flags); in fscache_report_unexpected_submission() 158 unsigned long flags; in fscache_submit_exclusive_op() local 177 flags = READ_ONCE(object->flags); in fscache_submit_exclusive_op() 178 if (unlikely(!(flags & BIT(FSCACHE_OBJECT_IS_LIVE)))) { in fscache_submit_exclusive_op() 187 } else if (flags & BIT(FSCACHE_OBJECT_IS_AVAILABLE)) { in fscache_submit_exclusive_op() 207 clear_bit(FSCACHE_OBJECT_PENDING_WRITE, &object->flags); in fscache_submit_exclusive_op() [all …]
|
/fs/nfs/ |
D | delegation.c | 42 set_bit(NFS_DELEGATION_REFERENCED, &delegation->flags); in nfs_mark_delegation_referenced() 47 fmode_t flags) in nfs4_is_valid_delegation() argument 49 if (delegation != NULL && (delegation->type & flags) == flags && in nfs4_is_valid_delegation() 50 !test_bit(NFS_DELEGATION_REVOKED, &delegation->flags) && in nfs4_is_valid_delegation() 51 !test_bit(NFS_DELEGATION_RETURNING, &delegation->flags)) in nfs4_is_valid_delegation() 67 nfs4_do_check_delegation(struct inode *inode, fmode_t flags, bool mark) in nfs4_do_check_delegation() argument 72 flags &= FMODE_READ|FMODE_WRITE; in nfs4_do_check_delegation() 75 if (nfs4_is_valid_delegation(delegation, flags)) { in nfs4_do_check_delegation() 91 int nfs4_have_delegation(struct inode *inode, fmode_t flags) in nfs4_have_delegation() argument 93 return nfs4_do_check_delegation(inode, flags, true); in nfs4_have_delegation() [all …]
|
/fs/overlayfs/ |
D | file.c | 30 int flags = file->f_flags | O_NOATIME | FMODE_NONOTIFY; in ovl_open_realfile() local 33 realfile = open_with_fake_path(&file->f_path, flags, realinode, in ovl_open_realfile() 46 static int ovl_change_flags(struct file *file, unsigned int flags) in ovl_change_flags() argument 52 flags |= O_NOATIME | FMODE_NONOTIFY; in ovl_change_flags() 55 if (WARN_ON((file->f_flags ^ flags) & ~OVL_SETFL_MASK)) in ovl_change_flags() 58 flags &= OVL_SETFL_MASK; in ovl_change_flags() 60 if (((flags ^ file->f_flags) & O_APPEND) && IS_APPEND(inode)) in ovl_change_flags() 63 if (flags & O_DIRECT) { in ovl_change_flags() 70 err = file->f_op->check_flags(flags); in ovl_change_flags() 76 file->f_flags = (file->f_flags & ~OVL_SETFL_MASK) | flags; in ovl_change_flags() [all …]
|
/fs/dlm/ |
D | ast.c | 27 lkb->lkb_last_bast.flags, in dlm_dump_lkb_callbacks() 35 lkb->lkb_last_cast.flags, in dlm_dump_lkb_callbacks() 44 lkb->lkb_callbacks[i].flags, in dlm_dump_lkb_callbacks() 51 int dlm_add_lkb_callback(struct dlm_lkb *lkb, uint32_t flags, int mode, in dlm_add_lkb_callback() argument 70 if ((i > 0) && (flags & DLM_CB_BAST) && in dlm_add_lkb_callback() 71 (lkb->lkb_callbacks[i-1].flags & DLM_CB_BAST)) { in dlm_add_lkb_callback() 92 lkb->lkb_callbacks[i].flags = flags; in dlm_add_lkb_callback() 103 flags, mode, status, sbflags); in dlm_add_lkb_callback() 143 if ((cb->flags & DLM_CB_BAST) && lkb->lkb_last_cast.seq) { in dlm_rem_lkb_callback() 145 cb->flags |= DLM_CB_SKIP; in dlm_rem_lkb_callback() [all …]
|
/fs/ceph/ |
D | super.c | 262 fsopt->flags &= ~CEPH_MOUNT_OPT_CLEANRECOVER; in parse_fsopt_token() 265 fsopt->flags |= CEPH_MOUNT_OPT_CLEANRECOVER; in parse_fsopt_token() 278 fsopt->flags |= CEPH_MOUNT_OPT_FSCACHE; in parse_fsopt_token() 330 fsopt->flags |= CEPH_MOUNT_OPT_DIRSTAT; in parse_fsopt_token() 333 fsopt->flags &= ~CEPH_MOUNT_OPT_DIRSTAT; in parse_fsopt_token() 336 fsopt->flags |= CEPH_MOUNT_OPT_RBYTES; in parse_fsopt_token() 339 fsopt->flags &= ~CEPH_MOUNT_OPT_RBYTES; in parse_fsopt_token() 342 fsopt->flags &= ~CEPH_MOUNT_OPT_NOASYNCREADDIR; in parse_fsopt_token() 345 fsopt->flags |= CEPH_MOUNT_OPT_NOASYNCREADDIR; in parse_fsopt_token() 348 fsopt->flags |= CEPH_MOUNT_OPT_DCACHE; in parse_fsopt_token() [all …]
|