/fs/ubifs/ |
D | ioctl.c | 39 unsigned int flags = ubifs_inode(inode)->flags; in ubifs_set_inode_flags() local 42 if (flags & UBIFS_SYNC_FL) in ubifs_set_inode_flags() 44 if (flags & UBIFS_APPEND_FL) in ubifs_set_inode_flags() 46 if (flags & UBIFS_IMMUTABLE_FL) in ubifs_set_inode_flags() 48 if (flags & UBIFS_DIRSYNC_FL) in ubifs_set_inode_flags() 102 static int setflags(struct inode *inode, int flags) in setflags() argument 119 oldflags = ubifs2ioctl(ui->flags); in setflags() 120 if ((flags ^ oldflags) & (FS_APPEND_FL | FS_IMMUTABLE_FL)) { in setflags() 127 ui->flags = ioctl2ubifs(flags); in setflags() 149 int flags, err; in ubifs_ioctl() local [all …]
|
D | lprops.c | 200 lp->flags &= ~LPROPS_CAT_MASK; in add_to_lpt_heap() 201 lp->flags |= LPROPS_UNCAT; in add_to_lpt_heap() 304 lprops->flags &= ~LPROPS_CAT_MASK; in ubifs_add_to_cat() 305 lprops->flags |= cat; in ubifs_add_to_cat() 360 cat = new_lprops->flags & LPROPS_CAT_MASK; in ubifs_replace_cat() 389 int cat = lprops->flags & LPROPS_CAT_MASK; in ubifs_ensure_cat() 413 if (lprops->flags & LPROPS_TAKEN) in ubifs_categorize_lprops() 417 ubifs_assert(!(lprops->flags & LPROPS_INDEX)); in ubifs_categorize_lprops() 422 if (lprops->flags & LPROPS_INDEX) in ubifs_categorize_lprops() 428 if (lprops->flags & LPROPS_INDEX) { in ubifs_categorize_lprops() [all …]
|
D | find.c | 56 int n, cat = lprops->flags & LPROPS_CAT_MASK; in valuable() 102 if (lprops->flags & LPROPS_TAKEN) in scan_for_dirty_cb() 111 if (data->exclude_index && lprops->flags & LPROPS_INDEX) in scan_for_dirty_cb() 163 if (lprops->flags & LPROPS_TAKEN) in scan_for_dirty() 167 if (exclude_index && (lprops->flags & LPROPS_INDEX)) in scan_for_dirty() 196 ubifs_assert(!(lprops->flags & LPROPS_TAKEN)); in scan_for_dirty() 197 ubifs_assert(!exclude_index || !(lprops->flags & LPROPS_INDEX)); in scan_for_dirty() 334 lp->lnum, lp->free, lp->dirty, lp->flags); in ubifs_find_dirty_leb() 337 lp->flags | LPROPS_TAKEN, 0); in ubifs_find_dirty_leb() 369 if (lprops->flags & LPROPS_TAKEN) in scan_for_free_cb() [all …]
|
/fs/xfs/ |
D | xfs_qm_syscalls.c | 53 uint flags) in xfs_qm_scall_quotaoff() argument 66 if ((mp->m_qflags & flags) == 0) in xfs_qm_scall_quotaoff() 70 flags &= (XFS_ALL_QUOTA_ACCT | XFS_ALL_QUOTA_ENFD); in xfs_qm_scall_quotaoff() 84 if ((flags & XFS_ALL_QUOTA_ACCT) == 0) { in xfs_qm_scall_quotaoff() 85 mp->m_qflags &= ~(flags); in xfs_qm_scall_quotaoff() 103 if (flags & XFS_UQUOTA_ACCT) { in xfs_qm_scall_quotaoff() 105 flags |= (XFS_UQUOTA_CHKD | XFS_UQUOTA_ENFD); in xfs_qm_scall_quotaoff() 108 if (flags & XFS_GQUOTA_ACCT) { in xfs_qm_scall_quotaoff() 110 flags |= (XFS_GQUOTA_CHKD | XFS_GQUOTA_ENFD); in xfs_qm_scall_quotaoff() 113 if (flags & XFS_PQUOTA_ACCT) { in xfs_qm_scall_quotaoff() [all …]
|
D | kmem.c | 28 kmem_alloc(size_t size, xfs_km_flags_t flags) in kmem_alloc() argument 31 gfp_t lflags = kmem_flags_convert(flags); in kmem_alloc() 36 if (ptr || (flags & (KM_MAYFAIL|KM_NOSLEEP))) in kmem_alloc() 48 kmem_zalloc_large(size_t size, xfs_km_flags_t flags) in kmem_zalloc_large() argument 54 ptr = kmem_zalloc(size, flags | KM_MAYFAIL); in kmem_zalloc_large() 65 if ((current->flags & PF_FSTRANS) || (flags & KM_NOFS)) in kmem_zalloc_large() 68 lflags = kmem_flags_convert(flags); in kmem_zalloc_large() 71 if ((current->flags & PF_FSTRANS) || (flags & KM_NOFS)) in kmem_zalloc_large() 79 xfs_km_flags_t flags) in kmem_realloc() argument 83 new = kmem_alloc(newsize, flags); in kmem_realloc() [all …]
|
D | xfs_xattr.c | 38 int xflags = handler->flags; in xfs_xattr_get() 81 const char *name, const void *value, size_t size, int flags) in xfs_xattr_set() argument 83 int xflags = handler->flags; in xfs_xattr_set() 91 if (flags & XATTR_CREATE) in xfs_xattr_set() 93 if (flags & XATTR_REPLACE) in xfs_xattr_set() 108 .flags = 0, /* no flags implies user namespace */ 115 .flags = ATTR_ROOT, 122 .flags = ATTR_SECURE, 138 static unsigned int xfs_xattr_prefix_len(int flags) in xfs_xattr_prefix_len() argument 140 if (flags & XFS_ATTR_SECURE) in xfs_xattr_prefix_len() [all …]
|
D | xfs_inode.h | 114 __xfs_iflags_set(xfs_inode_t *ip, unsigned short flags) in __xfs_iflags_set() argument 116 ip->i_flags |= flags; in __xfs_iflags_set() 120 xfs_iflags_set(xfs_inode_t *ip, unsigned short flags) in xfs_iflags_set() argument 123 __xfs_iflags_set(ip, flags); in xfs_iflags_set() 128 xfs_iflags_clear(xfs_inode_t *ip, unsigned short flags) in xfs_iflags_clear() argument 131 ip->i_flags &= ~flags; in xfs_iflags_clear() 136 __xfs_iflags_test(xfs_inode_t *ip, unsigned short flags) in __xfs_iflags_test() argument 138 return (ip->i_flags & flags); in __xfs_iflags_test() 142 xfs_iflags_test(xfs_inode_t *ip, unsigned short flags) in xfs_iflags_test() argument 146 ret = __xfs_iflags_test(ip, flags); in xfs_iflags_test() [all …]
|
D | xfs_quotaops.c | 50 tstate->flags |= QCI_SYSFILE; in xfs_qm_fill_state() 80 state->s_state[USRQUOTA].flags |= QCI_ACCT_ENABLED; in xfs_fs_get_quota_state() 82 state->s_state[USRQUOTA].flags |= QCI_LIMITS_ENFORCED; in xfs_fs_get_quota_state() 84 state->s_state[GRPQUOTA].flags |= QCI_ACCT_ENABLED; in xfs_fs_get_quota_state() 86 state->s_state[GRPQUOTA].flags |= QCI_LIMITS_ENFORCED; in xfs_fs_get_quota_state() 88 state->s_state[PRJQUOTA].flags |= QCI_ACCT_ENABLED; in xfs_fs_get_quota_state() 90 state->s_state[PRJQUOTA].flags |= QCI_LIMITS_ENFORCED; in xfs_fs_get_quota_state() 153 unsigned int flags = 0; in xfs_quota_flags() local 156 flags |= XFS_UQUOTA_ACCT; in xfs_quota_flags() 158 flags |= XFS_PQUOTA_ACCT; in xfs_quota_flags() [all …]
|
D | kmem.h | 43 kmem_flags_convert(xfs_km_flags_t flags) in kmem_flags_convert() argument 47 BUG_ON(flags & ~(KM_SLEEP|KM_NOSLEEP|KM_NOFS|KM_MAYFAIL|KM_ZERO)); in kmem_flags_convert() 49 if (flags & KM_NOSLEEP) { in kmem_flags_convert() 53 if ((current->flags & PF_FSTRANS) || (flags & KM_NOFS)) in kmem_flags_convert() 57 if (flags & KM_ZERO) in kmem_flags_convert() 73 kmem_zalloc(size_t size, xfs_km_flags_t flags) in kmem_zalloc() argument 75 return kmem_alloc(size, flags | KM_ZERO); in kmem_zalloc() 96 kmem_zone_init_flags(int size, char *zone_name, unsigned long flags, in kmem_zone_init_flags() argument 99 return kmem_cache_create(zone_name, size, 0, flags, construct); in kmem_zone_init_flags() 118 kmem_zone_zalloc(kmem_zone_t *zone, xfs_km_flags_t flags) in kmem_zone_zalloc() argument [all …]
|
/fs/ocfs2/dlm/ |
D | dlmunlock.c | 73 int flags, 102 int flags, int *call_ast, in dlmunlock_common() argument 112 flags & LKM_VALBLK); in dlmunlock_common() 124 if (in_use && !(flags & LKM_CANCEL)) { in dlmunlock_common() 133 if (master_node && !(flags & LKM_CANCEL)) { in dlmunlock_common() 156 if (flags & LKM_CANCEL) in dlmunlock_common() 165 if (flags & LKM_VALBLK) { in dlmunlock_common() 170 flags |= LKM_PUT_LVB; /* let the send function in dlmunlock_common() 177 if (flags & LKM_CANCEL) in dlmunlock_common() 184 flags, owner); in dlmunlock_common() [all …]
|
D | dlmconvert.c | 60 struct dlm_lock *lock, int flags, 65 struct dlm_lock *lock, int flags, int type); 78 struct dlm_lock *lock, int flags, int type) in dlmconvert_master() argument 89 status = __dlmconvert_master(dlm, res, lock, flags, type, in dlmconvert_master() 121 struct dlm_lock *lock, int flags, in __dlmconvert_master() argument 151 if (flags & LKM_VALBLK) { in __dlmconvert_master() 158 lock->lksb->flags |= DLM_LKSB_PUT_LVB; in __dlmconvert_master() 168 lock->lksb->flags |= DLM_LKSB_GET_LVB; in __dlmconvert_master() 174 flags &= ~(LKM_VALBLK); in __dlmconvert_master() 212 if (lock->lksb->flags & DLM_LKSB_PUT_LVB) in __dlmconvert_master() [all …]
|
D | dlmlock.c | 62 struct dlm_lock *lock, int flags); 121 struct dlm_lock *lock, int flags) in dlmlock_master() argument 167 if (flags & LKM_NOQUEUE) { in dlmlock_master() 204 lock->lksb->flags &= ~DLM_LKSB_GET_LVB; in dlm_revert_pending_lock() 217 struct dlm_lock *lock, int flags) in dlmlock_remote() argument 224 res->lockname.name, flags); in dlmlock_remote() 246 status = dlm_send_remote_lock_request(dlm, res, lock, flags); in dlmlock_remote() 303 struct dlm_lock *lock, int flags) in dlm_send_remote_lock_request() argument 314 create.flags = cpu_to_be32(flags); in dlm_send_remote_lock_request() 407 newlock->ml.flags = 0; in dlm_init_lock() [all …]
|
/fs/notify/fanotify/ |
D | fanotify_user.c | 453 struct path *path, unsigned int flags) in fanotify_find_path() argument 458 dfd, filename, flags); in fanotify_find_path() 468 if ((flags & FAN_MARK_ONLYDIR) && in fanotify_find_path() 480 if (!(flags & FAN_MARK_DONT_FOLLOW)) in fanotify_find_path() 482 if (flags & FAN_MARK_ONLYDIR) in fanotify_find_path() 500 unsigned int flags, in fanotify_mark_remove_from_mask() argument 506 if (!(flags & FAN_MARK_IGNORED_MASK)) { in fanotify_mark_remove_from_mask() 509 if (flags & FAN_MARK_ONDIR) in fanotify_mark_remove_from_mask() 516 if (flags & FAN_MARK_ONDIR) in fanotify_mark_remove_from_mask() 529 unsigned int flags) in fanotify_remove_vfsmount_mark() argument [all …]
|
/fs/jfs/ |
D | ioctl.c | 40 static long jfs_map_ext2(unsigned long flags, int from) in jfs_map_ext2() argument 47 if (jfs_map[index].ext2_flag & flags) in jfs_map_ext2() 50 if (jfs_map[index].jfs_flag & flags) in jfs_map_ext2() 63 unsigned int flags; in jfs_ioctl() local 68 flags = jfs_inode->mode2 & JFS_FL_USER_VISIBLE; in jfs_ioctl() 69 flags = jfs_map_ext2(flags, 0); in jfs_ioctl() 70 return put_user(flags, (int __user *) arg); in jfs_ioctl() 83 if (get_user(flags, (int __user *) arg)) { in jfs_ioctl() 88 flags = jfs_map_ext2(flags, 1); in jfs_ioctl() 90 flags &= ~JFS_DIRSYNC_FL; in jfs_ioctl() [all …]
|
/fs/nfs/ |
D | nfstrace.h | 106 __entry->nfsi_flags = nfsi->flags; 161 #define show_lookup_flags(flags) \ argument 162 __print_flags((unsigned long)flags, "|", \ 173 unsigned int flags 176 TP_ARGS(dir, dentry, flags), 179 __field(unsigned int, flags) 188 __entry->flags = flags; 194 __entry->flags, 195 show_lookup_flags(__entry->flags), 207 unsigned int flags \ [all …]
|
D | delegation.c | 41 set_bit(NFS_DELEGATION_REFERENCED, &delegation->flags); in nfs_mark_delegation_referenced() 46 fmode_t flags) in nfs4_is_valid_delegation() argument 48 if (delegation != NULL && (delegation->type & flags) == flags && in nfs4_is_valid_delegation() 49 !test_bit(NFS_DELEGATION_REVOKED, &delegation->flags) && in nfs4_is_valid_delegation() 50 !test_bit(NFS_DELEGATION_RETURNING, &delegation->flags)) in nfs4_is_valid_delegation() 66 nfs4_do_check_delegation(struct inode *inode, fmode_t flags, bool mark) in nfs4_do_check_delegation() argument 71 flags &= FMODE_READ|FMODE_WRITE; in nfs4_do_check_delegation() 74 if (nfs4_is_valid_delegation(delegation, flags)) { in nfs4_do_check_delegation() 90 int nfs4_have_delegation(struct inode *inode, fmode_t flags) in nfs4_have_delegation() argument 92 return nfs4_do_check_delegation(inode, flags, true); in nfs4_have_delegation() [all …]
|
/fs/btrfs/ |
D | async-thread.c | 111 __btrfs_alloc_workqueue(const char *name, unsigned int flags, int limit_active, in __btrfs_alloc_workqueue() argument 137 if (flags & WQ_HIGHPRI) in __btrfs_alloc_workqueue() 138 ret->normal_wq = alloc_workqueue("%s-%s-high", flags, in __btrfs_alloc_workqueue() 142 ret->normal_wq = alloc_workqueue("%s-%s", flags, in __btrfs_alloc_workqueue() 153 trace_btrfs_workqueue_alloc(ret, name, flags & WQ_HIGHPRI); in __btrfs_alloc_workqueue() 161 unsigned int flags, in btrfs_alloc_workqueue() argument 170 ret->normal = __btrfs_alloc_workqueue(name, flags & ~WQ_HIGHPRI, in btrfs_alloc_workqueue() 177 if (flags & WQ_HIGHPRI) { in btrfs_alloc_workqueue() 178 ret->high = __btrfs_alloc_workqueue(name, flags, limit_active, in btrfs_alloc_workqueue() 254 unsigned long flags; in run_ordered_work() local [all …]
|
/fs/ |
D | splice.c | 74 buf->flags |= PIPE_BUF_FLAG_LRU; in page_cache_pipe_buf_steal() 92 buf->flags &= ~PIPE_BUF_FLAG_LRU; in page_cache_pipe_buf_release() 148 if (!(buf->flags & PIPE_BUF_FLAG_GIFT)) in user_page_pipe_buf_steal() 151 buf->flags |= PIPE_BUF_FLAG_LRU; in user_page_pipe_buf_steal() 214 buf->flags = 0; in splice_to_pipe() 215 if (spd->flags & SPLICE_F_GIFT) in splice_to_pipe() 216 buf->flags |= PIPE_BUF_FLAG_GIFT; in splice_to_pipe() 233 if (spd->flags & SPLICE_F_NONBLOCK) { in splice_to_pipe() 310 unsigned int flags) in __generic_file_splice_read() argument 324 .flags = flags, in __generic_file_splice_read() [all …]
|
D | namei.c | 125 getname_flags(const char __user *filename, int flags, int *empty) in getname_flags() argument 191 if (!(flags & LOOKUP_EMPTY)) { in getname_flags() 517 unsigned int flags; member 563 if (nd->flags & LOOKUP_RCU) { in __nd_alloc_stack() 623 if (!(nd->flags & LOOKUP_RCU)) { in terminate_walk() 628 if (nd->root.mnt && !(nd->flags & LOOKUP_ROOT)) { in terminate_walk() 633 nd->flags &= ~LOOKUP_RCU; in terminate_walk() 634 if (!(nd->flags & LOOKUP_ROOT)) in terminate_walk() 701 BUG_ON(!(nd->flags & LOOKUP_RCU)); in unlazy_walk() 703 nd->flags &= ~LOOKUP_RCU; in unlazy_walk() [all …]
|
D | timerfd.c | 63 unsigned long flags; in timerfd_triggered() local 65 spin_lock_irqsave(&ctx->wqh.lock, flags); in timerfd_triggered() 69 spin_unlock_irqrestore(&ctx->wqh.lock, flags); in timerfd_triggered() 99 unsigned long flags; in timerfd_clock_was_set() local 105 spin_lock_irqsave(&ctx->wqh.lock, flags); in timerfd_clock_was_set() 111 spin_unlock_irqrestore(&ctx->wqh.lock, flags); in timerfd_clock_was_set() 141 static void timerfd_setup_cancel(struct timerfd_ctx *ctx, int flags) in timerfd_setup_cancel() argument 146 (flags & TFD_TIMER_ABSTIME) && (flags & TFD_TIMER_CANCEL_ON_SET)) { in timerfd_setup_cancel() 171 static int timerfd_setup(struct timerfd_ctx *ctx, int flags, in timerfd_setup() argument 178 htmode = (flags & TFD_TIMER_ABSTIME) ? in timerfd_setup() [all …]
|
/fs/fscache/ |
D | operation.c | 73 switch (op->flags & FSCACHE_OP_TYPE) { in fscache_enqueue_operation() 84 pr_err("Unexpected op type %lx", op->flags); in fscache_enqueue_operation() 101 if (test_and_clear_bit(FSCACHE_OP_WAITING, &op->flags)) in fscache_run_op() 102 wake_up_bit(&op->flags, FSCACHE_OP_WAITING); in fscache_run_op() 126 kdebug("objflags=%lx", object->flags); in fscache_report_unexpected_submission() 154 unsigned long flags; in fscache_submit_exclusive_op() local 171 flags = READ_ONCE(object->flags); in fscache_submit_exclusive_op() 172 if (unlikely(!(flags & BIT(FSCACHE_OBJECT_IS_LIVE)))) { in fscache_submit_exclusive_op() 181 } else if (flags & BIT(FSCACHE_OBJECT_IS_AVAILABLE)) { in fscache_submit_exclusive_op() 201 clear_bit(FSCACHE_OBJECT_PENDING_WRITE, &object->flags); in fscache_submit_exclusive_op() [all …]
|
D | cookie.c | 110 cookie->flags = (1 << FSCACHE_COOKIE_NO_DATA_YET); in __fscache_acquire_cookie() 134 set_bit(FSCACHE_COOKIE_ENABLED, &cookie->flags); in __fscache_acquire_cookie() 143 set_bit(FSCACHE_COOKIE_ENABLED, &cookie->flags); in __fscache_acquire_cookie() 162 wait_on_bit_lock(&cookie->flags, FSCACHE_COOKIE_ENABLEMENT_LOCK, in __fscache_enable_cookie() 165 if (test_bit(FSCACHE_COOKIE_ENABLED, &cookie->flags)) in __fscache_enable_cookie() 175 set_bit(FSCACHE_COOKIE_ENABLED, &cookie->flags); in __fscache_enable_cookie() 177 set_bit(FSCACHE_COOKIE_ENABLED, &cookie->flags); in __fscache_enable_cookie() 181 clear_bit_unlock(FSCACHE_COOKIE_ENABLEMENT_LOCK, &cookie->flags); in __fscache_enable_cookie() 182 wake_up_bit(&cookie->flags, FSCACHE_COOKIE_ENABLEMENT_LOCK); in __fscache_enable_cookie() 200 set_bit(FSCACHE_COOKIE_UNAVAILABLE, &cookie->flags); in fscache_acquire_non_index_cookie() [all …]
|
/fs/hfsplus/ |
D | ioctl.c | 63 unsigned int flags = 0; in hfsplus_ioctl_getflags() local 66 flags |= FS_IMMUTABLE_FL; in hfsplus_ioctl_getflags() 68 flags |= FS_APPEND_FL; in hfsplus_ioctl_getflags() 70 flags |= FS_NODUMP_FL; in hfsplus_ioctl_getflags() 72 return put_user(flags, user_flags); in hfsplus_ioctl_getflags() 79 unsigned int flags, new_fl = 0; in hfsplus_ioctl_setflags() local 91 if (get_user(flags, user_flags)) { in hfsplus_ioctl_setflags() 98 if ((flags & (FS_IMMUTABLE_FL|FS_APPEND_FL)) || in hfsplus_ioctl_setflags() 107 if (flags & ~(FS_IMMUTABLE_FL|FS_APPEND_FL|FS_NODUMP_FL)) { in hfsplus_ioctl_setflags() 112 if (flags & FS_IMMUTABLE_FL) in hfsplus_ioctl_setflags() [all …]
|
/fs/notify/ |
D | fdinfo.c | 79 if (!(mark->flags & FSNOTIFY_MARK_FLAG_ALIVE) || in inotify_fdinfo() 80 !(mark->flags & FSNOTIFY_MARK_FLAG_INODE)) in inotify_fdinfo() 116 if (!(mark->flags & FSNOTIFY_MARK_FLAG_ALIVE)) in fanotify_fdinfo() 119 if (mark->flags & FSNOTIFY_MARK_FLAG_IGNORED_SURV_MODIFY) in fanotify_fdinfo() 122 if (mark->flags & FSNOTIFY_MARK_FLAG_INODE) { in fanotify_fdinfo() 132 } else if (mark->flags & FSNOTIFY_MARK_FLAG_VFSMOUNT) { in fanotify_fdinfo() 143 unsigned int flags = 0; in fanotify_show_fdinfo() local 147 flags |= FAN_CLASS_NOTIF; in fanotify_show_fdinfo() 150 flags |= FAN_CLASS_CONTENT; in fanotify_show_fdinfo() 153 flags |= FAN_CLASS_PRE_CONTENT; in fanotify_show_fdinfo() [all …]
|
/fs/dlm/ |
D | ast.c | 29 lkb->lkb_last_bast.flags, in dlm_dump_lkb_callbacks() 37 lkb->lkb_last_cast.flags, in dlm_dump_lkb_callbacks() 46 lkb->lkb_callbacks[i].flags, in dlm_dump_lkb_callbacks() 53 int dlm_add_lkb_callback(struct dlm_lkb *lkb, uint32_t flags, int mode, in dlm_add_lkb_callback() argument 72 if ((i > 0) && (flags & DLM_CB_BAST) && in dlm_add_lkb_callback() 73 (lkb->lkb_callbacks[i-1].flags & DLM_CB_BAST)) { in dlm_add_lkb_callback() 94 lkb->lkb_callbacks[i].flags = flags; in dlm_add_lkb_callback() 105 flags, mode, status, sbflags); in dlm_add_lkb_callback() 145 if ((cb->flags & DLM_CB_BAST) && lkb->lkb_last_cast.seq) { in dlm_rem_lkb_callback() 147 cb->flags |= DLM_CB_SKIP; in dlm_rem_lkb_callback() [all …]
|