Home
last modified time | relevance | path

Searched refs:flags (Results 1 – 25 of 802) sorted by relevance

12345678910>>...33

/fs/xfs/
Dkmem.c12 kmem_alloc(size_t size, xfs_km_flags_t flags) in kmem_alloc() argument
15 gfp_t lflags = kmem_flags_convert(flags); in kmem_alloc()
18 trace_kmem_alloc(size, flags, _RET_IP_); in kmem_alloc()
22 if (ptr || (flags & KM_MAYFAIL)) in kmem_alloc()
42 __kmem_vmalloc(size_t size, xfs_km_flags_t flags) in __kmem_vmalloc() argument
46 gfp_t lflags = kmem_flags_convert(flags); in __kmem_vmalloc()
48 if (flags & KM_NOFS) in __kmem_vmalloc()
53 if (flags & KM_NOFS) in __kmem_vmalloc()
66 kmem_alloc_io(size_t size, int align_mask, xfs_km_flags_t flags) in kmem_alloc_io() argument
70 trace_kmem_alloc_io(size, flags, _RET_IP_); in kmem_alloc_io()
[all …]
Dxfs_qm_syscalls.c37 uint flags) in xfs_qm_scall_quotaoff() argument
50 if ((mp->m_qflags & flags) == 0) in xfs_qm_scall_quotaoff()
54 flags &= (XFS_ALL_QUOTA_ACCT | XFS_ALL_QUOTA_ENFD); in xfs_qm_scall_quotaoff()
68 if ((flags & XFS_ALL_QUOTA_ACCT) == 0) { in xfs_qm_scall_quotaoff()
69 mp->m_qflags &= ~(flags); in xfs_qm_scall_quotaoff()
87 if (flags & XFS_UQUOTA_ACCT) { in xfs_qm_scall_quotaoff()
89 flags |= (XFS_UQUOTA_CHKD | XFS_UQUOTA_ENFD); in xfs_qm_scall_quotaoff()
92 if (flags & XFS_GQUOTA_ACCT) { in xfs_qm_scall_quotaoff()
94 flags |= (XFS_GQUOTA_CHKD | XFS_GQUOTA_ENFD); in xfs_qm_scall_quotaoff()
97 if (flags & XFS_PQUOTA_ACCT) { in xfs_qm_scall_quotaoff()
[all …]
Dkmem.h29 kmem_flags_convert(xfs_km_flags_t flags) in kmem_flags_convert() argument
33 BUG_ON(flags & ~(KM_NOFS|KM_MAYFAIL|KM_ZERO)); in kmem_flags_convert()
36 if (flags & KM_NOFS) in kmem_flags_convert()
46 if (flags & KM_MAYFAIL) in kmem_flags_convert()
49 if (flags & KM_ZERO) in kmem_flags_convert()
56 extern void *kmem_alloc_io(size_t size, int align_mask, xfs_km_flags_t flags);
66 kmem_zalloc(size_t size, xfs_km_flags_t flags) in kmem_zalloc() argument
68 return kmem_alloc(size, flags | KM_ZERO); in kmem_zalloc()
72 kmem_zalloc_large(size_t size, xfs_km_flags_t flags) in kmem_zalloc_large() argument
74 return kmem_alloc_large(size, flags | KM_ZERO); in kmem_zalloc_large()
[all …]
Dxfs_inode.h117 __xfs_iflags_set(xfs_inode_t *ip, unsigned short flags) in __xfs_iflags_set() argument
119 ip->i_flags |= flags; in __xfs_iflags_set()
123 xfs_iflags_set(xfs_inode_t *ip, unsigned short flags) in xfs_iflags_set() argument
126 __xfs_iflags_set(ip, flags); in xfs_iflags_set()
131 xfs_iflags_clear(xfs_inode_t *ip, unsigned short flags) in xfs_iflags_clear() argument
134 ip->i_flags &= ~flags; in xfs_iflags_clear()
139 __xfs_iflags_test(xfs_inode_t *ip, unsigned short flags) in __xfs_iflags_test() argument
141 return (ip->i_flags & flags); in __xfs_iflags_test()
145 xfs_iflags_test(xfs_inode_t *ip, unsigned short flags) in xfs_iflags_test() argument
149 ret = __xfs_iflags_test(ip, flags); in xfs_iflags_test()
[all …]
Dxfs_quotaops.c37 tstate->flags |= QCI_SYSFILE; in xfs_qm_fill_state()
67 state->s_state[USRQUOTA].flags |= QCI_ACCT_ENABLED; in xfs_fs_get_quota_state()
69 state->s_state[USRQUOTA].flags |= QCI_LIMITS_ENFORCED; in xfs_fs_get_quota_state()
71 state->s_state[GRPQUOTA].flags |= QCI_ACCT_ENABLED; in xfs_fs_get_quota_state()
73 state->s_state[GRPQUOTA].flags |= QCI_LIMITS_ENFORCED; in xfs_fs_get_quota_state()
75 state->s_state[PRJQUOTA].flags |= QCI_ACCT_ENABLED; in xfs_fs_get_quota_state()
77 state->s_state[PRJQUOTA].flags |= QCI_LIMITS_ENFORCED; in xfs_fs_get_quota_state()
140 unsigned int flags = 0; in xfs_quota_flags() local
143 flags |= XFS_UQUOTA_ACCT; in xfs_quota_flags()
145 flags |= XFS_PQUOTA_ACCT; in xfs_quota_flags()
[all …]
/fs/ubifs/
Dioctl.c32 unsigned int flags = ubifs_inode(inode)->flags; in ubifs_set_inode_flags() local
36 if (flags & UBIFS_SYNC_FL) in ubifs_set_inode_flags()
38 if (flags & UBIFS_APPEND_FL) in ubifs_set_inode_flags()
40 if (flags & UBIFS_IMMUTABLE_FL) in ubifs_set_inode_flags()
42 if (flags & UBIFS_DIRSYNC_FL) in ubifs_set_inode_flags()
44 if (flags & UBIFS_CRYPT_FL) in ubifs_set_inode_flags()
98 static int setflags(struct inode *inode, int flags) in setflags() argument
111 oldflags = ubifs2ioctl(ui->flags); in setflags()
112 err = vfs_ioc_setflags_prepare(inode, oldflags, flags); in setflags()
116 ui->flags = ioctl2ubifs(flags); in setflags()
[all …]
Dlprops.c188 lp->flags &= ~LPROPS_CAT_MASK; in add_to_lpt_heap()
189 lp->flags |= LPROPS_UNCAT; in add_to_lpt_heap()
290 lprops->flags &= ~LPROPS_CAT_MASK; in ubifs_add_to_cat()
291 lprops->flags |= cat; in ubifs_add_to_cat()
346 cat = new_lprops->flags & LPROPS_CAT_MASK; in ubifs_replace_cat()
375 int cat = lprops->flags & LPROPS_CAT_MASK; in ubifs_ensure_cat()
399 if (lprops->flags & LPROPS_TAKEN) in ubifs_categorize_lprops()
403 ubifs_assert(c, !(lprops->flags & LPROPS_INDEX)); in ubifs_categorize_lprops()
408 if (lprops->flags & LPROPS_INDEX) in ubifs_categorize_lprops()
414 if (lprops->flags & LPROPS_INDEX) { in ubifs_categorize_lprops()
[all …]
Dfind.c44 int n, cat = lprops->flags & LPROPS_CAT_MASK; in valuable()
90 if (lprops->flags & LPROPS_TAKEN) in scan_for_dirty_cb()
99 if (data->exclude_index && lprops->flags & LPROPS_INDEX) in scan_for_dirty_cb()
151 if (lprops->flags & LPROPS_TAKEN) in scan_for_dirty()
155 if (exclude_index && (lprops->flags & LPROPS_INDEX)) in scan_for_dirty()
184 ubifs_assert(c, !(lprops->flags & LPROPS_TAKEN)); in scan_for_dirty()
185 ubifs_assert(c, !exclude_index || !(lprops->flags & LPROPS_INDEX)); in scan_for_dirty()
322 lp->lnum, lp->free, lp->dirty, lp->flags); in ubifs_find_dirty_leb()
325 lp->flags | LPROPS_TAKEN, 0); in ubifs_find_dirty_leb()
357 if (lprops->flags & LPROPS_TAKEN) in scan_for_free_cb()
[all …]
/fs/ocfs2/dlm/
Ddlmunlock.c58 int flags,
87 int flags, int *call_ast, in dlmunlock_common() argument
97 flags & LKM_VALBLK); in dlmunlock_common()
109 if (in_use && !(flags & LKM_CANCEL)) { in dlmunlock_common()
118 if (master_node && !(flags & LKM_CANCEL)) { in dlmunlock_common()
141 if (flags & LKM_CANCEL) in dlmunlock_common()
150 if (flags & LKM_VALBLK) { in dlmunlock_common()
155 flags |= LKM_PUT_LVB; /* let the send function in dlmunlock_common()
162 if (flags & LKM_CANCEL) in dlmunlock_common()
169 flags, owner); in dlmunlock_common()
[all …]
Ddlmconvert.c45 struct dlm_lock *lock, int flags,
50 struct dlm_lock *lock, int flags, int type);
63 struct dlm_lock *lock, int flags, int type) in dlmconvert_master() argument
74 status = __dlmconvert_master(dlm, res, lock, flags, type, in dlmconvert_master()
106 struct dlm_lock *lock, int flags, in __dlmconvert_master() argument
136 if (flags & LKM_VALBLK) { in __dlmconvert_master()
143 lock->lksb->flags |= DLM_LKSB_PUT_LVB; in __dlmconvert_master()
153 lock->lksb->flags |= DLM_LKSB_GET_LVB; in __dlmconvert_master()
159 flags &= ~(LKM_VALBLK); in __dlmconvert_master()
197 if (lock->lksb->flags & DLM_LKSB_PUT_LVB) in __dlmconvert_master()
[all …]
Ddlmlock.c47 struct dlm_lock *lock, int flags);
105 struct dlm_lock *lock, int flags) in dlmlock_master() argument
151 if (flags & LKM_NOQUEUE) { in dlmlock_master()
188 lock->lksb->flags &= ~DLM_LKSB_GET_LVB; in dlm_revert_pending_lock()
201 struct dlm_lock *lock, int flags) in dlmlock_remote() argument
208 res->lockname.name, flags); in dlmlock_remote()
230 status = dlm_send_remote_lock_request(dlm, res, lock, flags); in dlmlock_remote()
287 struct dlm_lock *lock, int flags) in dlm_send_remote_lock_request() argument
298 create.flags = cpu_to_be32(flags); in dlm_send_remote_lock_request()
391 newlock->ml.flags = 0; in dlm_init_lock()
[all …]
/fs/
Dsplice.c78 buf->flags |= PIPE_BUF_FLAG_LRU; in page_cache_pipe_buf_steal()
96 buf->flags &= ~PIPE_BUF_FLAG_LRU; in page_cache_pipe_buf_release()
151 if (!(buf->flags & PIPE_BUF_FLAG_GIFT)) in user_page_pipe_buf_steal()
154 buf->flags |= PIPE_BUF_FLAG_LRU; in user_page_pipe_buf_steal()
208 buf->flags = 0; in splice_to_pipe()
297 unsigned int flags) in generic_file_splice_read() argument
367 unsigned int flags) in default_file_splice_read() argument
444 more = (sd->flags & SPLICE_F_MORE) ? MSG_MORE : 0; in pipe_to_sendpage()
553 if (sd->flags & SPLICE_F_NONBLOCK) in splice_from_pipe_next()
647 loff_t *ppos, size_t len, unsigned int flags, in splice_from_pipe() argument
[all …]
Dnamei.c131 getname_flags(const char __user *filename, int flags, int *empty) in getname_flags() argument
197 if (!(flags & LOOKUP_EMPTY)) { in getname_flags()
506 unsigned int flags; member
550 if (nd->flags & LOOKUP_RCU) { in __nd_alloc_stack()
607 if (!(nd->flags & LOOKUP_RCU)) { in terminate_walk()
612 if (nd->flags & LOOKUP_ROOT_GRABBED) { in terminate_walk()
614 nd->flags &= ~LOOKUP_ROOT_GRABBED; in terminate_walk()
617 nd->flags &= ~LOOKUP_RCU; in terminate_walk()
657 if (!nd->root.mnt || (nd->flags & LOOKUP_ROOT)) in legitimize_root()
659 nd->flags |= LOOKUP_ROOT_GRABBED; in legitimize_root()
[all …]
Dtimerfd.c64 unsigned long flags; in timerfd_triggered() local
66 spin_lock_irqsave(&ctx->wqh.lock, flags); in timerfd_triggered()
70 spin_unlock_irqrestore(&ctx->wqh.lock, flags); in timerfd_triggered()
100 unsigned long flags; in timerfd_clock_was_set() local
106 spin_lock_irqsave(&ctx->wqh.lock, flags); in timerfd_clock_was_set()
112 spin_unlock_irqrestore(&ctx->wqh.lock, flags); in timerfd_clock_was_set()
142 static void timerfd_setup_cancel(struct timerfd_ctx *ctx, int flags) in timerfd_setup_cancel() argument
147 (flags & TFD_TIMER_ABSTIME) && (flags & TFD_TIMER_CANCEL_ON_SET)) { in timerfd_setup_cancel()
172 static int timerfd_setup(struct timerfd_ctx *ctx, int flags, in timerfd_setup() argument
179 htmode = (flags & TFD_TIMER_ABSTIME) ? in timerfd_setup()
[all …]
/fs/jfs/
Dioctl.c41 static long jfs_map_ext2(unsigned long flags, int from) in jfs_map_ext2() argument
48 if (jfs_map[index].ext2_flag & flags) in jfs_map_ext2()
51 if (jfs_map[index].jfs_flag & flags) in jfs_map_ext2()
64 unsigned int flags; in jfs_ioctl() local
68 flags = jfs_inode->mode2 & JFS_FL_USER_VISIBLE; in jfs_ioctl()
69 flags = jfs_map_ext2(flags, 0); in jfs_ioctl()
70 return put_user(flags, (int __user *) arg); in jfs_ioctl()
83 if (get_user(flags, (int __user *) arg)) { in jfs_ioctl()
88 flags = jfs_map_ext2(flags, 1); in jfs_ioctl()
90 flags &= ~JFS_DIRSYNC_FL; in jfs_ioctl()
[all …]
/fs/notify/fanotify/
Dfanotify_user.c531 struct path *path, unsigned int flags, __u64 mask, in fanotify_find_path() argument
537 dfd, filename, flags); in fanotify_find_path()
547 if ((flags & FAN_MARK_ONLYDIR) && in fanotify_find_path()
559 if (!(flags & FAN_MARK_DONT_FOLLOW)) in fanotify_find_path()
561 if (flags & FAN_MARK_ONLYDIR) in fanotify_find_path()
586 unsigned int flags, in fanotify_mark_remove_from_mask() argument
592 if (!(flags & FAN_MARK_IGNORED_MASK)) { in fanotify_mark_remove_from_mask()
606 unsigned int flags) in fanotify_remove_mark() argument
619 removed = fanotify_mark_remove_from_mask(fsn_mark, mask, flags, in fanotify_remove_mark()
636 unsigned int flags) in fanotify_remove_vfsmount_mark() argument
[all …]
/fs/btrfs/
Dasync-thread.c84 unsigned int flags, int limit_active, int thresh) in __btrfs_alloc_workqueue() argument
110 if (flags & WQ_HIGHPRI) in __btrfs_alloc_workqueue()
111 ret->normal_wq = alloc_workqueue("btrfs-%s-high", flags, in __btrfs_alloc_workqueue()
114 ret->normal_wq = alloc_workqueue("btrfs-%s", flags, in __btrfs_alloc_workqueue()
124 trace_btrfs_workqueue_alloc(ret, name, flags & WQ_HIGHPRI); in __btrfs_alloc_workqueue()
133 unsigned int flags, in btrfs_alloc_workqueue() argument
143 flags & ~WQ_HIGHPRI, in btrfs_alloc_workqueue()
150 if (flags & WQ_HIGHPRI) { in btrfs_alloc_workqueue()
151 ret->high = __btrfs_alloc_workqueue(fs_info, name, flags, in btrfs_alloc_workqueue()
228 unsigned long flags; in run_ordered_work() local
[all …]
/fs/afs/
Drotate.c33 fc->flags |= AFS_FS_CURSOR_INTR; in afs_begin_vnode_operation()
36 fc->flags |= AFS_FS_CURSOR_STOP; in afs_begin_vnode_operation()
44 fc->flags |= AFS_FS_CURSOR_CUR_ONLY; in afs_begin_vnode_operation()
80 if (fc->flags & AFS_FS_CURSOR_CUR_ONLY) { in afs_start_fs_iteration()
89 if (test_and_clear_bit(AFS_VNODE_CB_PROMISED, &vnode->flags)) in afs_start_fs_iteration()
123 if (fc->flags & AFS_FS_CURSOR_INTR) { in afs_sleep_and_retry()
154 if (fc->flags & AFS_FS_CURSOR_STOP) { in afs_select_fileserver()
170 fc->flags |= AFS_FS_CURSOR_STOP; in afs_select_fileserver()
185 if (fc->flags & AFS_FS_CURSOR_VNOVOL) { in afs_select_fileserver()
194 set_bit(AFS_VOLUME_NEEDS_UPDATE, &vnode->volume->flags); in afs_select_fileserver()
[all …]
/fs/iomap/
Ddirect-io.c32 unsigned flags; member
83 ret = dops->end_io(iocb, dio->size, ret, dio->flags); in iomap_dio_complete()
89 !(dio->flags & IOMAP_DIO_WRITE)) in iomap_dio_complete()
107 (dio->flags & IOMAP_DIO_WRITE) && inode->i_mapping->nrpages) { in iomap_dio_complete()
120 if (ret > 0 && (dio->flags & IOMAP_DIO_NEED_SYNC)) in iomap_dio_complete()
150 bool should_dirty = (dio->flags & IOMAP_DIO_DIRTY); in iomap_dio_bio_end_io()
160 } else if (dio->flags & IOMAP_DIO_WRITE) { in iomap_dio_bio_end_io()
184 int flags = REQ_SYNC | REQ_IDLE; in iomap_dio_zero() local
197 bio_set_op_attrs(bio, REQ_OP_WRITE, flags); in iomap_dio_zero()
219 dio->flags |= IOMAP_DIO_UNWRITTEN; in iomap_dio_bio_actor()
[all …]
/fs/hfsplus/
Dioctl.c63 unsigned int flags = 0; in hfsplus_getflags() local
66 flags |= FS_IMMUTABLE_FL; in hfsplus_getflags()
68 flags |= FS_APPEND_FL; in hfsplus_getflags()
70 flags |= FS_NODUMP_FL; in hfsplus_getflags()
71 return flags; in hfsplus_getflags()
77 unsigned int flags = hfsplus_getflags(inode); in hfsplus_ioctl_getflags() local
79 return put_user(flags, user_flags); in hfsplus_ioctl_getflags()
86 unsigned int flags, new_fl = 0; in hfsplus_ioctl_setflags() local
99 if (get_user(flags, user_flags)) { in hfsplus_ioctl_setflags()
106 err = vfs_ioc_setflags_prepare(inode, oldflags, flags); in hfsplus_ioctl_setflags()
[all …]
/fs/fscache/
Doperation.c73 switch (op->flags & FSCACHE_OP_TYPE) { in fscache_enqueue_operation()
86 pr_err("Unexpected op type %lx", op->flags); in fscache_enqueue_operation()
103 if (test_and_clear_bit(FSCACHE_OP_WAITING, &op->flags)) in fscache_run_op()
104 wake_up_bit(&op->flags, FSCACHE_OP_WAITING); in fscache_run_op()
130 kdebug("objflags=%lx", object->flags); in fscache_report_unexpected_submission()
158 unsigned long flags; in fscache_submit_exclusive_op() local
177 flags = READ_ONCE(object->flags); in fscache_submit_exclusive_op()
178 if (unlikely(!(flags & BIT(FSCACHE_OBJECT_IS_LIVE)))) { in fscache_submit_exclusive_op()
187 } else if (flags & BIT(FSCACHE_OBJECT_IS_AVAILABLE)) { in fscache_submit_exclusive_op()
207 clear_bit(FSCACHE_OBJECT_PENDING_WRITE, &object->flags); in fscache_submit_exclusive_op()
[all …]
/fs/nfs/
Ddelegation.c42 set_bit(NFS_DELEGATION_REFERENCED, &delegation->flags); in nfs_mark_delegation_referenced()
47 fmode_t flags) in nfs4_is_valid_delegation() argument
49 if (delegation != NULL && (delegation->type & flags) == flags && in nfs4_is_valid_delegation()
50 !test_bit(NFS_DELEGATION_REVOKED, &delegation->flags) && in nfs4_is_valid_delegation()
51 !test_bit(NFS_DELEGATION_RETURNING, &delegation->flags)) in nfs4_is_valid_delegation()
67 nfs4_do_check_delegation(struct inode *inode, fmode_t flags, bool mark) in nfs4_do_check_delegation() argument
72 flags &= FMODE_READ|FMODE_WRITE; in nfs4_do_check_delegation()
75 if (nfs4_is_valid_delegation(delegation, flags)) { in nfs4_do_check_delegation()
91 int nfs4_have_delegation(struct inode *inode, fmode_t flags) in nfs4_have_delegation() argument
93 return nfs4_do_check_delegation(inode, flags, true); in nfs4_have_delegation()
[all …]
/fs/overlayfs/
Dfile.c30 int flags = file->f_flags | O_NOATIME | FMODE_NONOTIFY; in ovl_open_realfile() local
33 realfile = open_with_fake_path(&file->f_path, flags, realinode, in ovl_open_realfile()
46 static int ovl_change_flags(struct file *file, unsigned int flags) in ovl_change_flags() argument
52 flags |= O_NOATIME | FMODE_NONOTIFY; in ovl_change_flags()
55 if (WARN_ON((file->f_flags ^ flags) & ~OVL_SETFL_MASK)) in ovl_change_flags()
58 flags &= OVL_SETFL_MASK; in ovl_change_flags()
60 if (((flags ^ file->f_flags) & O_APPEND) && IS_APPEND(inode)) in ovl_change_flags()
63 if (flags & O_DIRECT) { in ovl_change_flags()
70 err = file->f_op->check_flags(flags); in ovl_change_flags()
76 file->f_flags = (file->f_flags & ~OVL_SETFL_MASK) | flags; in ovl_change_flags()
[all …]
/fs/dlm/
Dast.c27 lkb->lkb_last_bast.flags, in dlm_dump_lkb_callbacks()
35 lkb->lkb_last_cast.flags, in dlm_dump_lkb_callbacks()
44 lkb->lkb_callbacks[i].flags, in dlm_dump_lkb_callbacks()
51 int dlm_add_lkb_callback(struct dlm_lkb *lkb, uint32_t flags, int mode, in dlm_add_lkb_callback() argument
70 if ((i > 0) && (flags & DLM_CB_BAST) && in dlm_add_lkb_callback()
71 (lkb->lkb_callbacks[i-1].flags & DLM_CB_BAST)) { in dlm_add_lkb_callback()
92 lkb->lkb_callbacks[i].flags = flags; in dlm_add_lkb_callback()
103 flags, mode, status, sbflags); in dlm_add_lkb_callback()
143 if ((cb->flags & DLM_CB_BAST) && lkb->lkb_last_cast.seq) { in dlm_rem_lkb_callback()
145 cb->flags |= DLM_CB_SKIP; in dlm_rem_lkb_callback()
[all …]
/fs/ceph/
Dsuper.c262 fsopt->flags &= ~CEPH_MOUNT_OPT_CLEANRECOVER; in parse_fsopt_token()
265 fsopt->flags |= CEPH_MOUNT_OPT_CLEANRECOVER; in parse_fsopt_token()
278 fsopt->flags |= CEPH_MOUNT_OPT_FSCACHE; in parse_fsopt_token()
330 fsopt->flags |= CEPH_MOUNT_OPT_DIRSTAT; in parse_fsopt_token()
333 fsopt->flags &= ~CEPH_MOUNT_OPT_DIRSTAT; in parse_fsopt_token()
336 fsopt->flags |= CEPH_MOUNT_OPT_RBYTES; in parse_fsopt_token()
339 fsopt->flags &= ~CEPH_MOUNT_OPT_RBYTES; in parse_fsopt_token()
342 fsopt->flags &= ~CEPH_MOUNT_OPT_NOASYNCREADDIR; in parse_fsopt_token()
345 fsopt->flags |= CEPH_MOUNT_OPT_NOASYNCREADDIR; in parse_fsopt_token()
348 fsopt->flags |= CEPH_MOUNT_OPT_DCACHE; in parse_fsopt_token()
[all …]

12345678910>>...33