/fs/ |
D | splice.c | 77 buf->flags |= PIPE_BUF_FLAG_LRU; in page_cache_pipe_buf_try_steal() 95 buf->flags &= ~PIPE_BUF_FLAG_LRU; in page_cache_pipe_buf_release() 150 if (!(buf->flags & PIPE_BUF_FLAG_GIFT)) in user_page_pipe_buf_try_steal() 153 buf->flags |= PIPE_BUF_FLAG_LRU; in user_page_pipe_buf_try_steal() 208 buf->flags = 0; in splice_to_pipe() 300 unsigned int flags) in generic_file_splice_read() argument 358 more = (sd->flags & SPLICE_F_MORE) ? MSG_MORE : 0; in pipe_to_sendpage() 488 if (sd->flags & SPLICE_F_NONBLOCK) in splice_from_pipe_next() 585 loff_t *ppos, size_t len, unsigned int flags, in splice_from_pipe() argument 591 .flags = flags, in splice_from_pipe() [all …]
|
D | open.c | 397 static long do_faccessat(int dfd, const char __user *filename, int mode, int flags) in do_faccessat() argument 408 if (flags & ~(AT_EACCESS | AT_SYMLINK_NOFOLLOW | AT_EMPTY_PATH)) in do_faccessat() 411 if (flags & AT_SYMLINK_NOFOLLOW) in do_faccessat() 413 if (flags & AT_EMPTY_PATH) in do_faccessat() 416 if (!(flags & AT_EACCESS)) { in do_faccessat() 475 int, flags) in SYSCALL_DEFINE4() argument 477 return do_faccessat(dfd, filename, mode, flags); in SYSCALL_DEFINE4() 961 struct file *dentry_open(const struct path *path, int flags, in dentry_open() argument 972 f = alloc_empty_file(flags, cred); in dentry_open() 984 struct file *open_with_fake_path(const struct path *path, int flags, in open_with_fake_path() argument [all …]
|
D | namei.c | 128 getname_flags(const char __user *filename, int flags, int *empty) in getname_flags() argument 194 if (!(flags & LOOKUP_EMPTY)) { in getname_flags() 209 int flags = (uflags & AT_EMPTY_PATH) ? LOOKUP_EMPTY : 0; in getname_uflags() local 211 return getname_flags(filename, flags, NULL); in getname_uflags() 570 unsigned int flags, state; member 634 nd->flags & LOOKUP_RCU ? GFP_ATOMIC : GFP_KERNEL); in nd_alloc_stack() 672 if (!(nd->flags & LOOKUP_RCU)) { in terminate_walk() 682 nd->flags &= ~LOOKUP_RCU; in terminate_walk() 716 if (unlikely(nd->flags & LOOKUP_CACHED)) { in legitimize_links() 739 if (!nd->root.mnt && (nd->flags & LOOKUP_IS_SCOPED)) in legitimize_root() [all …]
|
D | ioctl.c | 116 u64 phys, u64 len, u32 flags) in fiemap_fill_next_extent() argument 124 return (flags & FIEMAP_EXTENT_LAST) ? 1 : 0; in fiemap_fill_next_extent() 130 if (flags & SET_UNKNOWN_FLAGS) in fiemap_fill_next_extent() 131 flags |= FIEMAP_EXTENT_UNKNOWN; in fiemap_fill_next_extent() 132 if (flags & SET_NO_UNMOUNTED_IO_FLAGS) in fiemap_fill_next_extent() 133 flags |= FIEMAP_EXTENT_ENCODED; in fiemap_fill_next_extent() 134 if (flags & SET_NOT_ALIGNED_FLAGS) in fiemap_fill_next_extent() 135 flags |= FIEMAP_EXTENT_NOT_ALIGNED; in fiemap_fill_next_extent() 141 extent.fe_flags = flags; in fiemap_fill_next_extent() 150 return (flags & FIEMAP_EXTENT_LAST) ? 1 : 0; in fiemap_fill_next_extent() [all …]
|
D | timerfd.c | 67 unsigned long flags; in timerfd_triggered() local 69 spin_lock_irqsave(&ctx->wqh.lock, flags); in timerfd_triggered() 73 spin_unlock_irqrestore(&ctx->wqh.lock, flags); in timerfd_triggered() 103 unsigned long flags; in timerfd_clock_was_set() local 109 spin_lock_irqsave(&ctx->wqh.lock, flags); in timerfd_clock_was_set() 115 spin_unlock_irqrestore(&ctx->wqh.lock, flags); in timerfd_clock_was_set() 161 static void timerfd_setup_cancel(struct timerfd_ctx *ctx, int flags) in timerfd_setup_cancel() argument 166 (flags & TFD_TIMER_ABSTIME) && (flags & TFD_TIMER_CANCEL_ON_SET)) { in timerfd_setup_cancel() 191 static int timerfd_setup(struct timerfd_ctx *ctx, int flags, in timerfd_setup() argument 198 htmode = (flags & TFD_TIMER_ABSTIME) ? in timerfd_setup() [all …]
|
/fs/ubifs/ |
D | ioctl.c | 37 unsigned int flags = ubifs_inode(inode)->flags; in ubifs_set_inode_flags() local 41 if (flags & UBIFS_SYNC_FL) in ubifs_set_inode_flags() 43 if (flags & UBIFS_APPEND_FL) in ubifs_set_inode_flags() 45 if (flags & UBIFS_IMMUTABLE_FL) in ubifs_set_inode_flags() 47 if (flags & UBIFS_DIRSYNC_FL) in ubifs_set_inode_flags() 49 if (flags & UBIFS_CRYPT_FL) in ubifs_set_inode_flags() 105 static int setflags(struct inode *inode, int flags) in setflags() argument 118 ui->flags &= ~ioctl2ubifs(UBIFS_SETTABLE_IOCTL_FLAGS); in setflags() 119 ui->flags |= ioctl2ubifs(flags); in setflags() 136 int flags = ubifs2ioctl(ubifs_inode(inode)->flags); in ubifs_fileattr_get() local [all …]
|
D | lprops.c | 188 lp->flags &= ~LPROPS_CAT_MASK; in add_to_lpt_heap() 189 lp->flags |= LPROPS_UNCAT; in add_to_lpt_heap() 290 lprops->flags &= ~LPROPS_CAT_MASK; in ubifs_add_to_cat() 291 lprops->flags |= cat; in ubifs_add_to_cat() 346 cat = new_lprops->flags & LPROPS_CAT_MASK; in ubifs_replace_cat() 375 int cat = lprops->flags & LPROPS_CAT_MASK; in ubifs_ensure_cat() 399 if (lprops->flags & LPROPS_TAKEN) in ubifs_categorize_lprops() 403 ubifs_assert(c, !(lprops->flags & LPROPS_INDEX)); in ubifs_categorize_lprops() 408 if (lprops->flags & LPROPS_INDEX) in ubifs_categorize_lprops() 414 if (lprops->flags & LPROPS_INDEX) { in ubifs_categorize_lprops() [all …]
|
D | find.c | 44 int n, cat = lprops->flags & LPROPS_CAT_MASK; in valuable() 90 if (lprops->flags & LPROPS_TAKEN) in scan_for_dirty_cb() 99 if (data->exclude_index && lprops->flags & LPROPS_INDEX) in scan_for_dirty_cb() 151 if (lprops->flags & LPROPS_TAKEN) in scan_for_dirty() 155 if (exclude_index && (lprops->flags & LPROPS_INDEX)) in scan_for_dirty() 184 ubifs_assert(c, !(lprops->flags & LPROPS_TAKEN)); in scan_for_dirty() 185 ubifs_assert(c, !exclude_index || !(lprops->flags & LPROPS_INDEX)); in scan_for_dirty() 322 lp->lnum, lp->free, lp->dirty, lp->flags); in ubifs_find_dirty_leb() 325 lp->flags | LPROPS_TAKEN, 0); in ubifs_find_dirty_leb() 357 if (lprops->flags & LPROPS_TAKEN) in scan_for_free_cb() [all …]
|
/fs/ocfs2/dlm/ |
D | dlmunlock.c | 56 int flags, 85 int flags, int *call_ast, in dlmunlock_common() argument 95 flags & LKM_VALBLK); in dlmunlock_common() 107 if (in_use && !(flags & LKM_CANCEL)) { in dlmunlock_common() 116 if (master_node && !(flags & LKM_CANCEL)) { in dlmunlock_common() 139 if (flags & LKM_CANCEL) in dlmunlock_common() 148 if (flags & LKM_VALBLK) { in dlmunlock_common() 153 flags |= LKM_PUT_LVB; /* let the send function in dlmunlock_common() 160 if (flags & LKM_CANCEL) in dlmunlock_common() 167 flags, owner); in dlmunlock_common() [all …]
|
D | dlmconvert.c | 43 struct dlm_lock *lock, int flags, 48 struct dlm_lock *lock, int flags, int type); 61 struct dlm_lock *lock, int flags, int type) in dlmconvert_master() argument 72 status = __dlmconvert_master(dlm, res, lock, flags, type, in dlmconvert_master() 104 struct dlm_lock *lock, int flags, in __dlmconvert_master() argument 134 if (flags & LKM_VALBLK) { in __dlmconvert_master() 141 lock->lksb->flags |= DLM_LKSB_PUT_LVB; in __dlmconvert_master() 151 lock->lksb->flags |= DLM_LKSB_GET_LVB; in __dlmconvert_master() 157 flags &= ~(LKM_VALBLK); in __dlmconvert_master() 195 if (lock->lksb->flags & DLM_LKSB_PUT_LVB) in __dlmconvert_master() [all …]
|
D | dlmlock.c | 45 struct dlm_lock *lock, int flags); 103 struct dlm_lock *lock, int flags) in dlmlock_master() argument 149 if (flags & LKM_NOQUEUE) { in dlmlock_master() 186 lock->lksb->flags &= ~DLM_LKSB_GET_LVB; in dlm_revert_pending_lock() 199 struct dlm_lock *lock, int flags) in dlmlock_remote() argument 206 res->lockname.name, flags); in dlmlock_remote() 228 status = dlm_send_remote_lock_request(dlm, res, lock, flags); in dlmlock_remote() 285 struct dlm_lock *lock, int flags) in dlm_send_remote_lock_request() argument 296 create.flags = cpu_to_be32(flags); in dlm_send_remote_lock_request() 389 newlock->ml.flags = 0; in dlm_init_lock() [all …]
|
/fs/afs/ |
D | rotate.c | 52 if (op->flags & AFS_OPERATION_CUR_ONLY) { in afs_start_fs_iteration() 61 if (test_and_clear_bit(AFS_VNODE_CB_PROMISED, &vnode->flags)) in afs_start_fs_iteration() 92 if (!(op->flags & AFS_OPERATION_UNINTR)) { in afs_sleep_and_retry() 123 if (op->flags & AFS_OPERATION_STOP) { in afs_select_fileserver() 139 op->flags |= AFS_OPERATION_STOP; in afs_select_fileserver() 154 if (op->flags & AFS_OPERATION_VNOVOL) { in afs_select_fileserver() 163 set_bit(AFS_VOLUME_NEEDS_UPDATE, &op->volume->flags); in afs_select_fileserver() 168 if (test_bit(AFS_VOLUME_DELETED, &op->volume->flags)) { in afs_select_fileserver() 182 op->flags |= AFS_OPERATION_VNOVOL; in afs_select_fileserver() 196 if (!test_and_set_bit(AFS_VOLUME_OFFLINE, &op->volume->flags)) { in afs_select_fileserver() [all …]
|
D | vl_probe.c | 20 if (!(server->probe.flags & AFS_VLSERVER_PROBE_RESPONDED)) { in afs_finished_vl_probe() 22 clear_bit(AFS_VLSERVER_FL_RESPONDING, &server->flags); in afs_finished_vl_probe() 25 clear_bit_unlock(AFS_VLSERVER_FL_PROBING, &server->flags); in afs_finished_vl_probe() 26 wake_up_bit(&server->flags, AFS_VLSERVER_FL_PROBING); in afs_finished_vl_probe() 66 if (!(server->probe.flags & AFS_VLSERVER_PROBE_RESPONDED)) { in afs_vlserver_probe_result() 76 server->probe.flags |= AFS_VLSERVER_PROBE_LOCAL_FAILURE; in afs_vlserver_probe_result() 93 if (!(server->probe.flags & AFS_VLSERVER_PROBE_RESPONDED) && in afs_vlserver_probe_result() 107 server->probe.flags |= AFS_VLSERVER_PROBE_IS_YFS; in afs_vlserver_probe_result() 108 set_bit(AFS_VLSERVER_FL_IS_YFS, &server->flags); in afs_vlserver_probe_result() 111 server->probe.flags |= AFS_VLSERVER_PROBE_NOT_YFS; in afs_vlserver_probe_result() [all …]
|
/fs/nfs/ |
D | delegation.c | 42 if (!test_and_set_bit(NFS_DELEGATION_REVOKED, &delegation->flags)) { in nfs_mark_delegation_revoked() 45 if (!test_bit(NFS_DELEGATION_RETURNING, &delegation->flags)) in nfs_mark_delegation_revoked() 75 set_bit(NFS_DELEGATION_REFERENCED, &delegation->flags); in nfs_mark_delegation_referenced() 81 set_bit(NFS_DELEGATION_RETURN, &delegation->flags); in nfs_mark_return_delegation() 87 fmode_t flags) in nfs4_is_valid_delegation() argument 89 if (delegation != NULL && (delegation->type & flags) == flags && in nfs4_is_valid_delegation() 90 !test_bit(NFS_DELEGATION_REVOKED, &delegation->flags) && in nfs4_is_valid_delegation() 91 !test_bit(NFS_DELEGATION_RETURNING, &delegation->flags)) in nfs4_is_valid_delegation() 107 nfs4_do_check_delegation(struct inode *inode, fmode_t flags, bool mark) in nfs4_do_check_delegation() argument 112 flags &= FMODE_READ|FMODE_WRITE; in nfs4_do_check_delegation() [all …]
|
D | fs_context.c | 434 ctx->flags &= ~NFS_MOUNT_VER3; in nfs_parse_version_string() 440 ctx->flags |= NFS_MOUNT_VER3; in nfs_parse_version_string() 501 ctx->flags |= NFS_MOUNT_SOFT; in nfs_fs_context_parse_param() 502 ctx->flags &= ~NFS_MOUNT_SOFTERR; in nfs_fs_context_parse_param() 505 ctx->flags |= NFS_MOUNT_SOFTERR | NFS_MOUNT_SOFTREVAL; in nfs_fs_context_parse_param() 506 ctx->flags &= ~NFS_MOUNT_SOFT; in nfs_fs_context_parse_param() 509 ctx->flags &= ~(NFS_MOUNT_SOFT | in nfs_fs_context_parse_param() 515 ctx->flags &= ~NFS_MOUNT_SOFTREVAL; in nfs_fs_context_parse_param() 517 ctx->flags |= NFS_MOUNT_SOFTREVAL; in nfs_fs_context_parse_param() 521 ctx->flags &= ~NFS_MOUNT_POSIX; in nfs_fs_context_parse_param() [all …]
|
/fs/notify/fanotify/ |
D | fanotify_user.c | 859 struct path *path, unsigned int flags, __u64 mask, in fanotify_find_path() argument 865 dfd, filename, flags); in fanotify_find_path() 875 if ((flags & FAN_MARK_ONLYDIR) && in fanotify_find_path() 887 if (!(flags & FAN_MARK_DONT_FOLLOW)) in fanotify_find_path() 889 if (flags & FAN_MARK_ONLYDIR) in fanotify_find_path() 913 __u32 mask, unsigned int flags, in fanotify_mark_remove_from_mask() argument 921 if (!(flags & FAN_MARK_IGNORED_MASK)) { in fanotify_mark_remove_from_mask() 941 unsigned int flags, __u32 umask) in fanotify_remove_mark() argument 954 removed = fanotify_mark_remove_from_mask(fsn_mark, mask, flags, in fanotify_remove_mark() 971 unsigned int flags, __u32 umask) in fanotify_remove_vfsmount_mark() argument [all …]
|
/fs/btrfs/ |
D | async-thread.c | 82 unsigned int flags, int limit_active, int thresh) in __btrfs_alloc_workqueue() argument 108 if (flags & WQ_HIGHPRI) in __btrfs_alloc_workqueue() 109 ret->normal_wq = alloc_workqueue("btrfs-%s-high", flags, in __btrfs_alloc_workqueue() 112 ret->normal_wq = alloc_workqueue("btrfs-%s", flags, in __btrfs_alloc_workqueue() 122 trace_btrfs_workqueue_alloc(ret, name, flags & WQ_HIGHPRI); in __btrfs_alloc_workqueue() 131 unsigned int flags, in btrfs_alloc_workqueue() argument 141 flags & ~WQ_HIGHPRI, in btrfs_alloc_workqueue() 148 if (flags & WQ_HIGHPRI) { in btrfs_alloc_workqueue() 149 ret->high = __btrfs_alloc_workqueue(fs_info, name, flags, in btrfs_alloc_workqueue() 226 unsigned long flags; in run_ordered_work() local [all …]
|
D | subpage.c | 318 unsigned long flags; in btrfs_subpage_set_uptodate() local 320 spin_lock_irqsave(&subpage->lock, flags); in btrfs_subpage_set_uptodate() 324 spin_unlock_irqrestore(&subpage->lock, flags); in btrfs_subpage_set_uptodate() 332 unsigned long flags; in btrfs_subpage_clear_uptodate() local 334 spin_lock_irqsave(&subpage->lock, flags); in btrfs_subpage_clear_uptodate() 337 spin_unlock_irqrestore(&subpage->lock, flags); in btrfs_subpage_clear_uptodate() 345 unsigned long flags; in btrfs_subpage_set_error() local 347 spin_lock_irqsave(&subpage->lock, flags); in btrfs_subpage_set_error() 350 spin_unlock_irqrestore(&subpage->lock, flags); in btrfs_subpage_set_error() 358 unsigned long flags; in btrfs_subpage_clear_error() local [all …]
|
/fs/iomap/ |
D | direct-io.c | 33 unsigned flags; member 88 ret = dops->end_io(iocb, dio->size, ret, dio->flags); in iomap_dio_complete() 94 !(dio->flags & IOMAP_DIO_WRITE)) in iomap_dio_complete() 111 (dio->flags & IOMAP_DIO_WRITE) && inode->i_mapping->nrpages) { in iomap_dio_complete() 128 if (dio->flags & IOMAP_DIO_NEED_SYNC) in iomap_dio_complete() 159 bool should_dirty = (dio->flags & IOMAP_DIO_DIRTY); in iomap_dio_bio_end_io() 169 } else if (dio->flags & IOMAP_DIO_WRITE) { in iomap_dio_bio_end_io() 192 int flags = REQ_SYNC | REQ_IDLE; in iomap_dio_zero() local 205 bio_set_op_attrs(bio, REQ_OP_WRITE, flags); in iomap_dio_zero() 219 if (!(dio->flags & IOMAP_DIO_WRITE)) { in iomap_dio_bio_opflags() [all …]
|
D | fiemap.c | 12 const struct iomap *iomap, u32 flags) in iomap_to_fiemap() argument 19 flags |= FIEMAP_EXTENT_DELALLOC | FIEMAP_EXTENT_UNKNOWN; in iomap_to_fiemap() 24 flags |= FIEMAP_EXTENT_UNWRITTEN; in iomap_to_fiemap() 27 flags |= FIEMAP_EXTENT_DATA_INLINE; in iomap_to_fiemap() 31 if (iomap->flags & IOMAP_F_MERGED) in iomap_to_fiemap() 32 flags |= FIEMAP_EXTENT_MERGED; in iomap_to_fiemap() 33 if (iomap->flags & IOMAP_F_SHARED) in iomap_to_fiemap() 34 flags |= FIEMAP_EXTENT_SHARED; in iomap_to_fiemap() 38 iomap->length, flags); in iomap_to_fiemap() 68 .flags = IOMAP_REPORT, in iomap_fiemap() [all …]
|
/fs/fscache/ |
D | operation.c | 76 switch (op->flags & FSCACHE_OP_TYPE) { in fscache_enqueue_operation() 89 pr_err("Unexpected op type %lx", op->flags); in fscache_enqueue_operation() 106 if (test_and_clear_bit(FSCACHE_OP_WAITING, &op->flags)) in fscache_run_op() 107 wake_up_bit(&op->flags, FSCACHE_OP_WAITING); in fscache_run_op() 133 kdebug("objflags=%lx", object->flags); in fscache_report_unexpected_submission() 161 unsigned long flags; in fscache_submit_exclusive_op() local 180 flags = READ_ONCE(object->flags); in fscache_submit_exclusive_op() 181 if (unlikely(!(flags & BIT(FSCACHE_OBJECT_IS_LIVE)))) { in fscache_submit_exclusive_op() 190 } else if (flags & BIT(FSCACHE_OBJECT_IS_AVAILABLE)) { in fscache_submit_exclusive_op() 210 clear_bit(FSCACHE_OBJECT_PENDING_WRITE, &object->flags); in fscache_submit_exclusive_op() [all …]
|
/fs/xfs/ |
D | xfs_quotaops.c | 37 tstate->flags |= QCI_SYSFILE; in xfs_qm_fill_state() 67 state->s_state[USRQUOTA].flags |= QCI_ACCT_ENABLED; in xfs_fs_get_quota_state() 69 state->s_state[USRQUOTA].flags |= QCI_LIMITS_ENFORCED; in xfs_fs_get_quota_state() 71 state->s_state[GRPQUOTA].flags |= QCI_ACCT_ENABLED; in xfs_fs_get_quota_state() 73 state->s_state[GRPQUOTA].flags |= QCI_LIMITS_ENFORCED; in xfs_fs_get_quota_state() 75 state->s_state[PRJQUOTA].flags |= QCI_ACCT_ENABLED; in xfs_fs_get_quota_state() 77 state->s_state[PRJQUOTA].flags |= QCI_LIMITS_ENFORCED; in xfs_fs_get_quota_state() 138 unsigned int flags = 0; in xfs_quota_flags() local 141 flags |= XFS_UQUOTA_ACCT; in xfs_quota_flags() 143 flags |= XFS_PQUOTA_ACCT; in xfs_quota_flags() [all …]
|
D | xfs_inode.h | 122 __xfs_iflags_set(xfs_inode_t *ip, unsigned short flags) in __xfs_iflags_set() argument 124 ip->i_flags |= flags; in __xfs_iflags_set() 128 xfs_iflags_set(xfs_inode_t *ip, unsigned short flags) in xfs_iflags_set() argument 131 __xfs_iflags_set(ip, flags); in xfs_iflags_set() 136 xfs_iflags_clear(xfs_inode_t *ip, unsigned short flags) in xfs_iflags_clear() argument 139 ip->i_flags &= ~flags; in xfs_iflags_clear() 144 __xfs_iflags_test(xfs_inode_t *ip, unsigned short flags) in __xfs_iflags_test() argument 146 return (ip->i_flags & flags); in __xfs_iflags_test() 150 xfs_iflags_test(xfs_inode_t *ip, unsigned short flags) in xfs_iflags_test() argument 154 ret = __xfs_iflags_test(ip, flags); in xfs_iflags_test() [all …]
|
/fs/jfs/ |
D | ioctl.c | 42 static long jfs_map_ext2(unsigned long flags, int from) in jfs_map_ext2() argument 49 if (jfs_map[index].ext2_flag & flags) in jfs_map_ext2() 52 if (jfs_map[index].jfs_flag & flags) in jfs_map_ext2() 63 unsigned int flags = jfs_inode->mode2 & JFS_FL_USER_VISIBLE; in jfs_fileattr_get() local 68 fileattr_fill_flags(fa, jfs_map_ext2(flags, 0)); in jfs_fileattr_get() 78 unsigned int flags; in jfs_fileattr_set() local 86 flags = jfs_map_ext2(fa->flags, 1); in jfs_fileattr_set() 88 flags &= ~JFS_DIRSYNC_FL; in jfs_fileattr_set() 94 flags = flags & JFS_FL_USER_MODIFIABLE; in jfs_fileattr_set() 95 flags |= jfs_inode->mode2 & ~JFS_FL_USER_MODIFIABLE; in jfs_fileattr_set() [all …]
|
/fs/dlm/ |
D | ast.c | 27 lkb->lkb_last_bast.flags, in dlm_dump_lkb_callbacks() 35 lkb->lkb_last_cast.flags, in dlm_dump_lkb_callbacks() 44 lkb->lkb_callbacks[i].flags, in dlm_dump_lkb_callbacks() 51 int dlm_add_lkb_callback(struct dlm_lkb *lkb, uint32_t flags, int mode, in dlm_add_lkb_callback() argument 70 if ((i > 0) && (flags & DLM_CB_BAST) && in dlm_add_lkb_callback() 71 (lkb->lkb_callbacks[i-1].flags & DLM_CB_BAST)) { in dlm_add_lkb_callback() 92 lkb->lkb_callbacks[i].flags = flags; in dlm_add_lkb_callback() 103 flags, mode, status, sbflags); in dlm_add_lkb_callback() 143 if ((cb->flags & DLM_CB_BAST) && lkb->lkb_last_cast.seq) { in dlm_rem_lkb_callback() 145 cb->flags |= DLM_CB_SKIP; in dlm_rem_lkb_callback() [all …]
|