/fs/btrfs/ |
D | async-thread.c | 57 btrfs_workqueue_owner(const struct __btrfs_workqueue *wq) in btrfs_workqueue_owner() argument 59 return wq->fs_info; in btrfs_workqueue_owner() 65 return work->wq->fs_info; in btrfs_work_owner() 68 bool btrfs_workqueue_normal_congested(const struct btrfs_workqueue *wq) in btrfs_workqueue_normal_congested() argument 76 if (wq->normal->thresh == NO_THRESHOLD) in btrfs_workqueue_normal_congested() 79 return atomic_read(&wq->normal->pending) > wq->normal->thresh * 2; in btrfs_workqueue_normal_congested() 129 __btrfs_destroy_workqueue(struct __btrfs_workqueue *wq); 167 static inline void thresh_queue_hook(struct __btrfs_workqueue *wq) in thresh_queue_hook() argument 169 if (wq->thresh == NO_THRESHOLD) in thresh_queue_hook() 171 atomic_inc(&wq->pending); in thresh_queue_hook() [all …]
|
D | misc.h | 12 static inline void cond_wake_up(struct wait_queue_head *wq) in cond_wake_up() argument 18 if (wq_has_sleeper(wq)) in cond_wake_up() 19 wake_up(wq); in cond_wake_up() 22 static inline void cond_wake_up_nomb(struct wait_queue_head *wq) in cond_wake_up_nomb() argument 30 if (waitqueue_active(wq)) in cond_wake_up_nomb() 31 wake_up(wq); in cond_wake_up_nomb()
|
D | async-thread.h | 28 struct __btrfs_workqueue *wq; member 39 void btrfs_queue_work(struct btrfs_workqueue *wq, 41 void btrfs_destroy_workqueue(struct btrfs_workqueue *wq); 42 void btrfs_workqueue_set_max(struct btrfs_workqueue *wq, int max); 45 struct btrfs_fs_info *btrfs_workqueue_owner(const struct __btrfs_workqueue *wq); 46 bool btrfs_workqueue_normal_congested(const struct btrfs_workqueue *wq);
|
/fs/autofs/ |
D | waitq.c | 17 struct autofs_wait_queue *wq, *nwq; in autofs_catatonic_mode() local 28 wq = sbi->queues; in autofs_catatonic_mode() 30 while (wq) { in autofs_catatonic_mode() 31 nwq = wq->next; in autofs_catatonic_mode() 32 wq->status = -ENOENT; /* Magic is gone - report failure */ in autofs_catatonic_mode() 33 kfree(wq->name.name); in autofs_catatonic_mode() 34 wq->name.name = NULL; in autofs_catatonic_mode() 35 wq->wait_ctr--; in autofs_catatonic_mode() 36 wake_up_interruptible(&wq->queue); in autofs_catatonic_mode() 37 wq = nwq; in autofs_catatonic_mode() [all …]
|
/fs/xfs/ |
D | xfs_pwork.c | 73 pctl->wq = alloc_workqueue("%s-%d", WQ_FREEZABLE, nr_threads, tag, in xfs_pwork_init() 75 if (!pctl->wq) in xfs_pwork_init() 95 queue_work(pctl->wq, &pwork->work); in xfs_pwork_queue() 103 destroy_workqueue(pctl->wq); in xfs_pwork_destroy() 104 pctl->wq = NULL; in xfs_pwork_destroy()
|
D | xfs_log_priv.h | 545 static inline void xlog_wait(wait_queue_head_t *wq, spinlock_t *lock) in xlog_wait() argument 549 add_wait_queue_exclusive(wq, &wait); in xlog_wait() 553 remove_wait_queue(wq, &wait); in xlog_wait()
|
D | xfs_pwork.h | 18 struct workqueue_struct *wq; member
|
/fs/ |
D | userfaultfd.c | 96 wait_queue_entry_t wq; member 106 static int userfaultfd_wake_function(wait_queue_entry_t *wq, unsigned mode, in userfaultfd_wake_function() argument 114 uwq = container_of(wq, struct userfaultfd_wait_queue, wq); in userfaultfd_wake_function() 127 ret = wake_up_state(wq->private, mode); in userfaultfd_wake_function() 140 list_del_init(&wq->entry); in userfaultfd_wake_function() 458 init_waitqueue_func_entry(&uwq.wq, userfaultfd_wake_function); in handle_userfault() 459 uwq.wq.private = current; in handle_userfault() 476 __add_wait_queue(&ctx->fault_pending_wqh, &uwq.wq); in handle_userfault() 564 if (!list_empty_careful(&uwq.wq.entry)) { in handle_userfault() 570 list_del(&uwq.wq.entry); in handle_userfault() [all …]
|
D | eventpoll.c | 192 wait_queue_head_t wq; member 569 static void ep_poll_safewake(wait_queue_head_t *wq) in ep_poll_safewake() argument 574 ep_poll_wakeup_proc, NULL, wq, (void *) (long) this_cpu); in ep_poll_safewake() 581 static void ep_poll_safewake(wait_queue_head_t *wq) in ep_poll_safewake() argument 583 wake_up_poll(wq, EPOLLIN); in ep_poll_safewake() 748 if (waitqueue_active(&ep->wq)) in ep_scan_ready_list() 749 wake_up(&ep->wq); in ep_scan_ready_list() 1028 init_waitqueue_head(&ep->wq); in ep_alloc() 1262 if (waitqueue_active(&ep->wq)) { in ep_poll_callback() 1279 wake_up(&ep->wq); in ep_poll_callback() [all …]
|
D | dax.c | 188 wait_queue_head_t *wq; in dax_wake_entry() local 190 wq = dax_entry_waitqueue(xas, entry, &key); in dax_wake_entry() 198 if (waitqueue_active(wq)) in dax_wake_entry() 199 __wake_up(wq, TASK_NORMAL, wake_all ? 0 : 1, &key); in dax_wake_entry() 216 wait_queue_head_t *wq; in get_unlocked_entry() local 230 wq = dax_entry_waitqueue(xas, entry, &ewait.key); in get_unlocked_entry() 231 prepare_to_wait_exclusive(wq, &ewait.wait, in get_unlocked_entry() 236 finish_wait(wq, &ewait.wait); in get_unlocked_entry() 249 wait_queue_head_t *wq; in wait_entry_unlocked() local 254 wq = dax_entry_waitqueue(xas, entry, &ewait.key); in wait_entry_unlocked() [all …]
|
D | inode.c | 1955 wait_queue_head_t *wq; in __wait_on_freeing_inode() local 1957 wq = bit_waitqueue(&inode->i_state, __I_NEW); in __wait_on_freeing_inode() 1958 prepare_to_wait(wq, &wait.wq_entry, TASK_UNINTERRUPTIBLE); in __wait_on_freeing_inode() 1962 finish_wait(wq, &wait.wq_entry); in __wait_on_freeing_inode() 2097 wait_queue_head_t *wq = bit_waitqueue(&inode->i_state, __I_DIO_WAKEUP); in __inode_dio_wait() local 2101 prepare_to_wait(wq, &q.wq_entry, TASK_UNINTERRUPTIBLE); in __inode_dio_wait() 2105 finish_wait(wq, &q.wq_entry); in __inode_dio_wait()
|
D | direct-io.c | 620 struct workqueue_struct *wq = alloc_workqueue("dio/%s", in sb_init_dio_done_wq() local 623 if (!wq) in sb_init_dio_done_wq() 628 old = cmpxchg(&sb->s_dio_done_wq, NULL, wq); in sb_init_dio_done_wq() 631 destroy_workqueue(wq); in sb_init_dio_done_wq()
|
/fs/jfs/ |
D | jfs_lock.h | 22 #define __SLEEP_COND(wq, cond, lock_cmd, unlock_cmd) \ argument 26 add_wait_queue(&wq, &__wait); \ 36 remove_wait_queue(&wq, &__wait); \
|
/fs/nfs/blocklayout/ |
D | rpc_pipefs.c | 62 DECLARE_WAITQUEUE(wq, current); in bl_resolve_deviceid() 87 add_wait_queue(&nn->bl_wq, &wq); in bl_resolve_deviceid() 90 remove_wait_queue(&nn->bl_wq, &wq); in bl_resolve_deviceid() 96 remove_wait_queue(&nn->bl_wq, &wq); in bl_resolve_deviceid()
|
/fs/nfs/ |
D | callback.c | 111 DEFINE_WAIT(wq); in nfs41_callback_svc() 120 prepare_to_wait(&serv->sv_cb_waitq, &wq, TASK_INTERRUPTIBLE); in nfs41_callback_svc() 127 finish_wait(&serv->sv_cb_waitq, &wq); in nfs41_callback_svc() 136 finish_wait(&serv->sv_cb_waitq, &wq); in nfs41_callback_svc()
|
D | inode.c | 2144 struct workqueue_struct *wq; in nfsiod_start() local 2146 wq = alloc_workqueue("nfsiod", WQ_MEM_RECLAIM, 0); in nfsiod_start() 2147 if (wq == NULL) in nfsiod_start() 2149 nfsiod_workqueue = wq; in nfsiod_start() 2158 struct workqueue_struct *wq; in nfsiod_stop() local 2160 wq = nfsiod_workqueue; in nfsiod_stop() 2161 if (wq == NULL) in nfsiod_stop() 2164 destroy_workqueue(wq); in nfsiod_stop()
|
D | unlink.c | 123 alias = d_alloc_parallel(dentry->d_parent, &data->args.name, &data->wq); in nfs_call_unlink() 183 init_waitqueue_head(&data->wq); in nfs_async_unlink()
|
/fs/jffs2/ |
D | os-linux.h | 41 #define sleep_on_spinunlock(wq, s) \ argument 44 add_wait_queue((wq), &__wait); \ 48 remove_wait_queue((wq), &__wait); \
|
/fs/ocfs2/dlm/ |
D | dlmmaster.c | 265 init_waitqueue_head(&mle->wq); in dlm_init_mle() 541 init_waitqueue_head(&res->wq); in dlm_init_lockres() 655 wake_up(&res->wq); in dlm_lockres_drop_inflight_ref() 990 wake_up(&res->wq); in dlm_get_lock_resource() 1106 (void)wait_event_timeout(mle->wq, in dlm_wait_for_lock_mastery() 1747 wake_up(&res->wq); in dlm_do_assert_master() 1926 wake_up(&mle->wq); in dlm_assert_master_handler() 1947 wake_up(&res->wq); in dlm_assert_master_handler() 2045 wake_up(&res->wq); in dlm_assert_master_post_handler() 2398 wake_up(&res->wq); in dlm_deref_lockres_done_handler() [all …]
|
D | dlmconvert.c | 79 wake_up(&res->wq); in dlmconvert_master() 343 wake_up(&res->wq); in dlmconvert_remote() 530 wake_up(&res->wq); in dlm_convert_lock_handler()
|
/fs/afs/ |
D | dir_silly.c | 205 DECLARE_WAIT_QUEUE_HEAD_ONSTACK(wq); in afs_silly_iput() 211 alias = d_alloc_parallel(dentry->d_parent, &dentry->d_name, &wq); in afs_silly_iput()
|
/fs/ext4/ |
D | page-io.c | 193 struct workqueue_struct *wq; in ext4_add_complete_io() local 200 wq = sbi->rsv_conversion_wq; in ext4_add_complete_io() 202 queue_work(wq, &ei->i_rsv_conversion_work); in ext4_add_complete_io()
|
D | file.c | 108 wait_queue_head_t *wq = ext4_ioend_wq(inode); in ext4_unwritten_wait() local 110 wait_event(*wq, (atomic_read(&EXT4_I(inode)->i_unwritten) == 0)); in ext4_unwritten_wait()
|
/fs/cachefiles/ |
D | namei.c | 195 wait_queue_head_t *wq; in cachefiles_mark_object_active() local 212 wq = bit_waitqueue(&xobject->flags, CACHEFILES_OBJECT_ACTIVE); in cachefiles_mark_object_active() 216 prepare_to_wait(wq, &wait, TASK_UNINTERRUPTIBLE); in cachefiles_mark_object_active() 222 finish_wait(wq, &wait); in cachefiles_mark_object_active()
|
/fs/fscache/ |
D | page.c | 37 wait_queue_head_t *wq = bit_waitqueue(&cookie->flags, 0); in __fscache_wait_on_page_write() local 41 wait_event(*wq, !__fscache_check_page_write(cookie, page)); in __fscache_wait_on_page_write() 52 wait_queue_head_t *wq = bit_waitqueue(&cookie->flags, 0); in release_page_wait_timeout() local 54 return wait_event_timeout(*wq, !__fscache_check_page_write(cookie, page), in release_page_wait_timeout()
|