/fs/autofs/ |
D | waitq.c | 17 struct autofs_wait_queue *wq, *nwq; in autofs_catatonic_mode() local 28 wq = sbi->queues; in autofs_catatonic_mode() 30 while (wq) { in autofs_catatonic_mode() 31 nwq = wq->next; in autofs_catatonic_mode() 32 wq->status = -ENOENT; /* Magic is gone - report failure */ in autofs_catatonic_mode() 33 kfree(wq->name.name - wq->offset); in autofs_catatonic_mode() 34 wq->name.name = NULL; in autofs_catatonic_mode() 35 wake_up_interruptible(&wq->queue); in autofs_catatonic_mode() 36 if (!--wq->wait_ctr) in autofs_catatonic_mode() 37 kfree(wq); in autofs_catatonic_mode() [all …]
|
/fs/btrfs/ |
D | async-thread.c | 56 struct btrfs_fs_info * __pure btrfs_workqueue_owner(const struct __btrfs_workqueue *wq) in btrfs_workqueue_owner() argument 58 return wq->fs_info; in btrfs_workqueue_owner() 63 return work->wq->fs_info; in btrfs_work_owner() 66 bool btrfs_workqueue_normal_congested(const struct btrfs_workqueue *wq) in btrfs_workqueue_normal_congested() argument 74 if (wq->normal->thresh == NO_THRESHOLD) in btrfs_workqueue_normal_congested() 77 return atomic_read(&wq->normal->pending) > wq->normal->thresh * 2; in btrfs_workqueue_normal_congested() 127 __btrfs_destroy_workqueue(struct __btrfs_workqueue *wq); 165 static inline void thresh_queue_hook(struct __btrfs_workqueue *wq) in thresh_queue_hook() argument 167 if (wq->thresh == NO_THRESHOLD) in thresh_queue_hook() 169 atomic_inc(&wq->pending); in thresh_queue_hook() [all …]
|
D | async-thread.h | 28 struct __btrfs_workqueue *wq; member 39 void btrfs_queue_work(struct btrfs_workqueue *wq, 41 void btrfs_destroy_workqueue(struct btrfs_workqueue *wq); 42 void btrfs_workqueue_set_max(struct btrfs_workqueue *wq, int max); 45 struct btrfs_fs_info * __pure btrfs_workqueue_owner(const struct __btrfs_workqueue *wq); 46 bool btrfs_workqueue_normal_congested(const struct btrfs_workqueue *wq); 47 void btrfs_flush_workqueue(struct btrfs_workqueue *wq);
|
D | misc.h | 13 static inline void cond_wake_up(struct wait_queue_head *wq) in cond_wake_up() argument 19 if (wq_has_sleeper(wq)) in cond_wake_up() 20 wake_up(wq); in cond_wake_up() 23 static inline void cond_wake_up_nomb(struct wait_queue_head *wq) in cond_wake_up_nomb() argument 31 if (waitqueue_active(wq)) in cond_wake_up_nomb() 32 wake_up(wq); in cond_wake_up_nomb()
|
D | ordered-data.c | 323 struct btrfs_workqueue *wq; in btrfs_mark_ordered_io_finished() local 330 wq = fs_info->endio_freespace_worker; in btrfs_mark_ordered_io_finished() 332 wq = fs_info->endio_write_workers; in btrfs_mark_ordered_io_finished() 432 btrfs_queue_work(wq, &entry->work); in btrfs_mark_ordered_io_finished()
|
/fs/xfs/ |
D | xfs_pwork.c | 74 pctl->wq = alloc_workqueue("%s-%d", in xfs_pwork_init() 77 if (!pctl->wq) in xfs_pwork_init() 97 queue_work(pctl->wq, &pwork->work); in xfs_pwork_queue() 105 destroy_workqueue(pctl->wq); in xfs_pwork_destroy() 106 pctl->wq = NULL; in xfs_pwork_destroy()
|
D | xfs_log_priv.h | 613 struct wait_queue_head *wq, in xlog_wait() argument 619 add_wait_queue_exclusive(wq, &wait); in xlog_wait() 623 remove_wait_queue(wq, &wait); in xlog_wait()
|
D | xfs_pwork.h | 18 struct workqueue_struct *wq; member
|
/fs/ |
D | userfaultfd.c | 92 wait_queue_entry_t wq; member 110 static int userfaultfd_wake_function(wait_queue_entry_t *wq, unsigned mode, in userfaultfd_wake_function() argument 118 uwq = container_of(wq, struct userfaultfd_wait_queue, wq); in userfaultfd_wake_function() 131 ret = wake_up_state(wq->private, mode); in userfaultfd_wake_function() 144 list_del_init(&wq->entry); in userfaultfd_wake_function() 510 init_waitqueue_func_entry(&uwq.wq, userfaultfd_wake_function); in handle_userfault() 511 uwq.wq.private = current; in handle_userfault() 524 __add_wait_queue(&ctx->fault_pending_wqh, &uwq.wq); in handle_userfault() 562 if (!list_empty_careful(&uwq.wq.entry)) { in handle_userfault() 568 list_del(&uwq.wq.entry); in handle_userfault() [all …]
|
D | dax.c | 199 wait_queue_head_t *wq; in dax_wake_entry() local 201 wq = dax_entry_waitqueue(xas, entry, &key); in dax_wake_entry() 209 if (waitqueue_active(wq)) in dax_wake_entry() 210 __wake_up(wq, TASK_NORMAL, mode == WAKE_ALL ? 0 : 1, &key); in dax_wake_entry() 227 wait_queue_head_t *wq; in get_unlocked_entry() local 241 wq = dax_entry_waitqueue(xas, entry, &ewait.key); in get_unlocked_entry() 242 prepare_to_wait_exclusive(wq, &ewait.wait, in get_unlocked_entry() 247 finish_wait(wq, &ewait.wait); in get_unlocked_entry() 260 wait_queue_head_t *wq; in wait_entry_unlocked() local 265 wq = dax_entry_waitqueue(xas, entry, &ewait.key); in wait_entry_unlocked() [all …]
|
D | eventpoll.c | 190 wait_queue_head_t wq; member 667 if (waitqueue_active(&ep->wq)) in ep_done_scan() 668 wake_up(&ep->wq); in ep_done_scan() 948 init_waitqueue_head(&ep->wq); in ep_alloc() 1182 if (waitqueue_active(&ep->wq)) { in ep_poll_callback() 1199 wake_up(&ep->wq); in ep_poll_callback() 1549 if (waitqueue_active(&ep->wq)) in ep_insert() 1550 wake_up(&ep->wq); in ep_insert() 1625 if (waitqueue_active(&ep->wq)) in ep_modify() 1626 wake_up(&ep->wq); in ep_modify() [all …]
|
D | inode.c | 2069 wait_queue_head_t *wq; in __wait_on_freeing_inode() local 2071 wq = bit_waitqueue(&inode->i_state, __I_NEW); in __wait_on_freeing_inode() 2072 prepare_to_wait(wq, &wait.wq_entry, TASK_UNINTERRUPTIBLE); in __wait_on_freeing_inode() 2076 finish_wait(wq, &wait.wq_entry); in __wait_on_freeing_inode() 2224 wait_queue_head_t *wq = bit_waitqueue(&inode->i_state, __I_DIO_WAKEUP); in __inode_dio_wait() local 2228 prepare_to_wait(wq, &q.wq_entry, TASK_UNINTERRUPTIBLE); in __inode_dio_wait() 2232 finish_wait(wq, &q.wq_entry); in __inode_dio_wait()
|
D | direct-io.c | 579 struct workqueue_struct *wq = alloc_workqueue("dio/%s", in sb_init_dio_done_wq() local 582 if (!wq) in sb_init_dio_done_wq() 587 old = cmpxchg(&sb->s_dio_done_wq, NULL, wq); in sb_init_dio_done_wq() 590 destroy_workqueue(wq); in sb_init_dio_done_wq()
|
/fs/jfs/ |
D | jfs_lock.h | 22 #define __SLEEP_COND(wq, cond, lock_cmd, unlock_cmd) \ argument 26 add_wait_queue(&wq, &__wait); \ 36 remove_wait_queue(&wq, &__wait); \
|
/fs/nfs/blocklayout/ |
D | rpc_pipefs.c | 62 DECLARE_WAITQUEUE(wq, current); in bl_resolve_deviceid() 87 add_wait_queue(&nn->bl_wq, &wq); in bl_resolve_deviceid() 90 remove_wait_queue(&nn->bl_wq, &wq); in bl_resolve_deviceid() 96 remove_wait_queue(&nn->bl_wq, &wq); in bl_resolve_deviceid()
|
/fs/nfs/ |
D | callback.c | 111 DEFINE_WAIT(wq); in nfs41_callback_svc() 120 prepare_to_wait(&serv->sv_cb_waitq, &wq, TASK_INTERRUPTIBLE); in nfs41_callback_svc() 127 finish_wait(&serv->sv_cb_waitq, &wq); in nfs41_callback_svc() 136 finish_wait(&serv->sv_cb_waitq, &wq); in nfs41_callback_svc()
|
D | inode.c | 2335 struct workqueue_struct *wq; in nfsiod_start() local 2337 wq = alloc_workqueue("nfsiod", WQ_MEM_RECLAIM | WQ_UNBOUND, 0); in nfsiod_start() 2338 if (wq == NULL) in nfsiod_start() 2340 nfsiod_workqueue = wq; in nfsiod_start() 2349 struct workqueue_struct *wq; in nfsiod_stop() local 2351 wq = nfsiod_workqueue; in nfsiod_stop() 2352 if (wq == NULL) in nfsiod_stop() 2355 destroy_workqueue(wq); in nfsiod_stop()
|
/fs/jffs2/ |
D | os-linux.h | 41 #define sleep_on_spinunlock(wq, s) \ argument 44 add_wait_queue((wq), &__wait); \ 48 remove_wait_queue((wq), &__wait); \
|
/fs/ocfs2/dlm/ |
D | dlmmaster.c | 263 init_waitqueue_head(&mle->wq); in dlm_init_mle() 539 init_waitqueue_head(&res->wq); in dlm_init_lockres() 653 wake_up(&res->wq); in dlm_lockres_drop_inflight_ref() 988 wake_up(&res->wq); in dlm_get_lock_resource() 1104 (void)wait_event_timeout(mle->wq, in dlm_wait_for_lock_mastery() 1745 wake_up(&res->wq); in dlm_do_assert_master() 1924 wake_up(&mle->wq); in dlm_assert_master_handler() 1945 wake_up(&res->wq); in dlm_assert_master_handler() 2043 wake_up(&res->wq); in dlm_assert_master_post_handler() 2396 wake_up(&res->wq); in dlm_deref_lockres_done_handler() [all …]
|
D | dlmconvert.c | 77 wake_up(&res->wq); in dlmconvert_master() 341 wake_up(&res->wq); in dlmconvert_remote() 528 wake_up(&res->wq); in dlm_convert_lock_handler()
|
D | dlmthread.c | 48 add_wait_queue(&res->wq, &wait); in __dlm_wait_on_lockres_flags() 57 remove_wait_queue(&res->wq, &wait); in __dlm_wait_on_lockres_flags() 271 wake_up(&res->wq); in dlm_purge_lockres()
|
/fs/ext4/ |
D | page-io.c | 226 struct workqueue_struct *wq; in ext4_add_complete_io() local 233 wq = sbi->rsv_conversion_wq; in ext4_add_complete_io() 235 queue_work(wq, &ei->i_rsv_conversion_work); in ext4_add_complete_io()
|
/fs/afs/ |
D | dir_silly.c | 242 DECLARE_WAIT_QUEUE_HEAD_ONSTACK(wq); in afs_silly_iput() 248 alias = d_alloc_parallel(dentry->d_parent, &dentry->d_name, &wq); in afs_silly_iput()
|
/fs/cachefiles/ |
D | namei.c | 195 wait_queue_head_t *wq; in cachefiles_mark_object_active() local 212 wq = bit_waitqueue(&xobject->flags, CACHEFILES_OBJECT_ACTIVE); in cachefiles_mark_object_active() 216 prepare_to_wait(wq, &wait, TASK_UNINTERRUPTIBLE); in cachefiles_mark_object_active() 222 finish_wait(wq, &wait); in cachefiles_mark_object_active()
|
/fs/fscache/ |
D | page.c | 37 wait_queue_head_t *wq = bit_waitqueue(&cookie->flags, 0); in __fscache_wait_on_page_write() local 41 wait_event(*wq, !__fscache_check_page_write(cookie, page)); in __fscache_wait_on_page_write() 52 wait_queue_head_t *wq = bit_waitqueue(&cookie->flags, 0); in release_page_wait_timeout() local 54 return wait_event_timeout(*wq, !__fscache_check_page_write(cookie, page), in release_page_wait_timeout()
|