Home
last modified time | relevance | path

Searched refs:wq (Results 1 – 25 of 50) sorted by relevance

12

/fs/autofs/
Dwaitq.c17 struct autofs_wait_queue *wq, *nwq; in autofs_catatonic_mode() local
28 wq = sbi->queues; in autofs_catatonic_mode()
30 while (wq) { in autofs_catatonic_mode()
31 nwq = wq->next; in autofs_catatonic_mode()
32 wq->status = -ENOENT; /* Magic is gone - report failure */ in autofs_catatonic_mode()
33 kfree(wq->name.name - wq->offset); in autofs_catatonic_mode()
34 wq->name.name = NULL; in autofs_catatonic_mode()
35 wake_up_interruptible(&wq->queue); in autofs_catatonic_mode()
36 if (!--wq->wait_ctr) in autofs_catatonic_mode()
37 kfree(wq); in autofs_catatonic_mode()
[all …]
/fs/btrfs/
Dasync-thread.c56 struct btrfs_fs_info * __pure btrfs_workqueue_owner(const struct __btrfs_workqueue *wq) in btrfs_workqueue_owner() argument
58 return wq->fs_info; in btrfs_workqueue_owner()
63 return work->wq->fs_info; in btrfs_work_owner()
66 bool btrfs_workqueue_normal_congested(const struct btrfs_workqueue *wq) in btrfs_workqueue_normal_congested() argument
74 if (wq->normal->thresh == NO_THRESHOLD) in btrfs_workqueue_normal_congested()
77 return atomic_read(&wq->normal->pending) > wq->normal->thresh * 2; in btrfs_workqueue_normal_congested()
127 __btrfs_destroy_workqueue(struct __btrfs_workqueue *wq);
165 static inline void thresh_queue_hook(struct __btrfs_workqueue *wq) in thresh_queue_hook() argument
167 if (wq->thresh == NO_THRESHOLD) in thresh_queue_hook()
169 atomic_inc(&wq->pending); in thresh_queue_hook()
[all …]
Dasync-thread.h28 struct __btrfs_workqueue *wq; member
39 void btrfs_queue_work(struct btrfs_workqueue *wq,
41 void btrfs_destroy_workqueue(struct btrfs_workqueue *wq);
42 void btrfs_workqueue_set_max(struct btrfs_workqueue *wq, int max);
45 struct btrfs_fs_info * __pure btrfs_workqueue_owner(const struct __btrfs_workqueue *wq);
46 bool btrfs_workqueue_normal_congested(const struct btrfs_workqueue *wq);
47 void btrfs_flush_workqueue(struct btrfs_workqueue *wq);
Dmisc.h13 static inline void cond_wake_up(struct wait_queue_head *wq) in cond_wake_up() argument
19 if (wq_has_sleeper(wq)) in cond_wake_up()
20 wake_up(wq); in cond_wake_up()
23 static inline void cond_wake_up_nomb(struct wait_queue_head *wq) in cond_wake_up_nomb() argument
31 if (waitqueue_active(wq)) in cond_wake_up_nomb()
32 wake_up(wq); in cond_wake_up_nomb()
Dordered-data.c323 struct btrfs_workqueue *wq; in btrfs_mark_ordered_io_finished() local
330 wq = fs_info->endio_freespace_worker; in btrfs_mark_ordered_io_finished()
332 wq = fs_info->endio_write_workers; in btrfs_mark_ordered_io_finished()
432 btrfs_queue_work(wq, &entry->work); in btrfs_mark_ordered_io_finished()
/fs/xfs/
Dxfs_pwork.c74 pctl->wq = alloc_workqueue("%s-%d", in xfs_pwork_init()
77 if (!pctl->wq) in xfs_pwork_init()
97 queue_work(pctl->wq, &pwork->work); in xfs_pwork_queue()
105 destroy_workqueue(pctl->wq); in xfs_pwork_destroy()
106 pctl->wq = NULL; in xfs_pwork_destroy()
Dxfs_log_priv.h613 struct wait_queue_head *wq, in xlog_wait() argument
619 add_wait_queue_exclusive(wq, &wait); in xlog_wait()
623 remove_wait_queue(wq, &wait); in xlog_wait()
Dxfs_pwork.h18 struct workqueue_struct *wq; member
/fs/
Duserfaultfd.c92 wait_queue_entry_t wq; member
110 static int userfaultfd_wake_function(wait_queue_entry_t *wq, unsigned mode, in userfaultfd_wake_function() argument
118 uwq = container_of(wq, struct userfaultfd_wait_queue, wq); in userfaultfd_wake_function()
131 ret = wake_up_state(wq->private, mode); in userfaultfd_wake_function()
144 list_del_init(&wq->entry); in userfaultfd_wake_function()
510 init_waitqueue_func_entry(&uwq.wq, userfaultfd_wake_function); in handle_userfault()
511 uwq.wq.private = current; in handle_userfault()
524 __add_wait_queue(&ctx->fault_pending_wqh, &uwq.wq); in handle_userfault()
562 if (!list_empty_careful(&uwq.wq.entry)) { in handle_userfault()
568 list_del(&uwq.wq.entry); in handle_userfault()
[all …]
Ddax.c199 wait_queue_head_t *wq; in dax_wake_entry() local
201 wq = dax_entry_waitqueue(xas, entry, &key); in dax_wake_entry()
209 if (waitqueue_active(wq)) in dax_wake_entry()
210 __wake_up(wq, TASK_NORMAL, mode == WAKE_ALL ? 0 : 1, &key); in dax_wake_entry()
227 wait_queue_head_t *wq; in get_unlocked_entry() local
241 wq = dax_entry_waitqueue(xas, entry, &ewait.key); in get_unlocked_entry()
242 prepare_to_wait_exclusive(wq, &ewait.wait, in get_unlocked_entry()
247 finish_wait(wq, &ewait.wait); in get_unlocked_entry()
260 wait_queue_head_t *wq; in wait_entry_unlocked() local
265 wq = dax_entry_waitqueue(xas, entry, &ewait.key); in wait_entry_unlocked()
[all …]
Deventpoll.c190 wait_queue_head_t wq; member
667 if (waitqueue_active(&ep->wq)) in ep_done_scan()
668 wake_up(&ep->wq); in ep_done_scan()
948 init_waitqueue_head(&ep->wq); in ep_alloc()
1182 if (waitqueue_active(&ep->wq)) { in ep_poll_callback()
1199 wake_up(&ep->wq); in ep_poll_callback()
1549 if (waitqueue_active(&ep->wq)) in ep_insert()
1550 wake_up(&ep->wq); in ep_insert()
1625 if (waitqueue_active(&ep->wq)) in ep_modify()
1626 wake_up(&ep->wq); in ep_modify()
[all …]
Dinode.c2069 wait_queue_head_t *wq; in __wait_on_freeing_inode() local
2071 wq = bit_waitqueue(&inode->i_state, __I_NEW); in __wait_on_freeing_inode()
2072 prepare_to_wait(wq, &wait.wq_entry, TASK_UNINTERRUPTIBLE); in __wait_on_freeing_inode()
2076 finish_wait(wq, &wait.wq_entry); in __wait_on_freeing_inode()
2224 wait_queue_head_t *wq = bit_waitqueue(&inode->i_state, __I_DIO_WAKEUP); in __inode_dio_wait() local
2228 prepare_to_wait(wq, &q.wq_entry, TASK_UNINTERRUPTIBLE); in __inode_dio_wait()
2232 finish_wait(wq, &q.wq_entry); in __inode_dio_wait()
Ddirect-io.c579 struct workqueue_struct *wq = alloc_workqueue("dio/%s", in sb_init_dio_done_wq() local
582 if (!wq) in sb_init_dio_done_wq()
587 old = cmpxchg(&sb->s_dio_done_wq, NULL, wq); in sb_init_dio_done_wq()
590 destroy_workqueue(wq); in sb_init_dio_done_wq()
/fs/jfs/
Djfs_lock.h22 #define __SLEEP_COND(wq, cond, lock_cmd, unlock_cmd) \ argument
26 add_wait_queue(&wq, &__wait); \
36 remove_wait_queue(&wq, &__wait); \
/fs/nfs/blocklayout/
Drpc_pipefs.c62 DECLARE_WAITQUEUE(wq, current); in bl_resolve_deviceid()
87 add_wait_queue(&nn->bl_wq, &wq); in bl_resolve_deviceid()
90 remove_wait_queue(&nn->bl_wq, &wq); in bl_resolve_deviceid()
96 remove_wait_queue(&nn->bl_wq, &wq); in bl_resolve_deviceid()
/fs/nfs/
Dcallback.c111 DEFINE_WAIT(wq); in nfs41_callback_svc()
120 prepare_to_wait(&serv->sv_cb_waitq, &wq, TASK_INTERRUPTIBLE); in nfs41_callback_svc()
127 finish_wait(&serv->sv_cb_waitq, &wq); in nfs41_callback_svc()
136 finish_wait(&serv->sv_cb_waitq, &wq); in nfs41_callback_svc()
Dinode.c2335 struct workqueue_struct *wq; in nfsiod_start() local
2337 wq = alloc_workqueue("nfsiod", WQ_MEM_RECLAIM | WQ_UNBOUND, 0); in nfsiod_start()
2338 if (wq == NULL) in nfsiod_start()
2340 nfsiod_workqueue = wq; in nfsiod_start()
2349 struct workqueue_struct *wq; in nfsiod_stop() local
2351 wq = nfsiod_workqueue; in nfsiod_stop()
2352 if (wq == NULL) in nfsiod_stop()
2355 destroy_workqueue(wq); in nfsiod_stop()
/fs/jffs2/
Dos-linux.h41 #define sleep_on_spinunlock(wq, s) \ argument
44 add_wait_queue((wq), &__wait); \
48 remove_wait_queue((wq), &__wait); \
/fs/ocfs2/dlm/
Ddlmmaster.c263 init_waitqueue_head(&mle->wq); in dlm_init_mle()
539 init_waitqueue_head(&res->wq); in dlm_init_lockres()
653 wake_up(&res->wq); in dlm_lockres_drop_inflight_ref()
988 wake_up(&res->wq); in dlm_get_lock_resource()
1104 (void)wait_event_timeout(mle->wq, in dlm_wait_for_lock_mastery()
1745 wake_up(&res->wq); in dlm_do_assert_master()
1924 wake_up(&mle->wq); in dlm_assert_master_handler()
1945 wake_up(&res->wq); in dlm_assert_master_handler()
2043 wake_up(&res->wq); in dlm_assert_master_post_handler()
2396 wake_up(&res->wq); in dlm_deref_lockres_done_handler()
[all …]
Ddlmconvert.c77 wake_up(&res->wq); in dlmconvert_master()
341 wake_up(&res->wq); in dlmconvert_remote()
528 wake_up(&res->wq); in dlm_convert_lock_handler()
Ddlmthread.c48 add_wait_queue(&res->wq, &wait); in __dlm_wait_on_lockres_flags()
57 remove_wait_queue(&res->wq, &wait); in __dlm_wait_on_lockres_flags()
271 wake_up(&res->wq); in dlm_purge_lockres()
/fs/ext4/
Dpage-io.c226 struct workqueue_struct *wq; in ext4_add_complete_io() local
233 wq = sbi->rsv_conversion_wq; in ext4_add_complete_io()
235 queue_work(wq, &ei->i_rsv_conversion_work); in ext4_add_complete_io()
/fs/afs/
Ddir_silly.c242 DECLARE_WAIT_QUEUE_HEAD_ONSTACK(wq); in afs_silly_iput()
248 alias = d_alloc_parallel(dentry->d_parent, &dentry->d_name, &wq); in afs_silly_iput()
/fs/cachefiles/
Dnamei.c195 wait_queue_head_t *wq; in cachefiles_mark_object_active() local
212 wq = bit_waitqueue(&xobject->flags, CACHEFILES_OBJECT_ACTIVE); in cachefiles_mark_object_active()
216 prepare_to_wait(wq, &wait, TASK_UNINTERRUPTIBLE); in cachefiles_mark_object_active()
222 finish_wait(wq, &wait); in cachefiles_mark_object_active()
/fs/fscache/
Dpage.c37 wait_queue_head_t *wq = bit_waitqueue(&cookie->flags, 0); in __fscache_wait_on_page_write() local
41 wait_event(*wq, !__fscache_check_page_write(cookie, page)); in __fscache_wait_on_page_write()
52 wait_queue_head_t *wq = bit_waitqueue(&cookie->flags, 0); in release_page_wait_timeout() local
54 return wait_event_timeout(*wq, !__fscache_check_page_write(cookie, page), in release_page_wait_timeout()

12