Home
last modified time | relevance | path

Searched refs:wq (Results 1 – 25 of 52) sorted by relevance

123

/fs/btrfs/
Dasync-thread.c57 btrfs_workqueue_owner(const struct __btrfs_workqueue *wq) in btrfs_workqueue_owner() argument
59 return wq->fs_info; in btrfs_workqueue_owner()
65 return work->wq->fs_info; in btrfs_work_owner()
68 bool btrfs_workqueue_normal_congested(const struct btrfs_workqueue *wq) in btrfs_workqueue_normal_congested() argument
76 if (wq->normal->thresh == NO_THRESHOLD) in btrfs_workqueue_normal_congested()
79 return atomic_read(&wq->normal->pending) > wq->normal->thresh * 2; in btrfs_workqueue_normal_congested()
129 __btrfs_destroy_workqueue(struct __btrfs_workqueue *wq);
167 static inline void thresh_queue_hook(struct __btrfs_workqueue *wq) in thresh_queue_hook() argument
169 if (wq->thresh == NO_THRESHOLD) in thresh_queue_hook()
171 atomic_inc(&wq->pending); in thresh_queue_hook()
[all …]
Dmisc.h12 static inline void cond_wake_up(struct wait_queue_head *wq) in cond_wake_up() argument
18 if (wq_has_sleeper(wq)) in cond_wake_up()
19 wake_up(wq); in cond_wake_up()
22 static inline void cond_wake_up_nomb(struct wait_queue_head *wq) in cond_wake_up_nomb() argument
30 if (waitqueue_active(wq)) in cond_wake_up_nomb()
31 wake_up(wq); in cond_wake_up_nomb()
Dasync-thread.h28 struct __btrfs_workqueue *wq; member
39 void btrfs_queue_work(struct btrfs_workqueue *wq,
41 void btrfs_destroy_workqueue(struct btrfs_workqueue *wq);
42 void btrfs_workqueue_set_max(struct btrfs_workqueue *wq, int max);
45 struct btrfs_fs_info *btrfs_workqueue_owner(const struct __btrfs_workqueue *wq);
46 bool btrfs_workqueue_normal_congested(const struct btrfs_workqueue *wq);
/fs/autofs/
Dwaitq.c17 struct autofs_wait_queue *wq, *nwq; in autofs_catatonic_mode() local
28 wq = sbi->queues; in autofs_catatonic_mode()
30 while (wq) { in autofs_catatonic_mode()
31 nwq = wq->next; in autofs_catatonic_mode()
32 wq->status = -ENOENT; /* Magic is gone - report failure */ in autofs_catatonic_mode()
33 kfree(wq->name.name); in autofs_catatonic_mode()
34 wq->name.name = NULL; in autofs_catatonic_mode()
35 wq->wait_ctr--; in autofs_catatonic_mode()
36 wake_up_interruptible(&wq->queue); in autofs_catatonic_mode()
37 wq = nwq; in autofs_catatonic_mode()
[all …]
/fs/xfs/
Dxfs_pwork.c73 pctl->wq = alloc_workqueue("%s-%d", WQ_FREEZABLE, nr_threads, tag, in xfs_pwork_init()
75 if (!pctl->wq) in xfs_pwork_init()
95 queue_work(pctl->wq, &pwork->work); in xfs_pwork_queue()
103 destroy_workqueue(pctl->wq); in xfs_pwork_destroy()
104 pctl->wq = NULL; in xfs_pwork_destroy()
Dxfs_log_priv.h545 static inline void xlog_wait(wait_queue_head_t *wq, spinlock_t *lock) in xlog_wait() argument
549 add_wait_queue_exclusive(wq, &wait); in xlog_wait()
553 remove_wait_queue(wq, &wait); in xlog_wait()
Dxfs_pwork.h18 struct workqueue_struct *wq; member
/fs/
Duserfaultfd.c96 wait_queue_entry_t wq; member
106 static int userfaultfd_wake_function(wait_queue_entry_t *wq, unsigned mode, in userfaultfd_wake_function() argument
114 uwq = container_of(wq, struct userfaultfd_wait_queue, wq); in userfaultfd_wake_function()
127 ret = wake_up_state(wq->private, mode); in userfaultfd_wake_function()
140 list_del_init(&wq->entry); in userfaultfd_wake_function()
458 init_waitqueue_func_entry(&uwq.wq, userfaultfd_wake_function); in handle_userfault()
459 uwq.wq.private = current; in handle_userfault()
476 __add_wait_queue(&ctx->fault_pending_wqh, &uwq.wq); in handle_userfault()
564 if (!list_empty_careful(&uwq.wq.entry)) { in handle_userfault()
570 list_del(&uwq.wq.entry); in handle_userfault()
[all …]
Deventpoll.c192 wait_queue_head_t wq; member
569 static void ep_poll_safewake(wait_queue_head_t *wq) in ep_poll_safewake() argument
574 ep_poll_wakeup_proc, NULL, wq, (void *) (long) this_cpu); in ep_poll_safewake()
581 static void ep_poll_safewake(wait_queue_head_t *wq) in ep_poll_safewake() argument
583 wake_up_poll(wq, EPOLLIN); in ep_poll_safewake()
748 if (waitqueue_active(&ep->wq)) in ep_scan_ready_list()
749 wake_up(&ep->wq); in ep_scan_ready_list()
1028 init_waitqueue_head(&ep->wq); in ep_alloc()
1262 if (waitqueue_active(&ep->wq)) { in ep_poll_callback()
1279 wake_up(&ep->wq); in ep_poll_callback()
[all …]
Ddax.c188 wait_queue_head_t *wq; in dax_wake_entry() local
190 wq = dax_entry_waitqueue(xas, entry, &key); in dax_wake_entry()
198 if (waitqueue_active(wq)) in dax_wake_entry()
199 __wake_up(wq, TASK_NORMAL, wake_all ? 0 : 1, &key); in dax_wake_entry()
216 wait_queue_head_t *wq; in get_unlocked_entry() local
230 wq = dax_entry_waitqueue(xas, entry, &ewait.key); in get_unlocked_entry()
231 prepare_to_wait_exclusive(wq, &ewait.wait, in get_unlocked_entry()
236 finish_wait(wq, &ewait.wait); in get_unlocked_entry()
249 wait_queue_head_t *wq; in wait_entry_unlocked() local
254 wq = dax_entry_waitqueue(xas, entry, &ewait.key); in wait_entry_unlocked()
[all …]
Dinode.c1955 wait_queue_head_t *wq; in __wait_on_freeing_inode() local
1957 wq = bit_waitqueue(&inode->i_state, __I_NEW); in __wait_on_freeing_inode()
1958 prepare_to_wait(wq, &wait.wq_entry, TASK_UNINTERRUPTIBLE); in __wait_on_freeing_inode()
1962 finish_wait(wq, &wait.wq_entry); in __wait_on_freeing_inode()
2097 wait_queue_head_t *wq = bit_waitqueue(&inode->i_state, __I_DIO_WAKEUP); in __inode_dio_wait() local
2101 prepare_to_wait(wq, &q.wq_entry, TASK_UNINTERRUPTIBLE); in __inode_dio_wait()
2105 finish_wait(wq, &q.wq_entry); in __inode_dio_wait()
Ddirect-io.c620 struct workqueue_struct *wq = alloc_workqueue("dio/%s", in sb_init_dio_done_wq() local
623 if (!wq) in sb_init_dio_done_wq()
628 old = cmpxchg(&sb->s_dio_done_wq, NULL, wq); in sb_init_dio_done_wq()
631 destroy_workqueue(wq); in sb_init_dio_done_wq()
/fs/jfs/
Djfs_lock.h22 #define __SLEEP_COND(wq, cond, lock_cmd, unlock_cmd) \ argument
26 add_wait_queue(&wq, &__wait); \
36 remove_wait_queue(&wq, &__wait); \
/fs/nfs/blocklayout/
Drpc_pipefs.c62 DECLARE_WAITQUEUE(wq, current); in bl_resolve_deviceid()
87 add_wait_queue(&nn->bl_wq, &wq); in bl_resolve_deviceid()
90 remove_wait_queue(&nn->bl_wq, &wq); in bl_resolve_deviceid()
96 remove_wait_queue(&nn->bl_wq, &wq); in bl_resolve_deviceid()
/fs/nfs/
Dcallback.c111 DEFINE_WAIT(wq); in nfs41_callback_svc()
120 prepare_to_wait(&serv->sv_cb_waitq, &wq, TASK_INTERRUPTIBLE); in nfs41_callback_svc()
127 finish_wait(&serv->sv_cb_waitq, &wq); in nfs41_callback_svc()
136 finish_wait(&serv->sv_cb_waitq, &wq); in nfs41_callback_svc()
Dinode.c2144 struct workqueue_struct *wq; in nfsiod_start() local
2146 wq = alloc_workqueue("nfsiod", WQ_MEM_RECLAIM, 0); in nfsiod_start()
2147 if (wq == NULL) in nfsiod_start()
2149 nfsiod_workqueue = wq; in nfsiod_start()
2158 struct workqueue_struct *wq; in nfsiod_stop() local
2160 wq = nfsiod_workqueue; in nfsiod_stop()
2161 if (wq == NULL) in nfsiod_stop()
2164 destroy_workqueue(wq); in nfsiod_stop()
Dunlink.c123 alias = d_alloc_parallel(dentry->d_parent, &data->args.name, &data->wq); in nfs_call_unlink()
183 init_waitqueue_head(&data->wq); in nfs_async_unlink()
/fs/jffs2/
Dos-linux.h41 #define sleep_on_spinunlock(wq, s) \ argument
44 add_wait_queue((wq), &__wait); \
48 remove_wait_queue((wq), &__wait); \
/fs/ocfs2/dlm/
Ddlmmaster.c265 init_waitqueue_head(&mle->wq); in dlm_init_mle()
541 init_waitqueue_head(&res->wq); in dlm_init_lockres()
655 wake_up(&res->wq); in dlm_lockres_drop_inflight_ref()
990 wake_up(&res->wq); in dlm_get_lock_resource()
1106 (void)wait_event_timeout(mle->wq, in dlm_wait_for_lock_mastery()
1747 wake_up(&res->wq); in dlm_do_assert_master()
1926 wake_up(&mle->wq); in dlm_assert_master_handler()
1947 wake_up(&res->wq); in dlm_assert_master_handler()
2045 wake_up(&res->wq); in dlm_assert_master_post_handler()
2398 wake_up(&res->wq); in dlm_deref_lockres_done_handler()
[all …]
Ddlmconvert.c79 wake_up(&res->wq); in dlmconvert_master()
343 wake_up(&res->wq); in dlmconvert_remote()
530 wake_up(&res->wq); in dlm_convert_lock_handler()
/fs/afs/
Ddir_silly.c205 DECLARE_WAIT_QUEUE_HEAD_ONSTACK(wq); in afs_silly_iput()
211 alias = d_alloc_parallel(dentry->d_parent, &dentry->d_name, &wq); in afs_silly_iput()
/fs/ext4/
Dpage-io.c193 struct workqueue_struct *wq; in ext4_add_complete_io() local
200 wq = sbi->rsv_conversion_wq; in ext4_add_complete_io()
202 queue_work(wq, &ei->i_rsv_conversion_work); in ext4_add_complete_io()
Dfile.c108 wait_queue_head_t *wq = ext4_ioend_wq(inode); in ext4_unwritten_wait() local
110 wait_event(*wq, (atomic_read(&EXT4_I(inode)->i_unwritten) == 0)); in ext4_unwritten_wait()
/fs/cachefiles/
Dnamei.c195 wait_queue_head_t *wq; in cachefiles_mark_object_active() local
212 wq = bit_waitqueue(&xobject->flags, CACHEFILES_OBJECT_ACTIVE); in cachefiles_mark_object_active()
216 prepare_to_wait(wq, &wait, TASK_UNINTERRUPTIBLE); in cachefiles_mark_object_active()
222 finish_wait(wq, &wait); in cachefiles_mark_object_active()
/fs/fscache/
Dpage.c37 wait_queue_head_t *wq = bit_waitqueue(&cookie->flags, 0); in __fscache_wait_on_page_write() local
41 wait_event(*wq, !__fscache_check_page_write(cookie, page)); in __fscache_wait_on_page_write()
52 wait_queue_head_t *wq = bit_waitqueue(&cookie->flags, 0); in release_page_wait_timeout() local
54 return wait_event_timeout(*wq, !__fscache_check_page_write(cookie, page), in release_page_wait_timeout()

123