/net/sunrpc/ |
D | sched.c | 321 wait_queue_head_t *wq = bit_waitqueue(m, RPC_TASK_ACTIVE); in rpc_complete_task() local 328 spin_lock_irqsave(&wq->lock, flags); in rpc_complete_task() 331 if (waitqueue_active(wq)) in rpc_complete_task() 332 __wake_up_locked_key(wq, TASK_NORMAL, &k); in rpc_complete_task() 333 spin_unlock_irqrestore(&wq->lock, flags); in rpc_complete_task() 364 static void rpc_make_runnable(struct workqueue_struct *wq, in rpc_make_runnable() argument 374 queue_work(wq, &task->u.tk_work); in rpc_make_runnable() 502 static void __rpc_do_wake_up_task_on_wq(struct workqueue_struct *wq, in __rpc_do_wake_up_task_on_wq() argument 519 rpc_make_runnable(wq, task); in __rpc_do_wake_up_task_on_wq() 528 rpc_wake_up_task_on_wq_queue_action_locked(struct workqueue_struct *wq, in rpc_wake_up_task_on_wq_queue_action_locked() argument [all …]
|
D | xprtsock.c | 903 struct socket_wq *wq; in xs_nospace() local 906 wq = rcu_dereference(sk->sk_wq); in xs_nospace() 907 set_bit(SOCKWQ_ASYNC_NOSPACE, &wq->flags); in xs_nospace() 1577 struct socket_wq *wq; in xs_write_space() local 1589 wq = rcu_dereference(sk->sk_wq); in xs_write_space() 1590 if (!wq || test_and_clear_bit(SOCKWQ_ASYNC_NOSPACE, &wq->flags) == 0) in xs_write_space()
|
/net/core/ |
D | stream.c | 33 struct socket_wq *wq; in sk_stream_write_space() local 39 wq = rcu_dereference(sk->sk_wq); in sk_stream_write_space() 40 if (skwq_has_sleeper(wq)) in sk_stream_write_space() 41 wake_up_interruptible_poll(&wq->wait, EPOLLOUT | in sk_stream_write_space() 43 if (wq && wq->fasync_list && !(sk->sk_shutdown & SEND_SHUTDOWN)) in sk_stream_write_space() 44 sock_wake_async(wq, SOCK_WAKE_SPACE, POLL_OUT); in sk_stream_write_space()
|
D | sock.c | 2412 prepare_to_wait_exclusive(&sk->sk_lock.wq, &wait, in __lock_sock() 2420 finish_wait(&sk->sk_lock.wq, &wait); in __lock_sock() 2770 struct socket_wq *wq; in sock_def_wakeup() local 2773 wq = rcu_dereference(sk->sk_wq); in sock_def_wakeup() 2774 if (skwq_has_sleeper(wq)) in sock_def_wakeup() 2775 wake_up_interruptible_all(&wq->wait); in sock_def_wakeup() 2781 struct socket_wq *wq; in sock_def_error_report() local 2784 wq = rcu_dereference(sk->sk_wq); in sock_def_error_report() 2785 if (skwq_has_sleeper(wq)) in sock_def_error_report() 2786 wake_up_interruptible_poll(&wq->wait, EPOLLERR); in sock_def_error_report() [all …]
|
D | page_pool.c | 383 static void page_pool_release_retry(struct work_struct *wq) in page_pool_release_retry() argument 385 struct delayed_work *dwq = to_delayed_work(wq); in page_pool_release_retry()
|
/net/smc/ |
D | smc_tx.c | 44 struct socket_wq *wq; in smc_tx_write_space() local 50 wq = rcu_dereference(sk->sk_wq); in smc_tx_write_space() 51 if (skwq_has_sleeper(wq)) in smc_tx_write_space() 52 wake_up_interruptible_poll(&wq->wait, in smc_tx_write_space() 55 if (wq && wq->fasync_list && !(sk->sk_shutdown & SEND_SHUTDOWN)) in smc_tx_write_space() 56 sock_wake_async(wq, SOCK_WAKE_SPACE, POLL_OUT); in smc_tx_write_space()
|
D | smc_rx.c | 30 struct socket_wq *wq; in smc_rx_wake_up() local 35 wq = rcu_dereference(sk->sk_wq); in smc_rx_wake_up() 36 if (skwq_has_sleeper(wq)) in smc_rx_wake_up() 37 wake_up_interruptible_sync_poll(&wq->wait, EPOLLIN | EPOLLPRI | in smc_rx_wake_up()
|
/net/atm/ |
D | common.c | 91 struct socket_wq *wq; in vcc_def_wakeup() local 94 wq = rcu_dereference(sk->sk_wq); in vcc_def_wakeup() 95 if (skwq_has_sleeper(wq)) in vcc_def_wakeup() 96 wake_up(&wq->wait); in vcc_def_wakeup() 110 struct socket_wq *wq; in vcc_write_space() local 115 wq = rcu_dereference(sk->sk_wq); in vcc_write_space() 116 if (skwq_has_sleeper(wq)) in vcc_write_space() 117 wake_up_interruptible(&wq->wait); in vcc_write_space()
|
/net/9p/ |
D | trans_fd.c | 130 struct work_struct wq; member 443 m = container_of(work, struct p9_conn, wq); in p9_write_work() 507 schedule_work(&m->wq); in p9_write_work() 589 INIT_WORK(&m->wq, p9_write_work); in p9_conn_create() 640 schedule_work(&m->wq); in p9_poll_mux() 678 schedule_work(&m->wq); in p9_fd_request() 868 cancel_work_sync(&m->wq); in p9_conn_destroy()
|
D | trans_xen.c | 69 wait_queue_head_t wq; member 161 while (wait_event_killable(ring->wq, in p9_xen_request() 262 wake_up_interruptible(&ring->wq); in xen_9pfs_front_event_handler() 332 init_waitqueue_head(&ring->wq); in xen_9pfs_front_alloc_dataring()
|
D | client.c | 286 init_waitqueue_head(&req->wq); in p9_tag_alloc() 430 wake_up(&req->wq); in p9_client_cb() 757 err = wait_event_killable(req->wq, req->status >= REQ_STATUS_RCVD); in p9_client_rpc()
|
D | trans_virtio.c | 500 err = wait_event_killable(req->wq, req->status >= REQ_STATUS_RCVD); in p9_virtio_zc_request()
|
/net/dccp/ |
D | output.c | 197 struct socket_wq *wq; in dccp_write_space() local 200 wq = rcu_dereference(sk->sk_wq); in dccp_write_space() 201 if (skwq_has_sleeper(wq)) in dccp_write_space() 202 wake_up_interruptible(&wq->wait); in dccp_write_space()
|
/net/ |
D | socket.c | 242 init_waitqueue_head(&ei->socket.wq.wait); in sock_alloc_inode() 243 ei->socket.wq.fasync_list = NULL; in sock_alloc_inode() 244 ei->socket.wq.flags = 0; in sock_alloc_inode() 599 if (sock->wq.fasync_list) in __sock_release() 1288 struct socket_wq *wq = &sock->wq; in sock_fasync() local 1294 fasync_helper(fd, filp, on, &wq->fasync_list); in sock_fasync() 1296 if (!wq->fasync_list) in sock_fasync() 1307 int sock_wake_async(struct socket_wq *wq, int how, int band) in sock_wake_async() argument 1309 if (!wq || !wq->fasync_list) in sock_wake_async() 1314 if (test_bit(SOCKWQ_ASYNC_WAITDATA, &wq->flags)) in sock_wake_async() [all …]
|
/net/nfc/hci/ |
D | command.c | 52 wake_up(hcp_ew->wq); in nfc_hci_execute_cb() 61 hcp_ew.wq = &ew_wq; in nfc_hci_execute_cmd()
|
D | hci.h | 27 wait_queue_head_t *wq; member
|
/net/rxrpc/ |
D | af_rxrpc.c | 64 struct socket_wq *wq = rcu_dereference(sk->sk_wq); in rxrpc_write_space() local 66 if (skwq_has_sleeper(wq)) in rxrpc_write_space() 67 wake_up_interruptible(&wq->wait); in rxrpc_write_space()
|
/net/tipc/ |
D | socket.c | 1947 struct socket_wq *wq; in tipc_write_space() local 1950 wq = rcu_dereference(sk->sk_wq); in tipc_write_space() 1951 if (skwq_has_sleeper(wq)) in tipc_write_space() 1952 wake_up_interruptible_sync_poll(&wq->wait, EPOLLOUT | in tipc_write_space() 1964 struct socket_wq *wq; in tipc_data_ready() local 1967 wq = rcu_dereference(sk->sk_wq); in tipc_data_ready() 1968 if (skwq_has_sleeper(wq)) in tipc_data_ready() 1969 wake_up_interruptible_sync_poll(&wq->wait, EPOLLIN | in tipc_data_ready()
|
/net/iucv/ |
D | af_iucv.c | 304 struct socket_wq *wq; in iucv_sock_wake_msglim() local 307 wq = rcu_dereference(sk->sk_wq); in iucv_sock_wake_msglim() 308 if (skwq_has_sleeper(wq)) in iucv_sock_wake_msglim() 309 wake_up_interruptible_all(&wq->wait); in iucv_sock_wake_msglim()
|
/net/unix/ |
D | af_unix.c | 454 struct socket_wq *wq; in unix_write_space() local 458 wq = rcu_dereference(sk->sk_wq); in unix_write_space() 459 if (skwq_has_sleeper(wq)) in unix_write_space() 460 wake_up_interruptible_sync_poll(&wq->wait, in unix_write_space()
|
/net/sctp/ |
D | socket.c | 8908 struct socket_wq *wq; in __sctp_write_space() local 8911 wq = rcu_dereference(sk->sk_wq); in __sctp_write_space() 8912 if (wq) { in __sctp_write_space() 8913 if (waitqueue_active(&wq->wait)) in __sctp_write_space() 8914 wake_up_interruptible(&wq->wait); in __sctp_write_space() 8921 sock_wake_async(wq, SOCK_WAKE_SPACE, POLL_OUT); in __sctp_write_space() 9100 struct socket_wq *wq; in sctp_data_ready() local 9103 wq = rcu_dereference(sk->sk_wq); in sctp_data_ready() 9104 if (skwq_has_sleeper(wq)) in sctp_data_ready() 9105 wake_up_interruptible_sync_poll(&wq->wait, EPOLLIN | in sctp_data_ready()
|