/net/tipc/ |
D | discover.c | 187 static void disc_update(struct tipc_link_req *req) in disc_update() argument 189 if (!req->num_nodes) { in disc_update() 190 if ((req->timer_intv == TIPC_LINK_REQ_INACTIVE) || in disc_update() 191 (req->timer_intv > TIPC_LINK_REQ_FAST)) { in disc_update() 192 req->timer_intv = TIPC_LINK_REQ_INIT; in disc_update() 193 mod_timer(&req->timer, jiffies + req->timer_intv); in disc_update() 202 void tipc_disc_add_dest(struct tipc_link_req *req) in tipc_disc_add_dest() argument 204 spin_lock_bh(&req->lock); in tipc_disc_add_dest() 205 req->num_nodes++; in tipc_disc_add_dest() 206 spin_unlock_bh(&req->lock); in tipc_disc_add_dest() [all …]
|
/net/sunrpc/ |
D | backchannel_rqst.c | 59 static void xprt_free_allocation(struct rpc_rqst *req) in xprt_free_allocation() argument 63 dprintk("RPC: free allocations for req= %p\n", req); in xprt_free_allocation() 64 WARN_ON_ONCE(test_bit(RPC_BC_PA_IN_USE, &req->rq_bc_pa_state)); in xprt_free_allocation() 65 xbufp = &req->rq_rcv_buf; in xprt_free_allocation() 67 xbufp = &req->rq_snd_buf; in xprt_free_allocation() 69 kfree(req); in xprt_free_allocation() 86 struct rpc_rqst *req; in xprt_alloc_bc_req() local 89 req = kzalloc(sizeof(*req), gfp_flags); in xprt_alloc_bc_req() 90 if (req == NULL) in xprt_alloc_bc_req() 93 req->rq_xprt = xprt; in xprt_alloc_bc_req() [all …]
|
D | xprt.c | 185 struct rpc_rqst *req = task->tk_rqstp; in xprt_reserve_xprt() local 194 if (req != NULL) in xprt_reserve_xprt() 195 req->rq_ntrans++; in xprt_reserve_xprt() 204 if (req == NULL) in xprt_reserve_xprt() 206 else if (!req->rq_ntrans) in xprt_reserve_xprt() 236 struct rpc_rqst *req = task->tk_rqstp; in xprt_reserve_xprt_cong() local 244 if (req == NULL) { in xprt_reserve_xprt_cong() 250 req->rq_ntrans++; in xprt_reserve_xprt_cong() 255 if (req) in xprt_reserve_xprt_cong() 256 __xprt_put_cong(xprt, req); in xprt_reserve_xprt_cong() [all …]
|
/net/9p/ |
D | client.c | 270 struct p9_req_t *req; in p9_tag_alloc() local 301 req = &c->reqs[row][col]; in p9_tag_alloc() 302 if (!req->wq) { in p9_tag_alloc() 303 req->wq = kmalloc(sizeof(wait_queue_head_t), GFP_NOFS); in p9_tag_alloc() 304 if (!req->wq) in p9_tag_alloc() 306 init_waitqueue_head(req->wq); in p9_tag_alloc() 309 if (!req->tc) in p9_tag_alloc() 310 req->tc = p9_fcall_alloc(alloc_msize); in p9_tag_alloc() 311 if (!req->rc) in p9_tag_alloc() 312 req->rc = p9_fcall_alloc(alloc_msize); in p9_tag_alloc() [all …]
|
D | trans_fd.c | 134 struct p9_req_t *req; member 201 struct p9_req_t *req, *rtmp; in p9_conn_cancel() local 215 list_for_each_entry_safe(req, rtmp, &m->req_list, req_list) { in p9_conn_cancel() 216 list_move(&req->req_list, &cancel_list); in p9_conn_cancel() 218 list_for_each_entry_safe(req, rtmp, &m->unsent_req_list, req_list) { in p9_conn_cancel() 219 list_move(&req->req_list, &cancel_list); in p9_conn_cancel() 222 list_for_each_entry_safe(req, rtmp, &cancel_list, req_list) { in p9_conn_cancel() 223 p9_debug(P9_DEBUG_ERROR, "call back req %p\n", req); in p9_conn_cancel() 224 list_del(&req->req_list); in p9_conn_cancel() 225 if (!req->t_err) in p9_conn_cancel() [all …]
|
/net/ceph/ |
D | mon_client.c | 498 struct ceph_mon_generic_request *req = in DEFINE_RB_FUNCS() local 501 dout("%s greq %p request %p reply %p\n", __func__, req, req->request, in DEFINE_RB_FUNCS() 502 req->reply); in DEFINE_RB_FUNCS() 503 WARN_ON(!RB_EMPTY_NODE(&req->node)); in DEFINE_RB_FUNCS() 505 if (req->reply) in DEFINE_RB_FUNCS() 506 ceph_msg_put(req->reply); in DEFINE_RB_FUNCS() 507 if (req->request) in DEFINE_RB_FUNCS() 508 ceph_msg_put(req->request); in DEFINE_RB_FUNCS() 510 kfree(req); in DEFINE_RB_FUNCS() 513 static void put_generic_request(struct ceph_mon_generic_request *req) in put_generic_request() argument [all …]
|
D | osd_client.c | 48 static void link_request(struct ceph_osd *osd, struct ceph_osd_request *req); 49 static void unlink_request(struct ceph_osd *osd, struct ceph_osd_request *req); 408 static void request_release_checks(struct ceph_osd_request *req) in request_release_checks() argument 410 WARN_ON(!RB_EMPTY_NODE(&req->r_node)); in request_release_checks() 411 WARN_ON(!RB_EMPTY_NODE(&req->r_mc_node)); in request_release_checks() 412 WARN_ON(!list_empty(&req->r_unsafe_item)); in request_release_checks() 413 WARN_ON(req->r_osd); in request_release_checks() 418 struct ceph_osd_request *req = container_of(kref, in ceph_osdc_release_request() local 422 dout("%s %p (r_request %p r_reply %p)\n", __func__, req, in ceph_osdc_release_request() 423 req->r_request, req->r_reply); in ceph_osdc_release_request() [all …]
|
/net/bluetooth/ |
D | hci_request.c | 37 void hci_req_init(struct hci_request *req, struct hci_dev *hdev) in hci_req_init() argument 39 skb_queue_head_init(&req->cmd_q); in hci_req_init() 40 req->hdev = hdev; in hci_req_init() 41 req->err = 0; in hci_req_init() 44 static int req_run(struct hci_request *req, hci_req_complete_t complete, in req_run() argument 47 struct hci_dev *hdev = req->hdev; in req_run() 51 BT_DBG("length %u", skb_queue_len(&req->cmd_q)); in req_run() 56 if (req->err) { in req_run() 57 skb_queue_purge(&req->cmd_q); in req_run() 58 return req->err; in req_run() [all …]
|
D | hci_request.h | 38 void hci_req_init(struct hci_request *req, struct hci_dev *hdev); 39 int hci_req_run(struct hci_request *req, hci_req_complete_t complete); 40 int hci_req_run_skb(struct hci_request *req, hci_req_complete_skb_t complete); 41 void hci_req_add(struct hci_request *req, u16 opcode, u32 plen, 43 void hci_req_add_ev(struct hci_request *req, u16 opcode, u32 plen, 49 int hci_req_sync(struct hci_dev *hdev, int (*req)(struct hci_request *req, 52 int __hci_req_sync(struct hci_dev *hdev, int (*func)(struct hci_request *req, 62 void __hci_req_write_fast_connectable(struct hci_request *req, bool enable); 63 void __hci_req_update_name(struct hci_request *req); 64 void __hci_req_update_eir(struct hci_request *req); [all …]
|
D | a2mp.c | 140 struct a2mp_discov_req *req = (void *) skb->data; in a2mp_discover_req() local 147 if (len < sizeof(*req)) in a2mp_discover_req() 150 skb_pull(skb, sizeof(*req)); in a2mp_discover_req() 152 ext_feat = le16_to_cpu(req->ext_feat); in a2mp_discover_req() 154 BT_DBG("mtu %d efm 0x%4.4x", le16_to_cpu(req->mtu), ext_feat); in a2mp_discover_req() 233 struct a2mp_info_req req; in a2mp_discover_rsp() local 236 req.id = cl->id; in a2mp_discover_rsp() 238 sizeof(req), &req); in a2mp_discover_rsp() 301 struct a2mp_info_req *req = (void *) skb->data; in a2mp_getinfo_req() local 306 if (le16_to_cpu(hdr->len) < sizeof(*req)) in a2mp_getinfo_req() [all …]
|
D | ecdh_helper.c | 34 static void ecdh_complete(struct crypto_async_request *req, int err) in ecdh_complete() argument 36 struct ecdh_completion *res = req->data; in ecdh_complete() 57 struct kpp_request *req; in compute_ecdh_secret() local 76 req = kpp_request_alloc(tfm, GFP_KERNEL); in compute_ecdh_secret() 77 if (!req) in compute_ecdh_secret() 107 kpp_request_set_input(req, &src, 64); in compute_ecdh_secret() 108 kpp_request_set_output(req, &dst, 32); in compute_ecdh_secret() 109 kpp_request_set_callback(req, CRYPTO_TFM_REQ_MAY_BACKLOG, in compute_ecdh_secret() 111 err = crypto_kpp_compute_shared_secret(req); in compute_ecdh_secret() 128 kpp_request_free(req); in compute_ecdh_secret() [all …]
|
/net/ipv4/ |
D | udp_diag.c | 22 const struct inet_diag_req_v2 *req, in sk_diag_dump() argument 28 return inet_sk_diag_fill(sk, NULL, skb, req, in sk_diag_dump() 36 const struct inet_diag_req_v2 *req) in udp_dump_one() argument 44 if (req->sdiag_family == AF_INET) in udp_dump_one() 46 req->id.idiag_src[0], req->id.idiag_sport, in udp_dump_one() 47 req->id.idiag_dst[0], req->id.idiag_dport, in udp_dump_one() 48 req->id.idiag_if, 0, tbl, NULL); in udp_dump_one() 50 else if (req->sdiag_family == AF_INET6) in udp_dump_one() 52 (struct in6_addr *)req->id.idiag_src, in udp_dump_one() 53 req->id.idiag_sport, in udp_dump_one() [all …]
|
D | inet_connection_sock.c | 434 struct request_sock *req; in inet_csk_accept() local 460 req = reqsk_queue_remove(queue, sk); in inet_csk_accept() 461 newsk = req->sk; in inet_csk_accept() 464 tcp_rsk(req)->tfo_listener) { in inet_csk_accept() 466 if (tcp_rsk(req)->tfo_listener) { in inet_csk_accept() 473 req->sk = NULL; in inet_csk_accept() 474 req = NULL; in inet_csk_accept() 500 if (req) in inet_csk_accept() 501 reqsk_put(req); in inet_csk_accept() 505 req = NULL; in inet_csk_accept() [all …]
|
D | tcp_minisocks.c | 352 void tcp_openreq_init_rwin(struct request_sock *req, in tcp_openreq_init_rwin() argument 356 struct inet_request_sock *ireq = inet_rsk(req); in tcp_openreq_init_rwin() 367 req->rsk_window_clamp = window_clamp ? : dst_metric(dst, RTAX_WINDOW); in tcp_openreq_init_rwin() 371 (req->rsk_window_clamp > full_space || req->rsk_window_clamp == 0)) in tcp_openreq_init_rwin() 372 req->rsk_window_clamp = full_space; in tcp_openreq_init_rwin() 374 rcv_wnd = tcp_rwnd_init_bpf((struct sock *)req); in tcp_openreq_init_rwin() 383 &req->rsk_rcv_wnd, in tcp_openreq_init_rwin() 384 &req->rsk_window_clamp, in tcp_openreq_init_rwin() 393 const struct request_sock *req) in tcp_ecn_openreq_child() argument 395 tp->ecn_flags = inet_rsk(req)->ecn_ok ? TCP_ECN_OK : 0; in tcp_ecn_openreq_child() [all …]
|
D | syncookies.c | 69 u64 cookie_init_timestamp(struct request_sock *req) in cookie_init_timestamp() argument 75 ireq = inet_rsk(req); in cookie_init_timestamp() 206 struct request_sock *req, in tcp_get_cookie_sock() argument 213 child = icsk->icsk_af_ops->syn_recv_sock(sk, skb, req, dst, in tcp_get_cookie_sock() 216 refcount_set(&req->rsk_refcnt, 1); in tcp_get_cookie_sock() 219 if (!inet_csk_reqsk_queue_add(sk, req, child)) { in tcp_get_cookie_sock() 223 reqsk_put(req); in tcp_get_cookie_sock() 226 reqsk_free(req); in tcp_get_cookie_sock() 298 struct request_sock *req; in cookie_v4_check() local 334 req = inet_reqsk_alloc(&tcp_request_sock_ops, sk, false); /* for safety */ in cookie_v4_check() [all …]
|
/net/sunrpc/auth_gss/ |
D | gss_krb5_crypto.c | 65 SKCIPHER_REQUEST_ON_STACK(req, tfm); in krb5_encrypt() 82 skcipher_request_set_tfm(req, tfm); in krb5_encrypt() 83 skcipher_request_set_callback(req, 0, NULL, NULL); in krb5_encrypt() 84 skcipher_request_set_crypt(req, sg, sg, length, local_iv); in krb5_encrypt() 86 ret = crypto_skcipher_encrypt(req); in krb5_encrypt() 87 skcipher_request_zero(req); in krb5_encrypt() 104 SKCIPHER_REQUEST_ON_STACK(req, tfm); in krb5_decrypt() 120 skcipher_request_set_tfm(req, tfm); in krb5_decrypt() 121 skcipher_request_set_callback(req, 0, NULL, NULL); in krb5_decrypt() 122 skcipher_request_set_crypt(req, sg, sg, length, local_iv); in krb5_decrypt() [all …]
|
/net/sunrpc/xprtrdma/ |
D | backchannel.c | 25 struct rpcrdma_req *req = rpcr_to_rdmar(rqst); in rpcrdma_bc_free_rqst() local 28 list_del(&req->rl_all); in rpcrdma_bc_free_rqst() 31 rpcrdma_destroy_req(req); in rpcrdma_bc_free_rqst() 40 struct rpcrdma_req *req; in rpcrdma_bc_setup_rqst() local 43 req = rpcrdma_create_req(r_xprt); in rpcrdma_bc_setup_rqst() 44 if (IS_ERR(req)) in rpcrdma_bc_setup_rqst() 45 return PTR_ERR(req); in rpcrdma_bc_setup_rqst() 46 req->rl_backchannel = true; in rpcrdma_bc_setup_rqst() 52 req->rl_rdmabuf = rb; in rpcrdma_bc_setup_rqst() 53 xdr_buf_init(&req->rl_hdrbuf, rb->rg_base, rdmab_length(rb)); in rpcrdma_bc_setup_rqst() [all …]
|
D | transport.c | 551 rpcrdma_get_rdmabuf(struct rpcrdma_xprt *r_xprt, struct rpcrdma_req *req, in rpcrdma_get_rdmabuf() argument 557 if (req->rl_rdmabuf) in rpcrdma_get_rdmabuf() 565 req->rl_rdmabuf = rb; in rpcrdma_get_rdmabuf() 566 xdr_buf_init(&req->rl_hdrbuf, rb->rg_base, rdmab_length(rb)); in rpcrdma_get_rdmabuf() 571 rpcrdma_get_sendbuf(struct rpcrdma_xprt *r_xprt, struct rpcrdma_req *req, in rpcrdma_get_sendbuf() argument 576 if (req->rl_sendbuf && rdmab_length(req->rl_sendbuf) >= size) in rpcrdma_get_sendbuf() 583 rpcrdma_free_regbuf(req->rl_sendbuf); in rpcrdma_get_sendbuf() 585 req->rl_sendbuf = rb; in rpcrdma_get_sendbuf() 601 rpcrdma_get_recvbuf(struct rpcrdma_xprt *r_xprt, struct rpcrdma_req *req, in rpcrdma_get_recvbuf() argument 606 if (req->rl_recvbuf && rdmab_length(req->rl_recvbuf) >= size) in rpcrdma_get_recvbuf() [all …]
|
D | verbs.c | 448 struct rpcrdma_req *req; in rpcrdma_ia_remove() local 475 list_for_each_entry(req, &buf->rb_allreqs, rl_all) { in rpcrdma_ia_remove() 476 rpcrdma_dma_unmap_regbuf(req->rl_rdmabuf); in rpcrdma_ia_remove() 477 rpcrdma_dma_unmap_regbuf(req->rl_sendbuf); in rpcrdma_ia_remove() 478 rpcrdma_dma_unmap_regbuf(req->rl_recvbuf); in rpcrdma_ia_remove() 940 struct rpcrdma_req *req; in rpcrdma_create_req() local 942 req = kzalloc(sizeof(*req), GFP_KERNEL); in rpcrdma_create_req() 943 if (req == NULL) in rpcrdma_create_req() 947 list_add(&req->rl_all, &buffer->rb_allreqs); in rpcrdma_create_req() 949 req->rl_cqe.done = rpcrdma_wc_send; in rpcrdma_create_req() [all …]
|
D | rpc_rdma.c | 345 rpcrdma_encode_read_list(struct rpcrdma_xprt *r_xprt, struct rpcrdma_req *req, in rpcrdma_encode_read_list() argument 348 struct xdr_stream *xdr = &req->rl_stream; in rpcrdma_encode_read_list() 357 seg = req->rl_segments; in rpcrdma_encode_read_list() 368 rpcrdma_push_mw(mw, &req->rl_registered); in rpcrdma_encode_read_list() 401 rpcrdma_encode_write_list(struct rpcrdma_xprt *r_xprt, struct rpcrdma_req *req, in rpcrdma_encode_write_list() argument 404 struct xdr_stream *xdr = &req->rl_stream; in rpcrdma_encode_write_list() 410 seg = req->rl_segments; in rpcrdma_encode_write_list() 430 rpcrdma_push_mw(mw, &req->rl_registered); in rpcrdma_encode_write_list() 465 rpcrdma_encode_reply_chunk(struct rpcrdma_xprt *r_xprt, struct rpcrdma_req *req, in rpcrdma_encode_reply_chunk() argument 468 struct xdr_stream *xdr = &req->rl_stream; in rpcrdma_encode_reply_chunk() [all …]
|
/net/mac80211/ |
D | scan.c | 271 struct cfg80211_scan_request *req; in ieee80211_prep_hw_scan() local 276 req = rcu_dereference_protected(local->scan_req, in ieee80211_prep_hw_scan() 283 for (i = 0; i < req->n_channels; i++) { in ieee80211_prep_hw_scan() 284 local->hw_scan_req->req.channels[i] = req->channels[i]; in ieee80211_prep_hw_scan() 285 bands_used |= BIT(req->channels[i]->band); in ieee80211_prep_hw_scan() 288 n_chans = req->n_channels; in ieee80211_prep_hw_scan() 296 for (i = 0; i < req->n_channels; i++) { in ieee80211_prep_hw_scan() 297 if (req->channels[i]->band != in ieee80211_prep_hw_scan() 300 local->hw_scan_req->req.channels[n_chans] = in ieee80211_prep_hw_scan() 301 req->channels[i]; in ieee80211_prep_hw_scan() [all …]
|
/net/unix/ |
D | diag.c | 75 struct sock *req, *peer; in sk_diag_dump_icons() local 77 req = skb->sk; in sk_diag_dump_icons() 83 unix_state_lock_nested(req); in sk_diag_dump_icons() 84 peer = unix_sk(req)->peer; in sk_diag_dump_icons() 86 unix_state_unlock(req); in sk_diag_dump_icons() 113 static int sk_diag_fill(struct sock *sk, struct sk_buff *skb, struct unix_diag_req *req, in sk_diag_fill() argument 132 if ((req->udiag_show & UDIAG_SHOW_NAME) && in sk_diag_fill() 136 if ((req->udiag_show & UDIAG_SHOW_VFS) && in sk_diag_fill() 140 if ((req->udiag_show & UDIAG_SHOW_PEER) && in sk_diag_fill() 144 if ((req->udiag_show & UDIAG_SHOW_ICONS) && in sk_diag_fill() [all …]
|
/net/rxrpc/ |
D | rxkad.c | 107 SKCIPHER_REQUEST_ON_STACK(req, conn->cipher); in rxkad_prime_packet_security() 131 skcipher_request_set_tfm(req, conn->cipher); in rxkad_prime_packet_security() 132 skcipher_request_set_callback(req, 0, NULL, NULL); in rxkad_prime_packet_security() 133 skcipher_request_set_crypt(req, &sg, &sg, tmpsize, iv.x); in rxkad_prime_packet_security() 134 crypto_skcipher_encrypt(req); in rxkad_prime_packet_security() 135 skcipher_request_zero(req); in rxkad_prime_packet_security() 152 SKCIPHER_REQUEST_ON_STACK(req, call->conn->cipher); in rxkad_secure_packet_auth() 170 skcipher_request_set_tfm(req, call->conn->cipher); in rxkad_secure_packet_auth() 171 skcipher_request_set_callback(req, 0, NULL, NULL); in rxkad_secure_packet_auth() 172 skcipher_request_set_crypt(req, &sg, &sg, 8, iv.x); in rxkad_secure_packet_auth() [all …]
|
/net/core/ |
D | request_sock.c | 95 void reqsk_fastopen_remove(struct sock *sk, struct request_sock *req, in reqsk_fastopen_remove() argument 98 struct sock *lsk = req->rsk_listener; in reqsk_fastopen_remove() 106 tcp_rsk(req)->tfo_listener = false; in reqsk_fastopen_remove() 107 if (req->sk) /* the child socket hasn't been accepted yet */ in reqsk_fastopen_remove() 115 reqsk_put(req); in reqsk_fastopen_remove() 125 req->rsk_timer.expires = jiffies + 60*HZ; in reqsk_fastopen_remove() 127 fastopenq->rskq_rst_head = req; in reqsk_fastopen_remove() 129 fastopenq->rskq_rst_tail->dl_next = req; in reqsk_fastopen_remove() 131 req->dl_next = NULL; in reqsk_fastopen_remove() 132 fastopenq->rskq_rst_tail = req; in reqsk_fastopen_remove()
|
/net/dccp/ |
D | minisocks.c | 82 const struct request_sock *req, in dccp_create_openreq_child() argument 91 struct sock *newsk = inet_csk_clone_lock(sk, req, GFP_ATOMIC); in dccp_create_openreq_child() 94 struct dccp_request_sock *dreq = dccp_rsk(req); in dccp_create_openreq_child() 145 struct request_sock *req) in dccp_check_req() argument 148 struct dccp_request_sock *dreq = dccp_rsk(req); in dccp_check_req() 169 inet_rtx_syn_ack(sk, req); in dccp_check_req() 196 child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL, in dccp_check_req() 197 req, &own_req); in dccp_check_req() 199 child = inet_csk_complete_hashdance(sk, child, req, own_req); in dccp_check_req() 206 req->rsk_ops->send_reset(sk, skb); in dccp_check_req() [all …]
|