/net/sunrpc/ |
D | backchannel_rqst.c | 54 static void xprt_free_allocation(struct rpc_rqst *req) in xprt_free_allocation() argument 58 dprintk("RPC: free allocations for req= %p\n", req); in xprt_free_allocation() 59 WARN_ON_ONCE(test_bit(RPC_BC_PA_IN_USE, &req->rq_bc_pa_state)); in xprt_free_allocation() 60 xbufp = &req->rq_rcv_buf; in xprt_free_allocation() 62 xbufp = &req->rq_snd_buf; in xprt_free_allocation() 64 kfree(req); in xprt_free_allocation() 92 struct rpc_rqst *req; in xprt_alloc_bc_req() local 95 req = kzalloc(sizeof(*req), gfp_flags); in xprt_alloc_bc_req() 96 if (req == NULL) in xprt_alloc_bc_req() 99 req->rq_xprt = xprt; in xprt_alloc_bc_req() [all …]
|
D | xprt.c | 77 static unsigned long xprt_request_timeout(const struct rpc_rqst *req) in xprt_request_timeout() argument 79 unsigned long timeout = jiffies + req->rq_timeout; in xprt_request_timeout() 81 if (time_before(timeout, req->rq_majortimeo)) in xprt_request_timeout() 83 return req->rq_majortimeo; in xprt_request_timeout() 237 struct rpc_rqst *req = task->tk_rqstp; in xprt_reserve_xprt() local 258 xprt_request_timeout(req)); in xprt_reserve_xprt() 301 struct rpc_rqst *req = task->tk_rqstp; in xprt_reserve_xprt_cong() local 308 if (req == NULL) { in xprt_reserve_xprt_cong() 325 xprt_request_timeout(req)); in xprt_reserve_xprt_cong() 427 __xprt_get_cong(struct rpc_xprt *xprt, struct rpc_rqst *req) in __xprt_get_cong() argument [all …]
|
/net/9p/ |
D | client.c | 270 struct p9_req_t *req = kmem_cache_alloc(p9_req_cache, GFP_NOFS); in p9_tag_alloc() local 274 if (!req) in p9_tag_alloc() 277 if (p9_fcall_init(c, &req->tc, alloc_msize)) in p9_tag_alloc() 279 if (p9_fcall_init(c, &req->rc, alloc_msize)) in p9_tag_alloc() 282 p9pdu_reset(&req->tc); in p9_tag_alloc() 283 p9pdu_reset(&req->rc); in p9_tag_alloc() 284 req->t_err = 0; in p9_tag_alloc() 285 req->status = REQ_STATUS_ALLOC; in p9_tag_alloc() 286 init_waitqueue_head(&req->wq); in p9_tag_alloc() 287 INIT_LIST_HEAD(&req->req_list); in p9_tag_alloc() [all …]
|
D | trans_virtio.c | 129 struct p9_req_t *req; in req_done() local 136 while ((req = virtqueue_get_buf(chan->vq, &len)) != NULL) { in req_done() 143 req->rc.size = len; in req_done() 144 p9_client_cb(chan->client, req, REQ_STATUS_RCVD); in req_done() 190 static int p9_virtio_cancel(struct p9_client *client, struct p9_req_t *req) in p9_virtio_cancel() argument 196 static int p9_virtio_cancelled(struct p9_client *client, struct p9_req_t *req) in p9_virtio_cancelled() argument 198 p9_req_put(req); in p9_virtio_cancelled() 251 p9_virtio_request(struct p9_client *client, struct p9_req_t *req) in p9_virtio_request() argument 261 req->status = REQ_STATUS_SENT; in p9_virtio_request() 268 VIRTQUEUE_NUM, req->tc.sdata, req->tc.size); in p9_virtio_request() [all …]
|
/net/ceph/ |
D | mon_client.c | 510 struct ceph_mon_generic_request *req = in DEFINE_RB_FUNCS() local 513 dout("%s greq %p request %p reply %p\n", __func__, req, req->request, in DEFINE_RB_FUNCS() 514 req->reply); in DEFINE_RB_FUNCS() 515 WARN_ON(!RB_EMPTY_NODE(&req->node)); in DEFINE_RB_FUNCS() 517 if (req->reply) in DEFINE_RB_FUNCS() 518 ceph_msg_put(req->reply); in DEFINE_RB_FUNCS() 519 if (req->request) in DEFINE_RB_FUNCS() 520 ceph_msg_put(req->request); in DEFINE_RB_FUNCS() 522 kfree(req); in DEFINE_RB_FUNCS() 525 static void put_generic_request(struct ceph_mon_generic_request *req) in put_generic_request() argument [all …]
|
D | osd_client.c | 49 static void link_request(struct ceph_osd *osd, struct ceph_osd_request *req); 50 static void unlink_request(struct ceph_osd *osd, struct ceph_osd_request *req); 470 static void request_release_checks(struct ceph_osd_request *req) in request_release_checks() argument 472 WARN_ON(!RB_EMPTY_NODE(&req->r_node)); in request_release_checks() 473 WARN_ON(!RB_EMPTY_NODE(&req->r_mc_node)); in request_release_checks() 474 WARN_ON(!list_empty(&req->r_private_item)); in request_release_checks() 475 WARN_ON(req->r_osd); in request_release_checks() 480 struct ceph_osd_request *req = container_of(kref, in ceph_osdc_release_request() local 484 dout("%s %p (r_request %p r_reply %p)\n", __func__, req, in ceph_osdc_release_request() 485 req->r_request, req->r_reply); in ceph_osdc_release_request() [all …]
|
/net/bluetooth/ |
D | hci_request.c | 37 void hci_req_init(struct hci_request *req, struct hci_dev *hdev) in hci_req_init() argument 39 skb_queue_head_init(&req->cmd_q); in hci_req_init() 40 req->hdev = hdev; in hci_req_init() 41 req->err = 0; in hci_req_init() 44 void hci_req_purge(struct hci_request *req) in hci_req_purge() argument 46 skb_queue_purge(&req->cmd_q); in hci_req_purge() 54 static int req_run(struct hci_request *req, hci_req_complete_t complete, in req_run() argument 57 struct hci_dev *hdev = req->hdev; in req_run() 61 BT_DBG("length %u", skb_queue_len(&req->cmd_q)); in req_run() 66 if (req->err) { in req_run() [all …]
|
D | hci_request.h | 38 void hci_req_init(struct hci_request *req, struct hci_dev *hdev); 39 void hci_req_purge(struct hci_request *req); 41 int hci_req_run(struct hci_request *req, hci_req_complete_t complete); 42 int hci_req_run_skb(struct hci_request *req, hci_req_complete_skb_t complete); 43 void hci_req_add(struct hci_request *req, u16 opcode, u32 plen, 45 void hci_req_add_ev(struct hci_request *req, u16 opcode, u32 plen, 51 int hci_req_sync(struct hci_dev *hdev, int (*req)(struct hci_request *req, 54 int __hci_req_sync(struct hci_dev *hdev, int (*func)(struct hci_request *req, 64 void __hci_req_write_fast_connectable(struct hci_request *req, bool enable); 65 void __hci_req_update_name(struct hci_request *req); [all …]
|
D | a2mp.c | 133 struct a2mp_discov_req *req = (void *) skb->data; in a2mp_discover_req() local 140 if (len < sizeof(*req)) in a2mp_discover_req() 143 skb_pull(skb, sizeof(*req)); in a2mp_discover_req() 145 ext_feat = le16_to_cpu(req->ext_feat); in a2mp_discover_req() 147 BT_DBG("mtu %d efm 0x%4.4x", le16_to_cpu(req->mtu), ext_feat); in a2mp_discover_req() 226 struct a2mp_info_req req; in a2mp_discover_rsp() local 230 memset(&req, 0, sizeof(req)); in a2mp_discover_rsp() 232 req.id = cl->id; in a2mp_discover_rsp() 234 sizeof(req), &req); in a2mp_discover_rsp() 297 struct a2mp_info_req *req = (void *) skb->data; in a2mp_getinfo_req() local [all …]
|
D | ecdh_helper.c | 33 static void ecdh_complete(struct crypto_async_request *req, int err) in ecdh_complete() argument 35 struct ecdh_completion *res = req->data; in ecdh_complete() 63 struct kpp_request *req; in compute_ecdh_secret() local 73 req = kpp_request_alloc(tfm, GFP_KERNEL); in compute_ecdh_secret() 74 if (!req) { in compute_ecdh_secret() 86 kpp_request_set_input(req, &src, 64); in compute_ecdh_secret() 87 kpp_request_set_output(req, &dst, 32); in compute_ecdh_secret() 88 kpp_request_set_callback(req, CRYPTO_TFM_REQ_MAY_BACKLOG, in compute_ecdh_secret() 90 err = crypto_kpp_compute_shared_secret(req); in compute_ecdh_secret() 105 kpp_request_free(req); in compute_ecdh_secret() [all …]
|
/net/ipv4/ |
D | udp_diag.c | 18 const struct inet_diag_req_v2 *req, in sk_diag_dump() argument 24 return inet_sk_diag_fill(sk, NULL, skb, req, in sk_diag_dump() 32 const struct inet_diag_req_v2 *req) in udp_dump_one() argument 40 if (req->sdiag_family == AF_INET) in udp_dump_one() 43 req->id.idiag_src[0], req->id.idiag_sport, in udp_dump_one() 44 req->id.idiag_dst[0], req->id.idiag_dport, in udp_dump_one() 45 req->id.idiag_if, 0, tbl, NULL); in udp_dump_one() 47 else if (req->sdiag_family == AF_INET6) in udp_dump_one() 49 (struct in6_addr *)req->id.idiag_src, in udp_dump_one() 50 req->id.idiag_sport, in udp_dump_one() [all …]
|
D | inet_connection_sock.c | 455 struct request_sock *req; in inet_csk_accept() local 481 req = reqsk_queue_remove(queue, sk); in inet_csk_accept() 482 newsk = req->sk; in inet_csk_accept() 485 tcp_rsk(req)->tfo_listener) { in inet_csk_accept() 487 if (tcp_rsk(req)->tfo_listener) { in inet_csk_accept() 494 req->sk = NULL; in inet_csk_accept() 495 req = NULL; in inet_csk_accept() 521 if (req) in inet_csk_accept() 522 reqsk_put(req); in inet_csk_accept() 526 req = NULL; in inet_csk_accept() [all …]
|
D | tcp_minisocks.c | 360 void tcp_openreq_init_rwin(struct request_sock *req, in tcp_openreq_init_rwin() argument 364 struct inet_request_sock *ireq = inet_rsk(req); in tcp_openreq_init_rwin() 375 req->rsk_window_clamp = window_clamp ? : dst_metric(dst, RTAX_WINDOW); in tcp_openreq_init_rwin() 379 (req->rsk_window_clamp > full_space || req->rsk_window_clamp == 0)) in tcp_openreq_init_rwin() 380 req->rsk_window_clamp = full_space; in tcp_openreq_init_rwin() 382 rcv_wnd = tcp_rwnd_init_bpf((struct sock *)req); in tcp_openreq_init_rwin() 391 &req->rsk_rcv_wnd, in tcp_openreq_init_rwin() 392 &req->rsk_window_clamp, in tcp_openreq_init_rwin() 401 const struct request_sock *req) in tcp_ecn_openreq_child() argument 403 tp->ecn_flags = inet_rsk(req)->ecn_ok ? TCP_ECN_OK : 0; in tcp_ecn_openreq_child() [all …]
|
D | syncookies.c | 64 u64 cookie_init_timestamp(struct request_sock *req, u64 now) in cookie_init_timestamp() argument 66 const struct inet_request_sock *ireq = inet_rsk(req); in cookie_init_timestamp() 196 struct request_sock *req, in tcp_get_cookie_sock() argument 203 child = icsk->icsk_af_ops->syn_recv_sock(sk, skb, req, dst, in tcp_get_cookie_sock() 206 refcount_set(&req->rsk_refcnt, 1); in tcp_get_cookie_sock() 209 if (inet_csk_reqsk_queue_add(sk, req, child)) in tcp_get_cookie_sock() 215 __reqsk_free(req); in tcp_get_cookie_sock() 287 struct request_sock *req; in cookie_v4_check() local 324 req = inet_reqsk_alloc(&tcp_request_sock_ops, sk, false); /* for safety */ in cookie_v4_check() 325 if (!req) in cookie_v4_check() [all …]
|
/net/sunrpc/auth_gss/ |
D | gss_krb5_crypto.c | 65 SYNC_SKCIPHER_REQUEST_ON_STACK(req, tfm); in krb5_encrypt() 82 skcipher_request_set_sync_tfm(req, tfm); in krb5_encrypt() 83 skcipher_request_set_callback(req, 0, NULL, NULL); in krb5_encrypt() 84 skcipher_request_set_crypt(req, sg, sg, length, local_iv); in krb5_encrypt() 86 ret = crypto_skcipher_encrypt(req); in krb5_encrypt() 87 skcipher_request_zero(req); in krb5_encrypt() 104 SYNC_SKCIPHER_REQUEST_ON_STACK(req, tfm); in krb5_decrypt() 120 skcipher_request_set_sync_tfm(req, tfm); in krb5_decrypt() 121 skcipher_request_set_callback(req, 0, NULL, NULL); in krb5_decrypt() 122 skcipher_request_set_crypt(req, sg, sg, length, local_iv); in krb5_decrypt() [all …]
|
/net/wireless/ |
D | pmsr.c | 138 struct nlattr *req[NL80211_PMSR_REQ_ATTR_MAX + 1]; in pmsr_parse_peer() local 170 nla_parse_nested_deprecated(req, NL80211_PMSR_REQ_ATTR_MAX, in pmsr_parse_peer() 174 if (!req[NL80211_PMSR_REQ_ATTR_DATA]) { in pmsr_parse_peer() 181 if (req[NL80211_PMSR_REQ_ATTR_GET_AP_TSF]) in pmsr_parse_peer() 186 req[NL80211_PMSR_REQ_ATTR_GET_AP_TSF], in pmsr_parse_peer() 191 nla_for_each_nested(treq, req[NL80211_PMSR_REQ_ATTR_DATA], rem) { in pmsr_parse_peer() 214 struct cfg80211_pmsr_request *req; in nl80211_pmsr_start() local 240 req = kzalloc(struct_size(req, peers, count), GFP_KERNEL); in nl80211_pmsr_start() 241 if (!req) in nl80211_pmsr_start() 245 req->timeout = nla_get_u32(info->attrs[NL80211_ATTR_TIMEOUT]); in nl80211_pmsr_start() [all …]
|
/net/rxrpc/ |
D | rxkad.c | 103 struct skcipher_request *req; in rxkad_prime_packet_security() local 119 req = skcipher_request_alloc(&conn->cipher->base, GFP_NOFS); in rxkad_prime_packet_security() 120 if (!req) { in rxkad_prime_packet_security() 134 skcipher_request_set_sync_tfm(req, conn->cipher); in rxkad_prime_packet_security() 135 skcipher_request_set_callback(req, 0, NULL, NULL); in rxkad_prime_packet_security() 136 skcipher_request_set_crypt(req, &sg, &sg, tmpsize, iv.x); in rxkad_prime_packet_security() 137 crypto_skcipher_encrypt(req); in rxkad_prime_packet_security() 138 skcipher_request_free(req); in rxkad_prime_packet_security() 182 struct skcipher_request *req) in rxkad_secure_packet_auth() argument 202 skcipher_request_set_sync_tfm(req, call->conn->cipher); in rxkad_secure_packet_auth() [all …]
|
/net/sunrpc/xprtrdma/ |
D | rpc_rdma.c | 362 struct rpcrdma_req *req, in rpcrdma_mr_prepare() argument 367 *mr = rpcrdma_mr_pop(&req->rl_free_mrs); in rpcrdma_mr_prepare() 372 trace_xprtrdma_mr_get(req); in rpcrdma_mr_prepare() 373 (*mr)->mr_req = req; in rpcrdma_mr_prepare() 376 rpcrdma_mr_push(*mr, &req->rl_registered); in rpcrdma_mr_prepare() 377 return frwr_map(r_xprt, seg, nsegs, writing, req->rl_slot.rq_xid, *mr); in rpcrdma_mr_prepare() 380 trace_xprtrdma_nomrs(req); in rpcrdma_mr_prepare() 402 struct rpcrdma_req *req, in rpcrdma_encode_read_list() argument 406 struct xdr_stream *xdr = &req->rl_stream; in rpcrdma_encode_read_list() 418 seg = req->rl_segments; in rpcrdma_encode_read_list() [all …]
|
D | backchannel.c | 63 struct rpcrdma_req *req = rpcr_to_rdmar(rqst); in rpcrdma_bc_marshal_reply() local 66 rpcrdma_set_xdrlen(&req->rl_hdrbuf, 0); in rpcrdma_bc_marshal_reply() 67 xdr_init_encode(&req->rl_stream, &req->rl_hdrbuf, in rpcrdma_bc_marshal_reply() 68 rdmab_data(req->rl_rdmabuf), rqst); in rpcrdma_bc_marshal_reply() 70 p = xdr_reserve_space(&req->rl_stream, 28); in rpcrdma_bc_marshal_reply() 81 if (rpcrdma_prepare_send_sges(r_xprt, req, RPCRDMA_HDRLEN_MIN, in rpcrdma_bc_marshal_reply() 105 struct rpcrdma_req *req = rpcr_to_rdmar(rqst); in xprt_rdma_bc_send_reply() local 118 if (rpcrdma_ep_post(&r_xprt->rx_ia, &r_xprt->rx_ep, req)) in xprt_rdma_bc_send_reply() 157 struct rpcrdma_req *req = rpcr_to_rdmar(rqst); in xprt_rdma_bc_free_rqst() local 160 rpcrdma_recv_buffer_put(req->rl_reply); in xprt_rdma_bc_free_rqst() [all …]
|
D | verbs.c | 408 struct rpcrdma_req *req; in rpcrdma_ia_remove() local 433 list_for_each_entry(req, &buf->rb_allreqs, rl_all) { in rpcrdma_ia_remove() 434 rpcrdma_regbuf_dma_unmap(req->rl_rdmabuf); in rpcrdma_ia_remove() 435 rpcrdma_regbuf_dma_unmap(req->rl_sendbuf); in rpcrdma_ia_remove() 436 rpcrdma_regbuf_dma_unmap(req->rl_recvbuf); in rpcrdma_ia_remove() 1003 struct rpcrdma_req *req; in rpcrdma_req_create() local 1006 req = kzalloc(sizeof(*req), flags); in rpcrdma_req_create() 1007 if (req == NULL) in rpcrdma_req_create() 1018 req->rl_rdmabuf = rb; in rpcrdma_req_create() 1019 xdr_buf_init(&req->rl_hdrbuf, rdmab_data(rb), rdmab_length(rb)); in rpcrdma_req_create() [all …]
|
/net/mac80211/ |
D | scan.c | 325 struct cfg80211_scan_request *req; in ieee80211_prep_hw_scan() local 331 req = rcu_dereference_protected(local->scan_req, in ieee80211_prep_hw_scan() 338 for (i = 0; i < req->n_channels; i++) { in ieee80211_prep_hw_scan() 339 local->hw_scan_req->req.channels[i] = req->channels[i]; in ieee80211_prep_hw_scan() 340 bands_used |= BIT(req->channels[i]->band); in ieee80211_prep_hw_scan() 343 n_chans = req->n_channels; in ieee80211_prep_hw_scan() 351 for (i = 0; i < req->n_channels; i++) { in ieee80211_prep_hw_scan() 352 if (req->channels[i]->band != in ieee80211_prep_hw_scan() 355 local->hw_scan_req->req.channels[n_chans] = in ieee80211_prep_hw_scan() 356 req->channels[i]; in ieee80211_prep_hw_scan() [all …]
|
/net/unix/ |
D | diag.c | 78 struct sock *req, *peer; in sk_diag_dump_icons() local 80 req = skb->sk; in sk_diag_dump_icons() 86 unix_state_lock_nested(req, U_LOCK_DIAG); in sk_diag_dump_icons() 87 peer = unix_sk(req)->peer; in sk_diag_dump_icons() 89 unix_state_unlock(req); in sk_diag_dump_icons() 123 static int sk_diag_fill(struct sock *sk, struct sk_buff *skb, struct unix_diag_req *req, in sk_diag_fill() argument 143 if ((req->udiag_show & UDIAG_SHOW_NAME) && in sk_diag_fill() 147 if ((req->udiag_show & UDIAG_SHOW_VFS) && in sk_diag_fill() 151 if ((req->udiag_show & UDIAG_SHOW_PEER) && in sk_diag_fill() 155 if ((req->udiag_show & UDIAG_SHOW_ICONS) && in sk_diag_fill() [all …]
|
/net/core/ |
D | request_sock.c | 91 void reqsk_fastopen_remove(struct sock *sk, struct request_sock *req, in reqsk_fastopen_remove() argument 94 struct sock *lsk = req->rsk_listener; in reqsk_fastopen_remove() 102 tcp_rsk(req)->tfo_listener = false; in reqsk_fastopen_remove() 103 if (req->sk) /* the child socket hasn't been accepted yet */ in reqsk_fastopen_remove() 111 reqsk_put(req); in reqsk_fastopen_remove() 121 req->rsk_timer.expires = jiffies + 60*HZ; in reqsk_fastopen_remove() 123 fastopenq->rskq_rst_head = req; in reqsk_fastopen_remove() 125 fastopenq->rskq_rst_tail->dl_next = req; in reqsk_fastopen_remove() 127 req->dl_next = NULL; in reqsk_fastopen_remove() 128 fastopenq->rskq_rst_tail = req; in reqsk_fastopen_remove()
|
/net/dccp/ |
D | minisocks.c | 78 const struct request_sock *req, in dccp_create_openreq_child() argument 87 struct sock *newsk = inet_csk_clone_lock(sk, req, GFP_ATOMIC); in dccp_create_openreq_child() 90 struct dccp_request_sock *dreq = dccp_rsk(req); in dccp_create_openreq_child() 143 struct request_sock *req) in dccp_check_req() argument 146 struct dccp_request_sock *dreq = dccp_rsk(req); in dccp_check_req() 167 inet_rtx_syn_ack(sk, req); in dccp_check_req() 194 child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL, in dccp_check_req() 195 req, &own_req); in dccp_check_req() 197 child = inet_csk_complete_hashdance(sk, child, req, own_req); in dccp_check_req() 204 req->rsk_ops->send_reset(sk, skb); in dccp_check_req() [all …]
|
/net/ipv6/ |
D | syncookies.c | 138 struct request_sock *req; in cookie_v6_check() local 174 req = inet_reqsk_alloc(&tcp6_request_sock_ops, sk, false); in cookie_v6_check() 175 if (!req) in cookie_v6_check() 178 ireq = inet_rsk(req); in cookie_v6_check() 179 treq = tcp_rsk(req); in cookie_v6_check() 183 req->mss = mss; in cookie_v6_check() 189 if (security_inet_conn_request(sk, skb, req)) in cookie_v6_check() 207 req->num_retrans = 0; in cookie_v6_check() 212 req->ts_recent = tcp_opt.saw_tstamp ? tcp_opt.rcv_tsval : 0; in cookie_v6_check() 239 security_req_classify_flow(req, flowi6_to_flowi(&fl6)); in cookie_v6_check() [all …]
|