/net/dccp/ |
D | minisocks.c | 145 struct sock *child = NULL; in dccp_check_req() local 194 child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL, in dccp_check_req() 196 if (child) { in dccp_check_req() 197 child = inet_csk_complete_hashdance(sk, child, req, own_req); in dccp_check_req() 209 return child; in dccp_check_req() 219 int dccp_child_process(struct sock *parent, struct sock *child, in dccp_child_process() argument 223 const int state = child->sk_state; in dccp_child_process() 225 if (!sock_owned_by_user(child)) { in dccp_child_process() 226 ret = dccp_rcv_state_process(child, skb, dccp_hdr(skb), in dccp_child_process() 230 if (state == DCCP_RESPOND && child->sk_state != state) in dccp_child_process() [all …]
|
D | dccp.h | 284 int dccp_child_process(struct sock *parent, struct sock *child,
|
/net/sched/ |
D | sch_red.c | 61 struct Qdisc *child = q->qdisc; in red_enqueue() local 67 child->qstats.backlog); in red_enqueue() 99 ret = qdisc_enqueue(skb, child, to_free); in red_enqueue() 118 struct Qdisc *child = q->qdisc; in red_dequeue() local 120 skb = child->dequeue(child); in red_dequeue() 135 struct Qdisc *child = q->qdisc; in red_peek() local 137 return child->ops->peek(child); in red_peek() 196 struct Qdisc *old_child = NULL, *child = NULL; in red_change() local 225 child = fifo_create_dflt(sch, &bfifo_qdisc_ops, ctl->limit, in red_change() 227 if (IS_ERR(child)) in red_change() [all …]
|
D | sch_sfb.c | 285 struct Qdisc *child = q->qdisc; in sfb_enqueue() local 405 ret = qdisc_enqueue(skb, child, to_free); in sfb_enqueue() 429 struct Qdisc *child = q->qdisc; in sfb_dequeue() local 432 skb = child->dequeue(q->qdisc); in sfb_dequeue() 447 struct Qdisc *child = q->qdisc; in sfb_peek() local 449 return child->ops->peek(child); in sfb_peek() 495 struct Qdisc *child, *old; in sfb_change() local 517 child = fifo_create_dflt(sch, &pfifo_qdisc_ops, limit, extack); in sfb_change() 518 if (IS_ERR(child)) in sfb_change() 519 return PTR_ERR(child); in sfb_change() [all …]
|
D | sch_multiq.c | 198 struct Qdisc *child = q->queues[i]; in multiq_tune() local 201 qdisc_purge_queue(child); in multiq_tune() 202 removed[n_removed++] = child; in multiq_tune() 214 struct Qdisc *child, *old; in multiq_tune() local 215 child = qdisc_create_dflt(sch->dev_queue, in multiq_tune() 219 if (child) { in multiq_tune() 222 q->queues[i] = child; in multiq_tune() 223 if (child != &noop_qdisc) in multiq_tune() 224 qdisc_hash_add(child, true); in multiq_tune()
|
D | sch_taprio.c | 422 struct Qdisc *child; in taprio_enqueue() local 427 child = q->qdiscs[queue]; in taprio_enqueue() 428 if (unlikely(!child)) in taprio_enqueue() 444 return qdisc_enqueue(skb, child, to_free); in taprio_enqueue() 465 struct Qdisc *child = q->qdiscs[i]; in taprio_peek_soft() local 469 if (unlikely(!child)) in taprio_peek_soft() 472 skb = child->ops->peek(child); in taprio_peek_soft() 499 struct Qdisc *child = q->qdiscs[i]; in taprio_peek_offload() local 501 if (unlikely(!child)) in taprio_peek_offload() 504 skb = child->ops->peek(child); in taprio_peek_offload() [all …]
|
D | sch_tbf.c | 299 struct Qdisc *child = NULL; in tbf_change() local 378 child = fifo_create_dflt(sch, &bfifo_qdisc_ops, qopt->limit, in tbf_change() 380 if (IS_ERR(child)) { in tbf_change() 381 err = PTR_ERR(child); in tbf_change() 386 qdisc_hash_add(child, true); in tbf_change() 390 if (child) { in tbf_change() 393 q->qdisc = child; in tbf_change()
|
D | sch_cbs.c | 88 struct Qdisc *child, in cbs_child_enqueue() argument 94 err = child->ops->enqueue(skb, child, to_free); in cbs_child_enqueue() 160 static struct sk_buff *cbs_child_dequeue(struct Qdisc *sch, struct Qdisc *child) in cbs_child_dequeue() argument 164 skb = child->ops->dequeue(child); in cbs_child_dequeue()
|
/net/ipv4/ |
D | tcp_minisocks.c | 576 struct sock *child; in tcp_check_req() local 773 child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL, in tcp_check_req() 775 if (!child) in tcp_check_req() 778 sock_rps_save_rxhash(child, skb); in tcp_check_req() 779 tcp_synack_rtt_meas(child, req); in tcp_check_req() 781 return inet_csk_complete_hashdance(sk, child, req, own_req); in tcp_check_req() 824 int tcp_child_process(struct sock *parent, struct sock *child, in tcp_child_process() argument 828 int state = child->sk_state; in tcp_child_process() 831 sk_mark_napi_id(child, skb); in tcp_child_process() 833 tcp_segs_in(tcp_sk(child), skb); in tcp_child_process() [all …]
|
D | tcp_fastopen.c | 261 struct sock *child; in tcp_fastopen_create_child() local 264 child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL, in tcp_fastopen_create_child() 266 if (!child) in tcp_fastopen_create_child() 277 tp = tcp_sk(child); in tcp_fastopen_create_child() 292 inet_csk_reset_xmit_timer(child, ICSK_TIME_RETRANS, in tcp_fastopen_create_child() 298 tcp_init_transfer(child, BPF_SOCK_OPS_PASSIVE_ESTABLISHED_CB); in tcp_fastopen_create_child() 302 tcp_fastopen_add_skb(child, skb); in tcp_fastopen_create_child() 309 return child; in tcp_fastopen_create_child() 371 struct sock *child; in tcp_try_fastopen() local 407 child = tcp_fastopen_create_child(sk, skb, req); in tcp_try_fastopen() [all …]
|
D | inet_connection_sock.c | 967 struct sock *child) in inet_child_forget() argument 969 sk->sk_prot->disconnect(child, O_NONBLOCK); in inet_child_forget() 971 sock_orphan(child); in inet_child_forget() 976 BUG_ON(rcu_access_pointer(tcp_sk(child)->fastopen_rsk) != req); in inet_child_forget() 985 RCU_INIT_POINTER(tcp_sk(child)->fastopen_rsk, NULL); in inet_child_forget() 987 inet_csk_destroy_sock(child); in inet_child_forget() 992 struct sock *child) in inet_csk_reqsk_queue_add() argument 998 inet_child_forget(sk, req, child); in inet_csk_reqsk_queue_add() 999 child = NULL; in inet_csk_reqsk_queue_add() 1001 req->sk = child; in inet_csk_reqsk_queue_add() [all …]
|
D | syncookies.c | 200 struct sock *child; in tcp_get_cookie_sock() local 203 child = icsk->icsk_af_ops->syn_recv_sock(sk, skb, req, dst, in tcp_get_cookie_sock() 205 if (child) { in tcp_get_cookie_sock() 207 tcp_sk(child)->tsoffset = tsoff; in tcp_get_cookie_sock() 208 sock_rps_save_rxhash(child, skb); in tcp_get_cookie_sock() 209 if (inet_csk_reqsk_queue_add(sk, req, child)) in tcp_get_cookie_sock() 210 return child; in tcp_get_cookie_sock() 212 bh_unlock_sock(child); in tcp_get_cookie_sock() 213 sock_put(child); in tcp_get_cookie_sock()
|
D | inet_hashtables.c | 131 int __inet_inherit_port(const struct sock *sk, struct sock *child) in __inet_inherit_port() argument 134 unsigned short port = inet_sk(child)->inet_num; in __inet_inherit_port() 169 inet_csk_update_fastreuse(tb, child); in __inet_inherit_port() 171 inet_bind_hash(child, tb, port); in __inet_inherit_port()
|
/net/core/ |
D | dst.c | 105 struct dst_entry *child = NULL; in dst_destroy() local 113 child = xdst->child; in dst_destroy() 131 dst = child; in dst_destroy()
|
/net/vmw_vsock/ |
D | virtio_transport_common.c | 1013 struct sock *child; in virtio_transport_recv_listen() local 1025 child = __vsock_create(sock_net(sk), NULL, sk, GFP_KERNEL, in virtio_transport_recv_listen() 1027 if (!child) { in virtio_transport_recv_listen() 1034 lock_sock_nested(child, SINGLE_DEPTH_NESTING); in virtio_transport_recv_listen() 1036 child->sk_state = TCP_ESTABLISHED; in virtio_transport_recv_listen() 1038 vchild = vsock_sk(child); in virtio_transport_recv_listen() 1045 vsock_enqueue_accept(sk, child); in virtio_transport_recv_listen() 1048 release_sock(child); in virtio_transport_recv_listen()
|
/net/sunrpc/ |
D | rpcb_clnt.c | 685 struct rpc_task *child; in rpcb_getport_async() local 796 child = rpcb_call_async(rpcb_clnt, map, proc); in rpcb_getport_async() 798 if (IS_ERR(child)) { in rpcb_getport_async() 806 rpc_put_task(child); in rpcb_getport_async() 823 static void rpcb_getport_done(struct rpc_task *child, void *data) in rpcb_getport_done() argument 827 int status = child->tk_status; in rpcb_getport_done() 852 child->tk_pid, status, map->r_port); in rpcb_getport_done()
|
D | sysctl.c | 182 .child = debug_table
|
D | xprtsock.c | 157 .child = xs_tunables_table
|
/net/sunrpc/xprtrdma/ |
D | svc_rdma.c | 214 .child = svcrdma_parm_table 223 .child = svcrdma_table
|
D | transport.c | 150 .child = xr_tunables_table
|
/net/ipv6/ |
D | ip6_fib.c | 1709 struct fib6_node *child; in fib6_repair_tree() local 1744 child = NULL; in fib6_repair_tree() 1746 child = fn_r, children |= 1; in fib6_repair_tree() 1748 child = fn_l, children |= 2; in fib6_repair_tree() 1777 rcu_assign_pointer(pn->right, child); in fib6_repair_tree() 1779 rcu_assign_pointer(pn->left, child); in fib6_repair_tree() 1784 if (child) in fib6_repair_tree() 1785 rcu_assign_pointer(child->parent, pn); in fib6_repair_tree() 1793 if (!child) { in fib6_repair_tree() 1801 w->node = child; in fib6_repair_tree()
|
/net/netfilter/ |
D | xt_policy.c | 94 dst = ((struct xfrm_dst *)dst)->child, i++) { in match_policy_out()
|
D | nft_xfrm.c | 181 dst = ((const struct xfrm_dst *)dst)->child, i++) { in nft_xfrm_get_eval_out()
|
/net/xfrm/ |
D | xfrm_output.c | 47 struct dst_entry *child = dst_clone(xfrm_dst_child(skb_dst(skb))); in skb_dst_pop() local 50 return child; in skb_dst_pop()
|
D | xfrm_device.c | 278 (!xdst->child->xfrm)) { in xfrm_dev_offload_ok()
|