/net/ipv4/ |
D | tcp_fastopen.c | 134 struct sock *child; in tcp_fastopen_create_child() local 141 child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL); in tcp_fastopen_create_child() 142 if (child == NULL) in tcp_fastopen_create_child() 153 tp = tcp_sk(child); in tcp_fastopen_create_child() 172 inet_csk_reset_xmit_timer(child, ICSK_TIME_RETRANS, in tcp_fastopen_create_child() 176 inet_csk_reqsk_queue_add(sk, req, child); in tcp_fastopen_create_child() 179 inet_csk(child)->icsk_af_ops->rebuild_header(child); in tcp_fastopen_create_child() 180 tcp_init_congestion_control(child); in tcp_fastopen_create_child() 181 tcp_mtup_init(child); in tcp_fastopen_create_child() 182 tcp_init_metrics(child); in tcp_fastopen_create_child() [all …]
|
D | tcp_minisocks.c | 539 struct sock *child; in tcp_check_req() local 721 child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL); in tcp_check_req() 722 if (child == NULL) in tcp_check_req() 728 inet_csk_reqsk_queue_add(sk, req, child); in tcp_check_req() 729 return child; in tcp_check_req() 769 int tcp_child_process(struct sock *parent, struct sock *child, in tcp_child_process() argument 773 int state = child->sk_state; in tcp_child_process() 775 if (!sock_owned_by_user(child)) { in tcp_child_process() 776 ret = tcp_rcv_state_process(child, skb, tcp_hdr(skb), in tcp_child_process() 779 if (state == TCP_SYN_RECV && child->sk_state != state) in tcp_child_process() [all …]
|
D | inet_connection_sock.c | 807 struct sock *child = req->sk; in inet_csk_listen_stop() local 812 bh_lock_sock(child); in inet_csk_listen_stop() 813 WARN_ON(sock_owned_by_user(child)); in inet_csk_listen_stop() 814 sock_hold(child); in inet_csk_listen_stop() 816 sk->sk_prot->disconnect(child, O_NONBLOCK); in inet_csk_listen_stop() 818 sock_orphan(child); in inet_csk_listen_stop() 823 BUG_ON(tcp_sk(child)->fastopen_rsk != req); in inet_csk_listen_stop() 832 tcp_sk(child)->fastopen_rsk = NULL; in inet_csk_listen_stop() 835 inet_csk_destroy_sock(child); in inet_csk_listen_stop() 837 bh_unlock_sock(child); in inet_csk_listen_stop() [all …]
|
D | syncookies.c | 205 struct sock *child; in get_cookie_sock() local 207 child = icsk->icsk_af_ops->syn_recv_sock(sk, skb, req, dst); in get_cookie_sock() 208 if (child) in get_cookie_sock() 209 inet_csk_reqsk_queue_add(sk, req, child); in get_cookie_sock() 213 return child; in get_cookie_sock()
|
D | fib_trie.c | 129 struct rt_trie_node __rcu *child[0]; member 220 return rtnl_dereference(tn->child[i]); in tnode_get_child() 230 return rcu_dereference_rtnl(tn->child[i]); in tnode_get_child_rcu() 496 struct rt_trie_node *chi = rtnl_dereference(tn->child[i]); in tnode_put_child_reorg() 520 rcu_assign_pointer(tn->child[i], n); in tnode_put_child_reorg() 674 n = rtnl_dereference(tn->child[i]); in resize() 695 tofree = (struct tnode *)rtnl_dereference(tn->child[i]); in tnode_clean_free() 775 put_child(tn, 2*i, rtnl_dereference(inode->child[0])); in inflate() 776 put_child(tn, 2*i+1, rtnl_dereference(inode->child[1])); in inflate() 817 put_child(left, j, rtnl_dereference(inode->child[j])); in inflate() [all …]
|
D | inet_hashtables.c | 131 int __inet_inherit_port(struct sock *sk, struct sock *child) in __inet_inherit_port() argument 134 unsigned short port = inet_sk(child)->inet_num; in __inet_inherit_port() 162 inet_bind_hash(child, tb, port); in __inet_inherit_port()
|
D | xfrm4_mode_tunnel.c | 62 top_iph->ttl = ip4_dst_hoplimit(dst->child); in xfrm4_mode_tunnel_output()
|
/net/sched/ |
D | sch_red.c | 62 struct Qdisc *child = q->qdisc; in red_enqueue() local 67 child->qstats.backlog); in red_enqueue() 98 ret = qdisc_enqueue(skb, child); in red_enqueue() 116 struct Qdisc *child = q->qdisc; in red_dequeue() local 118 skb = child->dequeue(child); in red_dequeue() 132 struct Qdisc *child = q->qdisc; in red_peek() local 134 return child->ops->peek(child); in red_peek() 140 struct Qdisc *child = q->qdisc; in red_drop() local 143 if (child->ops->drop && (len = child->ops->drop(child)) > 0) { in red_drop() 184 struct Qdisc *child = NULL; in red_change() local [all …]
|
D | sch_prio.c | 191 struct Qdisc *child = q->queues[i]; in prio_tune() local 193 if (child != &noop_qdisc) { in prio_tune() 194 qdisc_tree_reduce_backlog(child, child->q.qlen, child->qstats.backlog); in prio_tune() 195 qdisc_destroy(child); in prio_tune() 202 struct Qdisc *child, *old; in prio_tune() local 204 child = qdisc_create_dflt(sch->dev_queue, in prio_tune() 207 if (child) { in prio_tune() 210 q->queues[i] = child; in prio_tune()
|
D | sch_multiq.c | 219 struct Qdisc *child = q->queues[i]; in multiq_tune() local 221 qdisc_tree_reduce_backlog(child, child->q.qlen, in multiq_tune() 222 child->qstats.backlog); in multiq_tune() 223 qdisc_destroy(child); in multiq_tune() 231 struct Qdisc *child, *old; in multiq_tune() local 232 child = qdisc_create_dflt(sch->dev_queue, in multiq_tune() 236 if (child) { in multiq_tune() 239 q->queues[i] = child; in multiq_tune()
|
D | sch_sfb.c | 283 struct Qdisc *child = q->qdisc; in sfb_enqueue() local 408 ret = qdisc_enqueue(skb, child); in sfb_enqueue() 431 struct Qdisc *child = q->qdisc; in sfb_dequeue() local 434 skb = child->dequeue(q->qdisc); in sfb_dequeue() 448 struct Qdisc *child = q->qdisc; in sfb_peek() local 450 return child->ops->peek(child); in sfb_peek() 494 struct Qdisc *child; in sfb_change() local 515 child = fifo_create_dflt(sch, &pfifo_qdisc_ops, limit); in sfb_change() 516 if (IS_ERR(child)) in sfb_change() 517 return PTR_ERR(child); in sfb_change() [all …]
|
D | sch_tbf.c | 320 struct Qdisc *child = NULL; in tbf_change() local 395 child = fifo_create_dflt(sch, &bfifo_qdisc_ops, qopt->limit); in tbf_change() 396 if (IS_ERR(child)) { in tbf_change() 397 err = PTR_ERR(child); in tbf_change() 403 if (child) { in tbf_change() 407 q->qdisc = child; in tbf_change()
|
/net/dccp/ |
D | minisocks.c | 159 struct sock *child = NULL; in dccp_check_req() local 200 child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL); in dccp_check_req() 201 if (child == NULL) in dccp_check_req() 206 inet_csk_reqsk_queue_add(sk, req, child); in dccp_check_req() 208 return child; in dccp_check_req() 227 int dccp_child_process(struct sock *parent, struct sock *child, in dccp_child_process() argument 231 const int state = child->sk_state; in dccp_child_process() 233 if (!sock_owned_by_user(child)) { in dccp_child_process() 234 ret = dccp_rcv_state_process(child, skb, dccp_hdr(skb), in dccp_child_process() 238 if (state == DCCP_RESPOND && child->sk_state != state) in dccp_child_process() [all …]
|
D | dccp.h | 286 int dccp_child_process(struct sock *parent, struct sock *child,
|
/net/ipv6/ |
D | syncookies.c | 49 struct sock *child; in get_cookie_sock() local 51 child = icsk->icsk_af_ops->syn_recv_sock(sk, skb, req, dst); in get_cookie_sock() 52 if (child) in get_cookie_sock() 53 inet_csk_reqsk_queue_add(sk, req, child); in get_cookie_sock() 57 return child; in get_cookie_sock()
|
D | ip6_fib.c | 1204 struct fib6_node *child, *pn; in fib6_repair_tree() local 1217 child = NULL; in fib6_repair_tree() 1219 child = fn->right, children |= 1; in fib6_repair_tree() 1221 child = fn->left, children |= 2; in fib6_repair_tree() 1250 pn->right = child; in fib6_repair_tree() 1252 pn->left = child; in fib6_repair_tree() 1257 if (child) in fib6_repair_tree() 1258 child->parent = pn; in fib6_repair_tree() 1266 if (!child) { in fib6_repair_tree() 1277 w->root = child; in fib6_repair_tree() [all …]
|
D | xfrm6_mode_tunnel.c | 60 top_iph->hop_limit = ip6_dst_hoplimit(dst->child); in xfrm6_mode_tunnel_output()
|
D | xfrm6_policy.c | 281 xdst = (struct xfrm_dst *)xdst->u.dst.child; in xfrm6_dst_ifdown()
|
/net/core/ |
D | dst.c | 174 dst->child = NULL; in dst_alloc() 237 struct dst_entry *child; in dst_destroy() local 242 child = dst->child; in dst_destroy() 253 dst = child; in dst_destroy()
|
/net/sunrpc/ |
D | rpcb_clnt.c | 679 struct rpc_task *child; in rpcb_getport_async() local 784 child = rpcb_call_async(rpcb_clnt, map, proc); in rpcb_getport_async() 786 if (IS_ERR(child)) { in rpcb_getport_async() 794 rpc_put_task(child); in rpcb_getport_async() 809 static void rpcb_getport_done(struct rpc_task *child, void *data) in rpcb_getport_done() argument 813 int status = child->tk_status; in rpcb_getport_done() 838 child->tk_pid, status, map->r_port); in rpcb_getport_done()
|
D | sysctl.c | 180 .child = debug_table
|
/net/dsa/ |
D | dsa.c | 442 struct device_node *child, *mdio, *ethernet, *port, *link; in dsa_of_probe() local 486 for_each_available_child_of_node(np, child) { in dsa_of_probe() 490 cd->of_node = child; in dsa_of_probe() 493 sw_addr = of_get_property(child, "reg", NULL); in dsa_of_probe() 501 for_each_available_child_of_node(child, port) { in dsa_of_probe()
|
/net/sunrpc/xprtrdma/ |
D | svc_rdma.c | 221 .child = svcrdma_parm_table 230 .child = svcrdma_table
|
/net/netfilter/ |
D | xt_policy.c | 96 for (i = 0; dst && dst->xfrm; dst = dst->child, i++) { in match_policy_out()
|
/net/xfrm/ |
D | xfrm_policy.c | 1682 dst_prev->child = dst_clone(dst1); in xfrm_bundle_create() 1719 dst_prev->child = dst; in xfrm_bundle_create() 1730 for (dst_prev = dst0; dst_prev != dst; dst_prev = dst_prev->child) { in xfrm_bundle_create() 1990 dst1->child = dst; in xfrm_create_dummy_bundle() 2613 while ((dst = dst->child) && dst->xfrm && dst->dev == dev) { in xfrm_dst_ifdown() 2654 pmtu = dst_mtu(dst->child); in xfrm_init_pmtu() 2699 mtu = dst_mtu(dst->child); in xfrm_bundle_ok() 2713 dst = dst->child; in xfrm_bundle_ok()
|