Home
last modified time | relevance | path

Searched refs:child (Results 1 – 25 of 26) sorted by relevance

12

/net/dccp/
Dminisocks.c145 struct sock *child = NULL; in dccp_check_req() local
194 child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL, in dccp_check_req()
196 if (child) { in dccp_check_req()
197 child = inet_csk_complete_hashdance(sk, child, req, own_req); in dccp_check_req()
209 return child; in dccp_check_req()
219 int dccp_child_process(struct sock *parent, struct sock *child, in dccp_child_process() argument
223 const int state = child->sk_state; in dccp_child_process()
225 if (!sock_owned_by_user(child)) { in dccp_child_process()
226 ret = dccp_rcv_state_process(child, skb, dccp_hdr(skb), in dccp_child_process()
230 if (state == DCCP_RESPOND && child->sk_state != state) in dccp_child_process()
[all …]
Ddccp.h284 int dccp_child_process(struct sock *parent, struct sock *child,
/net/sched/
Dsch_red.c61 struct Qdisc *child = q->qdisc; in red_enqueue() local
67 child->qstats.backlog); in red_enqueue()
99 ret = qdisc_enqueue(skb, child, to_free); in red_enqueue()
118 struct Qdisc *child = q->qdisc; in red_dequeue() local
120 skb = child->dequeue(child); in red_dequeue()
135 struct Qdisc *child = q->qdisc; in red_peek() local
137 return child->ops->peek(child); in red_peek()
196 struct Qdisc *old_child = NULL, *child = NULL; in red_change() local
225 child = fifo_create_dflt(sch, &bfifo_qdisc_ops, ctl->limit, in red_change()
227 if (IS_ERR(child)) in red_change()
[all …]
Dsch_sfb.c285 struct Qdisc *child = q->qdisc; in sfb_enqueue() local
405 ret = qdisc_enqueue(skb, child, to_free); in sfb_enqueue()
429 struct Qdisc *child = q->qdisc; in sfb_dequeue() local
432 skb = child->dequeue(q->qdisc); in sfb_dequeue()
447 struct Qdisc *child = q->qdisc; in sfb_peek() local
449 return child->ops->peek(child); in sfb_peek()
495 struct Qdisc *child, *old; in sfb_change() local
517 child = fifo_create_dflt(sch, &pfifo_qdisc_ops, limit, extack); in sfb_change()
518 if (IS_ERR(child)) in sfb_change()
519 return PTR_ERR(child); in sfb_change()
[all …]
Dsch_multiq.c198 struct Qdisc *child = q->queues[i]; in multiq_tune() local
201 qdisc_purge_queue(child); in multiq_tune()
202 removed[n_removed++] = child; in multiq_tune()
214 struct Qdisc *child, *old; in multiq_tune() local
215 child = qdisc_create_dflt(sch->dev_queue, in multiq_tune()
219 if (child) { in multiq_tune()
222 q->queues[i] = child; in multiq_tune()
223 if (child != &noop_qdisc) in multiq_tune()
224 qdisc_hash_add(child, true); in multiq_tune()
Dsch_taprio.c422 struct Qdisc *child; in taprio_enqueue() local
427 child = q->qdiscs[queue]; in taprio_enqueue()
428 if (unlikely(!child)) in taprio_enqueue()
444 return qdisc_enqueue(skb, child, to_free); in taprio_enqueue()
465 struct Qdisc *child = q->qdiscs[i]; in taprio_peek_soft() local
469 if (unlikely(!child)) in taprio_peek_soft()
472 skb = child->ops->peek(child); in taprio_peek_soft()
499 struct Qdisc *child = q->qdiscs[i]; in taprio_peek_offload() local
501 if (unlikely(!child)) in taprio_peek_offload()
504 skb = child->ops->peek(child); in taprio_peek_offload()
[all …]
Dsch_tbf.c299 struct Qdisc *child = NULL; in tbf_change() local
378 child = fifo_create_dflt(sch, &bfifo_qdisc_ops, qopt->limit, in tbf_change()
380 if (IS_ERR(child)) { in tbf_change()
381 err = PTR_ERR(child); in tbf_change()
386 qdisc_hash_add(child, true); in tbf_change()
390 if (child) { in tbf_change()
393 q->qdisc = child; in tbf_change()
Dsch_cbs.c88 struct Qdisc *child, in cbs_child_enqueue() argument
94 err = child->ops->enqueue(skb, child, to_free); in cbs_child_enqueue()
160 static struct sk_buff *cbs_child_dequeue(struct Qdisc *sch, struct Qdisc *child) in cbs_child_dequeue() argument
164 skb = child->ops->dequeue(child); in cbs_child_dequeue()
/net/ipv4/
Dtcp_minisocks.c576 struct sock *child; in tcp_check_req() local
773 child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL, in tcp_check_req()
775 if (!child) in tcp_check_req()
778 sock_rps_save_rxhash(child, skb); in tcp_check_req()
779 tcp_synack_rtt_meas(child, req); in tcp_check_req()
781 return inet_csk_complete_hashdance(sk, child, req, own_req); in tcp_check_req()
824 int tcp_child_process(struct sock *parent, struct sock *child, in tcp_child_process() argument
828 int state = child->sk_state; in tcp_child_process()
831 sk_mark_napi_id(child, skb); in tcp_child_process()
833 tcp_segs_in(tcp_sk(child), skb); in tcp_child_process()
[all …]
Dtcp_fastopen.c261 struct sock *child; in tcp_fastopen_create_child() local
264 child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL, in tcp_fastopen_create_child()
266 if (!child) in tcp_fastopen_create_child()
277 tp = tcp_sk(child); in tcp_fastopen_create_child()
292 inet_csk_reset_xmit_timer(child, ICSK_TIME_RETRANS, in tcp_fastopen_create_child()
298 tcp_init_transfer(child, BPF_SOCK_OPS_PASSIVE_ESTABLISHED_CB); in tcp_fastopen_create_child()
302 tcp_fastopen_add_skb(child, skb); in tcp_fastopen_create_child()
309 return child; in tcp_fastopen_create_child()
371 struct sock *child; in tcp_try_fastopen() local
407 child = tcp_fastopen_create_child(sk, skb, req); in tcp_try_fastopen()
[all …]
Dinet_connection_sock.c967 struct sock *child) in inet_child_forget() argument
969 sk->sk_prot->disconnect(child, O_NONBLOCK); in inet_child_forget()
971 sock_orphan(child); in inet_child_forget()
976 BUG_ON(rcu_access_pointer(tcp_sk(child)->fastopen_rsk) != req); in inet_child_forget()
985 RCU_INIT_POINTER(tcp_sk(child)->fastopen_rsk, NULL); in inet_child_forget()
987 inet_csk_destroy_sock(child); in inet_child_forget()
992 struct sock *child) in inet_csk_reqsk_queue_add() argument
998 inet_child_forget(sk, req, child); in inet_csk_reqsk_queue_add()
999 child = NULL; in inet_csk_reqsk_queue_add()
1001 req->sk = child; in inet_csk_reqsk_queue_add()
[all …]
Dsyncookies.c200 struct sock *child; in tcp_get_cookie_sock() local
203 child = icsk->icsk_af_ops->syn_recv_sock(sk, skb, req, dst, in tcp_get_cookie_sock()
205 if (child) { in tcp_get_cookie_sock()
207 tcp_sk(child)->tsoffset = tsoff; in tcp_get_cookie_sock()
208 sock_rps_save_rxhash(child, skb); in tcp_get_cookie_sock()
209 if (inet_csk_reqsk_queue_add(sk, req, child)) in tcp_get_cookie_sock()
210 return child; in tcp_get_cookie_sock()
212 bh_unlock_sock(child); in tcp_get_cookie_sock()
213 sock_put(child); in tcp_get_cookie_sock()
Dinet_hashtables.c131 int __inet_inherit_port(const struct sock *sk, struct sock *child) in __inet_inherit_port() argument
134 unsigned short port = inet_sk(child)->inet_num; in __inet_inherit_port()
169 inet_csk_update_fastreuse(tb, child); in __inet_inherit_port()
171 inet_bind_hash(child, tb, port); in __inet_inherit_port()
/net/core/
Ddst.c105 struct dst_entry *child = NULL; in dst_destroy() local
113 child = xdst->child; in dst_destroy()
131 dst = child; in dst_destroy()
/net/vmw_vsock/
Dvirtio_transport_common.c1013 struct sock *child; in virtio_transport_recv_listen() local
1025 child = __vsock_create(sock_net(sk), NULL, sk, GFP_KERNEL, in virtio_transport_recv_listen()
1027 if (!child) { in virtio_transport_recv_listen()
1034 lock_sock_nested(child, SINGLE_DEPTH_NESTING); in virtio_transport_recv_listen()
1036 child->sk_state = TCP_ESTABLISHED; in virtio_transport_recv_listen()
1038 vchild = vsock_sk(child); in virtio_transport_recv_listen()
1045 vsock_enqueue_accept(sk, child); in virtio_transport_recv_listen()
1048 release_sock(child); in virtio_transport_recv_listen()
/net/sunrpc/
Drpcb_clnt.c685 struct rpc_task *child; in rpcb_getport_async() local
796 child = rpcb_call_async(rpcb_clnt, map, proc); in rpcb_getport_async()
798 if (IS_ERR(child)) { in rpcb_getport_async()
806 rpc_put_task(child); in rpcb_getport_async()
823 static void rpcb_getport_done(struct rpc_task *child, void *data) in rpcb_getport_done() argument
827 int status = child->tk_status; in rpcb_getport_done()
852 child->tk_pid, status, map->r_port); in rpcb_getport_done()
Dsysctl.c182 .child = debug_table
Dxprtsock.c157 .child = xs_tunables_table
/net/sunrpc/xprtrdma/
Dsvc_rdma.c214 .child = svcrdma_parm_table
223 .child = svcrdma_table
Dtransport.c150 .child = xr_tunables_table
/net/ipv6/
Dip6_fib.c1709 struct fib6_node *child; in fib6_repair_tree() local
1744 child = NULL; in fib6_repair_tree()
1746 child = fn_r, children |= 1; in fib6_repair_tree()
1748 child = fn_l, children |= 2; in fib6_repair_tree()
1777 rcu_assign_pointer(pn->right, child); in fib6_repair_tree()
1779 rcu_assign_pointer(pn->left, child); in fib6_repair_tree()
1784 if (child) in fib6_repair_tree()
1785 rcu_assign_pointer(child->parent, pn); in fib6_repair_tree()
1793 if (!child) { in fib6_repair_tree()
1801 w->node = child; in fib6_repair_tree()
/net/netfilter/
Dxt_policy.c94 dst = ((struct xfrm_dst *)dst)->child, i++) { in match_policy_out()
Dnft_xfrm.c181 dst = ((const struct xfrm_dst *)dst)->child, i++) { in nft_xfrm_get_eval_out()
/net/xfrm/
Dxfrm_output.c47 struct dst_entry *child = dst_clone(xfrm_dst_child(skb_dst(skb))); in skb_dst_pop() local
50 return child; in skb_dst_pop()
Dxfrm_device.c278 (!xdst->child->xfrm)) { in xfrm_dev_offload_ok()

12