Home
last modified time | relevance | path

Searched refs:shinfo (Results 1 – 10 of 10) sorted by relevance

/net/tls/
Dtls_strp.c31 struct skb_shared_info *shinfo = skb_shinfo(strp->anchor); in tls_strp_anchor_free() local
33 DEBUG_NET_WARN_ON_ONCE(atomic_read(&shinfo->dataref) != 1); in tls_strp_anchor_free()
35 shinfo->frag_list = NULL; in tls_strp_anchor_free()
146 struct skb_shared_info *shinfo = skb_shinfo(strp->anchor); in tls_strp_msg_hold() local
151 WARN_ON_ONCE(!shinfo->nr_frags); in tls_strp_msg_hold()
166 iter = shinfo->frag_list; in tls_strp_msg_hold()
193 struct skb_shared_info *shinfo = skb_shinfo(strp->anchor); in tls_strp_flush_anchor_copy() local
196 DEBUG_NET_WARN_ON_ONCE(atomic_read(&shinfo->dataref) != 1); in tls_strp_flush_anchor_copy()
198 for (i = 0; i < shinfo->nr_frags; i++) in tls_strp_flush_anchor_copy()
199 __skb_frag_unref(&shinfo->frags[i], false); in tls_strp_flush_anchor_copy()
[all …]
/net/core/
Dskbuff.c276 struct skb_shared_info *shinfo; in __build_skb_around() local
292 shinfo = skb_shinfo(skb); in __build_skb_around()
293 memset(shinfo, 0, offsetof(struct skb_shared_info, dataref)); in __build_skb_around()
294 atomic_set(&shinfo->dataref, 1); in __build_skb_around()
768 struct skb_shared_info *shinfo = skb_shinfo(skb); in skb_release_data() local
773 &shinfo->dataref)) in skb_release_data()
777 bool skip_unref = shinfo->flags & SKBFL_MANAGED_FRAG_REFS; in skb_release_data()
784 for (i = 0; i < shinfo->nr_frags; i++) in skb_release_data()
785 __skb_frag_unref(&shinfo->frags[i], skb->pp_recycle); in skb_release_data()
788 if (shinfo->frag_list) in skb_release_data()
[all …]
Dlwt_bpf.c526 struct skb_shared_info *shinfo = skb_shinfo(skb); in handle_gso_type() local
529 shinfo->gso_type |= gso_type; in handle_gso_type()
530 skb_decrease_gso_size(shinfo, encap_len); in handle_gso_type()
531 shinfo->gso_segs = 0; in handle_gso_type()
Dfilter.c3278 struct skb_shared_info *shinfo = skb_shinfo(skb); in bpf_skb_proto_4_to_6() local
3281 if (shinfo->gso_type & SKB_GSO_TCPV4) { in bpf_skb_proto_4_to_6()
3282 shinfo->gso_type &= ~SKB_GSO_TCPV4; in bpf_skb_proto_4_to_6()
3283 shinfo->gso_type |= SKB_GSO_TCPV6; in bpf_skb_proto_4_to_6()
3308 struct skb_shared_info *shinfo = skb_shinfo(skb); in bpf_skb_proto_6_to_4() local
3311 if (shinfo->gso_type & SKB_GSO_TCPV6) { in bpf_skb_proto_6_to_4()
3312 shinfo->gso_type &= ~SKB_GSO_TCPV6; in bpf_skb_proto_6_to_4()
3313 shinfo->gso_type |= SKB_GSO_TCPV4; in bpf_skb_proto_6_to_4()
3511 struct skb_shared_info *shinfo = skb_shinfo(skb); in bpf_skb_net_grow() local
3515 skb_decrease_gso_size(shinfo, len_diff); in bpf_skb_net_grow()
[all …]
Ddev.c3784 const struct skb_shared_info *shinfo = skb_shinfo(skb); in qdisc_pkt_len_init() local
3791 if (shinfo->gso_size && skb_transport_header_was_set(skb)) { in qdisc_pkt_len_init()
3793 u16 gso_segs = shinfo->gso_segs; in qdisc_pkt_len_init()
3799 if (likely(shinfo->gso_type & (SKB_GSO_TCPV4 | SKB_GSO_TCPV6))) { in qdisc_pkt_len_init()
3815 if (shinfo->gso_type & SKB_GSO_DODGY) in qdisc_pkt_len_init()
3817 shinfo->gso_size); in qdisc_pkt_len_init()
/net/ipv4/
Dtcp_output.c1491 struct skb_shared_info *shinfo = skb_shinfo(skb); in tcp_fragment_tstamp() local
1494 !before(shinfo->tskey, TCP_SKB_CB(skb2)->seq)) { in tcp_fragment_tstamp()
1496 u8 tsflags = shinfo->tx_flags & SKBTX_ANY_TSTAMP; in tcp_fragment_tstamp()
1498 shinfo->tx_flags &= ~tsflags; in tcp_fragment_tstamp()
1500 swap(shinfo->tskey, shinfo2->tskey); in tcp_fragment_tstamp()
1628 struct skb_shared_info *shinfo; in __pskb_trim_head() local
1640 shinfo = skb_shinfo(skb); in __pskb_trim_head()
1641 for (i = 0; i < shinfo->nr_frags; i++) { in __pskb_trim_head()
1642 int size = skb_frag_size(&shinfo->frags[i]); in __pskb_trim_head()
1648 shinfo->frags[k] = shinfo->frags[i]; in __pskb_trim_head()
[all …]
Dtcp_ipv4.c1769 struct skb_shared_info *shinfo; in tcp_add_backlog() local
1828 shinfo = skb_shinfo(skb); in tcp_add_backlog()
1829 gso_size = shinfo->gso_size ?: skb->len; in tcp_add_backlog()
1830 gso_segs = shinfo->gso_segs ?: 1; in tcp_add_backlog()
1832 shinfo = skb_shinfo(tail); in tcp_add_backlog()
1833 tail_gso_size = shinfo->gso_size ?: (tail->len - hdrlen); in tcp_add_backlog()
1834 tail_gso_segs = shinfo->gso_segs ?: 1; in tcp_add_backlog()
1862 shinfo->gso_size = max(gso_size, tail_gso_size); in tcp_add_backlog()
1863 shinfo->gso_segs = min_t(u32, gso_segs + tail_gso_segs, 0xFFFF); in tcp_add_backlog()
Dtcp.c471 struct skb_shared_info *shinfo = skb_shinfo(skb); in tcp_tx_timestamp() local
474 sock_tx_timestamp(sk, tsflags, &shinfo->tx_flags); in tcp_tx_timestamp()
478 shinfo->tskey = TCP_SKB_CB(skb)->seq + skb->len - 1; in tcp_tx_timestamp()
Dtcp_input.c3232 const struct skb_shared_info *shinfo; in tcp_ack_tstamp() local
3238 shinfo = skb_shinfo(skb); in tcp_ack_tstamp()
3239 if (!before(shinfo->tskey, prior_snd_una) && in tcp_ack_tstamp()
3240 before(shinfo->tskey, tcp_sk(sk)->snd_una)) { in tcp_ack_tstamp()
/net/sched/
Dsch_cake.c1351 const struct skb_shared_info *shinfo = skb_shinfo(skb); in cake_overhead() local
1359 if (!shinfo->gso_size) in cake_overhead()
1366 if (likely(shinfo->gso_type & (SKB_GSO_TCPV4 | in cake_overhead()
1383 if (unlikely(shinfo->gso_type & SKB_GSO_DODGY)) in cake_overhead()
1385 shinfo->gso_size); in cake_overhead()
1387 segs = shinfo->gso_segs; in cake_overhead()
1389 len = shinfo->gso_size + hdr_len; in cake_overhead()
1390 last_len = skb->len - shinfo->gso_size * (segs - 1); in cake_overhead()