/net/core/ |
D | skbuff.c | 619 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_coalesce_rx_frag() 673 __skb_frag_unref(&shinfo->frags[i], skb->pp_recycle); in skb_release_data() 874 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_dump() 1472 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; in skb_copy_ubufs() 1678 skb_shinfo(n)->frags[i] = skb_shinfo(skb)->frags[i]; in __pskb_copy_fclone() 1740 offsetof(struct skb_shared_info, frags[skb_shinfo(skb)->nr_frags])); in pskb_expand_head() 1843 offsetof(struct skb_shared_info, frags[shinfo->nr_frags])); in __skb_unclone_keeptruesize() 2124 int end = offset + skb_frag_size(&skb_shinfo(skb)->frags[i]); in ___pskb_trim() 2131 skb_frag_size_set(&skb_shinfo(skb)->frags[i++], len - offset); in ___pskb_trim() 2263 int size = skb_frag_size(&skb_shinfo(skb)->frags[i]); in __pskb_pull_tail() [all …]
|
D | tso.c | 62 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; in tso_build_data() 88 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; in tso_start()
|
D | datagram.c | 437 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in __skb_datagram_iter() 569 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_copy_datagram_from_iter() 661 skb_frag_t *last = &skb_shinfo(skb)->frags[frag - 1]; in __zerocopy_sg_from_iter()
|
D | pktgen.c | 2765 int frags = pkt_dev->nfrags; in pktgen_finalize_skb() local 2770 if (frags > MAX_SKB_FRAGS) in pktgen_finalize_skb() 2771 frags = MAX_SKB_FRAGS; in pktgen_finalize_skb() 2772 len = datalen - frags * PAGE_SIZE; in pktgen_finalize_skb() 2775 datalen = frags * PAGE_SIZE; in pktgen_finalize_skb() 2779 frag_len = (datalen/frags) < PAGE_SIZE ? in pktgen_finalize_skb() 2780 (datalen/frags) : PAGE_SIZE; in pktgen_finalize_skb() 2793 skb_frag_off_set(&skb_shinfo(skb)->frags[i], 0); in pktgen_finalize_skb() 2795 if (i == (frags - 1)) in pktgen_finalize_skb() 2796 skb_frag_size_set(&skb_shinfo(skb)->frags[i], in pktgen_finalize_skb() [all …]
|
D | dev.c | 3476 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in illegal_highdma() 6030 const skb_frag_t *frag0 = &pinfo->frags[0]; in skb_gro_reset_offset() 6057 skb_frag_off_add(&pinfo->frags[0], grow); in gro_pull_from_frag0() 6058 skb_frag_size_sub(&pinfo->frags[0], grow); in gro_pull_from_frag0() 6060 if (unlikely(!skb_frag_size(&pinfo->frags[0]))) { in gro_pull_from_frag0() 6062 memmove(pinfo->frags, pinfo->frags + 1, in gro_pull_from_frag0() 6063 --pinfo->nr_frags * sizeof(pinfo->frags[0])); in gro_pull_from_frag0()
|
/net/sunrpc/auth_gss/ |
D | gss_krb5_crypto.c | 427 struct scatterlist frags[4]; member 444 sg_set_page(&desc->frags[desc->fragno], sg_page(sg), sg->length, in decryptor() 455 sg_mark_end(&desc->frags[desc->fragno - 1]); in decryptor() 457 skcipher_request_set_crypt(desc->req, desc->frags, desc->frags, in decryptor() 464 sg_init_table(desc->frags, 4); in decryptor() 467 sg_set_page(&desc->frags[0], sg_page(sg), fraglen, in decryptor() 497 sg_init_table(desc.frags, 4); in gss_decrypt_xdr_buf() 771 sg_init_table(desc.frags, 4); in gss_krb5_aes_decrypt()
|
/net/sctp/ |
D | outqueue.c | 1225 union sctp_sack_variable *frags; in sctp_sack_update_unack_data() local 1231 frags = sack->variable; in sctp_sack_update_unack_data() 1233 unack_data -= ((ntohs(frags[i].gab.end) - in sctp_sack_update_unack_data() 1234 ntohs(frags[i].gab.start) + 1)); in sctp_sack_update_unack_data() 1252 union sctp_sack_variable *frags = sack->variable; in sctp_outq_sack() local 1314 highest_tsn += ntohs(frags[gap_ack_blocks - 1].gab.end); in sctp_outq_sack() 1770 union sctp_sack_variable *frags; in sctp_acked() local 1789 frags = sack->variable; in sctp_acked() 1793 if (tsn_offset >= ntohs(frags[i].gab.start) && in sctp_acked() 1794 tsn_offset <= ntohs(frags[i].gab.end)) in sctp_acked()
|
/net/tipc/ |
D | msg.c | 307 int pktmax, struct sk_buff_head *frags) in tipc_msg_fragment() argument 335 __skb_queue_tail(frags, _skb); in tipc_msg_fragment() 349 __skb_queue_purge(frags); in tipc_msg_fragment() 350 __skb_queue_head_init(frags); in tipc_msg_fragment()
|
D | link.c | 1979 struct sk_buff_head tmpxq, tnlq, frags; in tipc_link_tnl_prepare() local 2015 __skb_queue_head_init(&frags); in tipc_link_tnl_prepare() 2058 &frags); in tipc_link_tnl_prepare() 2064 pktcnt += skb_queue_len(&frags) - 1; in tipc_link_tnl_prepare() 2066 skb_queue_splice_tail_init(&frags, &tnlq); in tipc_link_tnl_prepare()
|
D | msg.h | 1193 int pktmax, struct sk_buff_head *frags);
|
/net/tls/ |
D | tls_device.c | 131 __skb_frag_unref(&record->frags[i], false); in destroy_record() 251 frag = &record->frags[record->num_frags - 1]; in tls_append_frag() 289 frag = &record->frags[i]; in tls_push_record() 326 tls_fill_prepend(ctx, skb_frag_address(&record->frags[0]), in tls_device_record_close() 342 frag = &record->frags[0]; in tls_create_new_record()
|
D | tls_device_fallback.c | 272 skb_frag_t *frag = &record->frags[i]; in fill_sg_in()
|
/net/ipv4/ |
D | tcp.c | 1008 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in tcp_build_frag() 1367 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in tcp_sendmsg_locked() 1786 frag = skb_shinfo(skb)->frags; in skb_advance_to_frag() 1849 if (frag == &info->frags[info->nr_frags - 1]) in tcp_zerocopy_set_hint_for_skb() 2076 const skb_frag_t *frags = NULL; in tcp_zerocopy_receive() local 2148 frags = skb_advance_to_frag(skb, offset, &offset_frag); in tcp_zerocopy_receive() 2149 if (!frags || offset_frag) in tcp_zerocopy_receive() 2153 mappable_offset = find_next_mappable_frag(frags, in tcp_zerocopy_receive() 2159 page = skb_frag_page(frags); in tcp_zerocopy_receive() 2164 frags++; in tcp_zerocopy_receive() [all …]
|
D | tcp_output.c | 1645 int size = skb_frag_size(&shinfo->frags[i]); in __pskb_trim_head() 1651 shinfo->frags[k] = shinfo->frags[i]; in __pskb_trim_head() 1653 skb_frag_off_add(&shinfo->frags[k], eat); in __pskb_trim_head() 1654 skb_frag_size_sub(&shinfo->frags[k], eat); in __pskb_trim_head()
|
D | inet_fragment.c | 493 plen += skb_frag_size(&skb_shinfo(head)->frags[i]); in inet_frag_reasm_prepare()
|
D | udp.c | 854 struct sk_buff *frags; in udp4_hwcsum() local 861 skb_walk_frags(skb, frags) { in udp4_hwcsum() 862 csum = csum_add(csum, frags->csum); in udp4_hwcsum() 863 hlen -= frags->len; in udp4_hwcsum()
|
/net/ipv6/ |
D | udp.c | 1176 struct sk_buff *frags = skb_shinfo(skb)->frag_list; in udp6_hwcsum_outgoing() local 1179 if (!frags) { in udp6_hwcsum_outgoing() 1197 csum = csum_add(csum, frags->csum); in udp6_hwcsum_outgoing() 1198 } while ((frags = frags->next)); in udp6_hwcsum_outgoing()
|
/net/xfrm/ |
D | xfrm_ipcomp.c | 71 frag = skb_shinfo(skb)->frags + skb_shinfo(skb)->nr_frags; in ipcomp_decompress()
|
/net/kcm/ |
D | kcmsock.c | 644 frag = &skb_shinfo(skb)->frags[fragidx]; in kcm_write_msgs() 793 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], size); in kcm_sendpage() 1003 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in kcm_sendmsg()
|
/net/mac80211/ |
D | sta_info.h | 677 struct ieee80211_fragment_cache frags; member
|
D | ieee80211_i.h | 869 struct sk_buff_head frags; member 919 struct ieee80211_fragment_cache frags; member 1641 return !(skb_queue_empty(&txqi->frags) && !txqi->tin.backlog_packets); in txq_has_queue()
|
D | iface.c | 689 ieee80211_destroy_frag_cache(&sdata->frags); in ieee80211_teardown_sdata() 1992 ieee80211_init_frag_cache(&sdata->frags); in ieee80211_if_add()
|
D | tx.c | 1435 __skb_queue_tail(&txqi->frags, skb); in ieee80211_txq_enqueue() 1484 __skb_queue_head_init(&txqi->frags); in ieee80211_txq_init() 1526 ieee80211_purge_tx_queue(&local->hw, &txqi->frags); in ieee80211_txq_purge() 3657 skb = __skb_dequeue(&txqi->frags); in ieee80211_tx_dequeue() 3754 skb_queue_splice_tail(&tx.skbs, &txqi->frags); in ieee80211_tx_dequeue()
|
D | sta_info.c | 395 ieee80211_init_frag_cache(&sta->frags); in sta_info_alloc() 1105 ieee80211_destroy_frag_cache(&sta->frags); in __sta_info_destroy_part2()
|
/net/iucv/ |
D | af_iucv.c | 1067 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in iucv_sock_sendmsg() 1169 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in iucv_process_message()
|