/net/core/ |
D | skbuff.c | 566 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_coalesce_rx_frag() 615 __skb_frag_unref(&shinfo->frags[i]); in skb_release_data() 794 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_dump() 1392 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; in skb_copy_ubufs() 1598 skb_shinfo(n)->frags[i] = skb_shinfo(skb)->frags[i]; in __pskb_copy_fclone() 1660 offsetof(struct skb_shared_info, frags[skb_shinfo(skb)->nr_frags])); in pskb_expand_head() 1962 int end = offset + skb_frag_size(&skb_shinfo(skb)->frags[i]); in ___pskb_trim() 1969 skb_frag_size_set(&skb_shinfo(skb)->frags[i++], len - offset); in ___pskb_trim() 2101 int size = skb_frag_size(&skb_shinfo(skb)->frags[i]); in __pskb_pull_tail() 2169 int size = skb_frag_size(&skb_shinfo(skb)->frags[i]); in __pskb_pull_tail() [all …]
|
D | tso.c | 55 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; in tso_build_data() 79 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; in tso_start()
|
D | pktgen.c | 2625 int frags = pkt_dev->nfrags; in pktgen_finalize_skb() local 2630 if (frags > MAX_SKB_FRAGS) in pktgen_finalize_skb() 2631 frags = MAX_SKB_FRAGS; in pktgen_finalize_skb() 2632 len = datalen - frags * PAGE_SIZE; in pktgen_finalize_skb() 2635 datalen = frags * PAGE_SIZE; in pktgen_finalize_skb() 2639 frag_len = (datalen/frags) < PAGE_SIZE ? in pktgen_finalize_skb() 2640 (datalen/frags) : PAGE_SIZE; in pktgen_finalize_skb() 2653 skb_frag_off_set(&skb_shinfo(skb)->frags[i], 0); in pktgen_finalize_skb() 2655 if (i == (frags - 1)) in pktgen_finalize_skb() 2656 skb_frag_size_set(&skb_shinfo(skb)->frags[i], in pktgen_finalize_skb() [all …]
|
D | datagram.c | 441 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in __skb_datagram_iter() 573 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_copy_datagram_from_iter()
|
D | dev.c | 3073 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in illegal_highdma() 5452 const skb_frag_t *frag0 = &pinfo->frags[0]; in skb_gro_reset_offset() 5480 skb_frag_off_add(&pinfo->frags[0], grow); in gro_pull_from_frag0() 5481 skb_frag_size_sub(&pinfo->frags[0], grow); in gro_pull_from_frag0() 5483 if (unlikely(!skb_frag_size(&pinfo->frags[0]))) { in gro_pull_from_frag0() 5485 memmove(pinfo->frags, pinfo->frags + 1, in gro_pull_from_frag0() 5486 --pinfo->nr_frags * sizeof(pinfo->frags[0])); in gro_pull_from_frag0()
|
/net/sunrpc/auth_gss/ |
D | gss_krb5_crypto.c | 561 struct scatterlist frags[4]; member 578 sg_set_page(&desc->frags[desc->fragno], sg_page(sg), sg->length, in decryptor() 589 sg_mark_end(&desc->frags[desc->fragno - 1]); in decryptor() 591 skcipher_request_set_crypt(desc->req, desc->frags, desc->frags, in decryptor() 598 sg_init_table(desc->frags, 4); in decryptor() 601 sg_set_page(&desc->frags[0], sg_page(sg), fraglen, in decryptor() 631 sg_init_table(desc.frags, 4); in gss_decrypt_xdr_buf() 905 sg_init_table(desc.frags, 4); in gss_krb5_aes_decrypt()
|
/net/sctp/ |
D | outqueue.c | 1202 union sctp_sack_variable *frags; in sctp_sack_update_unack_data() local 1208 frags = sack->variable; in sctp_sack_update_unack_data() 1210 unack_data -= ((ntohs(frags[i].gab.end) - in sctp_sack_update_unack_data() 1211 ntohs(frags[i].gab.start) + 1)); in sctp_sack_update_unack_data() 1229 union sctp_sack_variable *frags = sack->variable; in sctp_outq_sack() local 1290 highest_tsn += ntohs(frags[gap_ack_blocks - 1].gab.end); in sctp_outq_sack() 1746 union sctp_sack_variable *frags; in sctp_acked() local 1765 frags = sack->variable; in sctp_acked() 1769 if (tsn_offset >= ntohs(frags[i].gab.start) && in sctp_acked() 1770 tsn_offset <= ntohs(frags[i].gab.end)) in sctp_acked()
|
/net/ipv4/ |
D | tcp.c | 1040 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in do_tcp_sendpages() 1361 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in tcp_sendmsg_locked() 1760 const skb_frag_t *frags = NULL; in tcp_zerocopy_receive() local 1810 frags = skb_shinfo(skb)->frags; in tcp_zerocopy_receive() 1812 if (skb_frag_size(frags) > offset) in tcp_zerocopy_receive() 1814 offset -= skb_frag_size(frags); in tcp_zerocopy_receive() 1815 frags++; in tcp_zerocopy_receive() 1818 if (skb_frag_size(frags) != PAGE_SIZE || skb_frag_off(frags)) { in tcp_zerocopy_receive() 1821 while (remaining && (skb_frag_size(frags) != PAGE_SIZE || in tcp_zerocopy_receive() 1822 skb_frag_off(frags))) { in tcp_zerocopy_receive() [all …]
|
D | udp.c | 776 struct sk_buff *frags; in udp4_hwcsum() local 783 skb_walk_frags(skb, frags) { in udp4_hwcsum() 784 csum = csum_add(csum, frags->csum); in udp4_hwcsum() 785 hlen -= frags->len; in udp4_hwcsum()
|
D | tcp_output.c | 1419 int size = skb_frag_size(&shinfo->frags[i]); in __pskb_trim_head() 1425 shinfo->frags[k] = shinfo->frags[i]; in __pskb_trim_head() 1427 skb_frag_off_add(&shinfo->frags[k], eat); in __pskb_trim_head() 1428 skb_frag_size_sub(&shinfo->frags[k], eat); in __pskb_trim_head()
|
D | inet_fragment.c | 462 plen += skb_frag_size(&skb_shinfo(head)->frags[i]); in inet_frag_reasm_prepare()
|
/net/tipc/ |
D | msg.c | 254 int pktmax, struct sk_buff_head *frags) in tipc_msg_fragment() argument 282 __skb_queue_tail(frags, _skb); in tipc_msg_fragment() 296 __skb_queue_purge(frags); in tipc_msg_fragment() 297 __skb_queue_head_init(frags); in tipc_msg_fragment()
|
D | link.c | 1726 struct sk_buff_head tmpxq, tnlq, frags; in tipc_link_tnl_prepare() local 1737 __skb_queue_head_init(&frags); in tipc_link_tnl_prepare() 1806 &frags); in tipc_link_tnl_prepare() 1812 pktcnt += skb_queue_len(&frags) - 1; in tipc_link_tnl_prepare() 1814 skb_queue_splice_tail_init(&frags, &tnlq); in tipc_link_tnl_prepare()
|
D | msg.h | 1084 int pktmax, struct sk_buff_head *frags);
|
/net/tls/ |
D | tls_device.c | 131 __skb_frag_unref(&record->frags[i]); in destroy_record() 240 frag = &record->frags[record->num_frags - 1]; in tls_append_frag() 278 frag = &record->frags[i]; in tls_push_record() 317 tls_fill_prepend(ctx, skb_frag_address(&record->frags[0]), in tls_device_record_close() 334 frag = &record->frags[0]; in tls_create_new_record()
|
D | tls_device_fallback.c | 272 skb_frag_t *frag = &record->frags[i]; in fill_sg_in()
|
/net/ipv6/ |
D | udp.c | 1084 struct sk_buff *frags = skb_shinfo(skb)->frag_list; in udp6_hwcsum_outgoing() local 1087 if (!frags) { in udp6_hwcsum_outgoing() 1105 csum = csum_add(csum, frags->csum); in udp6_hwcsum_outgoing() 1106 } while ((frags = frags->next)); in udp6_hwcsum_outgoing()
|
/net/xfrm/ |
D | xfrm_ipcomp.c | 75 frag = skb_shinfo(skb)->frags + skb_shinfo(skb)->nr_frags; in ipcomp_decompress()
|
/net/mac80211/ |
D | ieee80211_i.h | 854 struct sk_buff_head frags; member 906 struct ieee80211_fragment_cache frags; member 1563 return !(skb_queue_empty(&txqi->frags) && !txqi->tin.backlog_packets); in txq_has_queue()
|
D | sta_info.h | 657 struct ieee80211_fragment_cache frags; member
|
D | tx.c | 1459 __skb_queue_head_init(&txqi->frags); in ieee80211_txq_init() 1501 ieee80211_purge_tx_queue(&local->hw, &txqi->frags); in ieee80211_txq_purge() 3606 skb = __skb_dequeue(&txqi->frags); in ieee80211_tx_dequeue() 3680 skb_queue_splice_tail(&tx.skbs, &txqi->frags); in ieee80211_tx_dequeue() 3782 (force || !skb_queue_empty(&txqi->frags) || in __ieee80211_schedule_txq()
|
D | iface.c | 1116 ieee80211_destroy_frag_cache(&sdata->frags); in ieee80211_teardown_sdata() 1826 ieee80211_init_frag_cache(&sdata->frags); in ieee80211_if_add()
|
D | sta_info.c | 381 ieee80211_init_frag_cache(&sta->frags); in sta_info_alloc() 1093 ieee80211_destroy_frag_cache(&sta->frags); in __sta_info_destroy_part2()
|
/net/kcm/ |
D | kcmsock.c | 646 frag = &skb_shinfo(skb)->frags[fragidx]; in kcm_write_msgs() 795 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], size); in kcm_sendpage() 1005 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in kcm_sendmsg()
|
/net/iucv/ |
D | af_iucv.c | 1226 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in iucv_sock_sendmsg() 1326 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in iucv_process_message()
|