Home
last modified time | relevance | path

Searched refs:frags (Results 1 – 25 of 32) sorted by relevance

12

/net/core/
Dskbuff.c566 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_coalesce_rx_frag()
615 __skb_frag_unref(&shinfo->frags[i]); in skb_release_data()
794 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_dump()
1392 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; in skb_copy_ubufs()
1598 skb_shinfo(n)->frags[i] = skb_shinfo(skb)->frags[i]; in __pskb_copy_fclone()
1660 offsetof(struct skb_shared_info, frags[skb_shinfo(skb)->nr_frags])); in pskb_expand_head()
1962 int end = offset + skb_frag_size(&skb_shinfo(skb)->frags[i]); in ___pskb_trim()
1969 skb_frag_size_set(&skb_shinfo(skb)->frags[i++], len - offset); in ___pskb_trim()
2101 int size = skb_frag_size(&skb_shinfo(skb)->frags[i]); in __pskb_pull_tail()
2169 int size = skb_frag_size(&skb_shinfo(skb)->frags[i]); in __pskb_pull_tail()
[all …]
Dtso.c55 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; in tso_build_data()
79 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; in tso_start()
Dpktgen.c2625 int frags = pkt_dev->nfrags; in pktgen_finalize_skb() local
2630 if (frags > MAX_SKB_FRAGS) in pktgen_finalize_skb()
2631 frags = MAX_SKB_FRAGS; in pktgen_finalize_skb()
2632 len = datalen - frags * PAGE_SIZE; in pktgen_finalize_skb()
2635 datalen = frags * PAGE_SIZE; in pktgen_finalize_skb()
2639 frag_len = (datalen/frags) < PAGE_SIZE ? in pktgen_finalize_skb()
2640 (datalen/frags) : PAGE_SIZE; in pktgen_finalize_skb()
2653 skb_frag_off_set(&skb_shinfo(skb)->frags[i], 0); in pktgen_finalize_skb()
2655 if (i == (frags - 1)) in pktgen_finalize_skb()
2656 skb_frag_size_set(&skb_shinfo(skb)->frags[i], in pktgen_finalize_skb()
[all …]
Ddatagram.c441 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in __skb_datagram_iter()
573 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_copy_datagram_from_iter()
Ddev.c3073 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in illegal_highdma()
5452 const skb_frag_t *frag0 = &pinfo->frags[0]; in skb_gro_reset_offset()
5480 skb_frag_off_add(&pinfo->frags[0], grow); in gro_pull_from_frag0()
5481 skb_frag_size_sub(&pinfo->frags[0], grow); in gro_pull_from_frag0()
5483 if (unlikely(!skb_frag_size(&pinfo->frags[0]))) { in gro_pull_from_frag0()
5485 memmove(pinfo->frags, pinfo->frags + 1, in gro_pull_from_frag0()
5486 --pinfo->nr_frags * sizeof(pinfo->frags[0])); in gro_pull_from_frag0()
/net/sunrpc/auth_gss/
Dgss_krb5_crypto.c561 struct scatterlist frags[4]; member
578 sg_set_page(&desc->frags[desc->fragno], sg_page(sg), sg->length, in decryptor()
589 sg_mark_end(&desc->frags[desc->fragno - 1]); in decryptor()
591 skcipher_request_set_crypt(desc->req, desc->frags, desc->frags, in decryptor()
598 sg_init_table(desc->frags, 4); in decryptor()
601 sg_set_page(&desc->frags[0], sg_page(sg), fraglen, in decryptor()
631 sg_init_table(desc.frags, 4); in gss_decrypt_xdr_buf()
905 sg_init_table(desc.frags, 4); in gss_krb5_aes_decrypt()
/net/sctp/
Doutqueue.c1202 union sctp_sack_variable *frags; in sctp_sack_update_unack_data() local
1208 frags = sack->variable; in sctp_sack_update_unack_data()
1210 unack_data -= ((ntohs(frags[i].gab.end) - in sctp_sack_update_unack_data()
1211 ntohs(frags[i].gab.start) + 1)); in sctp_sack_update_unack_data()
1229 union sctp_sack_variable *frags = sack->variable; in sctp_outq_sack() local
1290 highest_tsn += ntohs(frags[gap_ack_blocks - 1].gab.end); in sctp_outq_sack()
1746 union sctp_sack_variable *frags; in sctp_acked() local
1765 frags = sack->variable; in sctp_acked()
1769 if (tsn_offset >= ntohs(frags[i].gab.start) && in sctp_acked()
1770 tsn_offset <= ntohs(frags[i].gab.end)) in sctp_acked()
/net/ipv4/
Dtcp.c1040 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in do_tcp_sendpages()
1361 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in tcp_sendmsg_locked()
1760 const skb_frag_t *frags = NULL; in tcp_zerocopy_receive() local
1810 frags = skb_shinfo(skb)->frags; in tcp_zerocopy_receive()
1812 if (skb_frag_size(frags) > offset) in tcp_zerocopy_receive()
1814 offset -= skb_frag_size(frags); in tcp_zerocopy_receive()
1815 frags++; in tcp_zerocopy_receive()
1818 if (skb_frag_size(frags) != PAGE_SIZE || skb_frag_off(frags)) { in tcp_zerocopy_receive()
1821 while (remaining && (skb_frag_size(frags) != PAGE_SIZE || in tcp_zerocopy_receive()
1822 skb_frag_off(frags))) { in tcp_zerocopy_receive()
[all …]
Dudp.c776 struct sk_buff *frags; in udp4_hwcsum() local
783 skb_walk_frags(skb, frags) { in udp4_hwcsum()
784 csum = csum_add(csum, frags->csum); in udp4_hwcsum()
785 hlen -= frags->len; in udp4_hwcsum()
Dtcp_output.c1419 int size = skb_frag_size(&shinfo->frags[i]); in __pskb_trim_head()
1425 shinfo->frags[k] = shinfo->frags[i]; in __pskb_trim_head()
1427 skb_frag_off_add(&shinfo->frags[k], eat); in __pskb_trim_head()
1428 skb_frag_size_sub(&shinfo->frags[k], eat); in __pskb_trim_head()
Dinet_fragment.c462 plen += skb_frag_size(&skb_shinfo(head)->frags[i]); in inet_frag_reasm_prepare()
/net/tipc/
Dmsg.c254 int pktmax, struct sk_buff_head *frags) in tipc_msg_fragment() argument
282 __skb_queue_tail(frags, _skb); in tipc_msg_fragment()
296 __skb_queue_purge(frags); in tipc_msg_fragment()
297 __skb_queue_head_init(frags); in tipc_msg_fragment()
Dlink.c1726 struct sk_buff_head tmpxq, tnlq, frags; in tipc_link_tnl_prepare() local
1737 __skb_queue_head_init(&frags); in tipc_link_tnl_prepare()
1806 &frags); in tipc_link_tnl_prepare()
1812 pktcnt += skb_queue_len(&frags) - 1; in tipc_link_tnl_prepare()
1814 skb_queue_splice_tail_init(&frags, &tnlq); in tipc_link_tnl_prepare()
Dmsg.h1084 int pktmax, struct sk_buff_head *frags);
/net/tls/
Dtls_device.c131 __skb_frag_unref(&record->frags[i]); in destroy_record()
240 frag = &record->frags[record->num_frags - 1]; in tls_append_frag()
278 frag = &record->frags[i]; in tls_push_record()
317 tls_fill_prepend(ctx, skb_frag_address(&record->frags[0]), in tls_device_record_close()
334 frag = &record->frags[0]; in tls_create_new_record()
Dtls_device_fallback.c272 skb_frag_t *frag = &record->frags[i]; in fill_sg_in()
/net/ipv6/
Dudp.c1084 struct sk_buff *frags = skb_shinfo(skb)->frag_list; in udp6_hwcsum_outgoing() local
1087 if (!frags) { in udp6_hwcsum_outgoing()
1105 csum = csum_add(csum, frags->csum); in udp6_hwcsum_outgoing()
1106 } while ((frags = frags->next)); in udp6_hwcsum_outgoing()
/net/xfrm/
Dxfrm_ipcomp.c75 frag = skb_shinfo(skb)->frags + skb_shinfo(skb)->nr_frags; in ipcomp_decompress()
/net/mac80211/
Dieee80211_i.h854 struct sk_buff_head frags; member
906 struct ieee80211_fragment_cache frags; member
1563 return !(skb_queue_empty(&txqi->frags) && !txqi->tin.backlog_packets); in txq_has_queue()
Dsta_info.h657 struct ieee80211_fragment_cache frags; member
Dtx.c1459 __skb_queue_head_init(&txqi->frags); in ieee80211_txq_init()
1501 ieee80211_purge_tx_queue(&local->hw, &txqi->frags); in ieee80211_txq_purge()
3606 skb = __skb_dequeue(&txqi->frags); in ieee80211_tx_dequeue()
3680 skb_queue_splice_tail(&tx.skbs, &txqi->frags); in ieee80211_tx_dequeue()
3782 (force || !skb_queue_empty(&txqi->frags) || in __ieee80211_schedule_txq()
Diface.c1116 ieee80211_destroy_frag_cache(&sdata->frags); in ieee80211_teardown_sdata()
1826 ieee80211_init_frag_cache(&sdata->frags); in ieee80211_if_add()
Dsta_info.c381 ieee80211_init_frag_cache(&sta->frags); in sta_info_alloc()
1093 ieee80211_destroy_frag_cache(&sta->frags); in __sta_info_destroy_part2()
/net/kcm/
Dkcmsock.c646 frag = &skb_shinfo(skb)->frags[fragidx]; in kcm_write_msgs()
795 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], size); in kcm_sendpage()
1005 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in kcm_sendmsg()
/net/iucv/
Daf_iucv.c1226 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in iucv_sock_sendmsg()
1326 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in iucv_process_message()

12