/drivers/net/ethernet/intel/iavf/ |
D | iavf_txrx.c | 224 total_packets += tx_buf->gso_segs; in iavf_clean_tx_irq() 1868 u16 gso_segs, gso_size; in iavf_tso() local 1936 gso_segs = skb_shinfo(skb)->gso_segs; in iavf_tso() 1939 first->gso_segs = gso_segs; in iavf_tso() 1940 first->bytecount += (first->gso_segs - 1) * *hdr_len; in iavf_tso() 2451 first->gso_segs = 1; in iavf_xmit_frame_ring()
|
D | iavf_txrx.h | 267 unsigned short gso_segs; member
|
/drivers/net/ethernet/mellanox/mlx5/core/ |
D | en_tx.c | 292 num_bytes = skb->len + (skb_shinfo(skb)->gso_segs - 1) * ihs; in mlx5e_sq_xmit() 293 stats->packets += skb_shinfo(skb)->gso_segs; in mlx5e_sq_xmit() 612 num_bytes = skb->len + (skb_shinfo(skb)->gso_segs - 1) * ihs; in mlx5i_sq_xmit() 613 stats->packets += skb_shinfo(skb)->gso_segs; in mlx5i_sq_xmit()
|
/drivers/net/ethernet/sun/ |
D | sunvnet_common.c | 1229 int gso_size, gso_type, gso_segs; in vnet_handle_offloads() local 1251 gso_segs = skb_shinfo(skb)->gso_segs; in vnet_handle_offloads() 1254 gso_segs = DIV_ROUND_UP(skb->len - hlen, datalen); in vnet_handle_offloads() 1256 if (unlikely(vnet_tx_dring_avail(dr) < gso_segs)) { in vnet_handle_offloads() 1261 if (vnet_tx_dring_avail(dr) < skb_shinfo(skb)->gso_segs) in vnet_handle_offloads() 1275 skb_shinfo(skb)->gso_segs = gso_segs; in vnet_handle_offloads() 1293 skb_shinfo(curr)->gso_segs = in vnet_handle_offloads()
|
/drivers/net/ethernet/intel/i40e/ |
D | i40e_txrx.c | 810 total_packets += tx_buf->gso_segs; in i40e_clean_tx_irq() 2902 u16 gso_segs, gso_size; in i40e_tso() local 2970 gso_segs = skb_shinfo(skb)->gso_segs; in i40e_tso() 2973 first->gso_segs = gso_segs; in i40e_tso() 2974 first->bytecount += (first->gso_segs - 1) * *hdr_len; in i40e_tso() 3523 tx_bi->gso_segs = 1; in i40e_xmit_xdp_ring() 3602 first->gso_segs = 1; in i40e_xmit_frame_ring()
|
D | i40e_txrx.h | 290 unsigned short gso_segs; member
|
/drivers/net/ethernet/intel/ice/ |
D | ice_txrx.h | 59 unsigned short gso_segs; member
|
D | ice_txrx.c | 137 total_pkts += tx_buf->gso_segs; in ice_clean_tx_irq() 1941 first->gso_segs = skb_shinfo(skb)->gso_segs; in ice_tso() 1942 first->bytecount += (first->gso_segs - 1) * off->header_len; in ice_tso() 2141 first->gso_segs = 1; in ice_xmit_frame_ring()
|
/drivers/net/ethernet/intel/fm10k/ |
D | fm10k_main.c | 782 first->gso_segs = skb_shinfo(skb)->gso_segs; in fm10k_tso() 783 first->bytecount += (first->gso_segs - 1) * hdrlen; in fm10k_tso() 1091 first->gso_segs = 1; in fm10k_xmit_frame_ring() 1225 total_packets += tx_buffer->gso_segs; in fm10k_clean_tx_irq()
|
D | fm10k.h | 66 u16 gso_segs; member
|
/drivers/net/ethernet/cavium/liquidio/ |
D | lio_vf_main.c | 54 u16 gso_segs; member 58 u16 gso_segs; 1578 tx_info->s.gso_segs = skb_shinfo(skb)->gso_segs; in liquidio_xmit() 1606 if (tx_info->s.gso_segs) in liquidio_xmit() 1607 stats->tx_done += tx_info->s.gso_segs; in liquidio_xmit()
|
D | lio_main.c | 120 u16 gso_segs; member 124 u16 gso_segs; 2514 tx_info->s.gso_segs = skb_shinfo(skb)->gso_segs; in liquidio_xmit() 2540 if (tx_info->s.gso_segs) in liquidio_xmit() 2541 stats->tx_done += tx_info->s.gso_segs; in liquidio_xmit()
|
/drivers/net/ethernet/intel/ixgbe/ |
D | ixgbe_xsk.c | 603 tx_bi->gso_segs = 1; in ixgbe_xmit_zc() 658 total_packets += tx_bi->gso_segs; in ixgbe_clean_xdp_tx_irq()
|
D | ixgbe_fcoe.c | 569 first->gso_segs = DIV_ROUND_UP(skb->len - *hdr_len, in ixgbe_fso() 571 first->bytecount += (first->gso_segs - 1) * *hdr_len; in ixgbe_fso()
|
D | ixgbe.h | 219 unsigned short gso_segs; member
|
/drivers/infiniband/sw/siw/ |
D | siw_qp_tx.c | 665 if (tp->gso_segs) { in siw_update_tcpseg() 667 c_tx->tcp_seglen = tp->mss_cache * tp->gso_segs; in siw_update_tcpseg() 671 min_t(u16, c_tx->gso_seg_limit, tp->gso_segs); in siw_update_tcpseg()
|
/drivers/net/ethernet/intel/ixgbevf/ |
D | ixgbevf.h | 38 unsigned short gso_segs; member
|
D | ixgbevf_main.c | 302 total_packets += tx_buffer->gso_segs; in ixgbevf_clean_tx_irq() 1009 tx_buffer->gso_segs = 1; in ixgbevf_xmit_xdp_ring() 3819 first->gso_segs = skb_shinfo(skb)->gso_segs; in ixgbevf_tso() 3820 first->bytecount += (first->gso_segs - 1) * *hdr_len; in ixgbevf_tso() 4158 first->gso_segs = 1; in ixgbevf_xmit_frame_ring()
|
/drivers/net/ethernet/intel/igc/ |
D | igc.h | 188 u16 gso_segs; member
|
/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
D | ipsec_rxtx.c | 178 MLX5E_IPSEC_ESN_SCOPE_MID < (oseq - skb_shinfo(skb)->gso_segs))) { in mlx5e_ipsec_set_iv_esn()
|
D | tls_rxtx.c | 171 skb_shinfo(nskb)->gso_segs = DIV_ROUND_UP(data_len, mss); in mlx5e_tls_complete_sync_skb()
|
D | ktls_tx.c | 480 stats->tls_encrypted_packets += skb_is_gso(skb) ? skb_shinfo(skb)->gso_segs : 1; in mlx5e_ktls_handle_tx_skb()
|
/drivers/net/ethernet/intel/igb/ |
D | igb.h | 205 u16 gso_segs; member
|
/drivers/infiniband/sw/rxe/ |
D | rxe_net.c | 315 skb_shinfo(skb)->gso_segs ?: 1); in prepare_ipv4_hdr()
|
/drivers/net/ethernet/sfc/ |
D | tx.c | 511 segments = skb_is_gso(skb) ? skb_shinfo(skb)->gso_segs : 0; in efx_enqueue_skb()
|