/drivers/net/ethernet/actions/ |
D | owl-emac.c | 206 ring->skbs[i] = skb; in owl_emac_ring_prepare_rx() 252 owl_emac_dma_unmap_rx(priv, ring->skbs[i], ring->skbs_dma[i]); in owl_emac_ring_unprepare_rx() 255 dev_kfree_skb(ring->skbs[i]); in owl_emac_ring_unprepare_rx() 256 ring->skbs[i] = NULL; in owl_emac_ring_unprepare_rx() 271 owl_emac_dma_unmap_tx(priv, ring->skbs[i], ring->skbs_dma[i]); in owl_emac_ring_unprepare_tx() 274 dev_kfree_skb(ring->skbs[i]); in owl_emac_ring_unprepare_tx() 275 ring->skbs[i] = NULL; in owl_emac_ring_unprepare_tx() 288 ring->skbs = devm_kcalloc(dev, size, sizeof(struct sk_buff *), in owl_emac_ring_alloc() 290 if (!ring->skbs) in owl_emac_ring_alloc() 530 ring->skbs[tx_head] = skb; in owl_emac_setup_frame_xmit() [all …]
|
D | owl-emac.h | 247 struct sk_buff **skbs; member
|
/drivers/net/ethernet/sfc/falcon/ |
D | selftest.c | 80 struct sk_buff **skbs; member 426 state->skbs[i] = skb; in ef4_begin_loopback() 480 skb = state->skbs[i]; in ef4_end_loopback() 536 state->skbs = kcalloc(state->packet_count, in ef4_test_loopback() 537 sizeof(state->skbs[0]), GFP_KERNEL); in ef4_test_loopback() 538 if (!state->skbs) in ef4_test_loopback() 559 kfree(state->skbs); in ef4_test_loopback()
|
/drivers/net/ethernet/sfc/ |
D | selftest.c | 83 struct sk_buff **skbs; member 424 state->skbs[i] = skb; in efx_begin_loopback() 478 skb = state->skbs[i]; in efx_end_loopback() 534 state->skbs = kcalloc(state->packet_count, in efx_test_loopback() 535 sizeof(state->skbs[0]), GFP_KERNEL); in efx_test_loopback() 536 if (!state->skbs) in efx_test_loopback() 557 kfree(state->skbs); in efx_test_loopback()
|
/drivers/net/ethernet/socionext/ |
D | sni_ave.c | 226 struct sk_buff *skbs; member 585 skb = priv->rx.desc[entry].skbs; in ave_rxdesc_prepare() 618 priv->rx.desc[entry].skbs = skb; in ave_rxdesc_prepare() 715 if (priv->tx.desc[done_idx].skbs) { in ave_tx_complete() 718 dev_consume_skb_any(priv->tx.desc[done_idx].skbs); in ave_tx_complete() 719 priv->tx.desc[done_idx].skbs = NULL; in ave_tx_complete() 776 skb = priv->rx.desc[proc_idx].skbs; in ave_rx_receive() 777 priv->rx.desc[proc_idx].skbs = NULL; in ave_rx_receive() 1371 if (!priv->tx.desc[entry].skbs) in ave_stop() 1375 dev_kfree_skb_any(priv->tx.desc[entry].skbs); in ave_stop() [all …]
|
/drivers/net/wireless/mediatek/mt7601u/ |
D | dma.c | 283 struct sk_buff_head skbs; in mt7601u_tx_tasklet() local 286 __skb_queue_head_init(&skbs); in mt7601u_tx_tasklet() 295 skb_queue_splice_init(&dev->tx_skb_done, &skbs); in mt7601u_tx_tasklet() 299 while (!skb_queue_empty(&skbs)) { in mt7601u_tx_tasklet() 300 struct sk_buff *skb = __skb_dequeue(&skbs); in mt7601u_tx_tasklet()
|
/drivers/net/ethernet/qlogic/qede/ |
D | qede_fp.c | 78 struct sk_buff *skb = txq->sw_tx_ring.skbs[idx].skb; in qede_free_tx_pkt() 83 bool data_split = txq->sw_tx_ring.skbs[idx].flags & QEDE_TSO_SPLIT_BD; in qede_free_tx_pkt() 123 txq->sw_tx_ring.skbs[idx].skb = NULL; in qede_free_tx_pkt() 124 txq->sw_tx_ring.skbs[idx].flags = 0; in qede_free_tx_pkt() 135 struct sk_buff *skb = txq->sw_tx_ring.skbs[idx].skb; in qede_free_failed_tx_pkt() 171 txq->sw_tx_ring.skbs[idx].skb = NULL; in qede_free_failed_tx_pkt() 172 txq->sw_tx_ring.skbs[idx].flags = 0; in qede_free_failed_tx_pkt() 1520 txq->sw_tx_ring.skbs[idx].skb = skb; in qede_start_xmit() 1642 txq->sw_tx_ring.skbs[idx].flags |= QEDE_TSO_SPLIT_BD; in qede_start_xmit()
|
D | qede.h | 433 struct sw_tx_bd *skbs; member
|
D | qede_main.c | 1582 kfree(txq->sw_tx_ring.skbs); in qede_free_mem_txq() 1609 size = sizeof(*txq->sw_tx_ring.skbs) * txq->num_tx_buffers; in qede_alloc_mem_txq() 1610 txq->sw_tx_ring.skbs = kzalloc(size, GFP_KERNEL); in qede_alloc_mem_txq() 1611 if (!txq->sw_tx_ring.skbs) in qede_alloc_mem_txq()
|
/drivers/net/wireless/intel/iwlwifi/dvm/ |
D | tx.c | 1128 struct sk_buff_head skbs; in iwlagn_rx_reply_tx() local 1149 __skb_queue_head_init(&skbs); in iwlagn_rx_reply_tx() 1178 iwl_trans_reclaim(priv->trans, txq_id, ssn, &skbs, false); in iwlagn_rx_reply_tx() 1183 skb_queue_walk(&skbs, skb) { in iwlagn_rx_reply_tx() 1254 while (!skb_queue_empty(&skbs)) { in iwlagn_rx_reply_tx() 1255 skb = __skb_dequeue(&skbs); in iwlagn_rx_reply_tx()
|
/drivers/net/ethernet/mediatek/ |
D | mtk_star_emac.c | 228 struct sk_buff *skbs[MTK_STAR_RING_NUM_DESCS]; member 301 desc_data->skb = ring->skbs[ring->tail]; in mtk_star_ring_pop_tail() 304 ring->skbs[ring->tail] = NULL; in mtk_star_ring_pop_tail() 325 ring->skbs[ring->head] = desc_data->skb; in mtk_star_ring_push_head() 676 ring->skbs[i] = skb; in mtk_star_prepare_rx_skbs() 696 desc_data.skb = ring->skbs[i]; in mtk_star_ring_free_skbs()
|
/drivers/net/ethernet/natsemi/ |
D | ns83820.c | 401 struct sk_buff *skbs[NR_RX_DESC]; member 524 BUG_ON(NULL != dev->rx_info.skbs[next_empty]); in ns83820_add_rx_skb() 525 dev->rx_info.skbs[next_empty] = skb; in ns83820_add_rx_skb() 790 struct sk_buff *skb = dev->rx_info.skbs[i]; in ns83820_cleanup_rx() 791 dev->rx_info.skbs[i] = NULL; in ns83820_cleanup_rx() 856 skb = info->skbs[next_rx]; 857 info->skbs[next_rx] = NULL;
|
/drivers/net/wireless/intel/iwlegacy/ |
D | 3945.c | 283 skb = txq->skbs[txq->q.read_ptr]; in il3945_tx_queue_reclaim() 285 txq->skbs[txq->q.read_ptr] = NULL; in il3945_tx_queue_reclaim() 332 info = IEEE80211_SKB_CB(txq->skbs[txq->q.read_ptr]); in il3945_hdl_tx() 667 if (txq->skbs) { in il3945_hw_txq_free_tfd() 668 struct sk_buff *skb = txq->skbs[txq->q.read_ptr]; in il3945_hw_txq_free_tfd() 673 txq->skbs[txq->q.read_ptr] = NULL; in il3945_hw_txq_free_tfd()
|
D | 4965-mac.c | 1767 txq->skbs[q->write_ptr] = skb; in il4965_tx_skb() 2470 skb = txq->skbs[txq->q.read_ptr]; in il4965_tx_queue_reclaim() 2481 txq->skbs[txq->q.read_ptr] = NULL; in il4965_tx_queue_reclaim() 2545 info = IEEE80211_SKB_CB(il->txq[scd_flow].skbs[agg->start_idx]); in il4965_tx_status_reply_compressed_ba() 2671 info = IEEE80211_SKB_CB(il->txq[txq_id].skbs[idx]); in il4965_tx_status_reply_tx() 2704 skb = il->txq[txq_id].skbs[idx]; in il4965_tx_status_reply_tx() 2782 skb = txq->skbs[txq->q.read_ptr]; in il4965_hdl_tx() 3942 if (txq->skbs) { in il4965_hw_txq_free_tfd() 3943 struct sk_buff *skb = txq->skbs[txq->q.read_ptr]; in il4965_hw_txq_free_tfd() 3948 txq->skbs[txq->q.read_ptr] = NULL; in il4965_hw_txq_free_tfd()
|
D | common.c | 2791 kfree(txq->skbs); in il_tx_queue_free() 2792 txq->skbs = NULL; in il_tx_queue_free() 2969 txq->skbs = kcalloc(TFD_QUEUE_SIZE_MAX, in il_tx_queue_alloc() 2972 if (!txq->skbs) { in il_tx_queue_alloc() 2977 txq->skbs = NULL; in il_tx_queue_alloc() 2991 kfree(txq->skbs); in il_tx_queue_alloc() 2992 txq->skbs = NULL; in il_tx_queue_alloc()
|
/drivers/net/wireless/intel/iwlwifi/ |
D | iwl-trans.h | 542 struct sk_buff_head *skbs, bool is_flush); 1125 int ssn, struct sk_buff_head *skbs, in iwl_trans_reclaim() argument 1133 trans->ops->reclaim(trans, queue, ssn, skbs, is_flush); in iwl_trans_reclaim()
|
/drivers/net/wireless/intel/iwlwifi/queue/ |
D | tx.h | 177 struct sk_buff_head *skbs, bool is_flush);
|
D | tx.c | 1523 struct sk_buff_head *skbs, bool is_flush) in iwl_txq_reclaim() argument 1565 if (WARN_ON(!skb_queue_empty(skbs))) in iwl_txq_reclaim() 1579 __skb_queue_tail(skbs, skb); in iwl_txq_reclaim() 1598 is_flush ? skbs : &overflow_skbs); in iwl_txq_reclaim()
|
/drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
D | chtls_main.c | 422 struct sk_buff **skbs, const __be64 *rsp) in chtls_recv() argument 424 struct sk_buff *skb = *skbs; in chtls_recv()
|
/drivers/net/ethernet/chelsio/cxgb3/ |
D | cxgb3_offload.c | 480 static int rx_offload_blackhole(struct t3cdev *dev, struct sk_buff **skbs, in rx_offload_blackhole() argument 484 dev_kfree_skb_any(skbs[n]); in rx_offload_blackhole() 1019 static int process_rx(struct t3cdev *dev, struct sk_buff **skbs, int n) in process_rx() argument 1022 struct sk_buff *skb = *skbs++; in process_rx()
|
D | sge.c | 1879 struct sk_buff *skbs[], int n) in deliver_partial_bundle() argument 1883 tdev->recv(tdev, skbs, n); in deliver_partial_bundle() 1906 struct sk_buff *skb, *tmp, *skbs[RX_BUNDLE_SIZE]; in ofld_poll() local 1928 skbs[ngathered] = skb; in ofld_poll() 1931 adapter->tdev.recv(&adapter->tdev, skbs, in ofld_poll() 1942 deliver_partial_bundle(&adapter->tdev, q, skbs, ngathered); in ofld_poll()
|
/drivers/net/wireless/intel/iwlwifi/mvm/ |
D | tx.c | 1425 struct sk_buff_head skbs; in iwl_mvm_rx_tx_cmd_single() local 1431 __skb_queue_head_init(&skbs); in iwl_mvm_rx_tx_cmd_single() 1439 iwl_trans_reclaim(mvm->trans, txq_id, ssn, &skbs, false); in iwl_mvm_rx_tx_cmd_single() 1441 while (!skb_queue_empty(&skbs)) { in iwl_mvm_rx_tx_cmd_single() 1442 struct sk_buff *skb = __skb_dequeue(&skbs); in iwl_mvm_rx_tx_cmd_single()
|
/drivers/net/ |
D | veth.c | 681 void *skbs[VETH_XDP_BATCH]; in veth_xdp_rcv_bulk_skb() local 684 if (xdp_alloc_skb_bulk(skbs, n_xdpf, in veth_xdp_rcv_bulk_skb() 694 struct sk_buff *skb = skbs[i]; in veth_xdp_rcv_bulk_skb()
|
/drivers/net/ethernet/sun/ |
D | cassini.c | 1827 struct sk_buff **skbs; in cas_tx_ringN() local 1833 skbs = cp->tx_skbs[ring]; in cas_tx_ringN() 1838 struct sk_buff *skb = skbs[entry]; in cas_tx_ringN() 1858 skbs[entry] = NULL; in cas_tx_ringN() 3869 struct sk_buff *skb, **skbs = cp->tx_skbs[ring]; in cas_clean_txd() local 3877 if (skbs[i] == NULL) in cas_clean_txd() 3880 skb = skbs[i]; in cas_clean_txd() 3881 skbs[i] = NULL; in cas_clean_txd()
|
/drivers/net/ethernet/pasemi/ |
D | pasemi_mac.c | 823 struct sk_buff *skbs[TX_CLEAN_BATCHSIZE]; in pasemi_mac_clean_tx() local 880 skbs[descr_count] = skb; in pasemi_mac_clean_tx() 894 pasemi_mac_unmap_tx_skb(mac, nf[i], skbs[i], dmas[i]); in pasemi_mac_clean_tx()
|