Home
last modified time | relevance | path

Searched refs:tx_q (Results 1 – 25 of 30) sorted by relevance

12

/kernel/linux/linux-5.10/drivers/net/ethernet/stmicro/stmmac/
Dring_mode.c19 struct stmmac_tx_queue *tx_q = (struct stmmac_tx_queue *)p; in jumbo_frm() local
21 struct stmmac_priv *priv = tx_q->priv_data; in jumbo_frm()
22 unsigned int entry = tx_q->cur_tx; in jumbo_frm()
27 desc = (struct dma_desc *)(tx_q->dma_etx + entry); in jumbo_frm()
29 desc = tx_q->dma_tx + entry; in jumbo_frm()
46 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm()
47 tx_q->tx_skbuff_dma[entry].len = bmax; in jumbo_frm()
48 tx_q->tx_skbuff_dma[entry].is_jumbo = true; in jumbo_frm()
53 tx_q->tx_skbuff[entry] = NULL; in jumbo_frm()
57 desc = (struct dma_desc *)(tx_q->dma_etx + entry); in jumbo_frm()
[all …]
Dchain_mode.c19 struct stmmac_tx_queue *tx_q = (struct stmmac_tx_queue *)p; in jumbo_frm() local
21 struct stmmac_priv *priv = tx_q->priv_data; in jumbo_frm()
22 unsigned int entry = tx_q->cur_tx; in jumbo_frm()
27 desc = tx_q->dma_tx + entry; in jumbo_frm()
41 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm()
42 tx_q->tx_skbuff_dma[entry].len = bmax; in jumbo_frm()
48 tx_q->tx_skbuff[entry] = NULL; in jumbo_frm()
50 desc = tx_q->dma_tx + entry; in jumbo_frm()
59 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm()
60 tx_q->tx_skbuff_dma[entry].len = bmax; in jumbo_frm()
[all …]
Dstmmac_main.c298 struct stmmac_tx_queue *tx_q = &priv->tx_queue[queue]; in stmmac_tx_avail() local
301 if (tx_q->dirty_tx > tx_q->cur_tx) in stmmac_tx_avail()
302 avail = tx_q->dirty_tx - tx_q->cur_tx - 1; in stmmac_tx_avail()
304 avail = priv->dma_tx_size - tx_q->cur_tx + tx_q->dirty_tx - 1; in stmmac_tx_avail()
340 struct stmmac_tx_queue *tx_q = &priv->tx_queue[queue]; in stmmac_enable_eee_mode() local
342 if (tx_q->dirty_tx != tx_q->cur_tx) in stmmac_enable_eee_mode()
1205 struct stmmac_tx_queue *tx_q = &priv->tx_queue[queue]; in stmmac_display_tx_rings() local
1210 head_tx = (void *)tx_q->dma_etx; in stmmac_display_tx_rings()
1212 } else if (tx_q->tbs & STMMAC_TBS_AVAIL) { in stmmac_display_tx_rings()
1213 head_tx = (void *)tx_q->dma_entx; in stmmac_display_tx_rings()
[all …]
/kernel/linux/linux-5.10/drivers/net/ethernet/qualcomm/emac/
Demac-mac.c305 writel(upper_32_bits(adpt->tx_q.tpd.dma_addr), in emac_mac_dma_rings_config()
308 writel(lower_32_bits(adpt->tx_q.tpd.dma_addr), in emac_mac_dma_rings_config()
311 writel(adpt->tx_q.tpd.count & TPD_RING_SIZE_BMSK, in emac_mac_dma_rings_config()
587 struct emac_tx_queue *tx_q = &adpt->tx_q; in emac_tx_q_descs_free() local
592 if (!tx_q->tpd.tpbuff) in emac_tx_q_descs_free()
595 for (i = 0; i < tx_q->tpd.count; i++) { in emac_tx_q_descs_free()
596 struct emac_buffer *tpbuf = GET_TPD_BUFFER(tx_q, i); in emac_tx_q_descs_free()
610 size = sizeof(struct emac_buffer) * tx_q->tpd.count; in emac_tx_q_descs_free()
611 memset(tx_q->tpd.tpbuff, 0, size); in emac_tx_q_descs_free()
614 memset(tx_q->tpd.v_addr, 0, tx_q->tpd.size); in emac_tx_q_descs_free()
[all …]
Demac-mac.h231 struct emac_tx_queue *tx_q,
233 void emac_mac_tx_process(struct emac_adapter *adpt, struct emac_tx_queue *tx_q);
Demac.h341 struct emac_tx_queue tx_q; member
Demac.c123 return emac_mac_tx_buf_send(adpt, &adpt->tx_q, skb); in emac_start_xmit()
161 emac_mac_tx_process(adpt, &adpt->tx_q); in emac_isr()
/kernel/linux/linux-5.10/drivers/net/ethernet/intel/ice/
Dice_dcb_lib.h46 static inline bool ice_find_q_in_range(u16 low, u16 high, unsigned int tx_q) in ice_find_q_in_range() argument
48 return (tx_q >= low) && (tx_q < high); in ice_find_q_in_range()
/kernel/linux/linux-5.10/net/nfc/nci/
Duart.c56 skb = skb_dequeue(&nu->tx_q); in nci_uart_dequeue()
68 return skb_queue_empty(&nu->tx_q); in nci_uart_queue_empty()
135 skb_queue_head_init(&nu->tx_q); in nci_uart_set_driver()
198 skb_queue_purge(&nu->tx_q); in nci_uart_tty_close()
316 skb_queue_tail(&nu->tx_q, skb); in nci_uart_send()
Ddata.c146 spin_lock_irqsave(&ndev->tx_q.lock, flags); in nci_queue_tx_data_frags()
149 __skb_queue_tail(&ndev->tx_q, skb_frag); in nci_queue_tx_data_frags()
151 spin_unlock_irqrestore(&ndev->tx_q.lock, flags); in nci_queue_tx_data_frags()
185 skb_queue_tail(&ndev->tx_q, skb); in nci_send_data()
Dcore.c530 skb_queue_purge(&ndev->tx_q); in nci_open_device()
563 skb_queue_purge(&ndev->tx_q); in nci_close_device()
1238 skb_queue_head_init(&ndev->tx_q); in nci_register_device()
1462 skb = skb_dequeue(&ndev->tx_q); in nci_tx_work()
Dntf.c61 if (!skb_queue_empty(&ndev->tx_q)) in nci_core_conn_credits_ntf_packet()
685 skb_queue_purge(&ndev->tx_q); in nci_rf_deactivate_ntf_packet()
/kernel/linux/linux-5.10/drivers/net/wireless/mediatek/mt7601u/
Ddma.c300 struct mt7601u_tx_queue *q = &dev->tx_q[ep]; in mt7601u_dma_submit_tx()
465 if (!dev->tx_q) in mt7601u_free_tx()
469 mt7601u_free_tx_queue(&dev->tx_q[i]); in mt7601u_free_tx()
493 dev->tx_q = devm_kcalloc(dev->dev, __MT_EP_OUT_MAX, in mt7601u_alloc_tx()
494 sizeof(*dev->tx_q), GFP_KERNEL); in mt7601u_alloc_tx()
495 if (!dev->tx_q) in mt7601u_alloc_tx()
499 if (mt7601u_alloc_tx_queue(dev, &dev->tx_q[i])) in mt7601u_alloc_tx()
Dmt7601u.h201 struct mt7601u_tx_queue *tx_q; member
/kernel/linux/linux-5.10/drivers/net/fddi/skfp/h/
Dhwmtm.h242 #define HWM_GET_TX_USED(smc,queue) (int) (smc)->hw.fp.tx_q[queue].tx_used
261 (smc)->hw.fp.tx_q[queue].tx_curr_put
Dfplustm.h191 struct s_smt_tx_queue tx_q[USED_QUEUES] ; member
/kernel/linux/linux-5.10/drivers/net/wireless/mediatek/mt76/mt7615/
Dmain.c34 if (msta && dev->pm.tx_q[i].msta != msta) in mt7615_free_pending_tx_skbs()
37 dev_kfree_skb(dev->pm.tx_q[i].skb); in mt7615_free_pending_tx_skbs()
38 dev->pm.tx_q[i].skb = NULL; in mt7615_free_pending_tx_skbs()
751 if (!dev->pm.tx_q[qid].skb) { in mt7615_tx()
753 dev->pm.tx_q[qid].msta = msta; in mt7615_tx()
754 dev->pm.tx_q[qid].skb = skb; in mt7615_tx()
Dmac.c1867 struct mt7615_sta *msta = dev->pm.tx_q[i].msta; in mt7615_pm_wake_work()
1871 if (!dev->pm.tx_q[i].skb) in mt7615_pm_wake_work()
1879 mt76_tx(mphy, sta, wcid, dev->pm.tx_q[i].skb); in mt7615_pm_wake_work()
1880 dev->pm.tx_q[i].skb = NULL; in mt7615_pm_wake_work()
/kernel/linux/linux-5.10/drivers/media/radio/wl128x/
Dfmdrv.h205 struct sk_buff_head tx_q; /* TX queue */ member
Dfmdrv_common.c327 if (!skb_queue_empty(&fmdev->tx_q)) in recv_tasklet()
351 skb = skb_dequeue(&fmdev->tx_q); in send_tasklet()
441 skb_queue_tail(&fmdev->tx_q, skb); in fm_send_cmd()
1541 skb_queue_head_init(&fmdev->tx_q); in fmc_prepare()
1595 skb_queue_purge(&fmdev->tx_q); in fmc_release()
/kernel/linux/linux-5.10/net/bluetooth/
Dl2cap_core.c470 skb_queue_head_init(&chan->tx_q); in l2cap_chan_create()
569 skb_queue_head_init(&chan->tx_q); in l2cap_le_flowctl_init()
690 skb_queue_purge(&chan->tx_q); in l2cap_chan_del()
705 skb_queue_purge(&chan->tx_q); in l2cap_chan_del()
1321 skb_queue_walk(&chan->tx_q, skb) { in l2cap_move_setup()
2073 skb_queue_splice_tail_init(skbs, &chan->tx_q); in l2cap_streaming_send()
2075 while (!skb_queue_empty(&chan->tx_q)) { in l2cap_streaming_send()
2077 skb = skb_dequeue(&chan->tx_q); in l2cap_streaming_send()
2156 if (skb_queue_is_last(&chan->tx_q, skb)) in l2cap_ertm_send()
2159 chan->tx_send_head = skb_queue_next(&chan->tx_q, skb); in l2cap_ertm_send()
[all …]
/kernel/linux/linux-5.10/include/net/nfc/
Dnci_core.h227 struct sk_buff_head tx_q; member
454 struct sk_buff_head tx_q; member
/kernel/linux/linux-5.10/drivers/net/ethernet/chelsio/cxgb3/
Dadapter.h208 struct netdev_queue *tx_q; /* associated netdev TX queue */ member
Dsge.c1994 netif_tx_wake_queue(qs->tx_q); in restart_tx()
2930 if (__netif_tx_trylock(qs->tx_q)) { in sge_timer_tx()
2933 __netif_tx_unlock(qs->tx_q); in sge_timer_tx()
3171 q->tx_q = netdevq; in t3_sge_alloc_qset()
/kernel/linux/linux-5.10/drivers/net/fddi/skfp/
Dfplustm.c305 smc->hw.fp.tx[QUEUE_S] = queue = &smc->hw.fp.tx_q[QUEUE_S] ; in init_tx()
316 smc->hw.fp.tx[QUEUE_A0] = queue = &smc->hw.fp.tx_q[QUEUE_A0] ; in init_tx()

12