/drivers/net/wireless/ath/ath9k/ |
D | xmit.c | 50 static void ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq, 53 int tx_flags, struct ath_txq *txq); 55 struct ath_txq *txq, struct list_head *bf_q, 57 static void ath_tx_txqaddbuf(struct ath_softc *sc, struct ath_txq *txq, 65 struct ath_txq *txq, 107 static void ath_txq_lock(struct ath_softc *sc, struct ath_txq *txq) in ath_txq_lock() argument 108 __acquires(&txq->axq_lock) in ath_txq_lock() 110 spin_lock_bh(&txq->axq_lock); in ath_txq_lock() 113 static void ath_txq_unlock(struct ath_softc *sc, struct ath_txq *txq) in ath_txq_unlock() argument 114 __releases(&txq->axq_lock) in ath_txq_unlock() [all …]
|
/drivers/net/wireless/iwmc3200wifi/ |
D | tx.c | 160 queue_work(iwm->txq[queue].wq, &iwm->txq[queue].worker); in iwm_tx_credit_inc() 337 struct iwm_tx_queue *txq) in iwm_tx_send_concat_packets() argument 341 if (!txq->concat_count) in iwm_tx_send_concat_packets() 345 txq->id, txq->concat_count); in iwm_tx_send_concat_packets() 348 iwm_udma_wifi_hdr_set_eop(iwm, txq->concat_ptr, 1); in iwm_tx_send_concat_packets() 350 trace_iwm_tx_packets(iwm, txq->concat_buf, txq->concat_count); in iwm_tx_send_concat_packets() 351 ret = iwm_bus_send_chunk(iwm, txq->concat_buf, txq->concat_count); in iwm_tx_send_concat_packets() 353 txq->concat_count = 0; in iwm_tx_send_concat_packets() 354 txq->concat_ptr = txq->concat_buf; in iwm_tx_send_concat_packets() 364 struct iwm_tx_queue *txq; in iwm_tx_worker() local [all …]
|
D | debugfs.c | 122 struct iwm_tx_queue *txq = &iwm->txq[i]; in iwm_debugfs_txq_read() local 127 spin_lock_irqsave(&txq->queue.lock, flags); in iwm_debugfs_txq_read() 129 skb = (struct sk_buff *)&txq->queue; in iwm_debugfs_txq_read() 134 txq->id)); in iwm_debugfs_txq_read() 136 txq->concat_count); in iwm_debugfs_txq_read() 138 skb_queue_len(&txq->queue)); in iwm_debugfs_txq_read() 139 for (j = 0; j < skb_queue_len(&txq->queue); j++) { in iwm_debugfs_txq_read() 155 spin_unlock_irqrestore(&txq->queue.lock, flags); in iwm_debugfs_txq_read() 157 spin_lock_irqsave(&txq->stopped_queue.lock, flags); in iwm_debugfs_txq_read() 161 skb_queue_len(&txq->stopped_queue)); in iwm_debugfs_txq_read() [all …]
|
D | main.c | 294 INIT_WORK(&iwm->txq[i].worker, iwm_tx_worker); in iwm_priv_init() 296 iwm->txq[i].id = i; in iwm_priv_init() 297 iwm->txq[i].wq = create_singlethread_workqueue(name); in iwm_priv_init() 298 if (!iwm->txq[i].wq) in iwm_priv_init() 301 skb_queue_head_init(&iwm->txq[i].queue); in iwm_priv_init() 302 skb_queue_head_init(&iwm->txq[i].stopped_queue); in iwm_priv_init() 303 spin_lock_init(&iwm->txq[i].lock); in iwm_priv_init() 335 destroy_workqueue(iwm->txq[i].wq); in iwm_priv_deinit() 589 skb_queue_purge(&iwm->txq[i].queue); in iwm_link_off() 590 skb_queue_purge(&iwm->txq[i].stopped_queue); in iwm_link_off() [all …]
|
/drivers/net/wireless/iwlwifi/ |
D | iwl-trans-pcie-tx.c | 85 struct iwl_tx_queue *txq, in iwl_trans_txq_update_byte_cnt_tbl() argument 91 int write_ptr = txq->q.write_ptr; in iwl_trans_txq_update_byte_cnt_tbl() 92 int txq_id = txq->q.id; in iwl_trans_txq_update_byte_cnt_tbl() 98 (struct iwl_tx_cmd *) txq->cmd[txq->q.write_ptr]->payload; in iwl_trans_txq_update_byte_cnt_tbl() 131 void iwl_txq_update_write_ptr(struct iwl_trans *trans, struct iwl_tx_queue *txq) in iwl_txq_update_write_ptr() argument 134 int txq_id = txq->q.id; in iwl_txq_update_write_ptr() 136 if (txq->need_update == 0) in iwl_txq_update_write_ptr() 142 txq->q.write_ptr | (txq_id << 8)); in iwl_txq_update_write_ptr() 161 txq->q.write_ptr | (txq_id << 8)); in iwl_txq_update_write_ptr() 170 txq->q.write_ptr | (txq_id << 8)); in iwl_txq_update_write_ptr() [all …]
|
D | iwl-trans-pcie.c | 303 struct iwl_tx_queue *txq, int slots_num, in iwl_trans_txq_alloc() argument 310 if (WARN_ON(txq->meta || txq->cmd || txq->skbs || txq->tfds)) in iwl_trans_txq_alloc() 313 txq->q.n_window = slots_num; in iwl_trans_txq_alloc() 315 txq->meta = kcalloc(slots_num, sizeof(txq->meta[0]), GFP_KERNEL); in iwl_trans_txq_alloc() 316 txq->cmd = kcalloc(slots_num, sizeof(txq->cmd[0]), GFP_KERNEL); in iwl_trans_txq_alloc() 318 if (!txq->meta || !txq->cmd) in iwl_trans_txq_alloc() 323 txq->cmd[i] = kmalloc(sizeof(struct iwl_device_cmd), in iwl_trans_txq_alloc() 325 if (!txq->cmd[i]) in iwl_trans_txq_alloc() 333 txq->skbs = kcalloc(TFD_QUEUE_SIZE_MAX, sizeof(txq->skbs[0]), in iwl_trans_txq_alloc() 335 if (!txq->skbs) { in iwl_trans_txq_alloc() [all …]
|
D | iwl-trans-pcie-int.h | 280 struct iwl_tx_queue *txq; member 323 struct iwl_tx_queue *txq); 325 struct iwl_tx_queue *txq, 332 struct iwl_tx_queue *txq, 338 struct iwl_tx_queue *txq, 344 void iwlagn_txq_free_tfd(struct iwl_trans *trans, struct iwl_tx_queue *txq, 402 static inline void iwl_set_swq_id(struct iwl_tx_queue *txq, u8 ac, u8 hwq) in iwl_set_swq_id() argument 407 txq->swq_id = (hwq << 2) | ac; in iwl_set_swq_id() 410 static inline u8 iwl_get_queue_ac(struct iwl_tx_queue *txq) in iwl_get_queue_ac() argument 412 return txq->swq_id & 0x3; in iwl_get_queue_ac() [all …]
|
/drivers/net/ethernet/marvell/ |
D | mv643xx_eth.c | 433 struct tx_queue txq[8]; member 465 static struct mv643xx_eth_private *txq_to_mp(struct tx_queue *txq) in txq_to_mp() argument 467 return container_of(txq, struct mv643xx_eth_private, txq[txq->index]); in txq_to_mp() 486 static void txq_reset_hw_ptr(struct tx_queue *txq) in txq_reset_hw_ptr() argument 488 struct mv643xx_eth_private *mp = txq_to_mp(txq); in txq_reset_hw_ptr() 491 addr = (u32)txq->tx_desc_dma; in txq_reset_hw_ptr() 492 addr += txq->tx_curr_desc * sizeof(struct tx_desc); in txq_reset_hw_ptr() 493 wrlp(mp, TXQ_CURRENT_DESC_PTR(txq->index), addr); in txq_reset_hw_ptr() 496 static void txq_enable(struct tx_queue *txq) in txq_enable() argument 498 struct mv643xx_eth_private *mp = txq_to_mp(txq); in txq_enable() [all …]
|
/drivers/net/ethernet/chelsio/cxgb4vf/ |
D | sge.c | 1048 static void txq_stop(struct sge_eth_txq *txq) in txq_stop() argument 1050 netif_tx_stop_queue(txq->txq); in txq_stop() 1051 txq->q.stops++; in txq_stop() 1079 struct sge_eth_txq *txq; in t4vf_eth_xmit() local 1106 txq = &adapter->sge.ethtxq[pi->first_qset + qidx]; in t4vf_eth_xmit() 1112 reclaim_completed_tx(adapter, &txq->q, true); in t4vf_eth_xmit() 1121 credits = txq_avail(&txq->q) - ndesc; in t4vf_eth_xmit() 1130 txq_stop(txq); in t4vf_eth_xmit() 1144 txq->mapping_err++; in t4vf_eth_xmit() 1159 txq_stop(txq); in t4vf_eth_xmit() [all …]
|
/drivers/net/wireless/ath/ath5k/ |
D | base.c | 664 struct ath5k_txq *txq, int padsize) in ath5k_txbuf_setup() argument 750 spin_lock_bh(&txq->lock); in ath5k_txbuf_setup() 751 list_add_tail(&bf->list, &txq->q); in ath5k_txbuf_setup() 752 txq->txq_len++; in ath5k_txbuf_setup() 753 if (txq->link == NULL) /* is this first packet? */ in ath5k_txbuf_setup() 754 ath5k_hw_set_txdp(ah, txq->qnum, bf->daddr); in ath5k_txbuf_setup() 756 *txq->link = bf->daddr; in ath5k_txbuf_setup() 758 txq->link = &ds->ds_link; in ath5k_txbuf_setup() 759 ath5k_hw_start_tx_dma(ah, txq->qnum); in ath5k_txbuf_setup() 761 spin_unlock_bh(&txq->lock); in ath5k_txbuf_setup() [all …]
|
/drivers/net/ethernet/chelsio/cxgb3/ |
D | sge.c | 179 return container_of(q, struct sge_qset, txq[qidx]); in txq_to_qset() 654 memset(q->txq, 0, sizeof(struct sge_txq) * SGE_TXQ_PER_SET); in t3_reset_qset() 691 if (q->txq[i].desc) { in t3_free_qset() 693 t3_sge_enable_ecntxt(adapter, q->txq[i].cntxt_id, 0); in t3_free_qset() 695 if (q->txq[i].sdesc) { in t3_free_qset() 696 free_tx_desc(adapter, &q->txq[i], in t3_free_qset() 697 q->txq[i].in_use); in t3_free_qset() 698 kfree(q->txq[i].sdesc); in t3_free_qset() 701 q->txq[i].size * in t3_free_qset() 703 q->txq[i].desc, q->txq[i].phys_addr); in t3_free_qset() [all …]
|
/drivers/atm/ |
D | ambassador.c | 628 amb_txq * txq = &dev->txq; in tx_give() local 636 spin_lock_irqsave (&txq->lock, flags); in tx_give() 638 if (txq->pending < txq->maximum) { in tx_give() 639 PRINTD (DBG_TX, "TX in slot %p", txq->in.ptr); in tx_give() 641 *txq->in.ptr = *tx; in tx_give() 642 txq->pending++; in tx_give() 643 txq->in.ptr = NEXTQ (txq->in.ptr, txq->in.start, txq->in.limit); in tx_give() 645 wr_mem (dev, offsetof(amb_mem, mb.adapter.tx_address), virt_to_bus (txq->in.ptr)); in tx_give() 648 if (txq->pending > txq->high) in tx_give() 649 txq->high = txq->pending; in tx_give() [all …]
|
/drivers/net/wireless/iwlegacy/ |
D | common.c | 383 il->txq[il->cmd_queue].meta[cmd_idx].flags &= ~CMD_WANT_SKB; in il_send_cmd_sync() 2673 il_txq_update_write_ptr(struct il_priv *il, struct il_tx_queue *txq) in il_txq_update_write_ptr() argument 2676 int txq_id = txq->q.id; in il_txq_update_write_ptr() 2678 if (txq->need_update == 0) in il_txq_update_write_ptr() 2696 il_wr(il, HBUS_TARG_WRPTR, txq->q.write_ptr | (txq_id << 8)); in il_txq_update_write_ptr() 2704 _il_wr(il, HBUS_TARG_WRPTR, txq->q.write_ptr | (txq_id << 8)); in il_txq_update_write_ptr() 2705 txq->need_update = 0; in il_txq_update_write_ptr() 2715 struct il_tx_queue *txq = &il->txq[txq_id]; in il_tx_queue_unmap() local 2716 struct il_queue *q = &txq->q; in il_tx_queue_unmap() 2722 il->ops->txq_free_tfd(il, txq); in il_tx_queue_unmap() [all …]
|
D | 3945.c | 290 struct il_tx_queue *txq = &il->txq[txq_id]; in il3945_tx_queue_reclaim() local 291 struct il_queue *q = &txq->q; in il3945_tx_queue_reclaim() 299 skb = txq->skbs[txq->q.read_ptr]; in il3945_tx_queue_reclaim() 301 txq->skbs[txq->q.read_ptr] = NULL; in il3945_tx_queue_reclaim() 302 il->ops->txq_free_tfd(il, txq); in il3945_tx_queue_reclaim() 307 il_wake_queue(il, txq); in il3945_tx_queue_reclaim() 320 struct il_tx_queue *txq = &il->txq[txq_id]; in il3945_hdl_tx() local 327 if (idx >= txq->q.n_bd || il_queue_used(&txq->q, idx) == 0) { in il3945_hdl_tx() 330 txq->q.n_bd, txq->q.write_ptr, txq->q.read_ptr); in il3945_hdl_tx() 334 txq->time_stamp = jiffies; in il3945_hdl_tx() [all …]
|
D | 4965-mac.c | 233 if (!il->txq) { in il4965_hw_nic_init() 1639 struct il_tx_queue *txq; in il4965_tx_skb() local 1740 txq = &il->txq[txq_id]; in il4965_tx_skb() 1741 q = &txq->q; in il4965_tx_skb() 1756 txq->skbs[q->write_ptr] = skb; in il4965_tx_skb() 1759 out_cmd = txq->cmd[q->write_ptr]; in il4965_tx_skb() 1760 out_meta = &txq->meta[q->write_ptr]; in il4965_tx_skb() 1817 il->ops->txq_attach_buf_to_tfd(il, txq, txcmd_phys, firstlen, 1, 0); in il4965_tx_skb() 1820 txq->need_update = 1; in il4965_tx_skb() 1823 txq->need_update = 0; in il4965_tx_skb() [all …]
|
/drivers/net/ethernet/chelsio/cxgb4/ |
D | sge.c | 858 netif_tx_stop_queue(q->txq); in eth_txq_stop() 1941 struct sge_ofld_txq *txq = s->egr_map[id]; in sge_tx_timer_cb() local 1944 tasklet_schedule(&txq->qresume_tsk); in sge_tx_timer_cb() 1953 time_after_eq(jiffies, q->txq->trans_start + HZ / 100) && in sge_tx_timer_cb() 1954 __netif_tx_trylock(q->txq)) { in sge_tx_timer_cb() 1965 __netif_tx_unlock(q->txq); in sge_tx_timer_cb() 2087 int t4_sge_alloc_eth_txq(struct adapter *adap, struct sge_eth_txq *txq, in t4_sge_alloc_eth_txq() argument 2096 nentries = txq->q.size + STAT_LEN / sizeof(struct tx_desc); in t4_sge_alloc_eth_txq() 2098 txq->q.desc = alloc_ring(adap->pdev_dev, txq->q.size, in t4_sge_alloc_eth_txq() 2100 &txq->q.phys_addr, &txq->q.sdesc, STAT_LEN, in t4_sge_alloc_eth_txq() [all …]
|
/drivers/net/ethernet/brocade/bna/ |
D | bna_tx_rx.c | 2929 struct bna_txq *txq; in bna_tx_sm_started_entry() local 2934 txq = (struct bna_txq *)qe; in bna_tx_sm_started_entry() 2935 txq->tcb->priority = txq->priority; in bna_tx_sm_started_entry() 2937 bna_ib_start(tx->bna, &txq->ib, is_regular); in bna_tx_sm_started_entry() 3157 struct bna_txq *txq = NULL; in bna_bfi_tx_enet_start() local 3170 txq = (struct bna_txq *)qe; in bna_bfi_tx_enet_start() 3172 bfi_enet_datapath_q_init(&cfg_req->q_cfg[i].q.q, &txq->qpt); in bna_bfi_tx_enet_start() 3173 cfg_req->q_cfg[i].q.priority = txq->priority; in bna_bfi_tx_enet_start() 3176 txq->ib.ib_seg_host_addr.lsb; in bna_bfi_tx_enet_start() 3178 txq->ib.ib_seg_host_addr.msb; in bna_bfi_tx_enet_start() [all …]
|
/drivers/bluetooth/ |
D | hci_h4.c | 55 struct sk_buff_head txq; member 76 skb_queue_head_init(&h4->txq); in h4_open() 89 skb_queue_purge(&h4->txq); in h4_flush() 103 skb_queue_purge(&h4->txq); in h4_close() 122 skb_queue_tail(&h4->txq, skb); in h4_enqueue() 168 return skb_dequeue(&h4->txq); in h4_dequeue()
|
D | hci_ath.c | 48 struct sk_buff_head txq; member 119 skb_queue_head_init(&ath->txq); in ath_open() 136 skb_queue_purge(&ath->txq); in ath_flush() 148 skb_queue_purge(&ath->txq); in ath_close() 186 skb_queue_tail(&ath->txq, skb); in ath_enqueue() 198 return skb_dequeue(&ath->txq); in ath_dequeue()
|
D | hci_ll.c | 83 struct sk_buff_head txq; member 116 skb_queue_tail(&ll->txq, skb); in send_hcill_cmd() 132 skb_queue_head_init(&ll->txq); in ll_open() 151 skb_queue_purge(&ll->txq); in ll_flush() 164 skb_queue_purge(&ll->txq); in ll_close() 187 skb_queue_tail(&ll->txq, skb); in __ll_do_awake() 320 skb_queue_tail(&ll->txq, skb); in ll_enqueue() 507 return skb_dequeue(&ll->txq); in ll_dequeue()
|
D | btsdio.c | 63 struct sk_buff_head txq; member 114 while ((skb = skb_dequeue(&data->txq))) { in btsdio_work() 118 skb_queue_head(&data->txq, skb); in btsdio_work() 253 skb_queue_purge(&data->txq); in btsdio_flush() 285 skb_queue_tail(&data->txq, skb); in btsdio_send_frame() 315 skb_queue_head_init(&data->txq); in btsdio_probe()
|
D | bluecard_cs.c | 72 struct sk_buff_head txq; member 260 if (!(skb = skb_dequeue(&(info->txq)))) in bluecard_write_wakeup() 328 skb_queue_head(&(info->txq), skb); in bluecard_write_wakeup() 598 skb_queue_tail(&(info->txq), skb); in bluecard_hci_set_baud_rate() 615 skb_queue_purge(&(info->txq)); in bluecard_hci_flush() 686 skb_queue_tail(&(info->txq), skb); in bluecard_hci_send_frame() 716 skb_queue_head_init(&(info->txq)); in bluecard_open() 800 skb_queue_purge(&(info->txq)); in bluecard_open()
|
/drivers/net/ |
D | ifb.c | 67 struct netdev_queue *txq; in ri_tasklet() local 70 txq = netdev_get_tx_queue(_dev, 0); in ri_tasklet() 72 if (__netif_tx_trylock(txq)) { in ri_tasklet() 74 __netif_tx_unlock(txq); in ri_tasklet() 114 if (__netif_tx_trylock(txq)) { in ri_tasklet() 120 __netif_tx_unlock(txq); in ri_tasklet() 123 __netif_tx_unlock(txq); in ri_tasklet()
|
/drivers/net/usb/ |
D | usbnet.c | 666 temp = unlink_urbs(dev, &dev->txq) + in usbnet_terminate_urbs() 671 && !skb_queue_empty(&dev->txq) in usbnet_terminate_urbs() 930 unlink_urbs (dev, &dev->txq); in kevent() 1066 (void) defer_bh(dev, skb, &dev->txq, tx_done); in tx_complete() 1075 unlink_urbs (dev, &dev->txq); in usbnet_tx_timeout() 1147 spin_lock_irqsave(&dev->txq.lock, flags); in usbnet_start_xmit() 1150 spin_unlock_irqrestore(&dev->txq.lock, flags); in usbnet_start_xmit() 1162 spin_unlock_irqrestore(&dev->txq.lock, flags); in usbnet_start_xmit() 1181 __usbnet_queue_skb(&dev->txq, skb, tx_start); in usbnet_start_xmit() 1182 if (dev->txq.qlen >= TX_QLEN (dev)) in usbnet_start_xmit() [all …]
|
/drivers/net/wireless/ |
D | mwl8k.c | 230 struct mwl8k_tx_queue txq[MWL8K_MAX_TX_QUEUES]; member 1410 struct mwl8k_tx_queue *txq = priv->txq + index; in mwl8k_txq_init() local 1414 txq->len = 0; in mwl8k_txq_init() 1415 txq->head = 0; in mwl8k_txq_init() 1416 txq->tail = 0; in mwl8k_txq_init() 1420 txq->txd = pci_alloc_consistent(priv->pdev, size, &txq->txd_dma); in mwl8k_txq_init() 1421 if (txq->txd == NULL) { in mwl8k_txq_init() 1425 memset(txq->txd, 0, size); in mwl8k_txq_init() 1427 txq->skb = kcalloc(MWL8K_TX_DESCS, sizeof(*txq->skb), GFP_KERNEL); in mwl8k_txq_init() 1428 if (txq->skb == NULL) { in mwl8k_txq_init() [all …]
|