Home
last modified time | relevance | path

Searched refs:tx_list (Results 1 – 25 of 31) sorted by relevance

12

/drivers/staging/lustre/lnet/klnds/socklnd/
Dsocklnd_proto.c48 list_add_tail(&tx_msg->tx_list, &conn->ksnc_tx_queue); in ksocknal_queue_tx_msg_v1()
62 if (tx->tx_list.next == &conn->ksnc_tx_queue) { in ksocknal_next_tx_carrier()
66 conn->ksnc_tx_carrier = list_entry(tx->tx_list.next, in ksocknal_next_tx_carrier()
67 ksock_tx_t, tx_list); in ksocknal_next_tx_carrier()
90 list_add_tail(&tx_ack->tx_list, in ksocknal_queue_tx_zcack_v2()
100 list_add_tail(&tx_ack->tx_list, in ksocknal_queue_tx_zcack_v2()
132 list_add_tail(&tx_msg->tx_list, &conn->ksnc_tx_queue); in ksocknal_queue_tx_msg_v2()
138 list_add_tail(&tx_msg->tx_list, &conn->ksnc_tx_queue); in ksocknal_queue_tx_msg_v2()
149 list_add(&tx_msg->tx_list, &tx->tx_list); in ksocknal_queue_tx_msg_v2()
150 list_del(&tx->tx_list); in ksocknal_queue_tx_msg_v2()
[all …]
Dsocklnd_cb.c42 next, ksock_tx_t, tx_list); in ksocknal_alloc_tx()
44 list_del(&tx->tx_list); in ksocknal_alloc_tx()
101 list_add(&tx->tx_list, &ksocknal_data.ksnd_idle_noop_txs); in ksocknal_free_tx()
413 tx = list_entry(txlist->next, ksock_tx_t, tx_list); in ksocknal_txlist_done()
425 list_del(&tx->tx_list); in ksocknal_txlist_done()
759 list_add_tail(&ztx->tx_list, &sched->kss_zombie_noop_txs); in ksocknal_queue_tx_locked()
908 list_add_tail (&tx->tx_list, &peer->ksnp_tx_queue); in ksocknal_launch_packet()
1473 ksock_tx_t, tx_list); in ksocknal_scheduler()
1479 list_del(&tx->tx_list); in ksocknal_scheduler()
1499 list_add(&tx->tx_list, in ksocknal_scheduler()
[all …]
/drivers/ps3/
Dps3-vuart.c84 } tx_list; member
498 spin_lock_irqsave(&priv->tx_list.lock, flags); in ps3_vuart_write()
500 if (list_empty(&priv->tx_list.head)) { in ps3_vuart_write()
505 spin_unlock_irqrestore(&priv->tx_list.lock, flags); in ps3_vuart_write()
523 spin_unlock_irqrestore(&priv->tx_list.lock, flags); in ps3_vuart_write()
535 spin_lock_irqsave(&priv->tx_list.lock, flags); in ps3_vuart_write()
536 list_add_tail(&lb->link, &priv->tx_list.head); in ps3_vuart_write()
538 spin_unlock_irqrestore(&priv->tx_list.lock, flags); in ps3_vuart_write()
740 spin_lock_irqsave(&priv->tx_list.lock, flags); in ps3_vuart_handle_interrupt_tx()
742 list_for_each_entry_safe(lb, n, &priv->tx_list.head, link) { in ps3_vuart_handle_interrupt_tx()
[all …]
/drivers/dma/
Dxgene-dma.c252 struct list_head tx_list; member
530 list_splice_tail_init(&desc->tx_list, &chan->ld_pending); in xgene_dma_tx_submit()
557 INIT_LIST_HEAD(&desc->tx_list); in xgene_dma_alloc_descriptor()
954 list_add_tail(&new->node, &first->tx_list); in xgene_dma_prep_sg()
993 list_splice(&first->tx_list, &new->tx_list); in xgene_dma_prep_sg()
1000 xgene_dma_free_desc_list(chan, &first->tx_list); in xgene_dma_prep_sg()
1035 list_add_tail(&new->node, &first->tx_list); in xgene_dma_prep_xor()
1040 list_splice(&first->tx_list, &new->tx_list); in xgene_dma_prep_xor()
1048 xgene_dma_free_desc_list(chan, &first->tx_list); in xgene_dma_prep_xor()
1093 list_add_tail(&new->node, &first->tx_list); in xgene_dma_prep_pq()
[all …]
Dmmp_pdma.c87 struct list_head tx_list; member
349 list_for_each_entry(child, &desc->tx_list, node) { in mmp_pdma_tx_submit()
354 list_splice_tail_init(&desc->tx_list, &chan->chain_pending); in mmp_pdma_tx_submit()
374 INIT_LIST_HEAD(&desc->tx_list); in mmp_pdma_alloc_descriptor()
504 list_add_tail(&new->node, &first->tx_list); in mmp_pdma_prep_memcpy()
520 mmp_pdma_free_desc_list(chan, &first->tx_list); in mmp_pdma_prep_memcpy()
576 list_add_tail(&new->node, &first->tx_list); in mmp_pdma_prep_slave_sg()
598 mmp_pdma_free_desc_list(chan, &first->tx_list); in mmp_pdma_prep_slave_sg()
670 list_add_tail(&new->node, &first->tx_list); in mmp_pdma_prep_dma_cyclic()
684 mmp_pdma_free_desc_list(chan, &first->tx_list); in mmp_pdma_prep_dma_cyclic()
Dtxx9dmac.c185 if (!list_empty(&desc->tx_list)) in txx9dmac_last_child()
186 desc = list_entry(desc->tx_list.prev, typeof(*desc), desc_node); in txx9dmac_last_child()
201 INIT_LIST_HEAD(&desc->tx_list); in txx9dmac_desc_alloc()
250 list_for_each_entry(child, &desc->tx_list, desc_node) in txx9dmac_sync_desc_for_cpu()
272 list_for_each_entry(child, &desc->tx_list, desc_node) in txx9dmac_desc_put()
276 list_splice_init(&desc->tx_list, &dc->free_list); in txx9dmac_desc_put()
418 list_splice_init(&desc->tx_list, &dc->free_list); in txx9dmac_descriptor_complete()
537 list_for_each_entry(child, &bad_desc->tx_list, desc_node) in txx9dmac_handle_error()
579 list_for_each_entry(child, &desc->tx_list, desc_node) in txx9dmac_scan_descriptors()
789 list_add_tail(&desc->desc_node, &first->tx_list); in txx9dmac_prep_dma_memcpy()
[all …]
Dep93xx_dma.c128 struct list_head tx_list; member
244 while (!list_empty(&desc->tx_list)) { in ep93xx_dma_set_active()
245 struct ep93xx_dma_desc *d = list_first_entry(&desc->tx_list, in ep93xx_dma_set_active()
702 list_splice_init(&desc->tx_list, &edmac->free_list); in ep93xx_dma_desc_put()
917 INIT_LIST_HEAD(&desc->tx_list); in ep93xx_dma_alloc_chan_resources()
1003 list_add_tail(&desc->node, &first->tx_list); in ep93xx_dma_prep_dma_memcpy()
1076 list_add_tail(&desc->node, &first->tx_list); in ep93xx_dma_prep_slave_sg()
1155 list_add_tail(&desc->node, &first->tx_list); in ep93xx_dma_prep_dma_cyclic()
Dfsldma.c405 list_splice_tail_init(&desc->tx_list, &chan->ld_pending); in append_ld_queue()
429 list_for_each_entry(child, &desc->tx_list, node) { in fsl_dma_tx_submit()
472 INIT_LIST_HEAD(&desc->tx_list); in fsl_dma_alloc_descriptor()
809 list_add_tail(&new->node, &first->tx_list); in fsl_dma_prep_memcpy()
824 fsldma_free_desc_list_reverse(chan, &first->tx_list); in fsl_dma_prep_memcpy()
888 list_add_tail(&new->node, &first->tx_list); in fsl_dma_prep_sg()
940 fsldma_free_desc_list_reverse(chan, &first->tx_list); in fsl_dma_prep_sg()
Dpch_dma.c95 struct list_head tx_list; member
344 if (list_empty(&desc->tx_list)) { in pdc_dostart()
363 list_splice_init(&desc->tx_list, &pd_chan->free_list); in pdc_chain_complete()
446 INIT_LIST_HEAD(&desc->tx_list); in pdc_alloc_desc()
495 list_splice_init(&desc->tx_list, &pd_chan->free_list); in pdc_desc_put()
642 list_add_tail(&desc->desc_node, &first->tx_list); in pd_prep_slave_sg()
Dtegra20-apb-dma.c173 struct list_head tx_list; member
317 if (!list_empty(&dma_desc->tx_list)) in tegra_dma_desc_put()
318 list_splice_init(&dma_desc->tx_list, &tdc->free_sg_req); in tegra_dma_desc_put()
714 list_splice_tail_init(&dma_desc->tx_list, &tdc->pending_sg_req); in tegra_dma_tx_submit()
984 INIT_LIST_HEAD(&dma_desc->tx_list); in tegra_dma_prep_slave_sg()
1027 list_add_tail(&sg_req->node, &dma_desc->tx_list); in tegra_dma_prep_slave_sg()
1125 INIT_LIST_HEAD(&dma_desc->tx_list); in tegra_dma_prep_dma_cyclic()
1154 list_add_tail(&sg_req->node, &dma_desc->tx_list); in tegra_dma_prep_dma_cyclic()
Dat_hdmac.c117 INIT_LIST_HEAD(&desc->tx_list); in atc_alloc_descriptor()
183 list_for_each_entry(child, &desc->tx_list, desc_node) in atc_desc_put()
187 list_splice_init(&desc->tx_list, &atchan->free_list); in atc_desc_put()
213 &(*first)->tx_list); in atc_desc_chain()
420 list_for_each_entry(desc, &desc_first->tx_list, desc_node) { in atc_get_bytes_left()
468 list_splice_init(&desc->tx_list, &atchan->free_list); in atc_chain_complete()
584 list_for_each_entry(child, &bad_desc->tx_list, desc_node) in atc_handle_error()
Diop-adma.c343 list_splice(&chain, &alloc_tail->tx_list); in iop_adma_alloc_slots()
390 list_splice_init(&sw_desc->tx_list, in iop_adma_tx_submit()
457 INIT_LIST_HEAD(&slot->tx_list); in iop_adma_alloc_chan_resources()
1461 list_splice_init(&sw_desc->tx_list, &iop_chan->chain); in iop_chan_start_null_memcpy()
1513 list_splice_init(&sw_desc->tx_list, &iop_chan->chain); in iop_chan_start_null_xor()
Dfsldma.h104 struct list_head tx_list; member
Dtxx9dmac.h233 struct list_head tx_list; member
Dat_hdmac_regs.h193 struct list_head tx_list; member
/drivers/staging/lustre/lnet/klnds/o2iblnd/
Do2iblnd_cb.c77 kiblnd_pool_free_node(&tx->tx_pool->tpo_pool, &tx->tx_list); in kiblnd_tx_done()
94 tx = list_entry(txlist->next, kib_tx_t, tx_list); in kiblnd_txlist_done()
96 list_del(&tx->tx_list); in kiblnd_txlist_done()
116 tx = container_of(node, kib_tx_t, tx_list); in kiblnd_get_idle_tx()
225 kib_tx_t *tx = list_entry(tmp, kib_tx_t, tx_list); in kiblnd_find_waiting_tx_locked()
274 list_del(&tx->tx_list); in kiblnd_handle_completion()
401 list_del(&tx->tx_list); in kiblnd_handle_rx()
796 list_del(&tx->tx_list); in kiblnd_post_tx_locked()
831 list_add(&tx->tx_list, &conn->ibc_active_txs); in kiblnd_post_tx_locked()
863 list_del(&tx->tx_list); in kiblnd_post_tx_locked()
[all …]
/drivers/net/ethernet/octeon/
Docteon_mgmt.c132 struct sk_buff_head tx_list; member
260 spin_lock_irqsave(&p->tx_list.lock, flags); in octeon_mgmt_clean_tx_buffers()
265 spin_unlock_irqrestore(&p->tx_list.lock, flags); in octeon_mgmt_clean_tx_buffers()
276 skb = __skb_dequeue(&p->tx_list); in octeon_mgmt_clean_tx_buffers()
285 spin_unlock_irqrestore(&p->tx_list.lock, flags); in octeon_mgmt_clean_tx_buffers()
1258 skb_queue_purge(&p->tx_list); in octeon_mgmt_stop()
1288 spin_lock_irqsave(&p->tx_list.lock, flags); in octeon_mgmt_xmit()
1291 spin_unlock_irqrestore(&p->tx_list.lock, flags); in octeon_mgmt_xmit()
1293 spin_lock_irqsave(&p->tx_list.lock, flags); in octeon_mgmt_xmit()
1298 spin_unlock_irqrestore(&p->tx_list.lock, flags); in octeon_mgmt_xmit()
[all …]
/drivers/dma/dw/
Dcore.c119 list_for_each_entry(child, &desc->tx_list, desc_node) in dwc_desc_put()
123 list_splice_init(&desc->tx_list, &dwc->free_list); in dwc_desc_put()
241 dwc->tx_node_active = &first->tx_list; in dwc_dostart()
293 list_for_each_entry(child, &desc->tx_list, desc_node) in dwc_descriptor_complete()
297 list_splice_init(&desc->tx_list, &dwc->free_list); in dwc_descriptor_complete()
369 head = &desc->tx_list; in dwc_scan_descriptors()
431 list_for_each_entry(child, &desc->tx_list, desc_node) { in dwc_scan_descriptors()
500 list_for_each_entry(child, &bad_desc->tx_list, desc_node) in dwc_handle_error()
740 &first->tx_list); in dwc_prep_dma_memcpy()
837 &first->tx_list); in dwc_prep_slave_sg()
[all …]
Dregs.h333 struct list_head tx_list; member
/drivers/net/wireless/p54/
Dp54spi.c447 struct p54s_tx_info, tx_list); in p54spi_wq_tx()
449 list_del_init(&entry->tx_list); in p54spi_wq_tx()
454 tx_list); in p54spi_wq_tx()
485 list_add_tail(&di->tx_list, &priv->tx_pending); in p54spi_op_tx()
Dp54spi.h102 struct list_head tx_list; member
/drivers/dma/sh/
Dshdma-base.c569 LIST_HEAD(tx_list); in shdma_prep_sg()
618 list_add_tail(&new->node, &tx_list); in shdma_prep_sg()
626 list_splice_tail(&tx_list, &schan->ld_free); in shdma_prep_sg()
633 list_for_each_entry(new, &tx_list, node) in shdma_prep_sg()
635 list_splice(&tx_list, &schan->ld_free); in shdma_prep_sg()
/drivers/net/ethernet/ti/
Dtlan.c861 priv->tx_list = priv->rx_list + TLAN_NUM_RX_LISTS; in tlan_init()
1075 tail_list = priv->tx_list + priv->tx_tail; in tlan_start_tx()
1114 (priv->tx_list + (TLAN_NUM_TX_LISTS - 1))->forward in tlan_start_tx()
1117 (priv->tx_list + (priv->tx_tail - 1))->forward in tlan_start_tx()
1249 tlan_print_list(priv->tx_list + i, "TX", i); in tlan_get_stats()
1379 head_list = priv->tx_list + priv->tx_head; in tlan_handle_tx_eof()
1402 head_list = priv->tx_list + priv->tx_head; in tlan_handle_tx_eof()
1413 head_list = priv->tx_list + priv->tx_head; in tlan_handle_tx_eof()
1658 head_list = priv->tx_list + priv->tx_head; in tlan_handle_tx_eoc()
1927 list = priv->tx_list + i; in tlan_reset_lists()
[all …]
Dtlan.h186 struct tlan_list *tx_list; member
/drivers/tty/
Dn_gsm.c251 struct list_head tx_list; /* Pending data packets */ member
691 list_for_each_entry_safe(msg, nmsg, &gsm->tx_list, list) { in gsm_data_kick()
777 list_add_tail(&msg->list, &gsm->tx_list); in __gsm_data_queue()
2101 list_for_each_entry_safe(txq, ntxq, &gsm->tx_list, list) in gsm_cleanup_mux()
2103 INIT_LIST_HEAD(&gsm->tx_list); in gsm_cleanup_mux()
2210 INIT_LIST_HEAD(&gsm->tx_list); in gsm_alloc_mux()

12