/kernel/linux/linux-5.10/drivers/net/ethernet/chelsio/cxgb4/ |
D | sge.c | 2755 spin_lock(&q->sendq.lock); in ctrl_xmit() 2759 __skb_queue_tail(&q->sendq, skb); in ctrl_xmit() 2760 spin_unlock(&q->sendq.lock); in ctrl_xmit() 2772 spin_unlock(&q->sendq.lock); in ctrl_xmit() 2790 spin_lock(&q->sendq.lock); in restart_ctrlq() 2794 while ((skb = __skb_dequeue(&q->sendq)) != NULL) { in restart_ctrlq() 2804 spin_unlock(&q->sendq.lock); in restart_ctrlq() 2814 spin_lock(&q->sendq.lock); in restart_ctrlq() 2822 spin_lock(&q->sendq.lock); in restart_ctrlq() 2828 spin_unlock(&q->sendq.lock); in restart_ctrlq() [all …]
|
D | cxgb4.h | 884 struct sk_buff_head sendq; /* list of backpressured packets */ member 894 struct sk_buff_head sendq; /* list of backpressured packets */ member
|
D | cxgb4_uld.c | 415 __skb_queue_purge(&txq->sendq); in free_sge_txq_uld()
|
/kernel/linux/linux-5.10/drivers/isdn/mISDN/ |
D | layer2.h | 22 struct sk_buff_head sendq; member
|
D | dsp_core.c | 887 skb_queue_tail(&dsp->sendq, skb); in dsp_function() 934 skb_queue_purge(&dsp->sendq); in dsp_function() 979 skb_queue_purge(&dsp->sendq); in dsp_ctrl() 1015 while ((skb = skb_dequeue(&dsp->sendq))) { in dsp_send_bh() 1066 skb_queue_head_init(&ndsp->sendq); in dspcreate()
|
D | dsp.h | 186 struct sk_buff_head sendq; member
|
D | tei.c | 335 struct sk_buff *skb = skb_dequeue(&mgr->sendq); in do_send() 359 skb = skb_dequeue(&mgr->sendq); in do_ack() 376 skb_queue_tail(&mgr->sendq, skb); in mgr_send_down() 399 skb_queue_tail(&mgr->sendq, skb); in dl_unit_data() 1253 skb_queue_purge(&mgr->sendq); in delete_teimanager() 1353 skb_queue_head_init(&mgr->sendq); in create_teimanager()
|
D | dsp_cmx.c | 1584 skb_queue_tail(&dsp->sendq, nskb); 1600 skb_queue_tail(&dsp->sendq, txskb); 1617 skb_queue_tail(&dsp->sendq, nskb); 1939 skb_queue_tail(&dsp->sendq, nskb); 1955 skb_queue_tail(&member->dsp->sendq, nskb);
|
/kernel/linux/linux-5.10/drivers/infiniband/sw/siw/ |
D | siw.h | 435 struct siw_sqe *sendq; /* send queue element array */ member 627 struct siw_sqe *sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size]; in siw_sq_empty() 634 struct siw_sqe *sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size]; in sq_get_next()
|
D | siw_verbs.c | 381 qp->sendq = vmalloc_user(num_sqe * sizeof(struct siw_sqe)); in siw_create_qp() 383 qp->sendq = vzalloc(num_sqe * sizeof(struct siw_sqe)); in siw_create_qp() 385 if (qp->sendq == NULL) { in siw_create_qp() 441 if (qp->sendq) { in siw_create_qp() 444 siw_mmap_entry_insert(uctx, qp->sendq, in siw_create_qp() 492 vfree(qp->sendq); in siw_create_qp() 815 struct siw_sqe *sqe = &qp->sendq[idx]; in siw_post_send()
|
D | siw_qp.c | 1234 sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size]; in siw_sq_flush() 1339 vfree(qp->sendq); in siw_free_qp()
|
/kernel/linux/linux-5.10/drivers/rpmsg/ |
D | virtio_rpmsg_bus.c | 68 wait_queue_head_t sendq; member 607 err = wait_event_interruptible_timeout(vrp->sendq, in rpmsg_send_offchannel_raw() 821 wake_up_interruptible(&vrp->sendq); in rpmsg_xmit_done() 900 init_waitqueue_head(&vrp->sendq); in rpmsg_probe()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/chelsio/cxgb3/ |
D | adapter.h | 180 struct sk_buff_head sendq; /* List of backpressured offload packets */ member
|
D | sge.c | 708 __skb_queue_purge(&q->txq[i].sendq); in t3_free_qset() 1427 if (unlikely(!skb_queue_empty(&q->sendq))) { in check_desc_avail() 1428 addq_exit:__skb_queue_tail(&q->sendq, skb); in check_desc_avail() 1535 (skb = __skb_dequeue(&q->sendq)) != NULL) { in restart_ctrlq() 1546 if (!skb_queue_empty(&q->sendq)) { in restart_ctrlq() 1755 while ((skb = skb_peek(&q->sendq)) != NULL) { in restart_offloadq() 1783 __skb_unlink(skb, &q->sendq); in restart_offloadq() 3085 skb_queue_head_init(&q->txq[i].sendq); in t3_sge_alloc_qset()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlxsw/ |
D | pci.c | 636 u8 sendq = mlxsw_pci_cqe_sr_get(q->u.cq.v, cqe); in mlxsw_pci_cq_tasklet() local 643 if (sendq) { in mlxsw_pci_cq_tasklet()
|
/kernel/linux/patches/linux-5.10/imx8mm_patch/patches/drivers/ |
D | 0032_linux_drivers_regulator_remoteproc_reset_rpmsg_rtc.patch | 2230 * @sendq: wait queue of sending contexts waiting for a tx buffers 2238 wait_queue_head_t sendq; 2440 wake_up_interruptible(&vrp->sendq);
|
/kernel/linux/linux-5.10/drivers/crypto/chelsio/ |
D | chcr_algo.c | 589 spin_lock(&txq->sendq.lock); in cxgb4_is_crypto_q_full() 592 spin_unlock(&txq->sendq.lock); in cxgb4_is_crypto_q_full()
|