/kernel/linux/linux-5.10/drivers/scsi/snic/ |
D | vnic_cq.c | 48 unsigned int color_enable, unsigned int cq_head, unsigned int cq_tail, in svnic_cq_init() argument 60 iowrite32(cq_head, &cq->ctrl->cq_head); in svnic_cq_init() 75 iowrite32(0, &cq->ctrl->cq_head); in svnic_cq_clean()
|
D | vnic_cq.h | 33 u32 cq_head; /* 0x20 */ member 105 unsigned int color_enable, unsigned int cq_head, unsigned int cq_tail,
|
/kernel/linux/linux-5.10/drivers/net/ethernet/cisco/enic/ |
D | vnic_cq.c | 58 unsigned int color_enable, unsigned int cq_head, unsigned int cq_tail, in vnic_cq_init() argument 70 iowrite32(cq_head, &cq->ctrl->cq_head); in vnic_cq_init() 87 iowrite32(0, &cq->ctrl->cq_head); in vnic_cq_clean()
|
D | vnic_cq.h | 35 u32 cq_head; /* 0x20 */ member 117 unsigned int color_enable, unsigned int cq_head, unsigned int cq_tail,
|
/kernel/linux/linux-5.10/drivers/scsi/fnic/ |
D | vnic_cq.c | 53 unsigned int color_enable, unsigned int cq_head, unsigned int cq_tail, in vnic_cq_init() argument 65 iowrite32(cq_head, &cq->ctrl->cq_head); in vnic_cq_init() 80 iowrite32(0, &cq->ctrl->cq_head); in vnic_cq_clean()
|
D | vnic_cq.h | 43 u32 cq_head; /* 0x20 */ member 115 unsigned int color_enable, unsigned int cq_head, unsigned int cq_tail,
|
/kernel/linux/linux-5.10/drivers/dma/ |
D | hisi_dma.c | 96 u32 cq_head; member 232 chan->cq_head = 0; in hisi_dma_free_chan_resources() 439 cqe = chan->cq + chan->cq_head; in hisi_dma_irq() 442 chan->cq_head = (chan->cq_head + 1) % in hisi_dma_irq() 446 chan->cq_head); in hisi_dma_irq()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/marvell/octeontx2/nic/ |
D | otx2_txrx.c | 27 cqe_hdr = (struct nix_cqe_hdr_s *)CQE_ADDR(cq, cq->cq_head); in otx2_get_next_cqe() 31 cq->cq_head++; in otx2_get_next_cqe() 32 cq->cq_head &= (cq->cqe_cnt - 1); in otx2_get_next_cqe() 310 cqe = (struct nix_cqe_rx_s *)CQE_ADDR(cq, cq->cq_head); in otx2_rx_napi_handler() 317 cq->cq_head++; in otx2_rx_napi_handler() 318 cq->cq_head &= (cq->cqe_cnt - 1); in otx2_rx_napi_handler()
|
D | otx2_txrx.h | 128 u32 cq_head; member
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/i40iw/ |
D | i40iw_uk.c | 1094 u32 cq_head; in i40iw_clean_cq() local 1097 cq_head = cq->cq_ring.head; in i40iw_clean_cq() 1101 cqe = (u64 *)&(((struct i40iw_extended_cqe *)cq->cq_base)[cq_head]); in i40iw_clean_cq() 1103 cqe = (u64 *)&cq->cq_base[cq_head]; in i40iw_clean_cq() 1114 cq_head = (cq_head + 1) % cq->cq_ring.size; in i40iw_clean_cq() 1115 if (!cq_head) in i40iw_clean_cq()
|
D | i40iw_verbs.h | 107 u16 cq_head; member
|
/kernel/linux/linux-5.10/drivers/nvme/host/ |
D | pci.c | 203 u16 cq_head; member 976 struct nvme_completion *hcqe = &nvmeq->cqes[nvmeq->cq_head]; in nvme_cqe_pending() 983 u16 head = nvmeq->cq_head; in nvme_ring_cq_doorbell() 1030 u32 tmp = nvmeq->cq_head + 1; in nvme_update_cq_head() 1033 nvmeq->cq_head = 0; in nvme_update_cq_head() 1036 nvmeq->cq_head = tmp; in nvme_update_cq_head() 1051 nvme_handle_cqe(nvmeq, nvmeq->cq_head); in nvme_process_cq() 1525 nvmeq->cq_head = 0; in nvme_alloc_queue() 1560 nvmeq->cq_head = 0; in nvme_init_queue()
|
/kernel/linux/linux-5.10/drivers/crypto/hisilicon/ |
D | qm.h | 257 u16 cq_head; member
|
D | qm.c | 611 if (qp->qp_status.cq_head == QM_Q_DEPTH - 1) { in qm_cq_head_update() 613 qp->qp_status.cq_head = 0; in qm_cq_head_update() 615 qp->qp_status.cq_head++; in qm_cq_head_update() 627 struct qm_cqe *cqe = qp->cqe + qp->qp_status.cq_head; in qm_poll_qp() 634 cqe = qp->cqe + qp->qp_status.cq_head; in qm_poll_qp() 636 qp->qp_status.cq_head, 0); in qm_poll_qp() 642 qp->qp_status.cq_head, 1); in qm_poll_qp() 758 qp_status->cq_head = 0; in qm_init_qp_status()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/cavium/thunder/ |
D | nicvf_main.c | 961 u64 cq_head; in nicvf_poll() local 974 cq_head = nicvf_queue_reg_read(nic, NIC_QSET_CQ_0_7_HEAD, in nicvf_poll() 978 cq->cq_idx, cq_head); in nicvf_poll()
|