/drivers/net/ethernet/cavium/thunder/ |
D | nicvf_queues.c | 31 static int nicvf_poll_reg(struct nicvf *nic, int qidx, in nicvf_poll_reg() argument 42 reg_val = nicvf_queue_reg_read(nic, reg, qidx); in nicvf_poll_reg() 392 struct queue_set *qs, int qidx) in nicvf_reclaim_snd_queue() argument 395 nicvf_queue_reg_write(nic, NIC_QSET_SQ_0_7_CFG, qidx, 0); in nicvf_reclaim_snd_queue() 397 if (nicvf_poll_reg(nic, qidx, NIC_QSET_SQ_0_7_STATUS, 21, 1, 0x01)) in nicvf_reclaim_snd_queue() 400 nicvf_queue_reg_write(nic, NIC_QSET_SQ_0_7_CFG, qidx, NICVF_SQ_RESET); in nicvf_reclaim_snd_queue() 404 struct queue_set *qs, int qidx) in nicvf_reclaim_rcv_queue() argument 414 struct queue_set *qs, int qidx) in nicvf_reclaim_cmp_queue() argument 417 nicvf_queue_reg_write(nic, NIC_QSET_CQ_0_7_CFG2, qidx, 0); in nicvf_reclaim_cmp_queue() 419 nicvf_queue_reg_write(nic, NIC_QSET_CQ_0_7_CFG, qidx, 0); in nicvf_reclaim_cmp_queue() [all …]
|
D | nicvf_main.c | 64 static inline u8 nicvf_netdev_qidx(struct nicvf *nic, u8 qidx) in nicvf_netdev_qidx() argument 67 return qidx + ((nic->sqs_id + 1) * MAX_CMP_QUEUES_PER_QS); in nicvf_netdev_qidx() 69 return qidx; in nicvf_netdev_qidx() 93 u64 qidx, u64 val) in nicvf_queue_reg_write() argument 97 writeq_relaxed(val, addr + (qidx << NIC_Q_NUM_SHIFT)); in nicvf_queue_reg_write() 100 u64 nicvf_queue_reg_read(struct nicvf *nic, u64 offset, u64 qidx) in nicvf_queue_reg_read() argument 104 return readq_relaxed(addr + (qidx << NIC_Q_NUM_SHIFT)); in nicvf_queue_reg_read() 760 int qidx; in nicvf_handle_qs_err() local 766 for (qidx = 0; qidx < qs->cq_cnt; qidx++) { in nicvf_handle_qs_err() 768 qidx); in nicvf_handle_qs_err() [all …]
|
D | nicvf_ethtool.c | 179 int stats, qidx; in nicvf_get_qset_strings() local 182 for (qidx = 0; qidx < nic->qs->rq_cnt; qidx++) { in nicvf_get_qset_strings() 184 sprintf(*data, "rxq%d: %s", qidx + start_qidx, in nicvf_get_qset_strings() 190 for (qidx = 0; qidx < nic->qs->sq_cnt; qidx++) { in nicvf_get_qset_strings() 192 sprintf(*data, "txq%d: %s", qidx + start_qidx, in nicvf_get_qset_strings() 266 int stat, qidx; in nicvf_get_qset_stats() local 271 for (qidx = 0; qidx < nic->qs->rq_cnt; qidx++) { in nicvf_get_qset_stats() 272 nicvf_update_rq_stats(nic, qidx); in nicvf_get_qset_stats() 274 *((*data)++) = ((u64 *)&nic->qs->rq[qidx].stats) in nicvf_get_qset_stats() 278 for (qidx = 0; qidx < nic->qs->sq_cnt; qidx++) { in nicvf_get_qset_stats() [all …]
|
D | nicvf_queues.h | 288 int qidx, bool enable); 290 void nicvf_sq_enable(struct nicvf *nic, struct snd_queue *sq, int qidx); 291 void nicvf_sq_disable(struct nicvf *nic, int qidx); 294 struct snd_queue *sq, int qidx); 312 u64 qidx, u64 val); 314 u64 offset, u64 qidx);
|
/drivers/scsi/csiostor/ |
D | csio_wr.c | 741 csio_wr_cleanup_eq_stpg(struct csio_hw *hw, int qidx) in csio_wr_cleanup_eq_stpg() argument 743 struct csio_q *q = csio_hw_to_wrm(hw)->q_arr[qidx]; in csio_wr_cleanup_eq_stpg() 758 csio_wr_cleanup_iq_ftr(struct csio_hw *hw, int qidx) in csio_wr_cleanup_iq_ftr() argument 761 struct csio_q *q = wrm->q_arr[qidx]; in csio_wr_cleanup_iq_ftr() 857 csio_wr_get(struct csio_hw *hw, int qidx, uint32_t size, in csio_wr_get() argument 861 struct csio_q *q = wrm->q_arr[qidx]; in csio_wr_get() 872 CSIO_DB_ASSERT((qidx >= 0) && (qidx < wrm->free_qidx)); in csio_wr_get() 976 csio_wr_issue(struct csio_hw *hw, int qidx, bool prio) in csio_wr_issue() argument 979 struct csio_q *q = wrm->q_arr[qidx]; in csio_wr_issue() 981 CSIO_DB_ASSERT((qidx >= 0) && (qidx < wrm->free_qidx)); in csio_wr_issue() [all …]
|
/drivers/scsi/lpfc/ |
D | lpfc_debugfs.c | 2560 uint32_t qidx, quetp, queid, index, count, offset, value; in lpfc_idiag_queacc_write() local 2598 for (qidx = 0; qidx < phba->cfg_fcp_io_channel; in lpfc_idiag_queacc_write() 2599 qidx++) { in lpfc_idiag_queacc_write() 2600 if (phba->sli4_hba.hba_eq[qidx] && in lpfc_idiag_queacc_write() 2601 phba->sli4_hba.hba_eq[qidx]->queue_id == in lpfc_idiag_queacc_write() 2605 phba->sli4_hba.hba_eq[qidx], in lpfc_idiag_queacc_write() 2610 phba->sli4_hba.hba_eq[qidx]; in lpfc_idiag_queacc_write() 2642 qidx = 0; in lpfc_idiag_queacc_write() 2644 if (phba->sli4_hba.fcp_cq[qidx] && in lpfc_idiag_queacc_write() 2645 phba->sli4_hba.fcp_cq[qidx]->queue_id == in lpfc_idiag_queacc_write() [all …]
|
D | lpfc_sli.c | 12511 uint32_t qidx) in lpfc_sli4_hba_handle_eqe() argument 12532 if (unlikely(cqid != phba->sli4_hba.fcp_cq_map[qidx])) { in lpfc_sli4_hba_handle_eqe() 12534 phba->sli4_hba.hba_eq[qidx]); in lpfc_sli4_hba_handle_eqe() 12544 cq = phba->sli4_hba.fcp_cq[qidx]; in lpfc_sli4_hba_handle_eqe() 12549 "(%d) does not exist\n", qidx); in lpfc_sli4_hba_handle_eqe()
|
/drivers/net/ethernet/chelsio/cxgb3/ |
D | cxgb3_main.c | 403 int i, j, err, qidx = 0; in request_msix_data_irqs() local 409 err = request_irq(adap->msix_info[qidx + 1].vec, in request_msix_data_irqs() 411 adap->sge.qs[qidx]. in request_msix_data_irqs() 413 adap->msix_info[qidx + 1].desc, in request_msix_data_irqs() 414 &adap->sge.qs[qidx]); in request_msix_data_irqs() 416 while (--qidx >= 0) in request_msix_data_irqs() 417 free_irq(adap->msix_info[qidx + 1].vec, in request_msix_data_irqs() 418 &adap->sge.qs[qidx]); in request_msix_data_irqs() 421 qidx++; in request_msix_data_irqs() 942 static int send_pktsched_cmd(struct adapter *adap, int sched, int qidx, int lo, in send_pktsched_cmd() argument [all …]
|
D | sge.c | 167 static inline struct sge_qset *fl_to_qset(const struct sge_fl *q, int qidx) in fl_to_qset() argument 169 return container_of(q, struct sge_qset, fl[qidx]); in fl_to_qset() 177 static inline struct sge_qset *txq_to_qset(const struct sge_txq *q, int qidx) in txq_to_qset() argument 179 return container_of(q, struct sge_qset, txq[qidx]); in txq_to_qset() 1223 int qidx; in t3_eth_xmit() local 1240 qidx = skb_get_queue_mapping(skb); in t3_eth_xmit() 1241 qs = &pi->qs[qidx]; in t3_eth_xmit() 1243 txq = netdev_get_tx_queue(dev, qidx); in t3_eth_xmit()
|
/drivers/crypto/ccp/ |
D | ccp-dev-v5.c | 204 n = head_idx + COMMANDS_PER_QUEUE - cmd_q->qidx - 1; in ccp5_get_free_slots() 224 mP = (u32 *) &cmd_q->qbase[cmd_q->qidx]; in ccp5_do_cmd() 229 cmd_q->qidx = (cmd_q->qidx + 1) % COMMANDS_PER_QUEUE; in ccp5_do_cmd() 235 tail = low_address(cmd_q->qdma_tail + cmd_q->qidx * Q_DESC_SIZE); in ccp5_do_cmd() 757 cmd_q->qidx = 0; in ccp5_init()
|
D | ccp-dev.h | 266 unsigned int qidx; member
|
/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_sriov.c | 1736 int qidx = 0, abs_vfid; in bnx2x_iov_eq_sp_event() local 1785 qidx = cid & ((1 << BNX2X_VF_CID_WND)-1); in bnx2x_iov_eq_sp_event() 1799 vf->abs_vfid, qidx); in bnx2x_iov_eq_sp_event() 1800 vfq_get(vf, qidx)->sp_obj.complete_cmd(bp, in bnx2x_iov_eq_sp_event() 1802 qidx)->sp_obj, in bnx2x_iov_eq_sp_event() 1807 vf->abs_vfid, qidx); in bnx2x_iov_eq_sp_event() 1808 bnx2x_vf_handle_classification_eqe(bp, vfq_get(vf, qidx), elem); in bnx2x_iov_eq_sp_event() 1812 vf->abs_vfid, qidx); in bnx2x_iov_eq_sp_event() 1817 vf->abs_vfid, qidx); in bnx2x_iov_eq_sp_event() 1822 vf->abs_vfid, qidx); in bnx2x_iov_eq_sp_event()
|
D | bnx2x_vfpf.c | 26 static int bnx2x_vfpf_teardown_queue(struct bnx2x *bp, int qidx); 684 static int bnx2x_vfpf_teardown_queue(struct bnx2x *bp, int qidx) in bnx2x_vfpf_teardown_queue() argument 694 req->vf_qid = qidx; in bnx2x_vfpf_teardown_queue() 706 BNX2X_ERR("Sending TEARDOWN for queue %d failed: %d\n", qidx, in bnx2x_vfpf_teardown_queue() 713 BNX2X_ERR("TEARDOWN for queue %d failed: %d\n", qidx, in bnx2x_vfpf_teardown_queue()
|
/drivers/net/ethernet/chelsio/cxgb4vf/ |
D | sge.c | 1166 int qidx, credits, max_pkt_len; in t4vf_eth_xmit() local 1201 qidx = skb_get_queue_mapping(skb); in t4vf_eth_xmit() 1202 BUG_ON(qidx >= pi->nqsets); in t4vf_eth_xmit() 1203 txq = &adapter->sge.ethtxq[pi->first_qset + qidx]; in t4vf_eth_xmit() 1230 dev->name, qidx); in t4vf_eth_xmit()
|
D | cxgb4vf_main.c | 2581 int q10g, n10g, qidx, pidx, qs; in cfg_queues() local 2616 qidx = 0; in cfg_queues() 2620 pi->first_qset = qidx; in cfg_queues() 2622 qidx += pi->nqsets; in cfg_queues() 2624 s->ethqsets = qidx; in cfg_queues()
|
/drivers/net/wireless/broadcom/brcm80211/brcmfmac/ |
D | fwsignal.c | 1974 int qidx, hslot; in brcmf_fws_rollback_toq() local 1979 qidx = 2 * fifo; in brcmf_fws_rollback_toq() 1981 qidx++; in brcmf_fws_rollback_toq() 1983 pktout = brcmu_pktq_penq_head(&entry->psq, qidx, skb); in brcmf_fws_rollback_toq() 1985 brcmf_err("%s queue %d full\n", entry->name, qidx); in brcmf_fws_rollback_toq()
|
/drivers/net/ethernet/netronome/nfp/ |
D | nfp_net_common.c | 768 u16 qidx; in nfp_net_tx() local 770 qidx = skb_get_queue_mapping(skb); in nfp_net_tx() 771 tx_ring = &nn->tx_rings[qidx]; in nfp_net_tx() 773 nd_q = netdev_get_tx_queue(nn->netdev, qidx); in nfp_net_tx() 779 qidx, tx_ring->wr_p, tx_ring->rd_p); in nfp_net_tx()
|
/drivers/net/ethernet/chelsio/cxgb4/ |
D | cxgb4_main.c | 4013 int i, n10g = 0, qidx = 0; in cfg_queues() local 4044 pi->first_qset = qidx; in cfg_queues() 4046 qidx += pi->nqsets; in cfg_queues() 4061 pi->first_qset = qidx; in cfg_queues() 4063 qidx += pi->nqsets; in cfg_queues() 4067 s->ethqsets = qidx; in cfg_queues() 4068 s->max_ethqsets = qidx; /* MSI-X may lower it later */ in cfg_queues()
|
D | sge.c | 1169 int qidx, credits; in t4_eth_xmit() local 1202 qidx = skb_get_queue_mapping(skb); in t4_eth_xmit() 1203 q = &adap->sge.ethtxq[qidx + pi->first_qset]; in t4_eth_xmit() 1222 dev->name, qidx); in t4_eth_xmit()
|
/drivers/scsi/sym53c8xx_2/ |
D | sym_hipd.c | 1482 u_short qidx; local 1515 qidx = np->squeueput + 2; 1516 if (qidx >= MAX_QUEUE*2) qidx = 0; 1518 np->squeue [qidx] = cpu_to_scr(np->idletask_ba); 1522 np->squeueput = qidx;
|
/drivers/scsi/ |
D | ncr53c8xx.c | 4451 u16 qidx; in ncr_put_start_queue() local 4457 qidx = np->squeueput + 2; in ncr_put_start_queue() 4458 if (qidx >= MAX_START + MAX_START) qidx = 1; in ncr_put_start_queue() 4460 np->scripth->tryloop [qidx] = cpu_to_scr(NCB_SCRIPT_PHYS (np, idle)); in ncr_put_start_queue() 4464 np->squeueput = qidx; in ncr_put_start_queue()
|