Home
last modified time | relevance | path

Searched refs:qidx (Results 1 – 21 of 21) sorted by relevance

/drivers/net/ethernet/cavium/thunder/
Dnicvf_queues.c31 static int nicvf_poll_reg(struct nicvf *nic, int qidx, in nicvf_poll_reg() argument
42 reg_val = nicvf_queue_reg_read(nic, reg, qidx); in nicvf_poll_reg()
392 struct queue_set *qs, int qidx) in nicvf_reclaim_snd_queue() argument
395 nicvf_queue_reg_write(nic, NIC_QSET_SQ_0_7_CFG, qidx, 0); in nicvf_reclaim_snd_queue()
397 if (nicvf_poll_reg(nic, qidx, NIC_QSET_SQ_0_7_STATUS, 21, 1, 0x01)) in nicvf_reclaim_snd_queue()
400 nicvf_queue_reg_write(nic, NIC_QSET_SQ_0_7_CFG, qidx, NICVF_SQ_RESET); in nicvf_reclaim_snd_queue()
404 struct queue_set *qs, int qidx) in nicvf_reclaim_rcv_queue() argument
414 struct queue_set *qs, int qidx) in nicvf_reclaim_cmp_queue() argument
417 nicvf_queue_reg_write(nic, NIC_QSET_CQ_0_7_CFG2, qidx, 0); in nicvf_reclaim_cmp_queue()
419 nicvf_queue_reg_write(nic, NIC_QSET_CQ_0_7_CFG, qidx, 0); in nicvf_reclaim_cmp_queue()
[all …]
Dnicvf_main.c64 static inline u8 nicvf_netdev_qidx(struct nicvf *nic, u8 qidx) in nicvf_netdev_qidx() argument
67 return qidx + ((nic->sqs_id + 1) * MAX_CMP_QUEUES_PER_QS); in nicvf_netdev_qidx()
69 return qidx; in nicvf_netdev_qidx()
93 u64 qidx, u64 val) in nicvf_queue_reg_write() argument
97 writeq_relaxed(val, addr + (qidx << NIC_Q_NUM_SHIFT)); in nicvf_queue_reg_write()
100 u64 nicvf_queue_reg_read(struct nicvf *nic, u64 offset, u64 qidx) in nicvf_queue_reg_read() argument
104 return readq_relaxed(addr + (qidx << NIC_Q_NUM_SHIFT)); in nicvf_queue_reg_read()
760 int qidx; in nicvf_handle_qs_err() local
766 for (qidx = 0; qidx < qs->cq_cnt; qidx++) { in nicvf_handle_qs_err()
768 qidx); in nicvf_handle_qs_err()
[all …]
Dnicvf_ethtool.c179 int stats, qidx; in nicvf_get_qset_strings() local
182 for (qidx = 0; qidx < nic->qs->rq_cnt; qidx++) { in nicvf_get_qset_strings()
184 sprintf(*data, "rxq%d: %s", qidx + start_qidx, in nicvf_get_qset_strings()
190 for (qidx = 0; qidx < nic->qs->sq_cnt; qidx++) { in nicvf_get_qset_strings()
192 sprintf(*data, "txq%d: %s", qidx + start_qidx, in nicvf_get_qset_strings()
266 int stat, qidx; in nicvf_get_qset_stats() local
271 for (qidx = 0; qidx < nic->qs->rq_cnt; qidx++) { in nicvf_get_qset_stats()
272 nicvf_update_rq_stats(nic, qidx); in nicvf_get_qset_stats()
274 *((*data)++) = ((u64 *)&nic->qs->rq[qidx].stats) in nicvf_get_qset_stats()
278 for (qidx = 0; qidx < nic->qs->sq_cnt; qidx++) { in nicvf_get_qset_stats()
[all …]
Dnicvf_queues.h288 int qidx, bool enable);
290 void nicvf_sq_enable(struct nicvf *nic, struct snd_queue *sq, int qidx);
291 void nicvf_sq_disable(struct nicvf *nic, int qidx);
294 struct snd_queue *sq, int qidx);
312 u64 qidx, u64 val);
314 u64 offset, u64 qidx);
/drivers/scsi/csiostor/
Dcsio_wr.c741 csio_wr_cleanup_eq_stpg(struct csio_hw *hw, int qidx) in csio_wr_cleanup_eq_stpg() argument
743 struct csio_q *q = csio_hw_to_wrm(hw)->q_arr[qidx]; in csio_wr_cleanup_eq_stpg()
758 csio_wr_cleanup_iq_ftr(struct csio_hw *hw, int qidx) in csio_wr_cleanup_iq_ftr() argument
761 struct csio_q *q = wrm->q_arr[qidx]; in csio_wr_cleanup_iq_ftr()
857 csio_wr_get(struct csio_hw *hw, int qidx, uint32_t size, in csio_wr_get() argument
861 struct csio_q *q = wrm->q_arr[qidx]; in csio_wr_get()
872 CSIO_DB_ASSERT((qidx >= 0) && (qidx < wrm->free_qidx)); in csio_wr_get()
976 csio_wr_issue(struct csio_hw *hw, int qidx, bool prio) in csio_wr_issue() argument
979 struct csio_q *q = wrm->q_arr[qidx]; in csio_wr_issue()
981 CSIO_DB_ASSERT((qidx >= 0) && (qidx < wrm->free_qidx)); in csio_wr_issue()
[all …]
/drivers/scsi/lpfc/
Dlpfc_debugfs.c2560 uint32_t qidx, quetp, queid, index, count, offset, value; in lpfc_idiag_queacc_write() local
2598 for (qidx = 0; qidx < phba->cfg_fcp_io_channel; in lpfc_idiag_queacc_write()
2599 qidx++) { in lpfc_idiag_queacc_write()
2600 if (phba->sli4_hba.hba_eq[qidx] && in lpfc_idiag_queacc_write()
2601 phba->sli4_hba.hba_eq[qidx]->queue_id == in lpfc_idiag_queacc_write()
2605 phba->sli4_hba.hba_eq[qidx], in lpfc_idiag_queacc_write()
2610 phba->sli4_hba.hba_eq[qidx]; in lpfc_idiag_queacc_write()
2642 qidx = 0; in lpfc_idiag_queacc_write()
2644 if (phba->sli4_hba.fcp_cq[qidx] && in lpfc_idiag_queacc_write()
2645 phba->sli4_hba.fcp_cq[qidx]->queue_id == in lpfc_idiag_queacc_write()
[all …]
Dlpfc_sli.c12511 uint32_t qidx) in lpfc_sli4_hba_handle_eqe() argument
12532 if (unlikely(cqid != phba->sli4_hba.fcp_cq_map[qidx])) { in lpfc_sli4_hba_handle_eqe()
12534 phba->sli4_hba.hba_eq[qidx]); in lpfc_sli4_hba_handle_eqe()
12544 cq = phba->sli4_hba.fcp_cq[qidx]; in lpfc_sli4_hba_handle_eqe()
12549 "(%d) does not exist\n", qidx); in lpfc_sli4_hba_handle_eqe()
/drivers/net/ethernet/chelsio/cxgb3/
Dcxgb3_main.c403 int i, j, err, qidx = 0; in request_msix_data_irqs() local
409 err = request_irq(adap->msix_info[qidx + 1].vec, in request_msix_data_irqs()
411 adap->sge.qs[qidx]. in request_msix_data_irqs()
413 adap->msix_info[qidx + 1].desc, in request_msix_data_irqs()
414 &adap->sge.qs[qidx]); in request_msix_data_irqs()
416 while (--qidx >= 0) in request_msix_data_irqs()
417 free_irq(adap->msix_info[qidx + 1].vec, in request_msix_data_irqs()
418 &adap->sge.qs[qidx]); in request_msix_data_irqs()
421 qidx++; in request_msix_data_irqs()
942 static int send_pktsched_cmd(struct adapter *adap, int sched, int qidx, int lo, in send_pktsched_cmd() argument
[all …]
Dsge.c167 static inline struct sge_qset *fl_to_qset(const struct sge_fl *q, int qidx) in fl_to_qset() argument
169 return container_of(q, struct sge_qset, fl[qidx]); in fl_to_qset()
177 static inline struct sge_qset *txq_to_qset(const struct sge_txq *q, int qidx) in txq_to_qset() argument
179 return container_of(q, struct sge_qset, txq[qidx]); in txq_to_qset()
1223 int qidx; in t3_eth_xmit() local
1240 qidx = skb_get_queue_mapping(skb); in t3_eth_xmit()
1241 qs = &pi->qs[qidx]; in t3_eth_xmit()
1243 txq = netdev_get_tx_queue(dev, qidx); in t3_eth_xmit()
/drivers/crypto/ccp/
Dccp-dev-v5.c204 n = head_idx + COMMANDS_PER_QUEUE - cmd_q->qidx - 1; in ccp5_get_free_slots()
224 mP = (u32 *) &cmd_q->qbase[cmd_q->qidx]; in ccp5_do_cmd()
229 cmd_q->qidx = (cmd_q->qidx + 1) % COMMANDS_PER_QUEUE; in ccp5_do_cmd()
235 tail = low_address(cmd_q->qdma_tail + cmd_q->qidx * Q_DESC_SIZE); in ccp5_do_cmd()
757 cmd_q->qidx = 0; in ccp5_init()
Dccp-dev.h266 unsigned int qidx; member
/drivers/net/ethernet/broadcom/bnx2x/
Dbnx2x_sriov.c1736 int qidx = 0, abs_vfid; in bnx2x_iov_eq_sp_event() local
1785 qidx = cid & ((1 << BNX2X_VF_CID_WND)-1); in bnx2x_iov_eq_sp_event()
1799 vf->abs_vfid, qidx); in bnx2x_iov_eq_sp_event()
1800 vfq_get(vf, qidx)->sp_obj.complete_cmd(bp, in bnx2x_iov_eq_sp_event()
1802 qidx)->sp_obj, in bnx2x_iov_eq_sp_event()
1807 vf->abs_vfid, qidx); in bnx2x_iov_eq_sp_event()
1808 bnx2x_vf_handle_classification_eqe(bp, vfq_get(vf, qidx), elem); in bnx2x_iov_eq_sp_event()
1812 vf->abs_vfid, qidx); in bnx2x_iov_eq_sp_event()
1817 vf->abs_vfid, qidx); in bnx2x_iov_eq_sp_event()
1822 vf->abs_vfid, qidx); in bnx2x_iov_eq_sp_event()
Dbnx2x_vfpf.c26 static int bnx2x_vfpf_teardown_queue(struct bnx2x *bp, int qidx);
684 static int bnx2x_vfpf_teardown_queue(struct bnx2x *bp, int qidx) in bnx2x_vfpf_teardown_queue() argument
694 req->vf_qid = qidx; in bnx2x_vfpf_teardown_queue()
706 BNX2X_ERR("Sending TEARDOWN for queue %d failed: %d\n", qidx, in bnx2x_vfpf_teardown_queue()
713 BNX2X_ERR("TEARDOWN for queue %d failed: %d\n", qidx, in bnx2x_vfpf_teardown_queue()
/drivers/net/ethernet/chelsio/cxgb4vf/
Dsge.c1166 int qidx, credits, max_pkt_len; in t4vf_eth_xmit() local
1201 qidx = skb_get_queue_mapping(skb); in t4vf_eth_xmit()
1202 BUG_ON(qidx >= pi->nqsets); in t4vf_eth_xmit()
1203 txq = &adapter->sge.ethtxq[pi->first_qset + qidx]; in t4vf_eth_xmit()
1230 dev->name, qidx); in t4vf_eth_xmit()
Dcxgb4vf_main.c2581 int q10g, n10g, qidx, pidx, qs; in cfg_queues() local
2616 qidx = 0; in cfg_queues()
2620 pi->first_qset = qidx; in cfg_queues()
2622 qidx += pi->nqsets; in cfg_queues()
2624 s->ethqsets = qidx; in cfg_queues()
/drivers/net/wireless/broadcom/brcm80211/brcmfmac/
Dfwsignal.c1974 int qidx, hslot; in brcmf_fws_rollback_toq() local
1979 qidx = 2 * fifo; in brcmf_fws_rollback_toq()
1981 qidx++; in brcmf_fws_rollback_toq()
1983 pktout = brcmu_pktq_penq_head(&entry->psq, qidx, skb); in brcmf_fws_rollback_toq()
1985 brcmf_err("%s queue %d full\n", entry->name, qidx); in brcmf_fws_rollback_toq()
/drivers/net/ethernet/netronome/nfp/
Dnfp_net_common.c768 u16 qidx; in nfp_net_tx() local
770 qidx = skb_get_queue_mapping(skb); in nfp_net_tx()
771 tx_ring = &nn->tx_rings[qidx]; in nfp_net_tx()
773 nd_q = netdev_get_tx_queue(nn->netdev, qidx); in nfp_net_tx()
779 qidx, tx_ring->wr_p, tx_ring->rd_p); in nfp_net_tx()
/drivers/net/ethernet/chelsio/cxgb4/
Dcxgb4_main.c4013 int i, n10g = 0, qidx = 0; in cfg_queues() local
4044 pi->first_qset = qidx; in cfg_queues()
4046 qidx += pi->nqsets; in cfg_queues()
4061 pi->first_qset = qidx; in cfg_queues()
4063 qidx += pi->nqsets; in cfg_queues()
4067 s->ethqsets = qidx; in cfg_queues()
4068 s->max_ethqsets = qidx; /* MSI-X may lower it later */ in cfg_queues()
Dsge.c1169 int qidx, credits; in t4_eth_xmit() local
1202 qidx = skb_get_queue_mapping(skb); in t4_eth_xmit()
1203 q = &adap->sge.ethtxq[qidx + pi->first_qset]; in t4_eth_xmit()
1222 dev->name, qidx); in t4_eth_xmit()
/drivers/scsi/sym53c8xx_2/
Dsym_hipd.c1482 u_short qidx; local
1515 qidx = np->squeueput + 2;
1516 if (qidx >= MAX_QUEUE*2) qidx = 0;
1518 np->squeue [qidx] = cpu_to_scr(np->idletask_ba);
1522 np->squeueput = qidx;
/drivers/scsi/
Dncr53c8xx.c4451 u16 qidx; in ncr_put_start_queue() local
4457 qidx = np->squeueput + 2; in ncr_put_start_queue()
4458 if (qidx >= MAX_START + MAX_START) qidx = 1; in ncr_put_start_queue()
4460 np->scripth->tryloop [qidx] = cpu_to_scr(NCB_SCRIPT_PHYS (np, idle)); in ncr_put_start_queue()
4464 np->squeueput = qidx; in ncr_put_start_queue()