Home
last modified time | relevance | path

Searched refs:qidx (Results 1 – 25 of 54) sorted by relevance

123

/drivers/net/ethernet/cavium/thunder/
Dnicvf_queues.c31 static int nicvf_poll_reg(struct nicvf *nic, int qidx, in nicvf_poll_reg() argument
42 reg_val = nicvf_queue_reg_read(nic, reg, qidx); in nicvf_poll_reg()
504 struct snd_queue *sq, int q_len, int qidx) in nicvf_init_snd_queue() argument
524 qidx += ((nic->sqs_id + 1) * MAX_SND_QUEUES_PER_QS); in nicvf_init_snd_queue()
525 if (qidx < nic->pnicvf->xdp_tx_queues) { in nicvf_init_snd_queue()
627 struct queue_set *qs, int qidx) in nicvf_reclaim_snd_queue() argument
630 nicvf_queue_reg_write(nic, NIC_QSET_SQ_0_7_CFG, qidx, 0); in nicvf_reclaim_snd_queue()
632 if (nicvf_poll_reg(nic, qidx, NIC_QSET_SQ_0_7_STATUS, 21, 1, 0x01)) in nicvf_reclaim_snd_queue()
635 nicvf_queue_reg_write(nic, NIC_QSET_SQ_0_7_CFG, qidx, NICVF_SQ_RESET); in nicvf_reclaim_snd_queue()
639 struct queue_set *qs, int qidx) in nicvf_reclaim_rcv_queue() argument
[all …]
Dnicvf_main.c75 static inline u8 nicvf_netdev_qidx(struct nicvf *nic, u8 qidx) in nicvf_netdev_qidx() argument
78 return qidx + ((nic->sqs_id + 1) * MAX_CMP_QUEUES_PER_QS); in nicvf_netdev_qidx()
80 return qidx; in nicvf_netdev_qidx()
104 u64 qidx, u64 val) in nicvf_queue_reg_write() argument
108 writeq_relaxed(val, addr + (qidx << NIC_Q_NUM_SHIFT)); in nicvf_queue_reg_write()
111 u64 nicvf_queue_reg_read(struct nicvf *nic, u64 offset, u64 qidx) in nicvf_queue_reg_read() argument
115 return readq_relaxed(addr + (qidx << NIC_Q_NUM_SHIFT)); in nicvf_queue_reg_read()
989 int qidx; in nicvf_handle_qs_err() local
995 for (qidx = 0; qidx < qs->cq_cnt; qidx++) { in nicvf_handle_qs_err()
997 qidx); in nicvf_handle_qs_err()
[all …]
Dnicvf_ethtool.c214 int stats, qidx; in nicvf_get_qset_strings() local
217 for (qidx = 0; qidx < nic->qs->rq_cnt; qidx++) { in nicvf_get_qset_strings()
219 sprintf(*data, "rxq%d: %s", qidx + start_qidx, in nicvf_get_qset_strings()
225 for (qidx = 0; qidx < nic->qs->sq_cnt; qidx++) { in nicvf_get_qset_strings()
227 sprintf(*data, "txq%d: %s", qidx + start_qidx, in nicvf_get_qset_strings()
301 int stat, qidx; in nicvf_get_qset_stats() local
306 for (qidx = 0; qidx < nic->qs->rq_cnt; qidx++) { in nicvf_get_qset_stats()
307 nicvf_update_rq_stats(nic, qidx); in nicvf_get_qset_stats()
309 *((*data)++) = ((u64 *)&nic->qs->rq[qidx].stats) in nicvf_get_qset_stats()
313 for (qidx = 0; qidx < nic->qs->sq_cnt; qidx++) { in nicvf_get_qset_stats()
[all …]
Dnicvf_queues.h336 int qidx, bool enable);
338 void nicvf_sq_enable(struct nicvf *nic, struct snd_queue *sq, int qidx);
339 void nicvf_sq_disable(struct nicvf *nic, int qidx);
342 struct snd_queue *sq, int qidx);
365 u64 qidx, u64 val);
367 u64 offset, u64 qidx);
/drivers/net/ethernet/marvell/octeontx2/nic/
Dotx2_common.c18 struct otx2_nic *pfvf, int qidx) in otx2_nix_rq_op_stats() argument
20 u64 incr = (u64)qidx << 32; in otx2_nix_rq_op_stats()
31 struct otx2_nic *pfvf, int qidx) in otx2_nix_sq_op_stats() argument
33 u64 incr = (u64)qidx << 32; in otx2_nix_sq_op_stats()
74 int otx2_update_rq_stats(struct otx2_nic *pfvf, int qidx) in otx2_update_rq_stats() argument
76 struct otx2_rcv_queue *rq = &pfvf->qset.rq[qidx]; in otx2_update_rq_stats()
81 otx2_nix_rq_op_stats(&rq->stats, pfvf, qidx); in otx2_update_rq_stats()
85 int otx2_update_sq_stats(struct otx2_nic *pfvf, int qidx) in otx2_update_sq_stats() argument
87 struct otx2_snd_queue *sq = &pfvf->qset.sq[qidx]; in otx2_update_sq_stats()
92 otx2_nix_sq_op_stats(&sq->stats, pfvf, qidx); in otx2_update_sq_stats()
[all …]
Dcn10k.c75 int cn10k_sq_aq_init(void *dev, u16 qidx, u16 sqb_aura) in cn10k_sq_aq_init() argument
85 aq->sq.cq = pfvf->hw.rx_queues + qidx; in cn10k_sq_aq_init()
103 aq->qidx = qidx; in cn10k_sq_aq_init()
137 void cn10k_sqe_flush(void *dev, struct otx2_snd_queue *sq, int size, int qidx) in cn10k_sqe_flush() argument
320 aq->qidx = rq_idx; in cn10k_map_unmap_rq_policer()
344 int qidx, rc; in cn10k_free_matchall_ipolicer() local
349 for (qidx = 0; qidx < hw->rx_queues; qidx++) in cn10k_free_matchall_ipolicer()
350 cn10k_map_unmap_rq_policer(pfvf, qidx, in cn10k_free_matchall_ipolicer()
453 aq->qidx = profile; in cn10k_set_ipolicer_rate()
464 int qidx, rc; in cn10k_set_matchall_ipolicer_rate() local
[all …]
Dotx2_pf.c1218 u64 qidx = 0; in otx2_q_intr_handler() local
1221 for (qidx = 0; qidx < pf->qset.cq_cnt; qidx++) { in otx2_q_intr_handler()
1223 val = otx2_atomic64_add((qidx << 44), ptr); in otx2_q_intr_handler()
1225 otx2_write64(pf, NIX_LF_CQ_OP_INT, (qidx << 44) | in otx2_q_intr_handler()
1233 qidx, otx2_read64(pf, NIX_LF_ERR_INT)); in otx2_q_intr_handler()
1237 qidx); in otx2_q_intr_handler()
1241 qidx); in otx2_q_intr_handler()
1248 for (qidx = 0; qidx < pf->hw.tx_queues; qidx++) { in otx2_q_intr_handler()
1258 val = otx2_atomic64_add((qidx << 44), ptr); in otx2_q_intr_handler()
1259 otx2_write64(pf, NIX_LF_SQ_OP_INT, (qidx << 44) | in otx2_q_intr_handler()
[all …]
Dotx2_ethtool.c86 int qidx, stats; in otx2_get_qset_strings() local
88 for (qidx = 0; qidx < pfvf->hw.rx_queues; qidx++) { in otx2_get_qset_strings()
90 sprintf(*data, "rxq%d: %s", qidx + start_qidx, in otx2_get_qset_strings()
95 for (qidx = 0; qidx < pfvf->hw.tx_queues; qidx++) { in otx2_get_qset_strings()
97 sprintf(*data, "txq%d: %s", qidx + start_qidx, in otx2_get_qset_strings()
145 int stat, qidx; in otx2_get_qset_stats() local
149 for (qidx = 0; qidx < pfvf->hw.rx_queues; qidx++) { in otx2_get_qset_stats()
150 if (!otx2_update_rq_stats(pfvf, qidx)) { in otx2_get_qset_stats()
156 *((*data)++) = ((u64 *)&pfvf->qset.rq[qidx].stats) in otx2_get_qset_stats()
160 for (qidx = 0; qidx < pfvf->hw.tx_queues; qidx++) { in otx2_get_qset_stats()
[all …]
Dotx2_txrx.c212 int qidx) in otx2_free_rcv_seg() argument
225 pfvf->hw_ops->aura_freeptr(pfvf, qidx, in otx2_free_rcv_seg()
232 struct nix_cqe_rx_s *cqe, int qidx) in otx2_check_rcv_errors() argument
240 qidx, parse->errlev, parse->errcode); in otx2_check_rcv_errors()
291 otx2_free_rcv_seg(pfvf, cqe, qidx); in otx2_check_rcv_errors()
490 int size, int qidx) in otx2_sqe_flush() argument
638 struct sk_buff *skb, u16 qidx) in otx2_sqe_add_hdr() argument
651 sqe_hdr->sq = qidx; in otx2_sqe_add_hdr()
766 struct sk_buff *skb, u16 qidx) in otx2_sq_append_tso() argument
768 struct netdev_queue *txq = netdev_get_tx_queue(pfvf->netdev, qidx); in otx2_sq_append_tso()
[all …]
Dotx2_txrx.h159 struct sk_buff *skb, u16 qidx);
161 int size, int qidx);
163 int size, int qidx);
Dcn10k.h28 void cn10k_sqe_flush(void *dev, struct otx2_snd_queue *sq, int size, int qidx);
29 int cn10k_sq_aq_init(void *dev, u16 qidx, u16 sqb_aura);
Dotx2_common.h309 int (*sq_aq_init)(void *dev, u16 qidx, u16 sqb_aura);
311 int size, int qidx);
764 void otx2_config_irq_coalescing(struct otx2_nic *pfvf, int qidx);
790 int otx2_sq_aq_init(void *dev, u16 qidx, u16 sqb_aura);
791 int cn10k_sq_aq_init(void *dev, u16 qidx, u16 sqb_aura);
824 int otx2_update_rq_stats(struct otx2_nic *pfvf, int qidx);
825 int otx2_update_sq_stats(struct otx2_nic *pfvf, int qidx);
Dotx2_vf.c380 int qidx = skb_get_queue_mapping(skb); in otx2vf_xmit() local
384 sq = &vf->qset.sq[qidx]; in otx2vf_xmit()
385 txq = netdev_get_tx_queue(netdev, qidx); in otx2vf_xmit()
387 if (!otx2_sq_append_skb(netdev, sq, skb, qidx)) { in otx2vf_xmit()
/drivers/net/ethernet/chelsio/inline_crypto/ch_ipsec/
Dchcr_ipsec.c418 u32 qidx; in copy_esn_pktxt() local
426 qidx = skb->queue_mapping; in copy_esn_pktxt()
427 q = &adap->sge.ethtxq[qidx + pi->first_qset]; in copy_esn_pktxt()
468 u32 ctrl0, qidx; in copy_cpltx_pktxt() local
474 qidx = skb->queue_mapping; in copy_cpltx_pktxt()
475 q = &adap->sge.ethtxq[qidx + pi->first_qset]; in copy_cpltx_pktxt()
513 unsigned int qidx; in copy_key_cpltx_pktxt() local
517 qidx = skb->queue_mapping; in copy_key_cpltx_pktxt()
518 q = &adap->sge.ethtxq[qidx + pi->first_qset]; in copy_key_cpltx_pktxt()
573 int qidx = skb_get_queue_mapping(skb); in ch_ipsec_crypto_wreq() local
[all …]
/drivers/dma/ptdma/
Dptdma-dev.c72 u8 *q_desc = (u8 *)&cmd_q->qbase[cmd_q->qidx]; in pt_core_execute_cmd()
84 cmd_q->qidx = (cmd_q->qidx + 1) % CMD_Q_LEN; in pt_core_execute_cmd()
90 tail = lower_32_bits(cmd_q->qdma_tail + cmd_q->qidx * Q_DESC_SIZE); in pt_core_execute_cmd()
140 tail = lower_32_bits(cmd_q->qdma_tail + cmd_q->qidx * Q_DESC_SIZE); in pt_do_cmd_complete()
214 cmd_q->qidx = 0; in pt_core_init()
/drivers/net/ethernet/broadcom/bnxt/
Dbnxt_dcb.c55 u8 qidx; in bnxt_hwrm_queue_pri2cos_cfg() local
60 qidx = bp->tc_to_qidx[ets->prio_tc[i]]; in bnxt_hwrm_queue_pri2cos_cfg()
61 pri2cos[i] = bp->q_info[qidx].queue_id; in bnxt_hwrm_queue_pri2cos_cfg()
109 u8 qidx = bp->tc_to_qidx[i]; in bnxt_hwrm_queue_cos2bw_cfg() local
113 qidx); in bnxt_hwrm_queue_cos2bw_cfg()
116 cos2bw.queue_id = bp->q_info[qidx].queue_id; in bnxt_hwrm_queue_cos2bw_cfg()
132 data = &req->unused_0 + qidx * (sizeof(cos2bw) - 4); in bnxt_hwrm_queue_cos2bw_cfg()
134 if (qidx == 0) { in bnxt_hwrm_queue_cos2bw_cfg()
269 u8 qidx = bp->tc_to_qidx[i]; in bnxt_hwrm_queue_pfc_cfg() local
271 if (!BNXT_LLQ(bp->q_info[qidx].queue_profile)) { in bnxt_hwrm_queue_pfc_cfg()
/drivers/scsi/csiostor/
Dcsio_wr.c745 csio_wr_cleanup_eq_stpg(struct csio_hw *hw, int qidx) in csio_wr_cleanup_eq_stpg() argument
747 struct csio_q *q = csio_hw_to_wrm(hw)->q_arr[qidx]; in csio_wr_cleanup_eq_stpg()
762 csio_wr_cleanup_iq_ftr(struct csio_hw *hw, int qidx) in csio_wr_cleanup_iq_ftr() argument
765 struct csio_q *q = wrm->q_arr[qidx]; in csio_wr_cleanup_iq_ftr()
863 csio_wr_get(struct csio_hw *hw, int qidx, uint32_t size, in csio_wr_get() argument
867 struct csio_q *q = wrm->q_arr[qidx]; in csio_wr_get()
878 CSIO_DB_ASSERT((qidx >= 0) && (qidx < wrm->free_qidx)); in csio_wr_get()
982 csio_wr_issue(struct csio_hw *hw, int qidx, bool prio) in csio_wr_issue() argument
985 struct csio_q *q = wrm->q_arr[qidx]; in csio_wr_issue()
987 CSIO_DB_ASSERT((qidx >= 0) && (qidx < wrm->free_qidx)); in csio_wr_issue()
[all …]
/drivers/scsi/qla2xxx/
Dqla_nvme.c88 unsigned int qidx, u16 qsize, void **handle) in qla_nvme_alloc_queue() argument
95 if (qidx) in qla_nvme_alloc_queue()
96 qidx--; in qla_nvme_alloc_queue()
103 __func__, handle, qidx, qsize); in qla_nvme_alloc_queue()
105 if (qidx > qla_nvme_fc_transport.max_hw_queues) { in qla_nvme_alloc_queue()
108 __func__, qidx, qla_nvme_fc_transport.max_hw_queues); in qla_nvme_alloc_queue()
116 if (ha->queue_pair_map[qidx]) { in qla_nvme_alloc_queue()
117 *handle = ha->queue_pair_map[qidx]; in qla_nvme_alloc_queue()
120 *handle, qidx); in qla_nvme_alloc_queue()
/drivers/gpu/drm/amd/amdgpu/
Damdgpu_amdkfd_gfx_v9.c810 int qidx; in kgd_gfx_v9_get_cu_occupancy() local
851 for (qidx = 0; qidx < max_queue_cnt; qidx++) { in kgd_gfx_v9_get_cu_occupancy()
856 if (!test_bit(qidx, cp_queue_bitmap)) in kgd_gfx_v9_get_cu_occupancy()
859 if (!(queue_map & (1 << qidx))) in kgd_gfx_v9_get_cu_occupancy()
863 get_wave_count(adev, qidx, &wave_cnt, &vmid); in kgd_gfx_v9_get_cu_occupancy()
/drivers/net/ethernet/marvell/octeontx2/af/
Drvu_nix.c890 if (!pfvf->rq_ctx || req->qidx >= pfvf->rq_ctx->qsize) in rvu_nix_blk_aq_enq_inst()
894 if (!pfvf->sq_ctx || req->qidx >= pfvf->sq_ctx->qsize) in rvu_nix_blk_aq_enq_inst()
898 if (!pfvf->cq_ctx || req->qidx >= pfvf->cq_ctx->qsize) in rvu_nix_blk_aq_enq_inst()
905 (req->qidx >= (256UL << (cfg & 0xF)))) in rvu_nix_blk_aq_enq_inst()
913 (req->qidx >= (256UL << (cfg & 0xF)))) in rvu_nix_blk_aq_enq_inst()
947 inst.cindex = req->qidx; in rvu_nix_blk_aq_enq_inst()
1023 __set_bit(req->qidx, pfvf->rq_bmap); in rvu_nix_blk_aq_enq_inst()
1025 __set_bit(req->qidx, pfvf->sq_bmap); in rvu_nix_blk_aq_enq_inst()
1027 __set_bit(req->qidx, pfvf->cq_bmap); in rvu_nix_blk_aq_enq_inst()
1033 (test_bit(req->qidx, pfvf->rq_bmap) & in rvu_nix_blk_aq_enq_inst()
[all …]
/drivers/scsi/lpfc/
Dlpfc_nvme.c83 unsigned int qidx, u16 qsize, in lpfc_nvme_create_queue() argument
101 qhandle->qidx = qidx; in lpfc_nvme_create_queue()
107 if (qidx) { in lpfc_nvme_create_queue()
109 qhandle->index = ((qidx - 1) % in lpfc_nvme_create_queue()
113 qhandle->index = qidx; in lpfc_nvme_create_queue()
119 qidx, qhandle->cpu_id, qhandle->index, qhandle); in lpfc_nvme_create_queue()
140 unsigned int qidx, in lpfc_nvme_delete_queue() argument
154 lport, qidx, handle); in lpfc_nvme_delete_queue()
1583 if (!lpfc_queue_info->qidx && !pnvme_fcreq->sg_cnt) { in lpfc_nvme_fcp_io_submit()
1658 lpfc_ncmd->qidx = lpfc_queue_info->qidx; in lpfc_nvme_fcp_io_submit()
Dlpfc_debugfs.h522 lpfc_debug_dump_hba_eq(struct lpfc_hba *phba, int qidx) in lpfc_debug_dump_hba_eq() argument
526 qp = phba->sli4_hba.hdwq[qidx].hba_eq; in lpfc_debug_dump_hba_eq()
528 pr_err("EQ[Idx:%d|Qid:%d]\n", qidx, qp->queue_id); in lpfc_debug_dump_hba_eq()
/drivers/scsi/mpi3mr/
Dmpi3mr.h175 #define REPLY_QUEUE_IDX_TO_MSIX_IDX(qidx, offset) (qidx + offset) argument
895 u64 *reply_dma, u16 qidx);
Dmpi3mr_fw.c1351 static int mpi3mr_delete_op_reply_q(struct mpi3mr_ioc *mrioc, u16 qidx) in mpi3mr_delete_op_reply_q() argument
1357 reply_qid = mrioc->op_reply_qinfo[qidx].qid; in mpi3mr_delete_op_reply_q()
1359 midx = REPLY_QUEUE_IDX_TO_MSIX_IDX(qidx, mrioc->op_reply_q_offset); in mpi3mr_delete_op_reply_q()
1413 mpi3mr_free_op_reply_q_segments(mrioc, qidx); in mpi3mr_delete_op_reply_q()
1432 static int mpi3mr_alloc_op_reply_q_segments(struct mpi3mr_ioc *mrioc, u16 qidx) in mpi3mr_alloc_op_reply_q_segments() argument
1434 struct op_reply_qinfo *op_reply_q = mrioc->op_reply_qinfo + qidx; in mpi3mr_alloc_op_reply_q_segments()
1489 static int mpi3mr_alloc_op_req_q_segments(struct mpi3mr_ioc *mrioc, u16 qidx) in mpi3mr_alloc_op_req_q_segments() argument
1491 struct op_req_qinfo *op_req_q = mrioc->req_qinfo + qidx; in mpi3mr_alloc_op_req_q_segments()
1547 static int mpi3mr_create_op_reply_q(struct mpi3mr_ioc *mrioc, u16 qidx) in mpi3mr_create_op_reply_q() argument
1550 struct op_reply_qinfo *op_reply_q = mrioc->op_reply_qinfo + qidx; in mpi3mr_create_op_reply_q()
[all …]
/drivers/net/ethernet/chelsio/cxgb3/
Dcxgb3_main.c402 int i, j, err, qidx = 0; in request_msix_data_irqs() local
408 err = request_irq(adap->msix_info[qidx + 1].vec, in request_msix_data_irqs()
410 adap->sge.qs[qidx]. in request_msix_data_irqs()
412 adap->msix_info[qidx + 1].desc, in request_msix_data_irqs()
413 &adap->sge.qs[qidx]); in request_msix_data_irqs()
415 while (--qidx >= 0) in request_msix_data_irqs()
416 free_irq(adap->msix_info[qidx + 1].vec, in request_msix_data_irqs()
417 &adap->sge.qs[qidx]); in request_msix_data_irqs()
420 qidx++; in request_msix_data_irqs()
941 static int send_pktsched_cmd(struct adapter *adap, int sched, int qidx, int lo, in send_pktsched_cmd() argument
[all …]

123