/drivers/net/ethernet/huawei/hinic/ |
D | hinic_hw_eqs.c | 33 HINIC_CSR_AEQ_CONS_IDX_ADDR((eq)->q_id) : \ 34 HINIC_CSR_CEQ_CONS_IDX_ADDR((eq)->q_id)) 37 HINIC_CSR_AEQ_PROD_IDX_ADDR((eq)->q_id) : \ 38 HINIC_CSR_CEQ_PROD_IDX_ADDR((eq)->q_id)) 41 HINIC_CSR_AEQ_HI_PHYS_ADDR_REG((eq)->q_id, pg_num) : \ 42 HINIC_CSR_CEQ_HI_PHYS_ADDR_REG((eq)->q_id, pg_num)) 45 HINIC_CSR_AEQ_LO_PHYS_ADDR_REG((eq)->q_id, pg_num) : \ 46 HINIC_CSR_CEQ_LO_PHYS_ADDR_REG((eq)->q_id, pg_num)) 80 container_of((eq) - (eq)->q_id, struct hinic_aeqs, aeq[0]) 83 container_of((eq) - (eq)->q_id, struct hinic_ceqs, ceq[0]) [all …]
|
D | hinic_hw_csr.h | 88 #define HINIC_CSR_AEQ_HI_PHYS_ADDR_REG(q_id, pg_num) \ argument 89 (HINIC_CSR_AEQ_MTT_OFF(q_id) + \ 92 #define HINIC_CSR_CEQ_HI_PHYS_ADDR_REG(q_id, pg_num) \ argument 93 (HINIC_CSR_CEQ_MTT_OFF(q_id) + \ 96 #define HINIC_CSR_AEQ_LO_PHYS_ADDR_REG(q_id, pg_num) \ argument 97 (HINIC_CSR_AEQ_MTT_OFF(q_id) + \ 100 #define HINIC_CSR_CEQ_LO_PHYS_ADDR_REG(q_id, pg_num) \ argument 101 (HINIC_CSR_CEQ_MTT_OFF(q_id) + \
|
D | hinic_hw_io.c | 30 #define CI_ADDR(base_addr, q_id) ((base_addr) + \ argument 31 (q_id) * CI_Q_ADDR_SIZE) 132 base_qpn + qp->q_id); in write_sq_ctxts() 176 base_qpn + qp->q_id); in write_rq_ctxts() 272 struct hinic_qp *qp, int q_id, in init_qp() argument 281 qp->q_id = q_id; in init_qp() 283 err = hinic_wq_allocate(&func_to_io->wqs, &func_to_io->sq_wq[q_id], in init_qp() 291 err = hinic_wq_allocate(&func_to_io->wqs, &func_to_io->rq_wq[q_id], in init_qp() 306 func_to_io->sq_db[q_id] = db_base; in init_qp() 308 qp->sq.qid = q_id; in init_qp() [all …]
|
D | hinic_tx.c | 492 u16 prod_idx, q_id = skb->queue_mapping; in hinic_lb_xmit_frame() local 500 txq = &nic_dev->txqs[q_id]; in hinic_lb_xmit_frame() 512 netif_stop_subqueue(netdev, qp->q_id); in hinic_lb_xmit_frame() 516 netif_wake_subqueue(nic_dev->netdev, qp->q_id); in hinic_lb_xmit_frame() 535 netdev_txq = netdev_get_tx_queue(netdev, q_id); in hinic_lb_xmit_frame() 553 u16 prod_idx, q_id = skb->queue_mapping; in hinic_xmit_frame() local 561 txq = &nic_dev->txqs[q_id]; in hinic_xmit_frame() 593 netif_stop_subqueue(netdev, qp->q_id); in hinic_xmit_frame() 600 netif_wake_subqueue(nic_dev->netdev, qp->q_id); in hinic_xmit_frame() 624 netdev_txq = netdev_get_tx_queue(netdev, q_id); in hinic_xmit_frame() [all …]
|
D | hinic_hw_cmdq.h | 118 u32 q_id; member 181 enum hinic_set_arm_qtype q_type, u32 q_id);
|
D | hinic_rx.c | 424 skb_record_rx_queue(skb, qp->q_id); in rxq_recv() 534 intr_coal = &nic_dev->rx_intr_coalesce[qp->q_id]; in rx_request_irq() 551 cpumask_set_cpu(qp->q_id % num_online_cpus(), &rq->affinity_mask); in rx_request_irq() 596 "%s_rxq%d", netdev->name, qp->q_id); in hinic_init_rxq()
|
D | hinic_main.c | 825 u16 num_sqs, q_id; in hinic_tx_timeout() local 831 for (q_id = 0; q_id < num_sqs; q_id++) { in hinic_tx_timeout() 832 if (!netif_xmit_stopped(netdev_get_tx_queue(netdev, q_id))) in hinic_tx_timeout() 835 sq = hinic_hwdev_get_sq(nic_dev->hwdev, q_id); in hinic_tx_timeout() 840 q_id, sw_pi, hw_ci, sw_ci, in hinic_tx_timeout() 841 nic_dev->txqs[q_id].napi.state); in hinic_tx_timeout()
|
D | hinic_hw_qp.c | 42 #define SQ_CTXT_OFFSET(max_sqs, max_rqs, q_id) \ argument 43 (((max_rqs) + (max_sqs)) * CTXT_RSVD + (q_id) * Q_CTXT_SIZE) 45 #define RQ_CTXT_OFFSET(max_sqs, max_rqs, q_id) \ argument 47 (max_sqs + (q_id)) * Q_CTXT_SIZE) 629 HINIC_SQ_DB_INFO_SET(qp->q_id, QID)); in sq_prepare_db()
|
D | hinic_hw_cmdq.c | 513 enum hinic_set_arm_qtype q_type, u32 q_id) in hinic_set_arm_bit() argument 522 arm_bit.q_id = q_id; in hinic_set_arm_bit() 526 dev_err(&pdev->dev, "Failed to set arm for qid %d\n", q_id); in hinic_set_arm_bit()
|
D | hinic_ethtool.c | 681 static int set_queue_coalesce(struct hinic_dev *nic_dev, u16 q_id, in set_queue_coalesce() argument 691 intr_coal = set_rx_coal ? &nic_dev->rx_intr_coalesce[q_id] : in set_queue_coalesce() 692 &nic_dev->tx_intr_coalesce[q_id]; in set_queue_coalesce() 701 q_id >= nic_dev->num_qps) in set_queue_coalesce() 704 msix_idx = set_rx_coal ? nic_dev->rxqs[q_id].rq->msix_entry : in set_queue_coalesce() 705 nic_dev->txqs[q_id].sq->msix_entry; in set_queue_coalesce() 715 set_rx_coal ? "rx" : "tx", q_id); in set_queue_coalesce()
|
D | hinic_hw_eqs.h | 177 int q_id; member
|
D | hinic_hw_qp.h | 125 u16 q_id; member
|
/drivers/net/ethernet/intel/ice/ |
D | ice_base.c | 579 u16 q_id, q_base; in ice_vsi_map_rings_to_vectors() local 589 for (q_id = q_base; q_id < (q_base + tx_rings_per_v); q_id++) { in ice_vsi_map_rings_to_vectors() 590 struct ice_ring *tx_ring = vsi->tx_rings[q_id]; in ice_vsi_map_rings_to_vectors() 606 for (q_id = q_base; q_id < (q_base + rx_rings_per_v); q_id++) { in ice_vsi_map_rings_to_vectors() 607 struct ice_ring *rx_ring = vsi->rx_rings[q_id]; in ice_vsi_map_rings_to_vectors() 830 &txq_meta->q_id, &txq_meta->q_teid, rst_src, in ice_vsi_stop_tx_ring() 871 txq_meta->q_id = ring->reg_idx; in ice_fill_txq_meta()
|
D | ice_switch.h | 104 u16 q_id:11; member
|
/drivers/net/wireless/intel/iwlwifi/pcie/ |
D | tx.c | 144 txq_id != trans->txqs.cmd.q_id && in iwl_pcie_txq_inc_wr_ptr() 312 if (txq_id != trans->txqs.cmd.q_id) { in iwl_pcie_txq_unmap() 325 if (txq_id == trans->txqs.cmd.q_id) in iwl_pcie_txq_unmap() 363 if (txq_id == trans->txqs.cmd.q_id) in iwl_pcie_txq_free() 426 iwl_trans_ac_txq_enable(trans, trans->txqs.cmd.q_id, in iwl_pcie_tx_start() 635 bool cmd_queue = (txq_id == trans->txqs.cmd.q_id); in iwl_pcie_tx_alloc() 689 bool cmd_queue = (txq_id == trans->txqs.cmd.q_id); in iwl_pcie_tx_init() 761 if (WARN_ON(txq_id == trans->txqs.cmd.q_id)) in iwl_trans_pcie_reclaim() 1018 if (txq_id == trans->txqs.cmd.q_id && in iwl_trans_pcie_txq_enable() 1026 if (txq_id != trans->txqs.cmd.q_id) in iwl_trans_pcie_txq_enable() [all …]
|
D | tx-gen2.c | 78 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; in iwl_pcie_gen2_enqueue_hcmd() 190 cpu_to_le16(QUEUE_TO_SEQ(trans->txqs.cmd.q_id) | in iwl_pcie_gen2_enqueue_hcmd() 238 cmd_size, txq->write_ptr, idx, trans->txqs.cmd.q_id); in iwl_pcie_gen2_enqueue_hcmd() 315 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; in iwl_pcie_gen2_send_hcmd_sync()
|
D | trans-gen2.c | 248 if (iwl_txq_gen2_init(trans, trans->txqs.cmd.q_id, queue_size)) in iwl_pcie_gen2_nic_init()
|
D | ctxt-info.c | 269 cpu_to_le64(trans->txqs.txq[trans->txqs.cmd.q_id]->dma_addr); in iwl_pcie_ctxt_info_init()
|
D | ctxt-info-gen3.c | 223 cpu_to_le64(trans->txqs.txq[trans->txqs.cmd.q_id]->dma_addr); in iwl_pcie_ctxt_info_gen3_init()
|
/drivers/crypto/hisilicon/sec/ |
D | sec_drv.c | 688 int q_id; in sec_isr_handle() local 693 q_id = outorder_msg->data & SEC_OUT_BD_INFO_Q_ID_M; in sec_isr_handle() 694 msg = msg_ring->vaddr + q_id; in sec_isr_handle() 701 set_bit(q_id, queue->unprocessed); in sec_isr_handle() 702 if (q_id == queue->expected) in sec_isr_handle() 719 q_id = outorder_msg->data & SEC_OUT_BD_INFO_Q_ID_M; in sec_isr_handle() 720 msg = msg_ring->vaddr + q_id; in sec_isr_handle()
|
/drivers/staging/qlge/ |
D | qlge.h | 1245 u8 q_id; member 2243 u16 q_id); 2322 void ql_dump_hw_cb(struct ql_adapter *qdev, int size, u32 bit, u16 q_id); 2328 #define QL_DUMP_HW_CB(qdev, size, bit, q_id) \ argument 2329 ql_dump_hw_cb(qdev, size, bit, q_id) 2336 #define QL_DUMP_HW_CB(qdev, size, bit, q_id) argument
|
/drivers/net/hyperv/ |
D | hyperv_net.h | 693 u16 q_id; member 704 u16 q_id; member 710 u16 q_id; member 755 u16 q_id; member
|
/drivers/net/wireless/intel/iwlwifi/queue/ |
D | tx.c | 886 if (txq_id != trans->txqs.cmd.q_id) { in iwl_txq_gen2_unmap() 956 if (txq_id == trans->txqs.cmd.q_id) in iwl_txq_gen2_free() 1360 (txq_id == trans->txqs.cmd.q_id)); in iwl_txq_gen2_init() 1518 if (txq_id != trans->txqs.cmd.q_id) in iwl_txq_gen1_inval_byte_cnt_tbl()
|
/drivers/net/ethernet/hisilicon/hns/ |
D | hns_dsaf_main.c | 250 u32 q_id, q_num_per_port; in hns_dsaf_mix_def_qid_cfg() local 256 for (i = 0, q_id = 0; i < DSAF_SERVICE_NW_NUM; i++) { in hns_dsaf_mix_def_qid_cfg() 259 0xff, 0, q_id); in hns_dsaf_mix_def_qid_cfg() 260 q_id += q_num_per_port; in hns_dsaf_mix_def_qid_cfg() 267 u32 q_id, q_num_per_port; in hns_dsaf_inner_qid_cfg() local 276 for (mac_id = 0, q_id = 0; mac_id < DSAF_SERVICE_NW_NUM; mac_id++) { in hns_dsaf_inner_qid_cfg() 281 q_id); in hns_dsaf_inner_qid_cfg() 282 q_id += q_num_per_port; in hns_dsaf_inner_qid_cfg()
|
/drivers/net/ethernet/hisilicon/hns3/hns3pf/ |
D | hclge_err.c | 1705 u16 *q_id) in hclge_query_over_8bd_err_info() argument 1718 *q_id = le16_to_cpu(req->over_8bd_no_fe_qid); in hclge_query_over_8bd_err_info() 1735 u16 q_id; in hclge_handle_over_8bd_err() local 1738 ret = hclge_query_over_8bd_err_info(hdev, &vf_id, &q_id); in hclge_handle_over_8bd_err() 1746 vf_id, q_id); in hclge_handle_over_8bd_err()
|