Home
last modified time | relevance | path

Searched refs:num_cqs (Results 1 – 25 of 38) sorted by relevance

12

/drivers/infiniband/hw/mthca/
Dmthca_cq.c221 cq = mthca_array_get(&dev->cq_table.cq, cqn & (dev->limits.num_cqs - 1)); in mthca_cq_completion()
241 cq = mthca_array_get(&dev->cq_table.cq, cqn & (dev->limits.num_cqs - 1)); in mthca_cq_event()
852 cq->cqn & (dev->limits.num_cqs - 1), cq); in mthca_init_cq()
929 cq->cqn & (dev->limits.num_cqs - 1)); in mthca_free_cq()
960 dev->limits.num_cqs, in mthca_init_cq_table()
967 dev->limits.num_cqs); in mthca_init_cq_table()
976 mthca_array_cleanup(&dev->cq_table.cq, dev->limits.num_cqs); in mthca_cleanup_cq_table()
Dmthca_profile.c188 dev->limits.num_cqs = profile[i].num; in mthca_make_profile()
Dmthca_dev.h154 int num_cqs; member
Dmthca_eq.c798 err = mthca_create_eq(dev, dev->limits.num_cqs + MTHCA_NUM_SPARE_EQE, in mthca_init_eq_table()
/drivers/scsi/elx/efct/
Defct_hw_queues.c193 u32 num_cqs, u32 entry_count) in efct_hw_new_cq_set() argument
203 for (i = 0; i < num_cqs; i++) in efct_hw_new_cq_set()
206 for (i = 0; i < num_cqs; i++) { in efct_hw_new_cq_set()
222 if (sli_cq_alloc_set(sli4, qs, num_cqs, entry_count, assefct)) { in efct_hw_new_cq_set()
227 for (i = 0; i < num_cqs; i++) { in efct_hw_new_cq_set()
236 for (i = 0; i < num_cqs; i++) { in efct_hw_new_cq_set()
Defct_hw.h741 u32 num_cqs, u32 entry_count);
/drivers/net/ethernet/mellanox/mlx4/
Dcq.c110 cqn & (dev->caps.num_cqs - 1)); in mlx4_cq_completion()
132 cq = radix_tree_lookup(&cq_table->tree, cqn & (dev->caps.num_cqs - 1)); in mlx4_cq_event()
472 return mlx4_bitmap_init(&cq_table->bitmap, dev->caps.num_cqs, in mlx4_init_cq_table()
473 dev->caps.num_cqs - 1, dev->caps.reserved_cqs, 0); in mlx4_init_cq_table()
Dprofile.c210 dev->caps.num_cqs = profile[i].num; in mlx4_make_profile()
/drivers/scsi/bfa/
Dbfa_core.c980 WARN_ON(cfg->fwcfg.num_cqs > BFI_IOC_MAX_CQS); in bfa_iocfc_send_cfg()
981 bfa_trc(bfa, cfg->fwcfg.num_cqs); in bfa_iocfc_send_cfg()
992 cfg_info->num_cqs = cfg->fwcfg.num_cqs; in bfa_iocfc_send_cfg()
1001 for (i = 0; i < cfg->fwcfg.num_cqs; i++) { in bfa_iocfc_send_cfg()
1109 for (i = 0; i < cfg->fwcfg.num_cqs; i++) { in bfa_iocfc_mem_claim()
1125 for (i = 0; i < cfg->fwcfg.num_cqs; i++) { in bfa_iocfc_mem_claim()
1285 fwcfg->num_cqs = fwcfg->num_cqs; in bfa_iocfc_cfgrsp()
1490 for (q = 0; q < cfg->fwcfg.num_cqs; q++) { in bfa_iocfc_meminfo()
1498 for (q = 0; q < cfg->fwcfg.num_cqs; q++) in bfa_iocfc_meminfo()
1979 cfg->fwcfg.num_cqs = DEF_CFG_NUM_CQS; in bfa_cfg_get_default()
Dbfi_ms.h36 u8 num_cqs; /* Number of CQs to be used */ member
Dbfa_defs_svc.h46 u8 num_cqs; member
/drivers/infiniband/hw/vmw_pvrdma/
Dpvrdma_cq.c129 if (!atomic_add_unless(&dev->num_cqs, 1, dev->dsr->caps.max_cq)) in pvrdma_create_cq()
221 atomic_dec(&dev->num_cqs); in pvrdma_create_cq()
266 atomic_dec(&dev->num_cqs); in pvrdma_destroy_cq()
Dpvrdma.h242 atomic_t num_cqs; member
/drivers/infiniband/hw/hns/
Dhns_roce_cq.c455 cqn & (hr_dev->caps.num_cqs - 1)); in hns_roce_cq_completion()
476 cqn & (hr_dev->caps.num_cqs - 1)); in hns_roce_cq_event()
522 cq_table->bank[i].max = hr_dev->caps.num_cqs / in hns_roce_init_cq_table()
Dhns_roce_main.c185 props->max_cq = hr_dev->caps.num_cqs; in hns_roce_query_device()
720 hr_dev->caps.num_cqs); in hns_roce_init_hem()
Dhns_roce_device.h744 u32 num_cqs; member
/drivers/net/ethernet/qlogic/qed/
Dqed_fcoe.c120 if (fcoe_pf_params->num_cqs > p_hwfn->hw_info.feat_num[QED_FCOE_CQ]) { in qed_sp_fcoe_func_start()
123 fcoe_pf_params->num_cqs, in qed_sp_fcoe_func_start()
166 p_data->q_params.num_queues = fcoe_pf_params->num_cqs; in qed_sp_fcoe_func_start()
171 for (i = 0; i < fcoe_pf_params->num_cqs; i++) { in qed_sp_fcoe_func_start()
740 info->num_cqs = FEAT_NUM(hwfn, QED_FCOE_CQ); in qed_fill_fcoe_dev_info()
Dqed_nvmetcp.c154 info->num_cqs = FEAT_NUM(hwfn, QED_NVMETCP_CQ); in qed_fill_nvmetcp_dev_info()
/drivers/net/ethernet/mellanox/mlxsw/
Dpci.c1047 u8 num_cqs; in mlxsw_pci_aqs_init() local
1063 num_cqs = mlxsw_cmd_mbox_query_aq_cap_max_num_cqs_get(mbox); in mlxsw_pci_aqs_init()
1069 if (num_sdqs + num_rdqs > num_cqs || in mlxsw_pci_aqs_init()
1071 num_cqs > MLXSW_PCI_CQS_MAX || num_eqs != MLXSW_PCI_EQS_COUNT) { in mlxsw_pci_aqs_init()
1096 num_cqs); in mlxsw_pci_aqs_init()
/drivers/scsi/qedi/
Dqedi.h55 #define MIN_NUM_CPUS_MSIX(x) min_t(u32, x->dev_info.num_cqs, \
/drivers/net/ethernet/broadcom/
Dcnic.h297 int num_cqs; member
Dcnic.c1477 cp->num_cqs = req1->num_cqs; in cnic_bnx2x_iscsi_init1()
1807 for (i = 1, j = 1; i < cp->num_cqs; i++, j++) { in cnic_setup_bnx2x_ctx()
1830 ictx->ustorm_st_context.num_cqs = cp->num_cqs; in cnic_setup_bnx2x_ctx()
1854 ictx->cstorm_st_context.cq_proc_en_bit_map = (1 << cp->num_cqs) - 1; in cnic_setup_bnx2x_ctx()
1855 for (i = 0; i < cp->num_cqs; i++) { in cnic_setup_bnx2x_ctx()
/drivers/scsi/elx/libefc_sli/
Dsli4.c817 struct sli4_queue *qs[], u32 num_cqs, in sli_cmd_cq_set_create() argument
847 (SZ_DMAADDR * num_pages_cq * num_cqs), in sli_cmd_cq_set_create()
863 SZ_DMAADDR * num_pages_cq * num_cqs); in sli_cmd_cq_set_create()
898 req->num_cq_req = cpu_to_le16(num_cqs); in sli_cmd_cq_set_create()
901 for (i = 0; i < num_cqs; i++) { in sli_cmd_cq_set_create()
918 u32 num_cqs, u32 n_entries, struct sli4_queue *eqs[]) in sli_cq_alloc_set() argument
926 for (i = 0; i < num_cqs; i++) { in sli_cq_alloc_set()
932 if (sli_cmd_cq_set_create(sli4, qs, num_cqs, eqs, &dma)) in sli_cq_alloc_set()
951 if (le16_to_cpu(res->num_q_allocated) != num_cqs) { in sli_cq_alloc_set()
956 for (i = 0; i < num_cqs; i++) { in sli_cq_alloc_set()
[all …]
/drivers/scsi/qedf/
Dqedf.h588 #define MIN_NUM_CPUS_MSIX(x) min_t(u32, x->dev_info.num_cqs, \
/drivers/scsi/bnx2i/
D57xx_iscsi_hsi.h555 u8 num_cqs; member
557 u8 num_cqs;

12