/drivers/scsi/snic/ |
D | snic_res.c | 110 snic->cq_count = svnic_dev_get_res_count(snic->vdev, RES_TYPE_CQ); in snic_get_res_counts() 111 SNIC_BUG_ON(snic->cq_count == 0); in snic_get_res_counts() 125 for (i = 0; i < snic->cq_count; i++) in snic_free_vnic_res() 157 snic->cq_count, in snic_alloc_vnic_res() 183 SNIC_BUG_ON(snic->cq_count != 2 * snic->wq_count); in snic_alloc_vnic_res() 185 for (i = snic->wq_count; i < snic->cq_count; i++) { in snic_alloc_vnic_res() 216 for (i = 0; i < snic->cq_count; i++) { in snic_alloc_vnic_res()
|
D | snic_isr.c | 156 if (snic->wq_count < n || snic->cq_count < n + m) in snic_set_intr_mode() 163 snic->cq_count = n + m; in snic_set_intr_mode()
|
D | snic.h | 279 unsigned int cq_count; member
|
D | snic_main.c | 238 for (i = 0; i < snic->cq_count; i++) in snic_cleanup()
|
/drivers/scsi/fnic/ |
D | fnic_isr.c | 241 fnic->cq_count >= n + m + o) { in fnic_set_intr_mode() 250 fnic->cq_count = n + m + o; in fnic_set_intr_mode() 269 fnic->cq_count >= 3 && in fnic_set_intr_mode() 276 fnic->cq_count = 3; in fnic_set_intr_mode() 297 fnic->cq_count >= 3 && in fnic_set_intr_mode() 303 fnic->cq_count = 3; in fnic_set_intr_mode()
|
D | fnic_res.c | 193 fnic->cq_count = vnic_dev_get_res_count(fnic->vdev, RES_TYPE_CQ); in fnic_get_res_counts() 211 for (i = 0; i < fnic->cq_count; i++) in fnic_free_vnic_resources() 240 fnic->rq_count, fnic->cq_count, fnic->intr_count); in fnic_alloc_vnic_resources() 366 for (i = 0; i < fnic->cq_count; i++) { in fnic_alloc_vnic_resources()
|
D | fnic.h | 237 unsigned int cq_count; member
|
/drivers/net/ethernet/cisco/enic/ |
D | enic_res.c | 182 for (i = 0; i < enic->cq_count; i++) in enic_free_vnic_resources() 192 enic->cq_count = vnic_dev_get_res_count(enic->vdev, RES_TYPE_CQ); in enic_get_res_counts() 199 enic->cq_count, enic->intr_count); in enic_get_res_counts() 256 for (i = 0; i < enic->cq_count; i++) { in enic_init_vnic_resources() 315 enic->cq_count, enic->intr_count, in enic_alloc_vnic_resources() 340 for (i = 0; i < enic->cq_count; i++) { in enic_alloc_vnic_resources()
|
D | enic_main.c | 2011 for (i = 0; i < enic->cq_count; i++) in enic_stop() 2389 enic->cq_count >= n + m && in enic_set_intr_mode() 2397 enic->cq_count = n + m; in enic_set_intr_mode() 2410 enic->cq_count >= 1 + m && in enic_set_intr_mode() 2417 enic->cq_count = 1 + m; in enic_set_intr_mode() 2435 enic->cq_count >= 2 && in enic_set_intr_mode() 2441 enic->cq_count = 2; in enic_set_intr_mode() 2460 enic->cq_count >= 2 && in enic_set_intr_mode() 2465 enic->cq_count = 2; in enic_set_intr_mode()
|
D | enic.h | 184 unsigned int cq_count; member
|
/drivers/net/ethernet/fungible/funcore/ |
D | fun_dev.c | 567 unsigned int cq_count, sq_count, num_dbs; in fun_get_dev_limits() local 573 cq_count = rc; in fun_get_dev_limits() 583 if (cq_count < 2 || sq_count < 2 + !!fdev->admin_q->rq_depth) in fun_get_dev_limits() 591 fdev->max_qid = min3(cq_count, sq_count, num_dbs / 2) - 1; in fun_get_dev_limits()
|
/drivers/infiniband/hw/bnxt_re/ |
D | bnxt_re.h | 167 atomic_t cq_count; member
|
D | ib_verbs.h | 100 u16 cq_count; member
|
D | qplib_res.h | 238 u32 cq_count; member
|
D | hw_counters.c | 267 stats->value[BNXT_RE_ACTIVE_CQ] = atomic_read(&rdev->cq_count); in bnxt_re_ib_get_hw_stats()
|
D | main.c | 172 ctx->cq_count = min_t(u32, BNXT_RE_MAX_CQ_COUNT, attr->max_cq); in bnxt_re_limit_pf_res() 196 vf_res->max_cq_per_vf = (qplib_ctx->cq_count * vf_pct) / num_vf; in bnxt_re_limit_vf_res() 784 atomic_set(&rdev->cq_count, 0); in bnxt_re_dev_add()
|
D | qplib_res.c | 537 hwq_attr.depth = ctx->cq_count; in bnxt_qplib_alloc_ctx()
|
/drivers/scsi/elx/efct/ |
D | efct_hw_queues.c | 25 hw->cq_count = 0; in efct_hw_init_queues() 169 cq->instance = eq->hw->cq_count++; in efct_hw_new_cq() 214 cq->instance = hw->cq_count++; in efct_hw_new_cq_set()
|
D | efct_hw.h | 393 u32 cq_count; member
|
/drivers/scsi/be2iscsi/ |
D | be.h | 88 u32 cq_count; member
|
/drivers/net/ethernet/cavium/thunder/ |
D | nicvf_main.c | 1803 u8 cq_count, txq_count; in nicvf_set_xdp_queues() local 1815 cq_count = max(nic->rx_queues, txq_count); in nicvf_set_xdp_queues() 1816 if (cq_count > MAX_CMP_QUEUES_PER_QS) { in nicvf_set_xdp_queues() 1817 nic->sqs_count = roundup(cq_count, MAX_CMP_QUEUES_PER_QS); in nicvf_set_xdp_queues()
|
/drivers/infiniband/hw/mlx4/ |
D | cq.c | 93 int mlx4_ib_modify_cq(struct ib_cq *cq, u16 cq_count, u16 cq_period) in mlx4_ib_modify_cq() argument 98 return mlx4_cq_modify(dev->dev, &mcq->mcq, cq_count, cq_period); in mlx4_ib_modify_cq()
|
D | mlx4_ib.h | 765 int mlx4_ib_modify_cq(struct ib_cq *cq, u16 cq_count, u16 cq_period);
|
/drivers/infiniband/hw/mlx5/ |
D | cq.c | 1129 int mlx5_ib_modify_cq(struct ib_cq *cq, u16 cq_count, u16 cq_period) in mlx5_ib_modify_cq() argument 1142 cq_period, cq_count); in mlx5_ib_modify_cq()
|
/drivers/net/ethernet/mellanox/mlxsw/ |
D | pci.c | 828 u8 cq_count = mlxsw_pci_cq_count(mlxsw_pci); in mlxsw_pci_eq_tasklet() local 868 for_each_set_bit(cqn, active_cqns, cq_count) { in mlxsw_pci_eq_tasklet()
|