Home
last modified time | relevance | path

Searched refs:qps (Results 1 – 25 of 32) sorted by relevance

12

/drivers/infiniband/hw/cxgb4/
Ddevice.c344 xa_for_each(&qpd->devp->qps, index, qp) in qp_open()
354 xa_lock_irq(&qpd->devp->qps); in qp_open()
355 xa_for_each(&qpd->devp->qps, index, qp) in qp_open()
357 xa_unlock_irq(&qpd->devp->qps); in qp_open()
940 WARN_ON(!xa_empty(&ctx->dev->qps)); in c4iw_dealloc()
1050 xa_init_flags(&devp->qps, XA_FLAGS_LOCK_IRQ); in c4iw_alloc()
1273 xa_lock_irqsave(&ctx->dev->qps, flags); in stop_queues()
1277 xa_for_each(&ctx->dev->qps, index, qp) in stop_queues()
1282 xa_unlock_irqrestore(&ctx->dev->qps, flags); in stop_queues()
1312 xa_lock_irq(&ctx->dev->qps); in resume_queues()
[all …]
Dev.c126 xa_lock_irq(&dev->qps); in c4iw_ev_dispatch()
127 qhp = xa_load(&dev->qps, CQE_QPID(err_cqe)); in c4iw_ev_dispatch()
134 xa_unlock_irq(&dev->qps); in c4iw_ev_dispatch()
149 xa_unlock_irq(&dev->qps); in c4iw_ev_dispatch()
155 xa_unlock_irq(&dev->qps); in c4iw_ev_dispatch()
Dqp.c66 xa_lock_irq(&dev->qps); in alloc_ird()
71 xa_unlock_irq(&dev->qps); in alloc_ird()
82 xa_lock_irq(&dev->qps); in free_ird()
84 xa_unlock_irq(&dev->qps); in free_ird()
916 xa_lock_irqsave(&qhp->rhp->qps, flags); in ring_kernel_sq_db()
925 xa_unlock_irqrestore(&qhp->rhp->qps, flags); in ring_kernel_sq_db()
933 xa_lock_irqsave(&qhp->rhp->qps, flags); in ring_kernel_rq_db()
942 xa_unlock_irqrestore(&qhp->rhp->qps, flags); in ring_kernel_rq_db()
2088 xa_lock_irq(&rhp->qps); in c4iw_destroy_qp()
2089 __xa_erase(&rhp->qps, qhp->wq.sq.qid); in c4iw_destroy_qp()
[all …]
Diw_cxgb4.h318 struct xarray qps; member
350 return xa_load(&rhp->qps, qpid); in get_qhp()
/drivers/net/ethernet/huawei/hinic/
Dhinic_hw_io.c129 qp = &func_to_io->qps[i]; in write_sq_ctxts()
173 qp = &func_to_io->qps[i]; in write_rq_ctxts()
381 func_to_io->qps = devm_kcalloc(&pdev->dev, num_qps, in hinic_io_create_qps()
382 sizeof(*func_to_io->qps), GFP_KERNEL); in hinic_io_create_qps()
383 if (!func_to_io->qps) in hinic_io_create_qps()
419 err = init_qp(func_to_io, &func_to_io->qps[i], i, in hinic_io_create_qps()
444 destroy_qp(func_to_io, &func_to_io->qps[j]); in hinic_io_create_qps()
459 devm_kfree(&pdev->dev, func_to_io->qps); in hinic_io_create_qps()
478 destroy_qp(func_to_io, &func_to_io->qps[i]); in hinic_io_destroy_qps()
488 devm_kfree(&pdev->dev, func_to_io->qps); in hinic_io_destroy_qps()
[all …]
Dhinic_hw_io.h70 struct hinic_qp *qps; member
Dhinic_hw_dev.c1061 struct hinic_qp *qp = &func_to_io->qps[i]; in hinic_hwdev_get_sq()
1079 struct hinic_qp *qp = &func_to_io->qps[i]; in hinic_hwdev_get_rq()
/drivers/crypto/hisilicon/sec2/
Dsec_main.c369 void sec_destroy_qps(struct hisi_qp **qps, int qp_num) in sec_destroy_qps() argument
371 hisi_qm_free_qps(qps, qp_num); in sec_destroy_qps()
372 kfree(qps); in sec_destroy_qps()
379 struct hisi_qp **qps; in sec_create_qps() local
382 qps = kcalloc(ctx_num, sizeof(struct hisi_qp *), GFP_KERNEL); in sec_create_qps()
383 if (!qps) in sec_create_qps()
386 ret = hisi_qm_alloc_qps_node(&sec_devices, ctx_num, 0, node, qps); in sec_create_qps()
388 return qps; in sec_create_qps()
390 kfree(qps); in sec_create_qps()
Dsec.h139 struct hisi_qp **qps; member
230 void sec_destroy_qps(struct hisi_qp **qps, int qp_num);
Dsec_crypto.c551 qp = ctx->qps[qp_ctx_id]; in sec_create_qp_ctx()
593 ctx->qps = sec_create_qps(); in sec_ctx_base_init()
594 if (!ctx->qps) { in sec_ctx_base_init()
599 sec = container_of(ctx->qps[0]->qm, struct sec_dev, qm); in sec_ctx_base_init()
607 ctx->fake_req_limit = ctx->qps[0]->sq_depth >> 1; in sec_ctx_base_init()
628 sec_destroy_qps(ctx->qps, sec->ctx_q_num); in sec_ctx_base_init()
639 sec_destroy_qps(ctx->qps, ctx->sec->ctx_q_num); in sec_ctx_base_uninit()
/drivers/gpu/drm/amd/display/dc/dml/dsc/
Drc_calc_fpu.c64 static void get_qp_set(qp_set qps, enum colour_mode cm, enum bits_per_comp bpc, in get_qp_set() argument
107 memcpy(qps, table[index].qps, sizeof(qp_set)); in get_qp_set()
Drc_calc_fpu.h76 const qp_set qps; member
/drivers/net/ethernet/mellanox/mlx4/
Den_rx.c1223 &rss_map->qps[i]); in mlx4_en_config_rss_steer()
1231 rss_map->indir_qp = &rss_map->qps[0]; in mlx4_en_config_rss_steer()
1307 MLX4_QP_STATE_RST, NULL, 0, 0, &rss_map->qps[i]); in mlx4_en_config_rss_steer()
1308 mlx4_qp_remove(mdev->dev, &rss_map->qps[i]); in mlx4_en_config_rss_steer()
1309 mlx4_qp_free(mdev->dev, &rss_map->qps[i]); in mlx4_en_config_rss_steer()
1333 MLX4_QP_STATE_RST, NULL, 0, 0, &rss_map->qps[i]); in mlx4_en_release_rss_steer()
1334 mlx4_qp_remove(mdev->dev, &rss_map->qps[i]); in mlx4_en_release_rss_steer()
1335 mlx4_qp_free(mdev->dev, &rss_map->qps[i]); in mlx4_en_release_rss_steer()
Den_main.c139 &priv->rss_map.qps[i], in mlx4_en_update_loopback_state()
Dmlx4_en.h443 struct mlx4_qp qps[MAX_RX_RINGS]; member
/drivers/crypto/hisilicon/zip/
Dzip_crypto.c558 struct hisi_qp *qps[HZIP_CTX_Q_NUM] = { NULL }; in hisi_zip_ctx_init() local
563 ret = zip_create_qps(qps, HZIP_CTX_Q_NUM, node); in hisi_zip_ctx_init()
569 hisi_zip = container_of(qps[0]->qm, struct hisi_zip, qm); in hisi_zip_ctx_init()
575 ret = hisi_zip_start_qp(qps[i], qp_ctx, i, req_type); in hisi_zip_ctx_init()
580 hisi_qm_free_qps(qps, HZIP_CTX_Q_NUM); in hisi_zip_ctx_init()
Dzip.h84 int zip_create_qps(struct hisi_qp **qps, int qp_num, int node);
Dzip_main.c453 int zip_create_qps(struct hisi_qp **qps, int qp_num, int node) in zip_create_qps() argument
458 return hisi_qm_alloc_qps_node(&zip_devices, qp_num, 0, node, qps); in zip_create_qps()
/drivers/infiniband/hw/qedr/
Dqedr_iw_cm.c522 xa_lock(&dev->qps); in qedr_iw_load_qp()
523 qp = xa_load(&dev->qps, qpn); in qedr_iw_load_qp()
526 xa_unlock(&dev->qps); in qedr_iw_load_qp()
817 return xa_load(&dev->qps, qpn); in qedr_iw_get_qp()
Dqedr.h171 struct xarray qps; member
/drivers/infiniband/hw/qib/
Dqib_verbs.c671 struct rvt_qp *qps[20]; in qib_verbs_sdma_desc_avail() local
684 if (n == ARRAY_SIZE(qps)) in qib_verbs_sdma_desc_avail()
691 qps[n++] = qp; in qib_verbs_sdma_desc_avail()
697 qp = qps[i]; in qib_verbs_sdma_desc_avail()
1147 struct rvt_qp *qps[5]; in qib_ib_piobufavail() local
1164 if (n == ARRAY_SIZE(qps)) in qib_ib_piobufavail()
1170 qps[n++] = qp; in qib_ib_piobufavail()
1177 qp = qps[i]; in qib_ib_piobufavail()
/drivers/infiniband/hw/hfi1/
Dpio.c1573 struct rvt_qp *qps[PIO_WAIT_BATCH_SIZE]; in sc_piobufavail() local
1593 if (n == ARRAY_SIZE(qps)) in sc_piobufavail()
1602 priv = qps[top_idx]->priv; in sc_piobufavail()
1609 qps[n++] = qp; in sc_piobufavail()
1624 hfi1_qp_wakeup(qps[top_idx], in sc_piobufavail()
1628 hfi1_qp_wakeup(qps[i], in sc_piobufavail()
/drivers/infiniband/hw/ocrdma/
Docrdma_hw.h104 enum ib_qp_state get_ibqp_state(enum ocrdma_qp_state qps);
/drivers/net/ethernet/intel/i40e/
Di40e_virtchnl_pf.c896 u16 vsi_id, qps; in i40e_map_pf_queues_to_vsi() local
904 qps = vf->ch[i].num_qps; in i40e_map_pf_queues_to_vsi()
907 qps = pf->vsi[vf->lan_vsi_idx]->alloc_queue_pairs; in i40e_map_pf_queues_to_vsi()
912 if (j * 2 >= qps) { in i40e_map_pf_queues_to_vsi()
943 u32 qps, num_tc = 1; /* VF has at least one traffic class */ in i40e_map_pf_to_vf_queues() local
952 qps = vf->ch[i].num_qps; in i40e_map_pf_to_vf_queues()
955 qps = pf->vsi[vf->lan_vsi_idx]->alloc_queue_pairs; in i40e_map_pf_to_vf_queues()
959 for (j = 0; j < qps; j++) { in i40e_map_pf_to_vf_queues()
/drivers/crypto/hisilicon/
Dqm.c3383 void hisi_qm_free_qps(struct hisi_qp **qps, int qp_num) in hisi_qm_free_qps() argument
3387 if (!qps || qp_num <= 0) in hisi_qm_free_qps()
3391 hisi_qm_release_qp(qps[i]); in hisi_qm_free_qps()
3453 u8 alg_type, int node, struct hisi_qp **qps) in hisi_qm_alloc_qps_node() argument
3460 if (!qps || !qm_list || qp_num <= 0) in hisi_qm_alloc_qps_node()
3471 qps[i] = hisi_qm_create_qp(tmp->qm, alg_type); in hisi_qm_alloc_qps_node()
3472 if (IS_ERR(qps[i])) { in hisi_qm_alloc_qps_node()
3473 hisi_qm_free_qps(qps, i); in hisi_qm_alloc_qps_node()

12