Searched refs:wqs (Results 1 – 11 of 11) sorted by relevance
/drivers/net/ethernet/huawei/hinic/ |
D | hinic_hw_wq.c | 31 #define WQS_FREE_BLOCKS_SIZE(wqs) (WQS_MAX_NUM_BLOCKS * \ argument 32 sizeof((wqs)->free_blocks[0])) 44 #define WQ_BASE_VADDR(wqs, wq) \ argument 45 ((void *)((wqs)->page_vaddr[(wq)->page_idx]) \ 48 #define WQ_BASE_PADDR(wqs, wq) \ argument 49 ((wqs)->page_paddr[(wq)->page_idx] \ 52 #define WQ_BASE_ADDR(wqs, wq) \ argument 53 ((void *)((wqs)->shadow_page_vaddr[(wq)->page_idx]) \ 137 static int wqs_allocate_page(struct hinic_wqs *wqs, int page_idx) in wqs_allocate_page() argument 139 return queue_alloc_page(wqs->hwif, &wqs->page_vaddr[page_idx], in wqs_allocate_page() [all …]
|
D | hinic_hw_wq.h | 85 int hinic_wqs_alloc(struct hinic_wqs *wqs, int num_wqs, 88 void hinic_wqs_free(struct hinic_wqs *wqs); 90 int hinic_wq_allocate(struct hinic_wqs *wqs, struct hinic_wq *wq, 94 void hinic_wq_free(struct hinic_wqs *wqs, struct hinic_wq *wq);
|
D | hinic_hw_io.c | 283 err = hinic_wq_allocate(&func_to_io->wqs, &func_to_io->sq_wq[q_id], in init_qp() 291 err = hinic_wq_allocate(&func_to_io->wqs, &func_to_io->rq_wq[q_id], in init_qp() 335 hinic_wq_free(&func_to_io->wqs, &func_to_io->rq_wq[q_id]); in init_qp() 338 hinic_wq_free(&func_to_io->wqs, &func_to_io->sq_wq[q_id]); in init_qp() 357 hinic_wq_free(&func_to_io->wqs, &func_to_io->rq_wq[q_id]); in destroy_qp() 358 hinic_wq_free(&func_to_io->wqs, &func_to_io->sq_wq[q_id]); in destroy_qp() 551 err = hinic_wqs_alloc(&func_to_io->wqs, 2 * max_qps, hwif); in hinic_io_init() 607 hinic_wqs_free(&func_to_io->wqs); in hinic_io_init() 633 hinic_wqs_free(&func_to_io->wqs); in hinic_io_free()
|
D | hinic_hw_io.h | 65 struct hinic_wqs wqs; member
|
/drivers/dma/idxd/ |
D | device.c | 331 struct idxd_wq *wq = idxd->wqs[i]; in idxd_wqs_unmap_portal() 722 struct idxd_wq *wq = idxd->wqs[i]; in idxd_device_wqs_clear_state() 781 iowrite64(group->grpcfg.wqs[i], idxd->reg_base + grpcfg_offset); in idxd_group_config_write() 926 struct idxd_wq *wq = idxd->wqs[i]; in idxd_wqs_config_write() 999 group->grpcfg.wqs[j] = 0; in idxd_wqs_setup() 1003 wq = idxd->wqs[i]; in idxd_wqs_setup() 1017 group->grpcfg.wqs[wq->id / 64] |= BIT(wq->id % 64); in idxd_wqs_setup() 1100 group->grpcfg.wqs[i] = ioread64(idxd->reg_base + grpcfg_offset); in idxd_group_load_config() 1102 group->id, i, grpcfg_offset, group->grpcfg.wqs[i]); in idxd_group_load_config() 1116 if (group->grpcfg.wqs[i] & BIT(j)) { in idxd_group_load_config() [all …]
|
D | irq.c | 41 struct idxd_wq *wq = idxd->wqs[i]; in idxd_device_reinit() 80 struct idxd_wq *wq = idxd->wqs[id]; in process_misc_interrupts() 88 struct idxd_wq *wq = idxd->wqs[i]; in process_misc_interrupts()
|
D | init.c | 211 idxd->wqs = kcalloc_node(idxd->max_wqs, sizeof(struct idxd_wq *), in idxd_setup_wqs() 213 if (!idxd->wqs) in idxd_setup_wqs() 248 idxd->wqs[i] = wq; in idxd_setup_wqs() 255 wq = idxd->wqs[i]; in idxd_setup_wqs() 371 put_device(wq_confdev(idxd->wqs[i])); in idxd_cleanup_internals() 417 put_device(wq_confdev(idxd->wqs[i])); in idxd_setup_internals() 743 wq = idxd->wqs[i]; in idxd_wqs_quiesce()
|
D | sysfs.c | 282 struct idxd_wq *wq = idxd->wqs[i]; in group_work_queues_show() 554 struct idxd_wq *wq = idxd->wqs[i]; in total_claimed_wq_size() 1112 struct idxd_wq *wq = idxd->wqs[i]; in clients_show() 1275 kfree(idxd->wqs); in idxd_conf_device_release() 1357 wq = idxd->wqs[i]; in idxd_register_wq_devices() 1368 wq = idxd->wqs[i]; in idxd_register_wq_devices() 1373 wq = idxd->wqs[j]; in idxd_register_wq_devices() 1413 device_unregister(wq_confdev(idxd->wqs[i])); in idxd_register_devices() 1424 struct idxd_wq *wq = idxd->wqs[i]; in idxd_unregister_devices()
|
D | registers.h | 301 u64 wqs[4]; member
|
D | idxd.h | 264 struct idxd_wq **wqs; member
|
/drivers/infiniband/core/ |
D | uverbs_cmd.c | 3055 struct ib_wq **wqs = NULL; in ib_uverbs_ex_create_rwq_ind_table() local 3088 wqs = kcalloc(num_wq_handles, sizeof(*wqs), GFP_KERNEL); in ib_uverbs_ex_create_rwq_ind_table() 3089 if (!wqs) { in ib_uverbs_ex_create_rwq_ind_table() 3103 wqs[num_read_wqs] = wq; in ib_uverbs_ex_create_rwq_ind_table() 3104 atomic_inc(&wqs[num_read_wqs]->usecnt); in ib_uverbs_ex_create_rwq_ind_table() 3120 init_attr.ind_tbl = wqs; in ib_uverbs_ex_create_rwq_ind_table() 3122 rwq_ind_tbl->ind_tbl = wqs; in ib_uverbs_ex_create_rwq_ind_table() 3135 rdma_lookup_put_uobject(&wqs[i]->uobject->uevent.uobject, in ib_uverbs_ex_create_rwq_ind_table() 3151 rdma_lookup_put_uobject(&wqs[i]->uobject->uevent.uobject, in ib_uverbs_ex_create_rwq_ind_table() 3153 atomic_dec(&wqs[i]->usecnt); in ib_uverbs_ex_create_rwq_ind_table() [all …]
|