Home
last modified time | relevance | path

Searched refs:dma_mem (Results 1 – 15 of 15) sorted by relevance

/drivers/infiniband/hw/irdma/
Dhmc.c487 struct irdma_dma_mem dma_mem; in irdma_add_sd_table_entry() local
498 dma_mem.size = ALIGN(alloc_len, IRDMA_HMC_PD_BP_BUF_ALIGNMENT); in irdma_add_sd_table_entry()
499 dma_mem.va = dma_alloc_coherent(hw->device, dma_mem.size, in irdma_add_sd_table_entry()
500 &dma_mem.pa, GFP_KERNEL); in irdma_add_sd_table_entry()
501 if (!dma_mem.va) in irdma_add_sd_table_entry()
510 dma_free_coherent(hw->device, dma_mem.size, in irdma_add_sd_table_entry()
511 dma_mem.va, dma_mem.pa); in irdma_add_sd_table_entry()
512 dma_mem.va = NULL; in irdma_add_sd_table_entry()
517 memcpy(&sd_entry->u.pd_table.pd_page_addr, &dma_mem, in irdma_add_sd_table_entry()
520 memcpy(&sd_entry->u.bp.addr, &dma_mem, in irdma_add_sd_table_entry()
Dverbs.h141 struct irdma_dma_mem dma_mem; member
Dutils.c1149 dma_free_coherent(rf->sc_dev.hw->device, iwqp->kqp.dma_mem.size, in irdma_free_qp_rsrc()
1150 iwqp->kqp.dma_mem.va, iwqp->kqp.dma_mem.pa); in irdma_free_qp_rsrc()
1151 iwqp->kqp.dma_mem.va = NULL; in irdma_free_qp_rsrc()
Dverbs.c568 struct irdma_dma_mem *mem = &iwqp->kqp.dma_mem; in irdma_setup_kmode_qp()
/drivers/scsi/be2iscsi/
Dbe.h31 struct be_dma_mem dma_mem; member
53 return q->dma_mem.va + q->head * q->entry_size; in queue_head_node()
58 return q->dma_mem.va + wrb_num * q->entry_size; in queue_get_wrb()
63 return q->dma_mem.va + q->tail * q->entry_size; in queue_tail_node()
Dbe_cmds.c741 struct be_dma_mem *q_mem = &eq->dma_mem; in beiscsi_cmd_eq_create()
783 struct be_dma_mem *q_mem = &cq->dma_mem; in beiscsi_cmd_cq_create()
856 struct be_dma_mem *q_mem = &mccq->dma_mem; in beiscsi_cmd_mccq_create()
982 struct be_dma_mem *q_mem = &dq->dma_mem; in be_cmd_create_default_pdu_queue()
Dbe_main.c2979 struct be_dma_mem *mem = &q->dma_mem; in be_fill_queue()
3011 mem = &eq->dma_mem; in beiscsi_create_eqs()
3048 mem = &eq->dma_mem; in beiscsi_create_eqs()
3077 mem = &cq->dma_mem; in beiscsi_create_cqs()
3111 mem = &cq->dma_mem; in beiscsi_create_cqs()
3136 mem = &dq->dma_mem; in beiscsi_create_def_hdr()
3188 mem = &dataq->dma_mem; in beiscsi_create_def_data()
3303 struct be_dma_mem *mem = &q->dma_mem; in be_queue_free()
3314 struct be_dma_mem *mem = &q->dma_mem; in be_queue_alloc()
/drivers/net/ethernet/qlogic/qed/
Dqed_cxt.c776 if (!p_t2 || !p_t2->dma_mem) in qed_cxt_src_t2_free()
780 if (p_t2->dma_mem[i].virt_addr) in qed_cxt_src_t2_free()
782 p_t2->dma_mem[i].size, in qed_cxt_src_t2_free()
783 p_t2->dma_mem[i].virt_addr, in qed_cxt_src_t2_free()
784 p_t2->dma_mem[i].phys_addr); in qed_cxt_src_t2_free()
786 kfree(p_t2->dma_mem); in qed_cxt_src_t2_free()
787 p_t2->dma_mem = NULL; in qed_cxt_src_t2_free()
797 if (!p_t2 || !p_t2->dma_mem) in qed_cxt_t2_alloc_pages()
802 p_virt = &p_t2->dma_mem[i].virt_addr; in qed_cxt_t2_alloc_pages()
806 &p_t2->dma_mem[i].phys_addr, in qed_cxt_t2_alloc_pages()
[all …]
Dqed_cxt.h303 struct phys_mem_desc *dma_mem; member
/drivers/scsi/
Dstex.c315 void *dma_mem; member
412 struct req_msg *req = hba->dma_mem + hba->req_head * hba->rq_size; in stex_alloc_req()
422 return (struct req_msg *)(hba->dma_mem + in stex_ss_alloc_req()
1150 msg_h = (struct st_msg_header *)hba->dma_mem; in stex_ss_handshake()
1740 hba->dma_mem = dma_alloc_coherent(&pdev->dev, in stex_probe()
1742 if (!hba->dma_mem) { in stex_probe()
1751 hba->dma_mem = dma_alloc_coherent(&pdev->dev, in stex_probe()
1755 if (!hba->dma_mem) { in stex_probe()
1772 hba->scratch = (__le32 *)(hba->dma_mem + scratch_offset); in stex_probe()
1773 hba->status_buffer = (struct status_msg *)(hba->dma_mem + sts_offset); in stex_probe()
[all …]
/drivers/net/ethernet/emulex/benet/
Dbe.h131 struct be_dma_mem dma_mem; member
153 return q->dma_mem.va + q->head * q->entry_size; in queue_head_node()
158 return q->dma_mem.va + q->tail * q->entry_size; in queue_tail_node()
163 return q->dma_mem.va + index * q->entry_size; in queue_index_node()
Dbe_cmds.c998 struct be_dma_mem *q_mem = &eqo->q.dma_mem; in be_cmd_eq_create()
1163 struct be_dma_mem *q_mem = &cq->dma_mem; in be_cmd_cq_create()
1241 struct be_dma_mem *q_mem = &mccq->dma_mem; in be_cmd_mccq_ext_create()
1306 struct be_dma_mem *q_mem = &mccq->dma_mem; in be_cmd_mccq_org_create()
1365 struct be_dma_mem *q_mem = &txq->dma_mem; in be_cmd_txq_create()
1413 struct be_dma_mem *q_mem = &rxq->dma_mem; in be_cmd_rxq_create()
Dbe_main.c147 struct be_dma_mem *mem = &q->dma_mem; in be_queue_free()
159 struct be_dma_mem *mem = &q->dma_mem; in be_queue_alloc()
1437 entry = txo->q.dma_mem.va; in be_tx_timeout()
1447 entry = txo->cq.dma_mem.va; in be_tx_timeout()
/drivers/net/ethernet/freescale/dpaa2/
Ddpaa2-eth.c3197 void *dma_mem, *key, *mask; in dpaa2_eth_set_vlan_qos() local
3211 dma_mem = kzalloc(DPAA2_CLASSIFIER_DMA_SIZE, GFP_KERNEL); in dpaa2_eth_set_vlan_qos()
3212 if (!dma_mem) in dpaa2_eth_set_vlan_qos()
3221 err = dpni_prepare_key_cfg(&kg_cfg, dma_mem); in dpaa2_eth_set_vlan_qos()
3230 qos_cfg.key_cfg_iova = dma_map_single(dev, dma_mem, in dpaa2_eth_set_vlan_qos()
3297 kfree(dma_mem); in dpaa2_eth_set_vlan_qos()
3743 u8 *dma_mem; in dpaa2_eth_set_dist_key() local
3773 dma_mem = kzalloc(DPAA2_CLASSIFIER_DMA_SIZE, GFP_KERNEL); in dpaa2_eth_set_dist_key()
3774 if (!dma_mem) in dpaa2_eth_set_dist_key()
3777 err = dpni_prepare_key_cfg(&cls_cfg, dma_mem); in dpaa2_eth_set_dist_key()
[all …]
Ddpaa2-switch.c834 u8 *dma_mem; in dpaa2_switch_fdb_iterate() local
838 dma_mem = kzalloc(fdb_dump_size, GFP_KERNEL); in dpaa2_switch_fdb_iterate()
839 if (!dma_mem) in dpaa2_switch_fdb_iterate()
842 fdb_dump_iova = dma_map_single(dev, dma_mem, fdb_dump_size, in dpaa2_switch_fdb_iterate()
860 fdb_entries = (struct fdb_dump_entry *)dma_mem; in dpaa2_switch_fdb_iterate()
870 kfree(dma_mem); in dpaa2_switch_fdb_iterate()
877 kfree(dma_mem); in dpaa2_switch_fdb_iterate()