Home
last modified time | relevance | path

Searched refs:dma_pool (Results 1 – 25 of 112) sorted by relevance

12345

/drivers/gpu/drm/ttm/
Dttm_page_alloc_dma.c97 struct dma_pool { struct
151 struct dma_pool *pool;
266 static int ttm_set_pages_caching(struct dma_pool *pool, in ttm_set_pages_caching()
286 static void __ttm_dma_free_page(struct dma_pool *pool, struct dma_page *d_page) in __ttm_dma_free_page()
299 static struct dma_page *__ttm_dma_alloc_page(struct dma_pool *pool) in __ttm_dma_alloc_page()
344 static void ttm_pool_update_free_locked(struct dma_pool *pool, in ttm_pool_update_free_locked()
353 static void ttm_dma_page_put(struct dma_pool *pool, struct dma_page *d_page) in ttm_dma_page_put()
370 static void ttm_dma_pages_put(struct dma_pool *pool, struct list_head *d_pages, in ttm_dma_pages_put()
404 static unsigned ttm_dma_page_pool_free(struct dma_pool *pool, unsigned nr_free, in ttm_dma_page_pool_free()
501 struct dma_pool *pool; in ttm_dma_free_pool()
[all …]
/drivers/scsi/megaraid/
Dmegaraid_mbox.h194 struct dma_pool *mbox_pool_handle;
196 struct dma_pool *epthru_pool_handle;
198 struct dma_pool *sg_pool_handle;
Dmegaraid_ioctl.h241 struct dma_pool *handle;
293 struct dma_pool *pthru_dma_pool;
Dmegaraid_sas_fusion.h1274 struct dma_pool *dma_pool_ptr;
1284 struct dma_pool *io_request_frames_pool;
1288 struct dma_pool *sg_dma_pool;
1289 struct dma_pool *sense_dma_pool;
1297 struct dma_pool *reply_frames_desc_pool;
1298 struct dma_pool *reply_frames_desc_pool_align;
/drivers/crypto/ccp/
Dccp-dev-v3.c373 struct dma_pool *dma_pool; in ccp_init() local
388 dma_pool = dma_pool_create(dma_pool_name, dev, in ccp_init()
391 if (!dma_pool) { in ccp_init()
402 cmd_q->dma_pool = dma_pool; in ccp_init()
513 dma_pool_destroy(ccp->cmd_q[i].dma_pool); in ccp_init()
551 dma_pool_destroy(ccp->cmd_q[i].dma_pool); in ccp_destroy()
Dccp-dev.h267 struct dma_pool *dma_pool; member
459 struct dma_pool *dma_pool; member
Dccp-dev-v5.c783 struct dma_pool *dma_pool; in ccp5_init() local
799 dma_pool = dma_pool_create(dma_pool_name, dev, in ccp5_init()
802 if (!dma_pool) { in ccp5_init()
813 cmd_q->dma_pool = dma_pool; in ccp5_init()
990 dma_pool_destroy(ccp->cmd_q[i].dma_pool); in ccp5_init()
/drivers/net/ethernet/intel/ixgbe/
Dixgbe_fcoe.h42 struct dma_pool *pool;
47 struct dma_pool *pool;
/drivers/firmware/google/
Dgsmi.c99 struct dma_pool *dma_pool; /* DMA buffer pool */ member
159 smibuf->start = dma_pool_alloc(gsmi_dev.dma_pool, GFP_KERNEL, in gsmi_buf_alloc()
178 dma_pool_free(gsmi_dev.dma_pool, smibuf->start, in gsmi_buf_free()
898 gsmi_dev.dma_pool = dma_pool_create("gsmi", &gsmi_dev.pdev->dev, in gsmi_init()
900 if (!gsmi_dev.dma_pool) in gsmi_init()
1016 dma_pool_destroy(gsmi_dev.dma_pool); in gsmi_init()
1038 dma_pool_destroy(gsmi_dev.dma_pool); in gsmi_exit()
/drivers/usb/musb/
Dtusb6010_omap.c54 struct tusb_dma_data dma_pool[MAX_DMAREQ]; member
400 chdat->dma_data = &chdat->tusb_dma->dma_pool[dmareq_nr]; in tusb_omap_dma_allocate_dmareq()
476 chdat->dma_data = &tusb_dma->dma_pool[0]; in tusb_omap_dma_allocate()
528 if (tusb_dma && tusb_dma->dma_pool[i].chan) in tusb_dma_controller_destroy()
529 dma_release_channel(tusb_dma->dma_pool[i].chan); in tusb_dma_controller_destroy()
543 struct tusb_dma_data *dma_data = &tusb_dma->dma_pool[i]; in tusb_omap_allocate_dma_pool()
573 struct tusb_dma_data *dma_data = &tusb_dma->dma_pool[i]; in tusb_omap_allocate_dma_pool()
/drivers/scsi/lpfc/
Dlpfc.h987 struct dma_pool *lpfc_sg_dma_buf_pool;
988 struct dma_pool *lpfc_mbuf_pool;
989 struct dma_pool *lpfc_hrb_pool; /* header receive buffer pool */
990 struct dma_pool *lpfc_drb_pool; /* data receive buffer pool */
991 struct dma_pool *lpfc_nvmet_drb_pool; /* data receive buffer pool */
992 struct dma_pool *lpfc_hbq_pool; /* SLI3 hbq buffer pool */
993 struct dma_pool *txrdy_payload_pool;
994 struct dma_pool *lpfc_cmd_rsp_buf_pool;
/drivers/usb/chipidea/
Dci.h102 struct dma_pool *td_pool;
225 struct dma_pool *qh_pool;
226 struct dma_pool *td_pool;
/drivers/net/wan/
Dixp4xx_hss.c324 static struct dma_pool *dma_pool; variable
970 dma_pool = dma_pool_create(DRV_NAME, &port->netdev->dev, in init_hdlc_queues()
972 if (!dma_pool) in init_hdlc_queues()
976 if (!(port->desc_tab = dma_pool_alloc(dma_pool, GFP_KERNEL, in init_hdlc_queues()
1033 dma_pool_free(dma_pool, port->desc_tab, port->desc_tab_phys); in destroy_hdlc_queues()
1037 if (!ports_open && dma_pool) { in destroy_hdlc_queues()
1038 dma_pool_destroy(dma_pool); in destroy_hdlc_queues()
1039 dma_pool = NULL; in destroy_hdlc_queues()
/drivers/dma/
Dfsl_raid.h261 struct dma_pool *cf_desc_pool;
262 struct dma_pool *hw_desc_pool;
Dcoh901318.h18 struct dma_pool *dmapool;
Dat_hdmac_regs.h333 struct dma_pool *dma_desc_pool;
334 struct dma_pool *memset_pool;
/drivers/usb/host/
Dehci.h195 struct dma_pool *qh_pool; /* qh per active urb */
196 struct dma_pool *qtd_pool; /* one or more per qh */
197 struct dma_pool *itd_pool; /* itd per iso urb */
198 struct dma_pool *sitd_pool; /* sitd per split iso urb */
Dfotg210.h162 struct dma_pool *qh_pool; /* qh per active urb */
163 struct dma_pool *qtd_pool; /* one or more per qh */
164 struct dma_pool *itd_pool; /* itd per iso urb */
/drivers/crypto/marvell/
Dcesa.h395 struct dma_pool *tdma_desc_pool;
396 struct dma_pool *op_pool;
397 struct dma_pool *cache_pool;
398 struct dma_pool *padding_pool;
/drivers/crypto/hisilicon/sec/
Dsec_drv.h394 struct dma_pool;
414 struct dma_pool *hw_sgl_pool;
/drivers/scsi/mpt3sas/
Dmpt3sas_base.h1348 struct dma_pool *pcie_sgl_dma_pool;
1355 struct dma_pool *chain_dma_pool;
1383 struct dma_pool *sense_dma_pool;
1391 struct dma_pool *reply_dma_pool;
1397 struct dma_pool *reply_free_dma_pool;
1406 struct dma_pool *reply_post_free_dma_pool;
1407 struct dma_pool *reply_post_free_array_dma_pool;
/drivers/net/ethernet/xscale/
Dixp4xx_eth.c253 static struct dma_pool *dma_pool; variable
1102 dma_pool = dma_pool_create(DRV_NAME, &port->netdev->dev, in init_queues()
1104 if (!dma_pool) in init_queues()
1108 if (!(port->desc_tab = dma_pool_alloc(dma_pool, GFP_KERNEL, in init_queues()
1166 dma_pool_free(dma_pool, port->desc_tab, port->desc_tab_phys); in destroy_queues()
1170 if (!ports_open && dma_pool) { in destroy_queues()
1171 dma_pool_destroy(dma_pool); in destroy_queues()
1172 dma_pool = NULL; in destroy_queues()
/drivers/crypto/ccree/
Dcc_buffer_mgr.h33 struct dma_pool *curr_pool;
/drivers/s390/cio/
Ddevice.c691 cio_gp_dma_free(cdev->private->dma_pool, cdev->private->dma_area, in ccw_device_release()
693 cio_gp_dma_destroy(cdev->private->dma_pool, &cdev->dev); in ccw_device_release()
703 struct gen_pool *dma_pool; in io_subchannel_allocate_dev() local
714 dma_pool = cio_gp_dma_create(&cdev->dev, 1); in io_subchannel_allocate_dev()
715 if (!dma_pool) in io_subchannel_allocate_dev()
717 cdev->private->dma_pool = dma_pool; in io_subchannel_allocate_dev()
718 cdev->private->dma_area = cio_gp_dma_zalloc(dma_pool, &cdev->dev, in io_subchannel_allocate_dev()
724 cio_gp_dma_destroy(dma_pool, &cdev->dev); in io_subchannel_allocate_dev()
/drivers/nvme/host/
Dlightnvm.c717 struct dma_pool *dma_pool = pool; in nvme_nvm_destroy_dma_pool() local
719 dma_pool_destroy(dma_pool); in nvme_nvm_destroy_dma_pool()
780 ppa_list = dma_pool_alloc(dev->dma_pool, GFP_KERNEL, &ppa_dma); in nvme_nvm_submit_user_cmd()
802 metadata = dma_pool_alloc(dev->dma_pool, GFP_KERNEL, in nvme_nvm_submit_user_cmd()
840 dma_pool_free(dev->dma_pool, metadata, metadata_dma); in nvme_nvm_submit_user_cmd()
846 dma_pool_free(dev->dma_pool, ppa_list, ppa_dma); in nvme_nvm_submit_user_cmd()

12345