/drivers/net/wireless/broadcom/brcm80211/brcmsmac/ |
D | dma.c | 208 struct dma_info { struct 313 static uint txd(struct dma_info *di, uint x) in txd() 318 static uint rxd(struct dma_info *di, uint x) in rxd() 323 static uint nexttxd(struct dma_info *di, uint i) in nexttxd() 328 static uint prevtxd(struct dma_info *di, uint i) in prevtxd() 333 static uint nextrxd(struct dma_info *di, uint i) in nextrxd() 338 static uint ntxdactive(struct dma_info *di, uint h, uint t) in ntxdactive() 343 static uint nrxdactive(struct dma_info *di, uint h, uint t) in nrxdactive() 348 static uint _dma_ctrlflags(struct dma_info *di, uint mask, uint flags) in _dma_ctrlflags() 383 static bool _dma64_addrext(struct dma_info *di, uint ctrl_offset) in _dma64_addrext() [all …]
|
/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
D | rx.c | 19 struct mlx5e_dma_info *dma_info) in mlx5e_xsk_page_alloc_umem() argument 27 dma_info->xsk.handle = xsk_umem_adjust_offset(umem, handle, in mlx5e_xsk_page_alloc_umem() 29 dma_info->xsk.data = xdp_umem_get_data(umem, dma_info->xsk.handle); in mlx5e_xsk_page_alloc_umem() 36 dma_info->addr = xdp_umem_get_dma(umem, handle); in mlx5e_xsk_page_alloc_umem() 40 dma_sync_single_for_device(rq->pdev, dma_info->addr, PAGE_SIZE, in mlx5e_xsk_page_alloc_umem() 56 struct mlx5e_dma_info *dma_info) in mlx5e_xsk_page_release() argument 58 mlx5e_xsk_recycle_frame(rq, dma_info->xsk.handle); in mlx5e_xsk_page_release() 93 struct mlx5e_dma_info *di = &wi->umr.dma_info[page_idx]; in mlx5e_xsk_skb_from_cqe_mpwrq_linear()
|
D | rx.h | 14 struct mlx5e_dma_info *dma_info); 16 struct mlx5e_dma_info *dma_info);
|
/drivers/net/ethernet/mellanox/mlx5/core/ |
D | en_rx.c | 193 struct mlx5e_dma_info *dma_info) in mlx5e_rx_cache_put() argument 204 if (unlikely(mlx5e_page_is_reserved(dma_info->page))) { in mlx5e_rx_cache_put() 209 cache->page_cache[cache->tail] = *dma_info; in mlx5e_rx_cache_put() 215 struct mlx5e_dma_info *dma_info) in mlx5e_rx_cache_get() argument 230 *dma_info = cache->page_cache[cache->head]; in mlx5e_rx_cache_get() 234 dma_sync_single_for_device(rq->pdev, dma_info->addr, in mlx5e_rx_cache_get() 241 struct mlx5e_dma_info *dma_info) in mlx5e_page_alloc_pool() argument 243 if (mlx5e_rx_cache_get(rq, dma_info)) in mlx5e_page_alloc_pool() 246 dma_info->page = page_pool_dev_alloc_pages(rq->page_pool); in mlx5e_page_alloc_pool() 247 if (unlikely(!dma_info->page)) in mlx5e_page_alloc_pool() [all …]
|
D | en.h | 576 struct mlx5e_dma_info dma_info[MLX5_MPWRQ_PAGES_PER_WQE]; member 943 void mlx5e_page_dma_unmap(struct mlx5e_rq *rq, struct mlx5e_dma_info *dma_info); 945 struct mlx5e_dma_info *dma_info,
|
D | en_main.c | 660 struct mlx5e_dma_info *dma_info = &rq->page_cache.page_cache[i]; in mlx5e_free_rq() local 666 mlx5e_page_release_dynamic(rq, dma_info, false); in mlx5e_free_rq()
|
/drivers/scsi/bfa/ |
D | bfad.c | 526 struct bfa_mem_dma_s *dma_info, *dma_elem; in bfad_hal_mem_release() local 530 dma_info = &hal_meminfo->dma_info; in bfad_hal_mem_release() 540 list_for_each(dm_qe, &dma_info->qe) { in bfad_hal_mem_release() 588 struct bfa_mem_dma_s *dma_info, *dma_elem; in bfad_hal_mem_alloc() local 599 dma_info = &hal_meminfo->dma_info; in bfad_hal_mem_alloc() 614 list_for_each(dm_qe, &dma_info->qe) { in bfad_hal_mem_alloc()
|
D | bfa_core.c | 1774 INIT_LIST_HEAD(&meminfo->dma_info.qe); in bfa_cfg_get_meminfo() 1831 struct bfa_mem_dma_s *dma_info, *dma_elem; in bfa_attach() local 1840 dma_info = &meminfo->dma_info; in bfa_attach() 1841 dma_info->kva_curp = dma_info->kva; in bfa_attach() 1842 dma_info->dma_curp = dma_info->dma; in bfa_attach() 1847 list_for_each(dm_qe, &dma_info->qe) { in bfa_attach()
|
D | bfa_ioc.h | 102 struct bfa_mem_dma_s dma_info; member 113 list_add_tail(&dm_ptr->qe, &meminfo->dma_info.qe); in bfa_mem_dma_setup()
|
/drivers/scsi/ |
D | aha1542.c | 734 char dma_info[] = "no DMA"; in aha1542_hw_init() local 763 snprintf(dma_info, sizeof(dma_info), "DMA %d", sh->dma_channel); in aha1542_hw_init() 765 sh->this_id, base_io, sh->irq, dma_info); in aha1542_hw_init()
|