Home
last modified time | relevance | path

Searched refs:dma_info (Results 1 – 10 of 10) sorted by relevance

/drivers/net/wireless/broadcom/brcm80211/brcmsmac/
Ddma.c208 struct dma_info { struct
313 static uint txd(struct dma_info *di, uint x) in txd()
318 static uint rxd(struct dma_info *di, uint x) in rxd()
323 static uint nexttxd(struct dma_info *di, uint i) in nexttxd()
328 static uint prevtxd(struct dma_info *di, uint i) in prevtxd()
333 static uint nextrxd(struct dma_info *di, uint i) in nextrxd()
338 static uint ntxdactive(struct dma_info *di, uint h, uint t) in ntxdactive()
343 static uint nrxdactive(struct dma_info *di, uint h, uint t) in nrxdactive()
348 static uint _dma_ctrlflags(struct dma_info *di, uint mask, uint flags) in _dma_ctrlflags()
383 static bool _dma64_addrext(struct dma_info *di, uint ctrl_offset) in _dma64_addrext()
[all …]
/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/
Drx.c19 struct mlx5e_dma_info *dma_info) in mlx5e_xsk_page_alloc_umem() argument
27 dma_info->xsk.handle = xsk_umem_adjust_offset(umem, handle, in mlx5e_xsk_page_alloc_umem()
29 dma_info->xsk.data = xdp_umem_get_data(umem, dma_info->xsk.handle); in mlx5e_xsk_page_alloc_umem()
36 dma_info->addr = xdp_umem_get_dma(umem, handle); in mlx5e_xsk_page_alloc_umem()
40 dma_sync_single_for_device(rq->pdev, dma_info->addr, PAGE_SIZE, in mlx5e_xsk_page_alloc_umem()
56 struct mlx5e_dma_info *dma_info) in mlx5e_xsk_page_release() argument
58 mlx5e_xsk_recycle_frame(rq, dma_info->xsk.handle); in mlx5e_xsk_page_release()
93 struct mlx5e_dma_info *di = &wi->umr.dma_info[page_idx]; in mlx5e_xsk_skb_from_cqe_mpwrq_linear()
Drx.h14 struct mlx5e_dma_info *dma_info);
16 struct mlx5e_dma_info *dma_info);
/drivers/net/ethernet/mellanox/mlx5/core/
Den_rx.c193 struct mlx5e_dma_info *dma_info) in mlx5e_rx_cache_put() argument
204 if (unlikely(mlx5e_page_is_reserved(dma_info->page))) { in mlx5e_rx_cache_put()
209 cache->page_cache[cache->tail] = *dma_info; in mlx5e_rx_cache_put()
215 struct mlx5e_dma_info *dma_info) in mlx5e_rx_cache_get() argument
230 *dma_info = cache->page_cache[cache->head]; in mlx5e_rx_cache_get()
234 dma_sync_single_for_device(rq->pdev, dma_info->addr, in mlx5e_rx_cache_get()
241 struct mlx5e_dma_info *dma_info) in mlx5e_page_alloc_pool() argument
243 if (mlx5e_rx_cache_get(rq, dma_info)) in mlx5e_page_alloc_pool()
246 dma_info->page = page_pool_dev_alloc_pages(rq->page_pool); in mlx5e_page_alloc_pool()
247 if (unlikely(!dma_info->page)) in mlx5e_page_alloc_pool()
[all …]
Den.h576 struct mlx5e_dma_info dma_info[MLX5_MPWRQ_PAGES_PER_WQE]; member
943 void mlx5e_page_dma_unmap(struct mlx5e_rq *rq, struct mlx5e_dma_info *dma_info);
945 struct mlx5e_dma_info *dma_info,
Den_main.c660 struct mlx5e_dma_info *dma_info = &rq->page_cache.page_cache[i]; in mlx5e_free_rq() local
666 mlx5e_page_release_dynamic(rq, dma_info, false); in mlx5e_free_rq()
/drivers/scsi/bfa/
Dbfad.c526 struct bfa_mem_dma_s *dma_info, *dma_elem; in bfad_hal_mem_release() local
530 dma_info = &hal_meminfo->dma_info; in bfad_hal_mem_release()
540 list_for_each(dm_qe, &dma_info->qe) { in bfad_hal_mem_release()
588 struct bfa_mem_dma_s *dma_info, *dma_elem; in bfad_hal_mem_alloc() local
599 dma_info = &hal_meminfo->dma_info; in bfad_hal_mem_alloc()
614 list_for_each(dm_qe, &dma_info->qe) { in bfad_hal_mem_alloc()
Dbfa_core.c1774 INIT_LIST_HEAD(&meminfo->dma_info.qe); in bfa_cfg_get_meminfo()
1831 struct bfa_mem_dma_s *dma_info, *dma_elem; in bfa_attach() local
1840 dma_info = &meminfo->dma_info; in bfa_attach()
1841 dma_info->kva_curp = dma_info->kva; in bfa_attach()
1842 dma_info->dma_curp = dma_info->dma; in bfa_attach()
1847 list_for_each(dm_qe, &dma_info->qe) { in bfa_attach()
Dbfa_ioc.h102 struct bfa_mem_dma_s dma_info; member
113 list_add_tail(&dm_ptr->qe, &meminfo->dma_info.qe); in bfa_mem_dma_setup()
/drivers/scsi/
Daha1542.c734 char dma_info[] = "no DMA"; in aha1542_hw_init() local
763 snprintf(dma_info, sizeof(dma_info), "DMA %d", sh->dma_channel); in aha1542_hw_init()
765 sh->this_id, base_io, sh->irq, dma_info); in aha1542_hw_init()