/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
D | umem.c | 10 struct xdp_umem *umem) in mlx5e_xsk_map_umem() 37 struct xdp_umem *umem) in mlx5e_xsk_unmap_umem() 72 static int mlx5e_xsk_add_umem(struct mlx5e_xsk *xsk, struct xdp_umem *umem, u16 ix) in mlx5e_xsk_add_umem() 91 static bool mlx5e_xsk_is_umem_sane(struct xdp_umem *umem) in mlx5e_xsk_is_umem_sane() 96 void mlx5e_build_xsk_param(struct xdp_umem *umem, struct mlx5e_xsk_param *xsk) in mlx5e_build_xsk_param() 103 struct xdp_umem *umem, u16 ix) in mlx5e_xsk_enable_locked() 182 struct xdp_umem *umem = mlx5e_xsk_get_umem(&priv->channels.params, in mlx5e_xsk_disable_locked() local 208 static int mlx5e_xsk_enable_umem(struct mlx5e_priv *priv, struct xdp_umem *umem, in mlx5e_xsk_enable_umem() 231 int mlx5e_xsk_setup_umem(struct net_device *dev, struct xdp_umem *umem, u16 qid) in mlx5e_xsk_setup_umem() 244 int mlx5e_xsk_resize_reuseq(struct xdp_umem *umem, u32 nentries) in mlx5e_xsk_resize_reuseq()
|
D | tx.c | 66 struct xdp_umem *umem = sq->umem; in mlx5e_xsk_tx() local
|
D | rx.c | 21 struct xdp_umem *umem = rq->umem; in mlx5e_xsk_page_alloc_umem() local
|
D | setup.c | 65 struct mlx5e_xsk_param *xsk, struct xdp_umem *umem, in mlx5e_open_xsk()
|
/drivers/infiniband/core/ |
D | umem.c | 46 static void __ib_umem_release(struct ib_device *dev, struct ib_umem *umem, int dirty) in __ib_umem_release() 144 unsigned long ib_umem_find_best_pgsz(struct ib_umem *umem, in ib_umem_find_best_pgsz() 194 struct ib_umem *umem; in ib_umem_get() local 330 void ib_umem_release(struct ib_umem *umem) in ib_umem_release() 345 int ib_umem_page_count(struct ib_umem *umem) in ib_umem_page_count() 367 int ib_umem_copy_from(void *dst, struct ib_umem *umem, size_t offset, in ib_umem_copy_from()
|
D | umem_odp.c | 299 struct ib_umem *umem; in ib_umem_odp_alloc_implicit() local 347 struct ib_umem *umem; in ib_umem_odp_alloc_child() local 777 struct ib_umem_odp *umem; in rbt_ib_umem_for_each_in_range() local
|
/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
D | umem.c | 40 struct nvkm_umem *umem; in nvkm_umem_search() local 66 struct nvkm_umem *umem = nvkm_umem(object); in nvkm_umem_unmap() local 90 struct nvkm_umem *umem = nvkm_umem(object); in nvkm_umem_map() local 127 struct nvkm_umem *umem = nvkm_umem(object); in nvkm_umem_dtor() local 150 struct nvkm_umem *umem; in nvkm_umem_new() local
|
/drivers/net/ethernet/intel/i40e/ |
D | i40e_xsk.c | 19 static int i40e_xsk_umem_dma_map(struct i40e_vsi *vsi, struct xdp_umem *umem) in i40e_xsk_umem_dma_map() 53 static void i40e_xsk_umem_dma_unmap(struct i40e_vsi *vsi, struct xdp_umem *umem) in i40e_xsk_umem_dma_unmap() 77 static int i40e_xsk_umem_enable(struct i40e_vsi *vsi, struct xdp_umem *umem, in i40e_xsk_umem_enable() 137 struct xdp_umem *umem; in i40e_xsk_umem_disable() local 175 int i40e_xsk_umem_setup(struct i40e_vsi *vsi, struct xdp_umem *umem, in i40e_xsk_umem_setup() 193 struct xdp_umem *umem = rx_ring->xsk_umem; in i40e_run_xdp_zc() local 248 struct xdp_umem *umem = rx_ring->xsk_umem; in i40e_alloc_buffer_zc() local 289 struct xdp_umem *umem = rx_ring->xsk_umem; in i40e_alloc_buffer_slow_zc() local 724 struct xdp_umem *umem = tx_ring->xsk_umem; in i40e_clean_xdp_tx_irq() local 844 struct xdp_umem *umem = tx_ring->xsk_umem; in i40e_xsk_clean_tx_ring() local
|
/drivers/net/ethernet/intel/ixgbe/ |
D | ixgbe_xsk.c | 24 struct xdp_umem *umem) in ixgbe_xsk_umem_dma_map() 52 struct xdp_umem *umem) in ixgbe_xsk_umem_dma_unmap() 66 struct xdp_umem *umem, in ixgbe_xsk_umem_enable() 113 struct xdp_umem *umem; in ixgbe_xsk_umem_disable() local 135 int ixgbe_xsk_umem_setup(struct ixgbe_adapter *adapter, struct xdp_umem *umem, in ixgbe_xsk_umem_setup() 146 struct xdp_umem *umem = rx_ring->xsk_umem; in ixgbe_run_xdp_zc() local 258 struct xdp_umem *umem = rx_ring->xsk_umem; in ixgbe_alloc_buffer_zc() local 287 struct xdp_umem *umem = rx_ring->xsk_umem; in ixgbe_alloc_buffer_slow_zc() local 645 struct xdp_umem *umem = tx_ring->xsk_umem; in ixgbe_clean_xdp_tx_irq() local 732 struct xdp_umem *umem = tx_ring->xsk_umem; in ixgbe_xsk_clean_tx_ring() local
|
/drivers/infiniband/hw/mlx5/ |
D | mem.c | 46 void mlx5_ib_cont_pages(struct ib_umem *umem, u64 addr, in mlx5_ib_cont_pages() 128 void __mlx5_ib_populate_pas(struct mlx5_ib_dev *dev, struct ib_umem *umem, in __mlx5_ib_populate_pas() 192 void mlx5_ib_populate_pas(struct mlx5_ib_dev *dev, struct ib_umem *umem, in mlx5_ib_populate_pas()
|
D | doorbell.c | 41 struct ib_umem *umem; member
|
D | mr.c | 757 struct ib_umem **umem, int *npages, int *page_shift, in mr_umem_get() 850 struct ib_pd *pd, struct ib_umem *umem, in alloc_mr_from_cache() 890 struct ib_umem *umem = mr->umem; in populate_xlt() local 1052 struct ib_umem *umem, int npages, in reg_create() 1257 struct ib_umem *umem; in mlx5_ib_reg_user_mr() local 1576 struct ib_umem *umem = mr->umem; in dereg_mr() local
|
/drivers/vhost/ |
D | vhost.c | 591 void vhost_dev_reset_owner(struct vhost_dev *dev, struct vhost_umem *umem) in vhost_dev_reset_owner() 621 static void vhost_umem_free(struct vhost_umem *umem, in vhost_umem_free() 630 static void vhost_umem_clean(struct vhost_umem *umem) in vhost_umem_clean() 718 static bool vq_memory_access_ok(void __user *log_base, struct vhost_umem *umem, in vq_memory_access_ok() 758 static bool memory_access_ok(struct vhost_dev *d, struct vhost_umem *umem, in memory_access_ok() 1023 static int vhost_new_umem_range(struct vhost_umem *umem, in vhost_new_umem_range() 1054 static void vhost_del_umem_range(struct vhost_umem *umem, in vhost_del_umem_range() 1328 struct vhost_umem *umem = vq->iotlb; in iotlb_access_ok() local 1413 struct vhost_umem *umem = kvzalloc(sizeof(*umem), GFP_KERNEL); in vhost_umem_alloc() local 1878 struct vhost_umem *umem = vq->umem; in log_write_hva() local [all …]
|
D | vhost.h | 131 struct vhost_umem *umem; member 167 struct vhost_umem *umem; member
|
/drivers/infiniband/sw/siw/ |
D | siw_mem.c | 69 void siw_umem_release(struct siw_umem *umem, bool dirty) in siw_umem_release() 370 struct siw_umem *umem; in siw_umem_get() local
|
D | siw_mem.h | 63 static inline struct page *siw_get_upage(struct siw_umem *umem, u64 addr) in siw_get_upage()
|
/drivers/infiniband/sw/rdmavt/ |
D | mr.h | 59 struct ib_umem *umem; member
|
/drivers/infiniband/hw/mlx4/ |
D | doorbell.c | 40 struct ib_umem *umem; member
|
D | mr.c | 183 struct ib_umem *umem) in mlx4_ib_umem_write_mtt() 257 int mlx4_ib_umem_calc_optimal_mtt_size(struct ib_umem *umem, u64 start_va, in mlx4_ib_umem_calc_optimal_mtt_size()
|
/drivers/infiniband/hw/bnxt_re/ |
D | ib_verbs.h | 76 struct ib_umem *umem; member 106 struct ib_umem *umem; member
|
/drivers/gpu/drm/nouveau/include/nvkm/core/ |
D | client.h | 20 struct list_head umem; member
|
/drivers/infiniband/hw/vmw_pvrdma/ |
D | pvrdma_mr.c | 117 struct ib_umem *umem; in pvrdma_reg_user_mr() local
|
D | pvrdma_misc.c | 183 struct ib_umem *umem, u64 offset) in pvrdma_page_dir_insert_umem()
|
D | pvrdma.h | 91 struct ib_umem *umem; member 142 struct ib_umem *umem; member 172 struct ib_umem *umem; member
|
/drivers/infiniband/hw/usnic/ |
D | usnic_ib.h | 70 struct usnic_uiom_reg *umem; member
|