Home
last modified time | relevance | path

Searched refs:nvkm_vma (Results 1 – 25 of 25) sorted by relevance

/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/
Dvmm.c747 static inline struct nvkm_vma *
750 struct nvkm_vma *vma = kzalloc(sizeof(*vma), GFP_KERNEL); in nvkm_vma_new()
760 struct nvkm_vma *
761 nvkm_vma_tail(struct nvkm_vma *vma, u64 tail) in nvkm_vma_tail()
763 struct nvkm_vma *new; in nvkm_vma_tail()
785 nvkm_vmm_free_remove(struct nvkm_vmm *vmm, struct nvkm_vma *vma) in nvkm_vmm_free_remove()
791 nvkm_vmm_free_delete(struct nvkm_vmm *vmm, struct nvkm_vma *vma) in nvkm_vmm_free_delete()
799 nvkm_vmm_free_insert(struct nvkm_vmm *vmm, struct nvkm_vma *vma) in nvkm_vmm_free_insert()
805 struct nvkm_vma *this = rb_entry(*ptr, typeof(*this), tree); in nvkm_vmm_free_insert()
827 nvkm_vmm_node_remove(struct nvkm_vmm *vmm, struct nvkm_vma *vma) in nvkm_vmm_node_remove()
[all …]
Dmem.h12 u64 *, u64 *, struct nvkm_vma **);
17 u64 *, u64 *, struct nvkm_vma **);
22 u64 *, u64 *, struct nvkm_vma **);
Dvmm.h166 struct nvkm_vma *nvkm_vmm_node_search(struct nvkm_vmm *, u64 addr);
167 struct nvkm_vma *nvkm_vmm_node_split(struct nvkm_vmm *, struct nvkm_vma *,
171 struct nvkm_vma **pvma);
172 void nvkm_vmm_put_locked(struct nvkm_vmm *, struct nvkm_vma *);
173 void nvkm_vmm_unmap_locked(struct nvkm_vmm *, struct nvkm_vma *, bool pfn);
174 void nvkm_vmm_unmap_region(struct nvkm_vmm *, struct nvkm_vma *);
188 struct nvkm_vma *nvkm_vma_tail(struct nvkm_vma *, u64 tail);
Duvmm.c116 struct nvkm_vma *vma; in nvkm_uvmm_mthd_unmap()
160 struct nvkm_vma *vma; in nvkm_uvmm_mthd_map()
238 struct nvkm_vma *vma; in nvkm_uvmm_mthd_put()
276 struct nvkm_vma *vma; in nvkm_uvmm_mthd_get()
Dumem.h19 struct nvkm_vma *bar;
Dmem.c71 struct nvkm_vma *vma, void *argv, u32 argc) in nvkm_mem_map_dma()
111 struct nvkm_vma *vma, void *argv, u32 argc) in nvkm_mem_map_sgl()
Dmemnv04.c32 u32 argc, u64 *paddr, u64 *psize, struct nvkm_vma **pvma) in nv04_mem_map()
Dpriv.h26 u32 argc, u64 *addr, u64 *size, struct nvkm_vma **);
Dmemnv50.c35 u32 argc, u64 *paddr, u64 *psize, struct nvkm_vma **pvma) in nv50_mem_map()
Dmemgf100.c35 u32 argc, u64 *paddr, u64 *psize, struct nvkm_vma **pvma) in gf100_mem_map()
/drivers/gpu/drm/nouveau/include/nvkm/subdev/
Dmmu.h6 struct nvkm_vma { struct
59 int nvkm_vmm_get(struct nvkm_vmm *, u8 page, u64 size, struct nvkm_vma **);
60 void nvkm_vmm_put(struct nvkm_vmm *, struct nvkm_vma **);
80 int nvkm_vmm_map(struct nvkm_vmm *, struct nvkm_vma *, void *argv, u32 argc,
82 void nvkm_vmm_unmap(struct nvkm_vmm *, struct nvkm_vma *);
Dbar.h5 struct nvkm_vma;
/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/
Dnv50.c49 struct nvkm_vma *bar;
127 struct nvkm_vma *bar = NULL, *ebar; in nv50_instobj_kmap()
184 struct nvkm_vma *vma, void *argv, u32 argc) in nv50_instobj_map()
315 struct nvkm_vma *bar; in nv50_instobj_dtor()
Dgk20a.c282 struct nvkm_vma *vma, void *argv, u32 argc) in gk20a_instobj_map()
/drivers/gpu/drm/nouveau/include/nvkm/core/
Dmemory.h6 struct nvkm_vma;
39 struct nvkm_vma *, void *argv, u32 argc);
Dgpuobj.h32 struct nvkm_vma *, void *argv, u32 argc);
/drivers/gpu/drm/nouveau/nvkm/engine/fifo/
Dchangf100.h17 struct nvkm_vma *vma;
Dgf100.h30 struct nvkm_vma *bar;
Dchangk104.h21 struct nvkm_vma *vma;
Dgk104.h43 struct nvkm_vma *bar;
/drivers/gpu/drm/nouveau/nvkm/core/
Dgpuobj.c47 struct nvkm_vmm *vmm, struct nvkm_vma *vma, in nvkm_gpuobj_heap_map()
109 struct nvkm_vmm *vmm, struct nvkm_vma *vma, in nvkm_gpuobj_map()
/drivers/gpu/drm/nouveau/nvkm/engine/gr/
Dgf100.h262 struct nvkm_vma *mmio_vma;
267 struct nvkm_vma *vma;
Dctxgf100.c1444 struct nvkm_vma *ctx = NULL; in gf100_grctx_generate()
/drivers/gpu/drm/nouveau/nvkm/subdev/acr/
Dpriv.h109 struct nvkm_vma *vma;
/drivers/gpu/drm/nouveau/nvkm/subdev/fb/
Dram.c39 struct nvkm_vma *vma, void *argv, u32 argc) in nvkm_vram_map()