Home
last modified time | relevance | path

Searched refs:pvma (Results 1 – 10 of 10) sorted by relevance

/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/
Dnouveau_vmm.c62 nouveau_vma_del(struct nouveau_vma **pvma) in nouveau_vma_del() argument
64 struct nouveau_vma *vma = *pvma; in nouveau_vma_del()
71 kfree(*pvma); in nouveau_vma_del()
73 *pvma = NULL; in nouveau_vma_del()
78 struct nouveau_vma **pvma) in nouveau_vma_new() argument
85 if ((vma = *pvma = nouveau_vma_find(nvbo, vmm))) { in nouveau_vma_new()
90 if (!(vma = *pvma = kmalloc(sizeof(*vma), GFP_KERNEL))) in nouveau_vma_new()
116 nouveau_vma_del(pvma); in nouveau_vma_new()
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/
Dmemgf100.c35 u32 argc, u64 *paddr, u64 *psize, struct nvkm_vma **pvma) in gf100_mem_map() argument
55 nvkm_memory_size(memory), pvma); in gf100_mem_map()
59 ret = nvkm_memory_map(memory, 0, bar, *pvma, &uvmm, sizeof(uvmm)); in gf100_mem_map()
63 *paddr = device->func->resource_addr(device, 1) + (*pvma)->addr; in gf100_mem_map()
64 *psize = (*pvma)->size; in gf100_mem_map()
Dmemnv50.c35 u32 argc, u64 *paddr, u64 *psize, struct nvkm_vma **pvma) in nv50_mem_map() argument
56 ret = nvkm_vmm_get(bar, 12, size, pvma); in nv50_mem_map()
60 *paddr = device->func->resource_addr(device, 1) + (*pvma)->addr; in nv50_mem_map()
61 *psize = (*pvma)->size; in nv50_mem_map()
62 return nvkm_memory_map(memory, 0, bar, *pvma, &uvmm, sizeof(uvmm)); in nv50_mem_map()
Dmemnv04.c32 u32 argc, u64 *paddr, u64 *psize, struct nvkm_vma **pvma) in nv04_mem_map() argument
46 *pvma = ERR_PTR(-ENODEV); in nv04_mem_map()
Dvmm.c1623 nvkm_vmm_put(struct nvkm_vmm *vmm, struct nvkm_vma **pvma) in nvkm_vmm_put() argument
1625 struct nvkm_vma *vma = *pvma; in nvkm_vmm_put()
1630 *pvma = NULL; in nvkm_vmm_put()
1636 u8 shift, u8 align, u64 size, struct nvkm_vma **pvma) in nvkm_vmm_get_locked() argument
1768 *pvma = vma; in nvkm_vmm_get_locked()
1773 nvkm_vmm_get(struct nvkm_vmm *vmm, u8 page, u64 size, struct nvkm_vma **pvma) in nvkm_vmm_get() argument
1777 ret = nvkm_vmm_get_locked(vmm, false, true, false, page, 0, size, pvma); in nvkm_vmm_get()
Dvmm.h171 struct nvkm_vma **pvma);
/kernel/linux/linux-5.10/mm/
Dnommu.c565 struct vm_area_struct *pvma, *prev; in add_vma_to_mm() local
590 pvma = rb_entry(parent, struct vm_area_struct, vm_rb); in add_vma_to_mm()
594 if (vma->vm_start < pvma->vm_start) in add_vma_to_mm()
596 else if (vma->vm_start > pvma->vm_start) { in add_vma_to_mm()
599 } else if (vma->vm_end < pvma->vm_end) in add_vma_to_mm()
601 else if (vma->vm_end > pvma->vm_end) { in add_vma_to_mm()
604 } else if (vma < pvma) in add_vma_to_mm()
606 else if (vma > pvma) { in add_vma_to_mm()
Drmap.c330 int anon_vma_fork(struct vm_area_struct *vma, struct vm_area_struct *pvma) in anon_vma_fork() argument
337 if (!pvma->anon_vma) in anon_vma_fork()
347 error = anon_vma_clone(vma, pvma); in anon_vma_fork()
368 anon_vma->root = pvma->anon_vma->root; in anon_vma_fork()
369 anon_vma->parent = pvma->anon_vma; in anon_vma_fork()
Dshmem.c1526 struct vm_area_struct pvma; in shmem_swapin() local
1530 shmem_pseudo_vma_init(&pvma, info, index); in shmem_swapin()
1531 vmf.vma = &pvma; in shmem_swapin()
1534 shmem_pseudo_vma_destroy(&pvma); in shmem_swapin()
1542 struct vm_area_struct pvma; in shmem_alloc_hugepage() local
1552 shmem_pseudo_vma_init(&pvma, info, hindex); in shmem_alloc_hugepage()
1554 HPAGE_PMD_ORDER, &pvma, 0, numa_node_id(), true); in shmem_alloc_hugepage()
1555 shmem_pseudo_vma_destroy(&pvma); in shmem_alloc_hugepage()
1566 struct vm_area_struct pvma; in shmem_alloc_page() local
1569 shmem_pseudo_vma_init(&pvma, info, index); in shmem_alloc_page()
[all …]
Dmempolicy.c2662 struct vm_area_struct pvma; in mpol_shared_policy_init() local
2680 vma_init(&pvma, NULL); in mpol_shared_policy_init()
2681 pvma.vm_end = TASK_SIZE; /* policy covers entire file */ in mpol_shared_policy_init()
2682 mpol_set_shared_policy(sp, &pvma, new); /* adds ref */ in mpol_shared_policy_init()