Home
last modified time | relevance | path

Searched refs:vaddr (Results 1 – 6 of 6) sorted by relevance

/mm/
Dhighmem.c154 struct page *kmap_to_page(void *vaddr) in kmap_to_page() argument
156 unsigned long addr = (unsigned long)vaddr; in kmap_to_page()
219 unsigned long vaddr; in map_new_virtual() local
261 vaddr = PKMAP_ADDR(last_pkmap_nr); in map_new_virtual()
262 set_pte_at(&init_mm, vaddr, in map_new_virtual()
266 set_page_address(page, (void *)vaddr); in map_new_virtual()
268 return vaddr; in map_new_virtual()
281 unsigned long vaddr; in kmap_high() local
288 vaddr = (unsigned long)page_address(page); in kmap_high()
289 if (!vaddr) in kmap_high()
[all …]
Ddmapool.c58 void *vaddr; member
217 *(int *)(page->vaddr + offset) = next; in pool_initialise_page()
229 page->vaddr = dma_alloc_coherent(pool->dev, pool->allocation, in pool_alloc_page()
231 if (page->vaddr) { in pool_alloc_page()
233 memset(page->vaddr, POOL_POISON_FREED, pool->allocation); in pool_alloc_page()
255 memset(page->vaddr, POOL_POISON_FREED, pool->allocation); in pool_free_page()
257 dma_free_coherent(pool->dev, pool->allocation, page->vaddr, dma); in pool_free_page()
292 pool->name, page->vaddr); in dma_pool_destroy()
296 pool->name, page->vaddr); in dma_pool_destroy()
347 page->offset = *(int *)(page->vaddr + offset); in dma_pool_alloc()
[all …]
Dzsmalloc.c917 void *vaddr; in init_zspage() local
928 vaddr = kmap_atomic(page); in init_zspage()
929 link = (struct link_free *)vaddr + off / sizeof(*link); in init_zspage()
943 kunmap_atomic(vaddr); in init_zspage()
1353 void *vaddr; in obj_malloc() local
1360 vaddr = kmap_atomic(m_page); in obj_malloc()
1361 link = (struct link_free *)vaddr + m_offset / sizeof(*link); in obj_malloc()
1369 kunmap_atomic(vaddr); in obj_malloc()
1439 void *vaddr; in obj_free() local
1452 vaddr = kmap_atomic(f_page); in obj_free()
[all …]
Dvmalloc.c1982 char *vaddr, *buf_start = buf; in vread() local
1999 vaddr = (char *) vm->addr; in vread()
2000 if (addr >= vaddr + get_vm_area_size(vm)) in vread()
2002 while (addr < vaddr) { in vread()
2010 n = vaddr + get_vm_area_size(vm) - addr; in vread()
2063 char *vaddr; in vwrite() local
2081 vaddr = (char *) vm->addr; in vwrite()
2082 if (addr >= vaddr + get_vm_area_size(vm)) in vwrite()
2084 while (addr < vaddr) { in vwrite()
2091 n = vaddr + get_vm_area_size(vm) - addr; in vwrite()
Dhugetlb.c3303 unsigned long vaddr = *position; in follow_hugetlb_page() local
3307 while (vaddr < vma->vm_end && remainder) { in follow_hugetlb_page()
3320 pte = huge_pte_offset(mm, vaddr & huge_page_mask(h)); in follow_hugetlb_page()
3333 !hugetlbfs_pagecache_present(h, vma, vaddr)) { in follow_hugetlb_page()
3357 ret = hugetlb_fault(mm, vma, vaddr, in follow_hugetlb_page()
3366 pfn_offset = (vaddr & ~huge_page_mask(h)) >> PAGE_SHIFT; in follow_hugetlb_page()
3377 vaddr += PAGE_SIZE; in follow_hugetlb_page()
3381 if (vaddr < vma->vm_end && remainder && in follow_hugetlb_page()
3392 *position = vaddr; in follow_hugetlb_page()
Diov_iter.c303 static size_t __iovec_copy_from_user_inatomic(char *vaddr, in __iovec_copy_from_user_inatomic() argument
313 left = __copy_from_user_inatomic(vaddr, buf, copy); in __iovec_copy_from_user_inatomic()
316 vaddr += copy; in __iovec_copy_from_user_inatomic()