Searched refs:vaddr (Results 1 – 6 of 6) sorted by relevance
/mm/ |
D | highmem.c | 154 struct page *kmap_to_page(void *vaddr) in kmap_to_page() argument 156 unsigned long addr = (unsigned long)vaddr; in kmap_to_page() 219 unsigned long vaddr; in map_new_virtual() local 261 vaddr = PKMAP_ADDR(last_pkmap_nr); in map_new_virtual() 262 set_pte_at(&init_mm, vaddr, in map_new_virtual() 266 set_page_address(page, (void *)vaddr); in map_new_virtual() 268 return vaddr; in map_new_virtual() 281 unsigned long vaddr; in kmap_high() local 288 vaddr = (unsigned long)page_address(page); in kmap_high() 289 if (!vaddr) in kmap_high() [all …]
|
D | dmapool.c | 58 void *vaddr; member 217 *(int *)(page->vaddr + offset) = next; in pool_initialise_page() 229 page->vaddr = dma_alloc_coherent(pool->dev, pool->allocation, in pool_alloc_page() 231 if (page->vaddr) { in pool_alloc_page() 233 memset(page->vaddr, POOL_POISON_FREED, pool->allocation); in pool_alloc_page() 255 memset(page->vaddr, POOL_POISON_FREED, pool->allocation); in pool_free_page() 257 dma_free_coherent(pool->dev, pool->allocation, page->vaddr, dma); in pool_free_page() 292 pool->name, page->vaddr); in dma_pool_destroy() 296 pool->name, page->vaddr); in dma_pool_destroy() 347 page->offset = *(int *)(page->vaddr + offset); in dma_pool_alloc() [all …]
|
D | zsmalloc.c | 917 void *vaddr; in init_zspage() local 928 vaddr = kmap_atomic(page); in init_zspage() 929 link = (struct link_free *)vaddr + off / sizeof(*link); in init_zspage() 943 kunmap_atomic(vaddr); in init_zspage() 1353 void *vaddr; in obj_malloc() local 1360 vaddr = kmap_atomic(m_page); in obj_malloc() 1361 link = (struct link_free *)vaddr + m_offset / sizeof(*link); in obj_malloc() 1369 kunmap_atomic(vaddr); in obj_malloc() 1439 void *vaddr; in obj_free() local 1452 vaddr = kmap_atomic(f_page); in obj_free() [all …]
|
D | vmalloc.c | 1982 char *vaddr, *buf_start = buf; in vread() local 1999 vaddr = (char *) vm->addr; in vread() 2000 if (addr >= vaddr + get_vm_area_size(vm)) in vread() 2002 while (addr < vaddr) { in vread() 2010 n = vaddr + get_vm_area_size(vm) - addr; in vread() 2063 char *vaddr; in vwrite() local 2081 vaddr = (char *) vm->addr; in vwrite() 2082 if (addr >= vaddr + get_vm_area_size(vm)) in vwrite() 2084 while (addr < vaddr) { in vwrite() 2091 n = vaddr + get_vm_area_size(vm) - addr; in vwrite()
|
D | hugetlb.c | 3303 unsigned long vaddr = *position; in follow_hugetlb_page() local 3307 while (vaddr < vma->vm_end && remainder) { in follow_hugetlb_page() 3320 pte = huge_pte_offset(mm, vaddr & huge_page_mask(h)); in follow_hugetlb_page() 3333 !hugetlbfs_pagecache_present(h, vma, vaddr)) { in follow_hugetlb_page() 3357 ret = hugetlb_fault(mm, vma, vaddr, in follow_hugetlb_page() 3366 pfn_offset = (vaddr & ~huge_page_mask(h)) >> PAGE_SHIFT; in follow_hugetlb_page() 3377 vaddr += PAGE_SIZE; in follow_hugetlb_page() 3381 if (vaddr < vma->vm_end && remainder && in follow_hugetlb_page() 3392 *position = vaddr; in follow_hugetlb_page()
|
D | iov_iter.c | 303 static size_t __iovec_copy_from_user_inatomic(char *vaddr, in __iovec_copy_from_user_inatomic() argument 313 left = __copy_from_user_inatomic(vaddr, buf, copy); in __iovec_copy_from_user_inatomic() 316 vaddr += copy; in __iovec_copy_from_user_inatomic()
|