Searched refs:vaddr_end (Results 1 – 5 of 5) sorted by relevance
/arch/x86/mm/ |
D | mem_encrypt.c | 225 unsigned long vaddr_end; member 234 pgd_end = ppd->vaddr_end & PGDIR_MASK; in sme_clear_pgd() 378 while (ppd->vaddr < ppd->vaddr_end) { in __sme_map_range_pmd() 388 while (ppd->vaddr < ppd->vaddr_end) { in __sme_map_range_pte() 399 unsigned long vaddr_end; in __sme_map_range() local 405 vaddr_end = ppd->vaddr_end; in __sme_map_range() 408 ppd->vaddr_end = ALIGN(ppd->vaddr, PMD_PAGE_SIZE); in __sme_map_range() 412 ppd->vaddr_end = vaddr_end & PMD_PAGE_MASK; in __sme_map_range() 416 ppd->vaddr_end = vaddr_end; in __sme_map_range() 591 ppd.vaddr_end = workarea_end; in sme_encrypt_kernel() [all …]
|
D | kaslr.c | 44 static const unsigned long vaddr_end = CPU_ENTRY_AREA_BASE; variable 97 BUILD_BUG_ON(vaddr_start >= vaddr_end); in kernel_randomize_memory() 98 BUILD_BUG_ON(vaddr_end != CPU_ENTRY_AREA_BASE); in kernel_randomize_memory() 99 BUILD_BUG_ON(vaddr_end > __START_KERNEL_map); in kernel_randomize_memory() 117 remain_entropy = vaddr_end - vaddr_start; in kernel_randomize_memory()
|
D | init_64.c | 381 unsigned long vaddr_end = __START_KERNEL_map + KERNEL_IMAGE_SIZE; in cleanup_highmap() local 391 vaddr_end = __START_KERNEL_map + (max_pfn_mapped << PAGE_SHIFT); in cleanup_highmap() 393 for (; vaddr + PMD_SIZE - 1 < vaddr_end; pmd++, vaddr += PMD_SIZE) { in cleanup_highmap() 685 unsigned long vaddr, vaddr_start, vaddr_end, vaddr_next, paddr_last; in kernel_physical_mapping_init() local 689 vaddr_end = (unsigned long)__va(paddr_end); in kernel_physical_mapping_init() 692 for (; vaddr < vaddr_end; vaddr = vaddr_next) { in kernel_physical_mapping_init() 701 __pa(vaddr_end), in kernel_physical_mapping_init() 707 paddr_last = phys_p4d_init(p4d, __pa(vaddr), __pa(vaddr_end), in kernel_physical_mapping_init() 720 sync_global_pgds(vaddr_start, vaddr_end - 1); in kernel_physical_mapping_init()
|
/arch/tile/kernel/ |
D | setup.c | 308 unsigned long vaddr_end; in setup_pa_va_mapping() local 318 vaddr_end = PAGE_OFFSET + in setup_pa_va_mapping() 323 vaddr_end = PAGE_OFFSET + (curr_pages << PAGE_SHIFT); in setup_pa_va_mapping() 325 for (j = 0; vaddr < vaddr_end; vaddr += HPAGE_SIZE, ++j) { in setup_pa_va_mapping()
|
/arch/x86/xen/ |
D | mmu_pv.c | 1055 unsigned long vaddr_end) in xen_cleanhighmap() argument 1062 for (; vaddr <= vaddr_end && (pmd < (level2_kernel_pgt + PTRS_PER_PMD)); in xen_cleanhighmap() 1080 void *vaddr_end = vaddr + size; in xen_free_ro_pages() local 1082 for (; vaddr < vaddr_end; vaddr += PAGE_SIZE) in xen_free_ro_pages()
|