/arch/s390/boot/ |
D | startup.c | 24 unsigned long __bootdata_preserved(MODULES_END); 166 MODULES_END = _REGION1_SIZE; in setup_kernel_memory_layout() 169 MODULES_END = _REGION2_SIZE; in setup_kernel_memory_layout() 177 adjust_to_uv_max(&MODULES_END); in setup_kernel_memory_layout() 180 MODULES_END = min(MODULES_END, KASAN_SHADOW_START); in setup_kernel_memory_layout() 182 MODULES_VADDR = MODULES_END - MODULES_LEN; in setup_kernel_memory_layout()
|
/arch/x86/include/asm/ |
D | pgtable_64_types.h | 147 # define MODULES_END _AC(0xffffffffff000000, UL) macro 149 # define MODULES_END _AC(0xfffffffffe000000, UL) macro 151 #define MODULES_LEN (MODULES_END - MODULES_VADDR)
|
D | pgtable_32_areas.h | 48 #define MODULES_END VMALLOC_END macro 49 #define MODULES_LEN (MODULES_VADDR - MODULES_END)
|
/arch/nds32/include/asm/ |
D | memory.h | 40 #define MODULES_END (PAGE_OFFSET) macro 41 #define MODULES_VADDR (MODULES_END - SZ_32M)
|
/arch/powerpc/kernel/ |
D | module.c | 117 if (MODULES_VADDR < PAGE_OFFSET && MODULES_END > limit) in module_alloc() 118 ptr = __module_alloc(size, limit, MODULES_END, true); in module_alloc() 121 ptr = __module_alloc(size, MODULES_VADDR, MODULES_END, false); in module_alloc()
|
/arch/arm/include/asm/ |
D | memory.h | 69 #define MODULES_END (PAGE_OFFSET - PMD_SIZE) macro 71 #define MODULES_END (PAGE_OFFSET) macro 121 #define MODULES_END (END_MEM) macro
|
/arch/s390/mm/ |
D | page-states.c | 172 next = pgd_addr_end(addr, MODULES_END); in mark_kernel_pgd() 181 } while (pgd++, addr = next, addr != MODULES_END); in mark_kernel_pgd()
|
D | kasan_init.c | 356 kasan_early_pgtable_populate(__sha(VMALLOC_START), __sha(MODULES_END), in kasan_early_init() 365 kasan_early_pgtable_populate(__sha(MODULES_END), __sha(_REGION1_SIZE), in kasan_early_init()
|
D | dump_pagetables.c | 272 address_markers[MODULES_END_NR].start_address = MODULES_END; in pt_dump_init()
|
/arch/arm/mm/ |
D | pageattr.c | 49 if (!in_range(start, size, MODULES_VADDR, MODULES_END) && in change_memory_common()
|
D | kasan_init.c | 280 create_mapping((void *)MODULES_VADDR, (void *)(MODULES_END)); in kasan_init()
|
/arch/arm64/include/asm/ |
D | memory.h | 46 #define KIMAGE_VADDR (MODULES_END) 47 #define MODULES_END (MODULES_VADDR + MODULES_VSIZE) macro
|
/arch/um/include/asm/ |
D | pgtable.h | 50 #define MODULES_END VMALLOC_END macro 51 #define MODULES_LEN (MODULES_VADDR - MODULES_END)
|
/arch/powerpc/include/asm/nohash/32/ |
D | mmu-8xx.h | 176 #define MODULES_END PAGE_OFFSET macro
|
/arch/arm64/mm/ |
D | kasan_init.c | 226 mod_shadow_end = (u64)kasan_mem_to_shadow((void *)MODULES_END); in kasan_init_shadow() 250 BUILD_BUG_ON(VMALLOC_START != MODULES_END); in kasan_init_shadow()
|
/arch/x86/kernel/ |
D | head64.c | 474 MAYBE_BUILD_BUG_ON(!(((MODULES_END - 1) & PGDIR_MASK) == in x86_64_start_kernel() 476 BUILD_BUG_ON(__fix_to_virt(__end_of_fixed_addresses) <= MODULES_END); in x86_64_start_kernel()
|
D | module.c | 78 MODULES_END, gfp_mask, in module_alloc()
|
/arch/powerpc/mm/kasan/ |
D | kasan_init_32.c | 121 k_end = (unsigned long)kasan_mem_to_shadow((void *)MODULES_END); in kasan_unmap_early_shadow_vmalloc()
|
/arch/powerpc/include/asm/book3s/32/ |
D | pgtable.h | 234 #define MODULES_END ALIGN_DOWN(PAGE_OFFSET, SZ_256M) macro 235 #define MODULES_VADDR (MODULES_END - SZ_256M)
|
/arch/sparc/kernel/ |
D | module.c | 32 return __vmalloc_node_range(size, 1, MODULES_VADDR, MODULES_END, in module_map()
|
/arch/nds32/kernel/ |
D | module.c | 12 return __vmalloc_node_range(size, 1, MODULES_VADDR, MODULES_END, in module_alloc()
|
/arch/riscv/mm/ |
D | fault.c | 252 if (unlikely(addr >= MODULES_VADDR && addr < MODULES_END)) { in do_page_fault()
|
/arch/riscv/include/asm/ |
D | pgtable.h | 34 #define BPF_JIT_REGION_END (MODULES_END) 43 #define MODULES_END (PFN_ALIGN((unsigned long)&_start)) macro
|
/arch/powerpc/mm/ |
D | mem.c | 337 MODULES_VADDR, MODULES_END); in mem_init()
|
/arch/powerpc/mm/ptdump/ |
D | ptdump.c | 266 address_markers[i++].start_address = MODULES_END; in populate_markers()
|