Home
last modified time | relevance | path

Searched refs:MODULES_END (Results 1 – 25 of 41) sorted by relevance

12

/arch/s390/boot/
Dstartup.c24 unsigned long __bootdata_preserved(MODULES_END);
166 MODULES_END = _REGION1_SIZE; in setup_kernel_memory_layout()
169 MODULES_END = _REGION2_SIZE; in setup_kernel_memory_layout()
177 adjust_to_uv_max(&MODULES_END); in setup_kernel_memory_layout()
180 MODULES_END = min(MODULES_END, KASAN_SHADOW_START); in setup_kernel_memory_layout()
182 MODULES_VADDR = MODULES_END - MODULES_LEN; in setup_kernel_memory_layout()
/arch/x86/include/asm/
Dpgtable_64_types.h147 # define MODULES_END _AC(0xffffffffff000000, UL) macro
149 # define MODULES_END _AC(0xfffffffffe000000, UL) macro
151 #define MODULES_LEN (MODULES_END - MODULES_VADDR)
Dpgtable_32_areas.h48 #define MODULES_END VMALLOC_END macro
49 #define MODULES_LEN (MODULES_VADDR - MODULES_END)
/arch/nds32/include/asm/
Dmemory.h40 #define MODULES_END (PAGE_OFFSET) macro
41 #define MODULES_VADDR (MODULES_END - SZ_32M)
/arch/powerpc/kernel/
Dmodule.c117 if (MODULES_VADDR < PAGE_OFFSET && MODULES_END > limit) in module_alloc()
118 ptr = __module_alloc(size, limit, MODULES_END, true); in module_alloc()
121 ptr = __module_alloc(size, MODULES_VADDR, MODULES_END, false); in module_alloc()
/arch/arm/include/asm/
Dmemory.h69 #define MODULES_END (PAGE_OFFSET - PMD_SIZE) macro
71 #define MODULES_END (PAGE_OFFSET) macro
121 #define MODULES_END (END_MEM) macro
/arch/s390/mm/
Dpage-states.c172 next = pgd_addr_end(addr, MODULES_END); in mark_kernel_pgd()
181 } while (pgd++, addr = next, addr != MODULES_END); in mark_kernel_pgd()
Dkasan_init.c356 kasan_early_pgtable_populate(__sha(VMALLOC_START), __sha(MODULES_END), in kasan_early_init()
365 kasan_early_pgtable_populate(__sha(MODULES_END), __sha(_REGION1_SIZE), in kasan_early_init()
Ddump_pagetables.c272 address_markers[MODULES_END_NR].start_address = MODULES_END; in pt_dump_init()
/arch/arm/mm/
Dpageattr.c49 if (!in_range(start, size, MODULES_VADDR, MODULES_END) && in change_memory_common()
Dkasan_init.c280 create_mapping((void *)MODULES_VADDR, (void *)(MODULES_END)); in kasan_init()
/arch/arm64/include/asm/
Dmemory.h46 #define KIMAGE_VADDR (MODULES_END)
47 #define MODULES_END (MODULES_VADDR + MODULES_VSIZE) macro
/arch/um/include/asm/
Dpgtable.h50 #define MODULES_END VMALLOC_END macro
51 #define MODULES_LEN (MODULES_VADDR - MODULES_END)
/arch/powerpc/include/asm/nohash/32/
Dmmu-8xx.h176 #define MODULES_END PAGE_OFFSET macro
/arch/arm64/mm/
Dkasan_init.c226 mod_shadow_end = (u64)kasan_mem_to_shadow((void *)MODULES_END); in kasan_init_shadow()
250 BUILD_BUG_ON(VMALLOC_START != MODULES_END); in kasan_init_shadow()
/arch/x86/kernel/
Dhead64.c474 MAYBE_BUILD_BUG_ON(!(((MODULES_END - 1) & PGDIR_MASK) == in x86_64_start_kernel()
476 BUILD_BUG_ON(__fix_to_virt(__end_of_fixed_addresses) <= MODULES_END); in x86_64_start_kernel()
Dmodule.c78 MODULES_END, gfp_mask, in module_alloc()
/arch/powerpc/mm/kasan/
Dkasan_init_32.c121 k_end = (unsigned long)kasan_mem_to_shadow((void *)MODULES_END); in kasan_unmap_early_shadow_vmalloc()
/arch/powerpc/include/asm/book3s/32/
Dpgtable.h234 #define MODULES_END ALIGN_DOWN(PAGE_OFFSET, SZ_256M) macro
235 #define MODULES_VADDR (MODULES_END - SZ_256M)
/arch/sparc/kernel/
Dmodule.c32 return __vmalloc_node_range(size, 1, MODULES_VADDR, MODULES_END, in module_map()
/arch/nds32/kernel/
Dmodule.c12 return __vmalloc_node_range(size, 1, MODULES_VADDR, MODULES_END, in module_alloc()
/arch/riscv/mm/
Dfault.c252 if (unlikely(addr >= MODULES_VADDR && addr < MODULES_END)) { in do_page_fault()
/arch/riscv/include/asm/
Dpgtable.h34 #define BPF_JIT_REGION_END (MODULES_END)
43 #define MODULES_END (PFN_ALIGN((unsigned long)&_start)) macro
/arch/powerpc/mm/
Dmem.c337 MODULES_VADDR, MODULES_END); in mem_init()
/arch/powerpc/mm/ptdump/
Dptdump.c266 address_markers[i++].start_address = MODULES_END; in populate_markers()

12