/arch/alpha/include/asm/ |
D | setup.h | 23 #define KERNEL_START (PAGE_OFFSET+KERNEL_START_PHYS) 25 #define INIT_STACK (PAGE_OFFSET+KERNEL_START_PHYS+0x02000) 26 #define EMPTY_PGT (PAGE_OFFSET+KERNEL_START_PHYS+0x04000) 27 #define EMPTY_PGE (PAGE_OFFSET+KERNEL_START_PHYS+0x08000) 28 #define ZERO_PGE (PAGE_OFFSET+KERNEL_START_PHYS+0x0A000) 30 #define START_ADDR (PAGE_OFFSET+KERNEL_START_PHYS+0x10000)
|
D | page.h | 69 #define PAGE_OFFSET 0xffff800000000000UL macro 71 #define PAGE_OFFSET 0xfffffc0000000000UL macro 77 #define PAGE_OFFSET 0xffff800000000000 macro 79 #define PAGE_OFFSET 0xfffffc0000000000 macro 84 #define __pa(x) ((unsigned long) (x) - PAGE_OFFSET) 85 #define __va(x) ((void *)((unsigned long) (x) + PAGE_OFFSET))
|
/arch/powerpc/include/asm/ |
D | page.h | 81 #define PAGE_OFFSET ASM_CONST(CONFIG_PAGE_OFFSET) macro 115 #define MEMORY_START (PHYSICAL_START + PAGE_OFFSET - KERNELBASE) 137 _addr >= PAGE_OFFSET && _addr < (unsigned long)high_memory && \ 229 VIRTUAL_WARN_ON((unsigned long)(x) >= PAGE_OFFSET); \ 230 (void *)(unsigned long)((phys_addr_t)(x) | PAGE_OFFSET); \ 235 VIRTUAL_WARN_ON((unsigned long)(x) < PAGE_OFFSET); \ 240 #define __va(x) ((void *)(unsigned long)((phys_addr_t)(x) + PAGE_OFFSET - MEMORY_START)) 241 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET + MEMORY_START) 266 #define is_kernel_addr(x) ((x) >= PAGE_OFFSET)
|
D | kdump.h | 26 #define KDUMP_TRAMPOLINE_START (0x0100 + PAGE_OFFSET) 27 #define KDUMP_TRAMPOLINE_END (0x3000 + PAGE_OFFSET)
|
/arch/x86/mm/ |
D | physaddr.c | 25 x = y + (__START_KERNEL_map - PAGE_OFFSET); in __phys_addr() 58 x = y + (__START_KERNEL_map - PAGE_OFFSET); in __virt_addr_valid() 74 unsigned long phys_addr = x - PAGE_OFFSET; in __phys_addr() 76 VIRTUAL_BUG_ON(x < PAGE_OFFSET); in __phys_addr() 90 if (x < PAGE_OFFSET) in __virt_addr_valid() 96 return pfn_valid((x - PAGE_OFFSET) >> PAGE_SHIFT); in __virt_addr_valid()
|
/arch/arm/include/asm/ |
D | memory.h | 29 #define PAGE_OFFSET UL(CONFIG_PAGE_OFFSET) macro 30 #define KERNEL_OFFSET (PAGE_OFFSET) 55 #define MODULES_VADDR (PAGE_OFFSET - SZ_16M) 58 #define MODULES_VADDR (PAGE_OFFSET - SZ_8M) 69 #define MODULES_END (PAGE_OFFSET - PMD_SIZE) 71 #define MODULES_END (PAGE_OFFSET) 122 #define MODULES_VADDR PAGE_OFFSET 281 return (phys_addr_t)x - PAGE_OFFSET + PHYS_OFFSET; in __virt_to_phys_nodebug() 286 return x - PHYS_OFFSET + PAGE_OFFSET; in __phys_to_virt() 292 ((((unsigned long)(kaddr) - PAGE_OFFSET) >> PAGE_SHIFT) + \ [all …]
|
/arch/nds32/include/asm/ |
D | memory.h | 24 #define PAGE_OFFSET (CONFIG_PAGE_OFFSET) macro 32 #define __virt_to_phys(x) ((x) - PAGE_OFFSET + PHYS_OFFSET) 33 #define __phys_to_virt(x) ((x) - PHYS_OFFSET + PAGE_OFFSET) 40 #define MODULES_END (PAGE_OFFSET) 83 #define virt_addr_valid(kaddr) ((unsigned long)(kaddr) >= PAGE_OFFSET && (unsigned long)(kaddr) < (…
|
/arch/openrisc/include/asm/ |
D | page.h | 29 #define PAGE_OFFSET 0xc0000000 macro 30 #define KERNELBASE PAGE_OFFSET 72 #define __va(x) ((void *)((unsigned long)(x) + PAGE_OFFSET)) 73 #define __pa(x) ((unsigned long) (x) - PAGE_OFFSET) 79 (mem_map + (((unsigned long)(addr)-PAGE_OFFSET) >> PAGE_SHIFT))
|
/arch/sh/include/asm/ |
D | page.h | 123 #define PAGE_OFFSET CONFIG_PAGE_OFFSET macro 135 #define ___pa(x) ((x)-PAGE_OFFSET+__MEMORY_START) 136 #define ___va(x) ((x)+PAGE_OFFSET-__MEMORY_START) 138 #define ___pa(x) ((x)-PAGE_OFFSET) 139 #define ___va(x) ((x)+PAGE_OFFSET) 152 #define UNCAC_ADDR(addr) ((addr) - PAGE_OFFSET + uncached_start) 153 #define CAC_ADDR(addr) ((addr) - uncached_start + PAGE_OFFSET)
|
/arch/csky/include/asm/ |
D | page.h | 27 #define PAGE_OFFSET CONFIG_PAGE_OFFSET macro 40 #define virt_addr_valid(kaddr) ((void *)(kaddr) >= (void *)PAGE_OFFSET && \ 79 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET + va_pa_offset) 80 #define __va(x) ((void *)((unsigned long)(x) + PAGE_OFFSET - va_pa_offset)) 84 #define MAP_NR(x) PFN_DOWN((unsigned long)(x) - PAGE_OFFSET - \
|
/arch/nios2/include/asm/ |
D | page.h | 31 #define PAGE_OFFSET \ macro 80 ((unsigned long)(x) - PAGE_OFFSET + PHYS_OFFSET) 82 ((void *)((unsigned long)(x) + PAGE_OFFSET - PHYS_OFFSET)) 85 ((void *)(((page) - mem_map) << PAGE_SHIFT) + PAGE_OFFSET)
|
/arch/hexagon/include/asm/ |
D | mem-layout.h | 19 #define PAGE_OFFSET _AC(0xc0000000, UL) macro 41 #define TASK_SIZE (PAGE_OFFSET) 54 #define MIN_KERNEL_SEG (PAGE_OFFSET >> PGDIR_SHIFT) /* L1 shift is 22 bits */
|
/arch/powerpc/mm/book3s32/ |
D | hash_low.S | 52 lis r8, (mmu_hash_lock - PAGE_OFFSET)@h 53 ori r8, r8, (mmu_hash_lock - PAGE_OFFSET)@l 149 lis r8, (mmu_hash_lock - PAGE_OFFSET)@ha 151 stw r0, (mmu_hash_lock - PAGE_OFFSET)@l(r8) 158 lis r8, (mmu_hash_lock - PAGE_OFFSET)@ha 160 stw r0, (mmu_hash_lock - PAGE_OFFSET)@l(r8) 199 lis r6, (mmu_hash_lock - PAGE_OFFSET)@ha 200 addi r6, r6, (mmu_hash_lock - PAGE_OFFSET)@l 251 lis r6, (mmu_hash_lock - PAGE_OFFSET)@ha 252 addi r6, r6, (mmu_hash_lock - PAGE_OFFSET)@l [all …]
|
D | mmu.c | 152 setbat(idx, PAGE_OFFSET + base, base, size, PAGE_KERNEL_X); in __mmu_mapin_ram() 162 unsigned long border = (unsigned long)__srwx_boundary - PAGE_OFFSET; in mmu_mapin_ram() 165 size = roundup_pow_of_two((unsigned long)_einittext - PAGE_OFFSET); in mmu_mapin_ram() 166 setibat(0, PAGE_OFFSET, 0, size, PAGE_KERNEL_X); in mmu_mapin_ram() 201 unsigned long base = (unsigned long)_stext - PAGE_OFFSET; in mmu_mark_initmem_nx() 202 unsigned long top = ALIGN((unsigned long)_etext - PAGE_OFFSET, SZ_128K); in mmu_mark_initmem_nx() 203 unsigned long border = (unsigned long)__init_begin - PAGE_OFFSET; in mmu_mark_initmem_nx() 208 setibat(i++, PAGE_OFFSET + base, base, size, PAGE_KERNEL_TEXT); in mmu_mark_initmem_nx() 219 setibat(i++, PAGE_OFFSET + base, base, size, PAGE_KERNEL_TEXT); in mmu_mark_initmem_nx() 397 unsigned int hash = (unsigned int)Hash - PAGE_OFFSET; in MMU_init_hw_patch()
|
/arch/m68k/include/asm/ |
D | page_mm.h | 100 if(x >= PAGE_OFFSET) in ___pa() 101 return (x-PAGE_OFFSET); in ___pa() 112 return (void *)(x+PAGE_OFFSET); in __va() 139 #define virt_addr_valid(kaddr) ((unsigned long)(kaddr) >= PAGE_OFFSET && (unsigned long)(kaddr) < (…
|
D | page_no.h | 26 #define virt_to_page(addr) (mem_map + (((unsigned long)(addr)-PAGE_OFFSET) >> PAGE_SHIFT)) 27 #define page_to_virt(page) __va(((((page) - mem_map) << PAGE_SHIFT) + PAGE_OFFSET)) 33 #define virt_addr_valid(kaddr) (((unsigned long)(kaddr) >= PAGE_OFFSET) && \
|
/arch/arm64/include/asm/ |
D | memory.h | 34 #define VMEMMAP_SIZE ((_PAGE_END(VA_BITS_MIN) - PAGE_OFFSET) >> VMEMMAP_SHIFT) 45 #define PAGE_OFFSET (_PAGE_OFFSET(VA_BITS)) macro 284 #define __is_lm_address(addr) (((u64)(addr) - PAGE_OFFSET) < (PAGE_END - PAGE_OFFSET)) 286 #define __lm_to_phys(addr) (((addr) - PAGE_OFFSET) + PHYS_OFFSET) 304 #define __phys_to_virt(x) ((unsigned long)((x) - PHYS_OFFSET) | PAGE_OFFSET) 358 u64 __addr = PAGE_OFFSET + (__idx * PAGE_SIZE); \ 363 u64 __idx = (__tag_reset((u64)x) - PAGE_OFFSET) / PAGE_SIZE; \
|
/arch/xtensa/include/asm/ |
D | page.h | 28 #define PAGE_OFFSET XCHAL_KSEG_CACHED_VADDR macro 33 #define PAGE_OFFSET _AC(CONFIG_DEFAULT_MEM_START, UL) macro 167 unsigned long off = va - PAGE_OFFSET; in ___pa() 188 ((unsigned long) (x) - PAGE_OFFSET + PHYS_OFFSET) 191 ((void *)((unsigned long) (x) - PHYS_OFFSET + PAGE_OFFSET))
|
/arch/hexagon/mm/ |
D | init.c | 21 #define bootmem_startpg (PFN_UP(((unsigned long) _end) - PAGE_OFFSET + PHYS_OFFSET)) 186 segtable = segtable + (PAGE_OFFSET >> 22); in setup_arch_memory() 219 (unsigned long) _K_init_devicetable-PAGE_OFFSET); in setup_arch_memory() 220 *segtable = ((u32) (unsigned long) _K_init_devicetable-PAGE_OFFSET) | in setup_arch_memory()
|
/arch/arm/mm/ |
D | physaddr.c | 21 if (!high_memory && x >= PAGE_OFFSET) in __virt_addr_valid() 24 if (high_memory && x >= PAGE_OFFSET && x < (unsigned long)high_memory) in __virt_addr_valid()
|
/arch/riscv/include/asm/ |
D | page.h | 34 #define PAGE_OFFSET _AC(CONFIG_PAGE_OFFSET, UL) macro 36 #define KERN_VIRT_SIZE (-PAGE_OFFSET) 85 #define ARCH_PFN_OFFSET (PAGE_OFFSET >> PAGE_SHIFT) 110 ((x) >= PAGE_OFFSET && (!IS_ENABLED(CONFIG_64BIT) || (x) < kernel_map.virt_addr)) 169 (unsigned long)(_addr) >= PAGE_OFFSET && pfn_valid(virt_to_pfn(_addr)); \
|
/arch/sparc/include/asm/ |
D | page_32.h | 119 #define PAGE_OFFSET 0xf0000000 macro 124 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET + phys_base) 125 #define __va(x) ((void *)((unsigned long) (x) - phys_base + PAGE_OFFSET)) 134 #define virt_addr_valid(kaddr) ((((unsigned long)(kaddr)-PAGE_OFFSET)>>PAGE_SHIFT) < max_mapnr)
|
/arch/arm/mach-zynq/ |
D | platsmp.c | 46 if (__pa(PAGE_OFFSET)) { in zynq_cpun_start() 53 zero = (__force u8 __iomem *)PAGE_OFFSET; in zynq_cpun_start() 70 if (__pa(PAGE_OFFSET)) in zynq_cpun_start()
|
/arch/x86/kernel/ |
D | head32.c | 77 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET) in mk_early_pgtbl_32() 101 *(pl2p + ((PAGE_OFFSET >> PGDIR_SHIFT))) = pl2; in mk_early_pgtbl_32() 117 *ptr = (unsigned long)ptep + PAGE_OFFSET; in mk_early_pgtbl_32()
|
/arch/arm/mach-footbridge/include/mach/ |
D | memory.h | 43 #define __virt_to_bus(x) ((x) + (BUS_OFFSET - PAGE_OFFSET)) 44 #define __bus_to_virt(x) ((x) - (BUS_OFFSET - PAGE_OFFSET))
|