/bionic/libc/private/ |
D | WriteProtected.h | 32 char padding[PAGE_SIZE]; 36 } __attribute__((aligned(PAGE_SIZE))); 43 static_assert(sizeof(T) < PAGE_SIZE, 58 if (mprotect(&contents, PAGE_SIZE, PROT_READ)) { in initialize() 73 if (mprotect(&contents, PAGE_SIZE, PROT_READ | PROT_WRITE) != 0) { in mutate() 78 if (mprotect(&contents, PAGE_SIZE, PROT_READ) != 0) { in mutate()
|
D | bionic_page.h | 31 #define PAGE_END(x) PAGE_START((x) + (PAGE_SIZE-1))
|
D | CFIShadow.h | 72 align_up((kMaxTargetAddr >> (kShadowGranularity - 1)), PAGE_SIZE);
|
/bionic/libc/bionic/ |
D | getpagesize.cpp | 34 return PAGE_SIZE; in getpagesize()
|
D | sysinfo.cpp | 82 return (static_cast<int64_t>(si.totalram) * si.mem_unit) / PAGE_SIZE; in get_phys_pages() 88 return ((static_cast<int64_t>(si.freeram) + si.bufferram) * si.mem_unit) / PAGE_SIZE; in get_avphys_pages()
|
D | mremap.cpp | 41 size_t rounded = __BIONIC_ALIGN(new_size, PAGE_SIZE); in mremap()
|
D | bionic_allocator.cpp | 102 blocks_per_page_((PAGE_SIZE - sizeof(small_object_page_info)) / in BionicSmallObjectAllocator() 160 munmap(page, PAGE_SIZE); in free_page() 195 void* const map_ptr = mmap(nullptr, PAGE_SIZE, PROT_READ | PROT_WRITE, in alloc_page() 201 prctl(PR_SET_VMA, PR_SET_VMA_ANON_NAME, map_ptr, PAGE_SIZE, in alloc_page() 320 align = MIN(align, PAGE_SIZE); in memalign()
|
D | mmap.cpp | 51 size_t rounded = __BIONIC_ALIGN(size, PAGE_SIZE); in mmap64()
|
D | pthread_create.cpp | 90 size_t allocation_size = __BIONIC_ALIGN(sizeof(bionic_tls), PAGE_SIZE); in __allocate_temp_bionic_tls() 103 munmap(tls, __BIONIC_ALIGN(sizeof(bionic_tls), PAGE_SIZE)); in __free_temp_bionic_tls() 229 mmap_size = __BIONIC_ALIGN(mmap_size, PAGE_SIZE); in __allocate_thread_mapping() 273 attr->guard_size = __BIONIC_ALIGN(attr->guard_size, PAGE_SIZE); in __allocate_thread()
|
D | pthread_attr.cpp | 146 if ((stack_size & (PAGE_SIZE - 1) || stack_size < PTHREAD_STACK_MIN)) { in pthread_attr_setstack() 149 if (reinterpret_cast<uintptr_t>(stack_base) & (PAGE_SIZE - 1)) { in pthread_attr_setstack()
|
D | pthread_internal.h | 202 #define PTHREAD_GUARD_SIZE PAGE_SIZE
|
D | ndk_cruft.cpp | 60 unsigned int __page_size = PAGE_SIZE;
|
/bionic/libc/kernel/uapi/linux/ |
D | resource.h | 60 #define MLOCK_LIMIT ((PAGE_SIZE > 64 * 1024) ? PAGE_SIZE : 64 * 1024)
|
D | binfmts.h | 23 #define MAX_ARG_STRLEN (PAGE_SIZE * 32)
|
D | a.out.h | 88 #define N_TXTADDR(x) (N_MAGIC(x) == QMAGIC ? PAGE_SIZE : 0)
|
/bionic/libdl/ |
D | libdl_cfi.cpp | 29 char padding[PAGE_SIZE - sizeof(v)]; 30 } shadow_base_storage alignas(PAGE_SIZE); 37 static_assert(sizeof(shadow_base_storage) == PAGE_SIZE, ""); in __cfi_init()
|
/bionic/libc/include/sys/ |
D | user.h | 38 #define PAGE_SIZE 4096 macro 39 #define PAGE_MASK (~(PAGE_SIZE - 1)) 107 #define HOST_STACK_END_ADDR (u.start_stack + u.u_ssize * PAGE_SIZE)
|
/bionic/libc/include/android/ |
D | legacy_sys_mman_inlines.h | 58 size_t __rounded = __BIONIC_ALIGN(__size, PAGE_SIZE); in mmap64()
|
/bionic/linker/ |
D | linker_block_allocator.cpp | 36 static constexpr size_t kAllocateSize = PAGE_SIZE * 100; 37 static_assert(kAllocateSize % PAGE_SIZE == 0, "Invalid kAllocateSize.");
|
D | linker_phdr.cpp | 528 if (align == PAGE_SIZE) { in ReserveAligned() 538 size_t mmap_size = align_up(size, align) + align - PAGE_SIZE; in ReserveAligned() 550 size_t n = is_first_stage_init() ? 0 : arc4random_uniform((last - first) / PAGE_SIZE + 1); in ReserveAligned() 551 uint8_t* start = first + n * PAGE_SIZE; in ReserveAligned() 664 memset(reinterpret_cast<void*>(seg_file_end), 0, PAGE_SIZE - PAGE_OFFSET(seg_file_end)); in LoadSegments() 932 memcmp(mem_base + match_offset, file_base + match_offset, PAGE_SIZE) != 0) { in phdr_table_map_gnu_relro() 933 match_offset += PAGE_SIZE; in phdr_table_map_gnu_relro() 939 memcmp(mem_base + mismatch_offset, file_base + mismatch_offset, PAGE_SIZE) == 0) { in phdr_table_map_gnu_relro() 940 mismatch_offset += PAGE_SIZE; in phdr_table_map_gnu_relro()
|
D | linker_utils.cpp | 172 constexpr off64_t kPageMask = ~static_cast<off64_t>(PAGE_SIZE-1); 189 return static_cast<size_t>(offset & (PAGE_SIZE-1)); in page_offset()
|
/bionic/tests/ |
D | sys_mman_test.cpp | 231 void* map = mmap(nullptr, PAGE_SIZE, PROT_NONE, MAP_PRIVATE | MAP_ANONYMOUS, -1, 0); in TEST() 233 ASSERT_EQ(MAP_FAILED, mremap(map, PAGE_SIZE, kHuge, MREMAP_MAYMOVE)); in TEST() 237 char* base = reinterpret_cast<char*>(mmap(nullptr, PAGE_SIZE * 2, PROT_EXEC | PROT_READ, in TEST() 240 __builtin___clear_cache(base, base + (PAGE_SIZE * 2)); in TEST()
|
D | dlext_test.cpp | 165 extinfo.library_fd_offset = (5LL<<48) + PAGE_SIZE; in TEST_F() 170 extinfo.library_fd_offset = 0LL - PAGE_SIZE; in TEST_F() 328 void* new_start = mmap(start, PAGE_SIZE, PROT_NONE, MAP_PRIVATE | MAP_ANONYMOUS, -1, 0); in TEST_F() 333 void* start = mmap(nullptr, PAGE_SIZE, PROT_NONE, MAP_PRIVATE | MAP_ANONYMOUS, -1, 0); in TEST_F() 338 extinfo.reserved_size = PAGE_SIZE; in TEST_F() 375 void* start = mmap(nullptr, PAGE_SIZE, PROT_NONE, MAP_PRIVATE | MAP_ANONYMOUS, -1, 0); in TEST_F() 380 extinfo.reserved_size = PAGE_SIZE; in TEST_F() 403 void* start = mmap(nullptr, PAGE_SIZE, PROT_NONE, MAP_PRIVATE | MAP_ANONYMOUS, -1, 0); in TEST_F() 408 extinfo.reserved_size = PAGE_SIZE; in TEST_F() 415 reinterpret_cast<char*>(start) + PAGE_SIZE)); in TEST_F()
|
/bionic/libc/include/ |
D | pthread.h | 75 #define PTHREAD_STACK_MIN (4 * PAGE_SIZE) 77 #define PTHREAD_STACK_MIN (2 * PAGE_SIZE)
|
/bionic/tests/headers/posix/ |
D | limits_h.c | 62 MACRO(PAGE_SIZE); in limits_h()
|