/bionic/libc/private/ |
D | WriteProtected.h | 32 char padding[PAGE_SIZE]; 36 } __attribute__((aligned(PAGE_SIZE))); 43 static_assert(sizeof(T) < PAGE_SIZE, 56 return mprotect(reinterpret_cast<void*>(addr), PAGE_SIZE, prot); in set_protection()
|
D | CFIShadow.h | 72 align_up((kMaxTargetAddr >> (kShadowGranularity - 1)), PAGE_SIZE);
|
/bionic/tests/ |
D | sys_mman_test.cpp | 231 void* map = mmap(nullptr, PAGE_SIZE, PROT_NONE, MAP_PRIVATE | MAP_ANONYMOUS, -1, 0); in TEST() 234 ASSERT_EQ(MAP_FAILED, mremap(map, PAGE_SIZE, kHuge, MREMAP_MAYMOVE)); in TEST() 236 ASSERT_EQ(0, munmap(map, PAGE_SIZE)); in TEST() 240 char* base = reinterpret_cast<char*>(mmap(nullptr, PAGE_SIZE * 2, PROT_EXEC | PROT_READ, in TEST() 243 __builtin___clear_cache(base, base + (PAGE_SIZE * 2)); in TEST() 247 void* map = mmap(nullptr, PAGE_SIZE, PROT_NONE, MAP_PRIVATE | MAP_ANONYMOUS, -1, 0); in TEST() 251 mlock(map, PAGE_SIZE); in TEST() 253 ASSERT_EQ(0, munmap(map, PAGE_SIZE)); in TEST() 260 void* map = mmap(nullptr, PAGE_SIZE, PROT_NONE, MAP_PRIVATE | MAP_ANONYMOUS, -1, 0); in TEST() 264 mlock2(map, PAGE_SIZE, MLOCK_ONFAULT); in TEST() [all …]
|
D | dlext_test.cpp | 162 extinfo.library_fd_offset = (5LL<<48) + PAGE_SIZE; in TEST_F() 167 extinfo.library_fd_offset = 0LL - PAGE_SIZE; in TEST_F() 325 void* new_start = mmap(start, PAGE_SIZE, PROT_NONE, MAP_PRIVATE | MAP_ANONYMOUS, -1, 0); in TEST_F() 330 void* start = mmap(nullptr, PAGE_SIZE, PROT_NONE, MAP_PRIVATE | MAP_ANONYMOUS, -1, 0); in TEST_F() 335 extinfo.reserved_size = PAGE_SIZE; in TEST_F() 374 void* start = mmap(nullptr, PAGE_SIZE, PROT_NONE, MAP_PRIVATE | MAP_ANONYMOUS, -1, 0); in TEST_F() 379 extinfo.reserved_size = PAGE_SIZE; in TEST_F() 402 void* start = mmap(nullptr, PAGE_SIZE, PROT_NONE, MAP_PRIVATE | MAP_ANONYMOUS, -1, 0); in TEST_F() 407 extinfo.reserved_size = PAGE_SIZE; in TEST_F() 414 reinterpret_cast<char*>(start) + PAGE_SIZE)); in TEST_F()
|
/bionic/libdl/ |
D | libdl_cfi.cpp | 29 char padding[PAGE_SIZE - sizeof(v)]; 30 } shadow_base_storage alignas(PAGE_SIZE); 37 static_assert(sizeof(shadow_base_storage) == PAGE_SIZE, ""); in __cfi_init()
|
/bionic/libc/kernel/uapi/linux/ |
D | resource.h | 60 #define MLOCK_LIMIT ((PAGE_SIZE > 64 * 1024) ? PAGE_SIZE : 64 * 1024)
|
D | binfmts.h | 23 #define MAX_ARG_STRLEN (PAGE_SIZE * 32)
|
D | a.out.h | 88 #define N_TXTADDR(x) (N_MAGIC(x) == QMAGIC ? PAGE_SIZE : 0)
|
/bionic/libc/bionic/ |
D | sysinfo.cpp | 82 return (static_cast<int64_t>(si.totalram) * si.mem_unit) / PAGE_SIZE; in get_phys_pages() 88 return ((static_cast<int64_t>(si.freeram) + si.bufferram) * si.mem_unit) / PAGE_SIZE; in get_avphys_pages()
|
D | bionic_allocator.cpp | 102 blocks_per_page_((PAGE_SIZE - sizeof(small_object_page_info)) / in BionicSmallObjectAllocator() 160 munmap(page, PAGE_SIZE); in free_page() 195 void* const map_ptr = mmap(nullptr, PAGE_SIZE, PROT_READ | PROT_WRITE, in alloc_page() 201 prctl(PR_SET_VMA, PR_SET_VMA_ANON_NAME, map_ptr, PAGE_SIZE, in alloc_page() 320 align = MIN(align, PAGE_SIZE); in memalign()
|
D | getpagesize.cpp | 34 return PAGE_SIZE; in getpagesize()
|
D | pthread_create.cpp | 71 size_t allocation_size = __BIONIC_ALIGN(sizeof(bionic_tls), PAGE_SIZE); in __allocate_temp_bionic_tls() 84 munmap(tls, __BIONIC_ALIGN(sizeof(bionic_tls), PAGE_SIZE)); in __free_temp_bionic_tls() 210 mmap_size = __BIONIC_ALIGN(mmap_size, PAGE_SIZE); in __allocate_thread_mapping() 256 attr->guard_size = __BIONIC_ALIGN(attr->guard_size, PAGE_SIZE); in __allocate_thread()
|
D | mremap.cpp | 41 size_t rounded = __BIONIC_ALIGN(new_size, PAGE_SIZE); in mremap()
|
D | pthread_attr.cpp | 146 if ((stack_size & (PAGE_SIZE - 1) || stack_size < PTHREAD_STACK_MIN)) { in pthread_attr_setstack() 149 if (reinterpret_cast<uintptr_t>(stack_base) & (PAGE_SIZE - 1)) { in pthread_attr_setstack()
|
D | mmap.cpp | 51 size_t rounded = __BIONIC_ALIGN(size, PAGE_SIZE); in mmap64()
|
D | pthread_internal.h | 218 #define PTHREAD_GUARD_SIZE PAGE_SIZE
|
D | ndk_cruft.cpp | 60 unsigned int __page_size = PAGE_SIZE;
|
/bionic/libc/include/sys/ |
D | user.h | 38 #define PAGE_SIZE 4096 macro 39 #define PAGE_MASK (~(PAGE_SIZE - 1)) 107 #define HOST_STACK_END_ADDR (u.start_stack + u.u_ssize * PAGE_SIZE)
|
/bionic/linker/ |
D | linker_phdr.cpp | 525 if (align == PAGE_SIZE) { in ReserveAligned() 535 size_t mmap_size = align_up(size, align) + align - PAGE_SIZE; in ReserveAligned() 547 size_t n = is_first_stage_init() ? 0 : arc4random_uniform((last - first) / PAGE_SIZE + 1); in ReserveAligned() 548 uint8_t* start = first + n * PAGE_SIZE; in ReserveAligned() 661 memset(reinterpret_cast<void*>(seg_file_end), 0, PAGE_SIZE - PAGE_OFFSET(seg_file_end)); in LoadSegments() 929 memcmp(mem_base + match_offset, file_base + match_offset, PAGE_SIZE) != 0) { in phdr_table_map_gnu_relro() 930 match_offset += PAGE_SIZE; in phdr_table_map_gnu_relro() 936 memcmp(mem_base + mismatch_offset, file_base + mismatch_offset, PAGE_SIZE) == 0) { in phdr_table_map_gnu_relro() 937 mismatch_offset += PAGE_SIZE; in phdr_table_map_gnu_relro()
|
D | linker_block_allocator.cpp | 36 static constexpr size_t kAllocateSize = PAGE_SIZE * 100; 37 static_assert(kAllocateSize % PAGE_SIZE == 0, "Invalid kAllocateSize.");
|
D | linker_utils.cpp | 172 constexpr off64_t kPageMask = ~static_cast<off64_t>(PAGE_SIZE-1); 189 return static_cast<size_t>(offset & (PAGE_SIZE-1)); in page_offset()
|
/bionic/libc/platform/bionic/ |
D | page.h | 30 #define PAGE_END(x) PAGE_START((x) + (PAGE_SIZE-1))
|
/bionic/libc/include/android/ |
D | legacy_sys_mman_inlines.h | 58 size_t __rounded = __BIONIC_ALIGN(__size, PAGE_SIZE); in mmap64()
|
/bionic/libc/include/ |
D | pthread.h | 75 #define PTHREAD_STACK_MIN (4 * PAGE_SIZE) 77 #define PTHREAD_STACK_MIN (2 * PAGE_SIZE)
|
/bionic/tests/headers/posix/ |
D | limits_h.c | 62 MACRO(PAGE_SIZE); in limits_h()
|