Home
last modified time | relevance | path

Searched refs:kPageSize (Results 1 – 25 of 68) sorted by relevance

123

/art/libartbase/base/
Dsafe_copy_test.cc34 DCHECK_EQ(kPageSize, static_cast<decltype(kPageSize)>(PAGE_SIZE)); in TEST()
37 void* map = mmap(nullptr, kPageSize * 4, PROT_READ | PROT_WRITE, in TEST()
41 char* page2 = page1 + kPageSize; in TEST()
42 char* page3 = page2 + kPageSize; in TEST()
43 char* page4 = page3 + kPageSize; in TEST()
44 ASSERT_EQ(0, mprotect(page1 + kPageSize, kPageSize, PROT_NONE)); in TEST()
45 ASSERT_EQ(0, munmap(page4, kPageSize)); in TEST()
48 page1[kPageSize - 1] = 'z'; in TEST()
51 page3[kPageSize - 1] = 'y'; in TEST()
53 char buf[kPageSize]; in TEST()
[all …]
Dmem_map_test.cc67 const size_t page_size = static_cast<size_t>(kPageSize); in RemapAtEndTest()
149 uintptr_t random_start = CreateStartPos(i * kPageSize); in TEST_F()
166 kPageSize, in TEST_F()
172 kPageSize, in TEST_F()
182 std::vector<uint8_t> data = RandomData(kPageSize); in TEST_F()
191 ASSERT_EQ(dest.Size(), static_cast<size_t>(kPageSize)); in TEST_F()
199 5 * kPageSize, // Need to make it larger in TEST_F()
209 3 * kPageSize, in TEST_F()
219 std::vector<uint8_t> data = RandomData(3 * kPageSize); in TEST_F()
223 dest.SetSize(kPageSize); in TEST_F()
[all …]
Dmem_map.cc129 constexpr uintptr_t mask = mask_ones & ~(kPageSize - 1); in CreateStartPos()
239 DCHECK_ALIGNED(reservation.Begin(), kPageSize); in CheckReservation()
320 size_t page_aligned_byte_count = RoundUp(byte_count, kPageSize); in MapAnonymous()
396 const size_t page_aligned_byte_count = RoundUp(byte_count, kPageSize); in MapDummy()
523 int page_offset = start % kPageSize; in MapFileAtAddress()
526 size_t page_aligned_byte_count = RoundUp(byte_count + page_offset, kPageSize); in MapFileAtAddress()
534 redzone_size = kPageSize; in MapFileAtAddress()
724 DCHECK_ALIGNED(begin_, kPageSize); in RemapAtEnd()
725 DCHECK_ALIGNED(base_begin_, kPageSize); in RemapAtEnd()
726 DCHECK_ALIGNED(reinterpret_cast<uint8_t*>(base_begin_) + base_size_, kPageSize); in RemapAtEnd()
[all …]
Dutils.cc90 CHECK_EQ(RoundDown(start, kPageSize), RoundDown(limit - 1, kPageSize)) << "range spans pages"; in TouchAndFlushCacheLinesWithinPage()
138 uintptr_t next_page = RoundUp(start + 1, kPageSize); in FlushCpuCaches()
145 next_page += kPageSize; in FlushCpuCaches()
341 uintptr_t vmstart = reinterpret_cast<uintptr_t>(AlignDown(addr, kPageSize)); in IsAddressKnownBackedByFileOrShared()
342 off_t index = (vmstart / kPageSize) * sizeof(uint64_t); in IsAddressKnownBackedByFileOrShared()
Dglobals.h39 static constexpr int kPageSize = 4096; variable
/art/runtime/jit/
Djit_memory_region_test.cc61 size_t size = kPageSize; in BasicTest()
67 mmap(nullptr, kPageSize, PROT_READ | PROT_WRITE, MAP_SHARED, fd.get(), 0)); in BasicTest()
85 mmap(nullptr, kPageSize, PROT_READ | PROT_WRITE, MAP_SHARED, fd.get(), 0)); in BasicTest()
95 addr2 = reinterpret_cast<int32_t*>(mremap(addr, 0, kPageSize, MREMAP_MAYMOVE)); in BasicTest()
102 addr2 = reinterpret_cast<int32_t*>(mremap(addr, kPageSize, 2 * kPageSize, MREMAP_MAYMOVE)); in BasicTest()
114 size_t size = kPageSize; in TestUnmapWritableAfterFork()
123 mmap(nullptr, kPageSize, PROT_READ | PROT_WRITE, MAP_SHARED, fd.get(), 0)); in TestUnmapWritableAfterFork()
133 mmap(nullptr, kPageSize, PROT_READ, MAP_SHARED, fd.get(), 0)); in TestUnmapWritableAfterFork()
147 mmap(nullptr, kPageSize, PROT_READ | PROT_WRITE, MAP_SHARED, fd2.get(), 0)); in TestUnmapWritableAfterFork()
166 munmap(addr, kPageSize); in TestUnmapWritableAfterFork()
[all …]
/art/runtime/gc/allocator/
Drosalloc.cc52 size_t RosAlloc::dedicated_full_run_storage_[kPageSize / sizeof(size_t)] = { 0 };
66 DCHECK_ALIGNED(base, kPageSize); in RosAlloc()
67 DCHECK_EQ(RoundUp(capacity, kPageSize), capacity); in RosAlloc()
68 DCHECK_EQ(RoundUp(max_capacity, kPageSize), max_capacity); in RosAlloc()
70 CHECK_ALIGNED(page_release_size_threshold_, kPageSize); in RosAlloc()
91 size_t num_of_pages = footprint_ / kPageSize; in RosAlloc()
92 size_t max_num_of_pages = max_capacity_ / kPageSize; in RosAlloc()
95 RoundUp(max_num_of_pages, kPageSize), in RosAlloc()
109 DCHECK_EQ(capacity_ % kPageSize, static_cast<size_t>(0)); in RosAlloc()
134 const size_t req_byte_size = num_pages * kPageSize; in AllocPages()
[all …]
Ddlmalloc.cc72 start = reinterpret_cast<void*>(art::RoundUp(reinterpret_cast<uintptr_t>(start), art::kPageSize)); in DlmallocMadviseCallback()
73 end = reinterpret_cast<void*>(art::RoundDown(reinterpret_cast<uintptr_t>(end), art::kPageSize)); in DlmallocMadviseCallback()
Drosalloc.h59 DCHECK_ALIGNED(byte_size, kPageSize); in ByteSize()
64 DCHECK_EQ(byte_size % kPageSize, static_cast<size_t>(0)); in SetByteSize()
105 DCHECK_EQ(byte_size % kPageSize, static_cast<size_t>(0)); in ReleasePages()
393 return reinterpret_cast<uint8_t*>(this) + kPageSize * numOfPages[size_bracket_idx_]; in End()
613 DCHECK_EQ(byte_offset % static_cast<size_t>(kPageSize), static_cast<size_t>(0)); in ToPageMapIndex()
614 return byte_offset / kPageSize; in ToPageMapIndex()
619 return (reinterpret_cast<uintptr_t>(addr) - reinterpret_cast<uintptr_t>(base_)) / kPageSize; in RoundDownToPageMapIndex()
875 return RoundUp(bytes, kPageSize); in UsableSize()
914 DCHECK_LT(idx, capacity_ / kPageSize); in IsFreePage()
/art/runtime/gc/collector/
Dimmune_spaces_test.cc74 reinterpret_cast<uint8_t*>(kPageSize), in ReserveBitmaps()
75 kPageSize)); in ReserveBitmaps()
191 constexpr size_t kImageSize = 123 * kPageSize; in TEST_F()
192 constexpr size_t kImageOatSize = 321 * kPageSize; in TEST_F()
193 constexpr size_t kOtherSpaceSize = 100 * kPageSize; in TEST_F()
250 constexpr size_t kImage1Size = kPageSize * 17; in TEST_F()
251 constexpr size_t kImage2Size = kPageSize * 13; in TEST_F()
252 constexpr size_t kImage3Size = kPageSize * 3; in TEST_F()
253 constexpr size_t kImage1OatSize = kPageSize * 5; in TEST_F()
254 constexpr size_t kImage2OatSize = kPageSize * 8; in TEST_F()
[all …]
Dsemi_space.cc378 if (LIKELY(size <= static_cast<size_t>(kPageSize))) { in CopyAvoidingDirtyingPages()
396 size_t page_remain = AlignUp(byte_dest, kPageSize) - byte_dest; in CopyAvoidingDirtyingPages()
401 DCHECK_ALIGNED(reinterpret_cast<uintptr_t>(byte_dest), kPageSize); in CopyAvoidingDirtyingPages()
404 while (byte_src + kPageSize < limit) { in CopyAvoidingDirtyingPages()
408 for (size_t i = 0; i < kPageSize / sizeof(*word_src); ++i) { in CopyAvoidingDirtyingPages()
417 saved_bytes += kPageSize; in CopyAvoidingDirtyingPages()
419 byte_src += kPageSize; in CopyAvoidingDirtyingPages()
420 byte_dest += kPageSize; in CopyAvoidingDirtyingPages()
/art/libartbase/arch/
Dinstruction_set.cc96 static_assert(IsAligned<kPageSize>(kArmStackOverflowReservedBytes), "ARM gap not page aligned");
97 static_assert(IsAligned<kPageSize>(kArm64StackOverflowReservedBytes), "ARM64 gap not page aligned");
98 static_assert(IsAligned<kPageSize>(kX86StackOverflowReservedBytes), "X86 gap not page aligned");
99 static_assert(IsAligned<kPageSize>(kX86_64StackOverflowReservedBytes),
/art/runtime/
Dimage.cc66 CHECK_EQ(image_begin, RoundUp(image_begin, kPageSize)); in ImageHeader()
67 CHECK_EQ(oat_file_begin, RoundUp(oat_file_begin, kPageSize)); in ImageHeader()
68 CHECK_EQ(oat_data_begin, RoundUp(oat_data_begin, kPageSize)); in ImageHeader()
80 CHECK_ALIGNED(delta, kPageSize) << "relocation delta must be page aligned"; in RelocateImageReferences()
90 CHECK_ALIGNED(delta, kPageSize) << "relocation delta must be page aligned"; in RelocateBootImageReferences()
104 return image_reservation_size_ == RoundUp(image_size_, kPageSize); in IsAppImage()
122 if (!IsAligned<kPageSize>(image_reservation_size_)) { in IsValid()
Druntime_globals.h30 return offset < kPageSize; in CanDoImplicitNullCheckOn()
36 static constexpr size_t kLargeObjectAlignment = kPageSize;
Doat.cc104 if (!IsAligned<kPageSize>(executable_offset_)) { in IsValid()
126 if (!IsAligned<kPageSize>(executable_offset_)) { in GetValidationErrorMessage()
187 DCHECK_ALIGNED(executable_offset_, kPageSize); in GetExecutableOffset()
193 DCHECK_ALIGNED(executable_offset, kPageSize); in SetExecutableOffset()
Dthread_pool.cc48 stack_size += kPageSize; in ThreadPoolWorker()
56 CHECK_ALIGNED(stack_.Begin(), kPageSize); in ThreadPoolWorker()
60 kPageSize, in ThreadPoolWorker()
/art/runtime/gc/space/
Dmalloc_space.cc105 *growth_limit = RoundUp(*growth_limit, kPageSize); in CreateMemMap()
106 *capacity = RoundUp(*capacity, kPageSize); in CreateMemMap()
143 growth_limit = RoundUp(growth_limit, kPageSize); in SetGrowthLimit()
186 SetEnd(reinterpret_cast<uint8_t*>(RoundUp(reinterpret_cast<uintptr_t>(End()), kPageSize))); in CreateZygoteSpace()
189 DCHECK_ALIGNED(begin_, kPageSize); in CreateZygoteSpace()
190 DCHECK_ALIGNED(End(), kPageSize); in CreateZygoteSpace()
191 size_t size = RoundUp(Size(), kPageSize); in CreateZygoteSpace()
203 SetGrowthLimit(RoundUp(size, kPageSize)); in CreateZygoteSpace()
/art/libdexfile/dex/
Ddex_file_layout.cc36 begin = AlignUp(begin, kPageSize); in MadviseLargestPageAlignedRegion()
37 end = AlignDown(end, kPageSize); in MadviseLargestPageAlignedRegion()
/art/dexlayout/
Ddexdiag.cc189 size_t first_page_of_section = section.offset / kPageSize; in FindSectionTypeForPage()
290 uint64_t start_page = (dex_file_start - vdex_start) / kPageSize; in ProcessOneDexMapping()
291 uint64_t start_address = start_page * kPageSize; in ProcessOneDexMapping()
292 uint64_t end_page = RoundUp(start_address + dex_file_size, kPageSize) / kPageSize; in ProcessOneDexMapping()
296 map_start + start_page * kPageSize, in ProcessOneDexMapping()
297 map_start + end_page * kPageSize) in ProcessOneDexMapping()
Ddex_visualize.cc73 fprintf(out_file_, "\"%s\" %d", s.name.c_str(), s.offset / kPageSize); in OpenAndPrintHeader()
101 const uint32_t low_page = from / kPageSize; in DumpAddressRange()
102 const uint32_t high_page = (size > 0) ? (from + size - 1) / kPageSize : low_page; in DumpAddressRange()
339 RoundUp(bytes, kPageSize) / kPageSize, in ShowDexSectionStatistics()
/art/dex2oat/
Dcommon_compiler_driver_test.cc115 inaccessible_page_ = mmap(nullptr, kPageSize, PROT_NONE, MAP_PRIVATE | MAP_ANONYMOUS, -1, 0); in SetUp()
121 munmap(inaccessible_page_, kPageSize); in TearDown()
/art/libelffile/elf/
Delf_builder.h205 header_.sh_addralign = kPageSize; // Page-align if R/W/X flags changed. in AddSection()
453 rodata_(this, ".rodata", SHT_PROGBITS, SHF_ALLOC, nullptr, 0, kPageSize, 0), in ElfBuilder()
454 text_(this, ".text", SHT_PROGBITS, SHF_ALLOC | SHF_EXECINSTR, nullptr, 0, kPageSize, 0), in ElfBuilder()
456 this, ".data.bimg.rel.ro", SHT_PROGBITS, SHF_ALLOC, nullptr, 0, kPageSize, 0), in ElfBuilder()
457 bss_(this, ".bss", SHT_NOBITS, SHF_ALLOC, nullptr, 0, kPageSize, 0), in ElfBuilder()
458 dex_(this, ".dex", SHT_NOBITS, SHF_ALLOC, nullptr, 0, kPageSize, 0), in ElfBuilder()
459 dynstr_(this, ".dynstr", SHF_ALLOC, kPageSize), in ElfBuilder()
462 dynamic_(this, ".dynamic", SHT_DYNAMIC, SHF_ALLOC, &dynstr_, 0, kPageSize, sizeof(Elf_Dyn)), in ElfBuilder()
533 CHECK(loaded_size_ == 0 || loaded_size_ == RoundUp(virtual_address_, kPageSize)) in End()
608 DCHECK_LE(section->header_.sh_offset, end + kPageSize) << "Large gap between sections"; in Strip()
[all …]
Dxz_utils.cc111 dst->resize(RoundUp(dst_offset + kPageSize / 4, kPageSize)); in XzDecompress()
/art/runtime/base/
Dmem_map_arena_pool.cc49 static_assert(ArenaAllocator::kArenaAlignment <= kPageSize, in MemMapArena()
58 size = RoundUp(size, kPageSize); in Allocate()
/art/runtime/gc/accounting/
Dspace_bitmap_test.cc121 for (uintptr_t i = range.second; i < range.second + kPageSize; i += kObjectAlignment) { in TEST_F()
220 RunTestCount<kPageSize>(); in TEST_F()
257 RunTestOrder<kPageSize>(); in TEST_F()

123