Lines Matching refs:kPageSize
129 constexpr uintptr_t mask = mask_ones & ~(kPageSize - 1); in CreateStartPos()
239 DCHECK_ALIGNED(reservation.Begin(), kPageSize); in CheckReservation()
320 size_t page_aligned_byte_count = RoundUp(byte_count, kPageSize); in MapAnonymous()
396 const size_t page_aligned_byte_count = RoundUp(byte_count, kPageSize); in MapDummy()
523 int page_offset = start % kPageSize; in MapFileAtAddress()
526 size_t page_aligned_byte_count = RoundUp(byte_count + page_offset, kPageSize); in MapFileAtAddress()
534 redzone_size = kPageSize; in MapFileAtAddress()
724 DCHECK_ALIGNED(begin_, kPageSize); in RemapAtEnd()
725 DCHECK_ALIGNED(base_begin_, kPageSize); in RemapAtEnd()
726 DCHECK_ALIGNED(reinterpret_cast<uint8_t*>(base_begin_) + base_size_, kPageSize); in RemapAtEnd()
727 DCHECK_ALIGNED(new_end, kPageSize); in RemapAtEnd()
742 DCHECK_ALIGNED(tail_base_size, kPageSize); in RemapAtEnd()
781 size_t base_size = RoundUp(byte_count, kPageSize); in TakeReservedMemory()
793 DCHECK_ALIGNED(begin_, kPageSize); in ReleaseReservedMemory()
794 DCHECK_ALIGNED(size_, kPageSize); in ReleaseReservedMemory()
799 byte_count = RoundUp(byte_count, kPageSize); in ReleaseReservedMemory()
925 CHECK_ALIGNED(size, kPageSize); in DumpMapsLocked()
933 os << "+0x" << std::hex << (size / kPageSize) << "P"; in DumpMapsLocked()
939 CHECK_ALIGNED(gap, kPageSize); in DumpMapsLocked()
940 os << "~0x" << std::hex << (gap / kPageSize) << "P"; in DumpMapsLocked()
944 CHECK_ALIGNED(it->second->BaseSize(), kPageSize); in DumpMapsLocked()
950 os << "+0x" << std::hex << (size / kPageSize) << "P"; in DumpMapsLocked()
1018 kPageSize); in SetSize()
1047 for (uintptr_t ptr = next_mem_pos_; ptr < 4 * GB; ptr += kPageSize) { in MapInternalArtLow4GBAllocator()
1056 CHECK_ALIGNED(ptr, kPageSize); in MapInternalArtLow4GBAllocator()
1067 CHECK_ALIGNED(ptr, kPageSize); in MapInternalArtLow4GBAllocator()
1082 ptr = LOW_MEM_START - kPageSize; in MapInternalArtLow4GBAllocator()
1095 for (tail_ptr = ptr; tail_ptr < ptr + length; tail_ptr += kPageSize) { in MapInternalArtLow4GBAllocator()
1096 if (msync(reinterpret_cast<void*>(tail_ptr), kPageSize, 0) == 0) { in MapInternalArtLow4GBAllocator()
1152 DCHECK_ALIGNED(length, kPageSize); in MapInternal()
1211 DCHECK(IsAligned<kPageSize>(begin)); in TryReadable()
1212 DCHECK(IsAligned<kPageSize>(end)); in TryReadable()
1215 for (volatile uint8_t* ptr = begin; ptr < end; ptr += kPageSize) { in TryReadable()
1228 uint8_t* const page_begin = AlignUp(mem_begin, kPageSize); in ZeroAndReleasePages()
1229 uint8_t* const page_end = AlignDown(mem_end, kPageSize); in ZeroAndReleasePages()
1251 CHECK_GT(size, static_cast<size_t>(kPageSize)); in AlignBy()
1252 CHECK_ALIGNED(size, kPageSize); in AlignBy()