Lines Matching refs:kPageSize
282 current_page_idx = entry_address / kPageSize + page_off; in IsEntryOnDirtyPage()
288 } while ((current_page_idx * kPageSize) < RoundUp(entry_address + size, kObjectAlignment)); in IsEntryOnDirtyPage()
987 RegionCommon<T>::GetDirtyEntryBytes() * 1.0f / (mapping_data.dirty_pages * kPageSize); in ProcessRegion()
1226 const uint8_t* image_end = AlignUp(image_begin + image_header.GetImageSize(), kPageSize); in Init()
1227 size_t virtual_page_idx_begin = reinterpret_cast<uintptr_t>(image_begin) / kPageSize; in Init()
1228 size_t virtual_page_idx_end = reinterpret_cast<uintptr_t>(image_end) / kPageSize; in Init()
1328 for (uintptr_t begin = boot_map.start; begin != boot_map.end; begin += kPageSize) { in ComputeDirtyBytes()
1337 if (memcmp(local_ptr, remote_ptr, kPageSize) != 0) { in ComputeDirtyBytes()
1341 for (size_t i = 0; i < kPageSize / sizeof(uint32_t); ++i) { in ComputeDirtyBytes()
1366 virtual_page_idx = reinterpret_cast<uintptr_t>(local_ptr) / kPageSize; in ComputeDirtyBytes()
1369 page_idx = (offset + page_off_begin) / kPageSize; in ComputeDirtyBytes()
1376 size_t remote_virtual_page_idx = begin / kPageSize; in ComputeDirtyBytes()
1480 boot_map.end = RoundUp(boot_map.start + image_header.GetImageSize(), kPageSize); in DumpImageDiffMap()
1494 zygote_boot_map.end = RoundUp(zygote_boot_map.start + image_header.GetImageSize(), kPageSize); in DumpImageDiffMap()
1514 const uint8_t* image_begin = AlignDown(image_begin_unaligned, kPageSize); in DumpImageDiffMap()
1515 const uint8_t* image_end = AlignUp(image_end_unaligned, kPageSize); in DumpImageDiffMap()
1528 DCHECK_ALIGNED(boot_map.start, kPageSize); in DumpImageDiffMap()
1529 DCHECK_ALIGNED(boot_map_size, kPageSize); in DumpImageDiffMap()