• Home
  • Raw
  • Download

Lines Matching refs:kRegionSize

53   CHECK_ALIGNED(capacity, kRegionSize);  in CreateMemMap()
61 capacity + kRegionSize, in CreateMemMap()
80 CHECK_EQ(mem_map.Size(), capacity + kRegionSize); in CreateMemMap()
83 if (IsAlignedParam(mem_map.Begin(), kRegionSize)) { in CreateMemMap()
89 mem_map.AlignBy(kRegionSize); in CreateMemMap()
91 CHECK_ALIGNED(mem_map.Begin(), kRegionSize); in CreateMemMap()
92 CHECK_ALIGNED(mem_map.End(), kRegionSize); in CreateMemMap()
112 num_regions_(mem_map_.Size() / kRegionSize), in RegionSpace()
120 CHECK_ALIGNED(mem_map_.Size(), kRegionSize); in RegionSpace()
121 CHECK_ALIGNED(mem_map_.Begin(), kRegionSize); in RegionSpace()
125 for (size_t i = 0; i < num_regions_; ++i, region_addr += kRegionSize) { in RegionSpace()
126 regions_[i].Init(i, region_addr, region_addr + kRegionSize); in RegionSpace()
134 CHECK_EQ(static_cast<size_t>(regions_[i].End() - regions_[i].Begin()), kRegionSize); in RegionSpace()
158 return num_regions * kRegionSize; in FromSpaceSize()
170 return num_regions * kRegionSize; in UnevacFromSpaceSize()
182 return num_regions * kRegionSize; in ToSpaceSize()
274 const size_t bytes_allocated = RoundUp(BytesAllocated(), kRegionSize); in ShouldBeEvacuated()
300 DCHECK_ALIGNED(obj, kRegionSize); in ZeroLiveBytesForLargeObject()
302 DCHECK_GT(obj_size, space::RegionSpace::kRegionSize); in ZeroLiveBytesForLargeObject()
304 size_t obj_alloc_size = RoundUp(obj_size, space::RegionSpace::kRegionSize); in ZeroLiveBytesForLargeObject()
307 DCHECK_ALIGNED(end_addr, kRegionSize); in ZeroLiveBytesForLargeObject()
311 for (uint8_t* addr = begin_addr; addr < end_addr; addr += kRegionSize) { in ZeroLiveBytesForLargeObject()
380 num_expected_large_tails = RoundUp(r->BytesAllocated(), kRegionSize) / kRegionSize - 1; in SetFromSpace()
543 reinterpret_cast<mirror::Object*>(r->Begin() + free_regions * kRegionSize)); in ClearFromSpace()
594 + regions_to_clear_bitmap * kRegionSize)); in ClearFromSpace()
748 max_contiguous_free_regions * kRegionSize); in LogFragmentationAllocFailure()
785 size_t new_num_regions = new_capacity / kRegionSize; in ClampGrowthLimit()
862 DCHECK_ALIGNED(tlab_start, kRegionSize); in RevokeThreadLocalBuffersLocked()
865 DCHECK_LE(thread->GetThreadLocalBytesAllocated(), kRegionSize); in RevokeThreadLocalBuffersLocked()
916 << (static_cast<float>(live_bytes_) / RoundUp(BytesAllocated(), kRegionSize)); in Dump()
929 return kRegionSize; in GetLongestConsecutiveFreeBytes()
954 if (LIKELY(num_bytes <= kRegionSize)) { in AllocationSizeNonvirtual()
959 *usable_size = RoundUp(num_bytes, kRegionSize); in AllocationSizeNonvirtual()
1023 CheckedCall(mprotect, __FUNCTION__, Begin(), kRegionSize, PROT_READ | PROT_WRITE); in MarkAsAllocated()