Home
last modified time | relevance | path

Searched refs:RegionSpace (Results 1 – 12 of 12) sorted by relevance

/art/runtime/gc/space/
Dregion_space.cc31 RegionSpace* RegionSpace::Create(const std::string& name, size_t capacity, in Create()
44 return new RegionSpace(name, mem_map.release()); in Create()
47 RegionSpace::RegionSpace(const std::string& name, MemMap* mem_map) in RegionSpace() function in art::gc::space::RegionSpace
82 size_t RegionSpace::FromSpaceSize() { in FromSpaceSize()
94 size_t RegionSpace::UnevacFromSpaceSize() { in UnevacFromSpaceSize()
106 size_t RegionSpace::ToSpaceSize() { in ToSpaceSize()
118 inline bool RegionSpace::Region::ShouldBeEvacuated() { in ShouldBeEvacuated()
148 void RegionSpace::SetFromSpace(accounting::ReadBarrierTable* rb_table, bool force_evacuate_all) { in SetFromSpace()
205 void RegionSpace::ClearFromSpace() { in ClearFromSpace()
219 void RegionSpace::AssertAllRegionLiveBytesZeroOrCleared() { in AssertAllRegionLiveBytesZeroOrCleared()
[all …]
Dregion_space-inl.h26 inline mirror::Object* RegionSpace::Alloc(Thread*, size_t num_bytes, size_t* bytes_allocated, in Alloc()
34 inline mirror::Object* RegionSpace::AllocThreadUnsafe(Thread* self, size_t num_bytes, in AllocThreadUnsafe()
43 inline mirror::Object* RegionSpace::AllocNonvirtual(size_t num_bytes, size_t* bytes_allocated, in AllocNonvirtual()
114 inline mirror::Object* RegionSpace::Region::Alloc(size_t num_bytes, size_t* bytes_allocated, in Alloc()
141 inline size_t RegionSpace::AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size) { in AllocationSizeNonvirtual()
155 template<RegionSpace::RegionType kRegionType>
156 uint64_t RegionSpace::GetBytesAllocatedInternal() { in GetBytesAllocatedInternal()
190 template<RegionSpace::RegionType kRegionType>
191 uint64_t RegionSpace::GetObjectsAllocatedInternal() { in GetObjectsAllocatedInternal()
226 void RegionSpace::WalkInternal(ObjectCallback* callback, void* arg) { in WalkInternal()
[all …]
Dregion_space.h30 class RegionSpace FINAL : public ContinuousMemMapAllocSpace {
41 static RegionSpace* Create(const std::string& name, size_t capacity, uint8_t* requested_begin);
147 RegionSpace* AsRegionSpace() OVERRIDE { in AsRegionSpace()
228 RegionSpace(const std::string& name, MemMap* mem_map);
478 friend class RegionSpace; variable
522 DISALLOW_COPY_AND_ASSIGN(RegionSpace);
525 std::ostream& operator<<(std::ostream& os, const RegionSpace::RegionState& value);
526 std::ostream& operator<<(std::ostream& os, const RegionSpace::RegionType& value);
Dspace.h53 class RegionSpace; variable
140 virtual RegionSpace* AsRegionSpace();
Dspace.cc61 RegionSpace* Space::AsRegionSpace() { in AsRegionSpace()
/art/runtime/gc/collector/
Dconcurrent_copying-inl.h51 space::RegionSpace::RegionType rtype = region_space_->GetRegionType(from_ref); in Mark()
53 case space::RegionSpace::RegionType::kRegionTypeToSpace: in Mark()
56 case space::RegionSpace::RegionType::kRegionTypeFromSpace: { in Mark()
70 case space::RegionSpace::RegionType::kRegionTypeUnevacFromSpace: { in Mark()
87 case space::RegionSpace::RegionType::kRegionTypeNone: in Mark()
Dconcurrent_copying.h47 class RegionSpace; variable
81 void SetRegionSpace(space::RegionSpace* region_space) { in SetRegionSpace()
85 space::RegionSpace* RegionSpace() { in RegionSpace() function
189 space::RegionSpace* region_space_; // The underlying region space.
Dconcurrent_copying.cc58 static_assert(space::RegionSpace::kRegionSize == accounting::ReadBarrierTable::kRegionSize, in ConcurrentCopying()
672 if (collector_->RegionSpace()->IsInToSpace(ref)) { in operator ()()
746 space::RegionSpace* region_space = collector->RegionSpace(); in ObjectCallback()
751 if (collector->RegionSpace()->IsInToSpace(obj)) { in ObjectCallback()
877 space::RegionSpace* region_space = collector->RegionSpace(); in ObjectCallback()
1389 size_t alloc_size = RoundUp(obj_size, space::RegionSpace::kAlignment); in operator ()()
1735 CHECK_ALIGNED(alloc_size, space::RegionSpace::kAlignment); in AllocateInSkippedBlock()
1737 size_t min_object_size = RoundUp(sizeof(mirror::Object), space::RegionSpace::kAlignment); in AllocateInSkippedBlock()
1754 CHECK_ALIGNED(it->first - alloc_size, space::RegionSpace::kAlignment); in AllocateInSkippedBlock()
1765 CHECK_ALIGNED(byte_size, space::RegionSpace::kAlignment); in AllocateInSkippedBlock()
[all …]
/art/runtime/gc/
Dheap-inl.h338 alloc_size = RoundUp(alloc_size, space::RegionSpace::kAlignment); in TryToAllocate()
345 DCHECK_ALIGNED(alloc_size, space::RegionSpace::kAlignment); in TryToAllocate()
347 if (space::RegionSpace::kRegionSize >= alloc_size) { in TryToAllocate()
349 … if (LIKELY(!IsOutOfMemoryOnAllocation<kGrow>(allocator_type, space::RegionSpace::kRegionSize))) { in TryToAllocate()
357 *bytes_tl_bulk_allocated = space::RegionSpace::kRegionSize; in TryToAllocate()
Dreference_queue.cc76 const bool is_moving = concurrent_copying->RegionSpace()->IsInToSpace(ref); in DequeuePendingReference()
Dheap.h90 class RegionSpace; variable
1247 space::RegionSpace* region_space_;
Dheap.cc445 region_space_ = space::RegionSpace::Create("Region space", capacity_ * 2, request_begin); in Heap()