Home
last modified time | relevance | path

Searched refs:arena (Results 1 – 12 of 12) sorted by relevance

/art/runtime/base/
Dmem_map_arena_pool.cc104 Arena* arena = free_arenas_; in ReclaimMemory() local
106 delete arena; in ReclaimMemory()
134 for (Arena* arena = free_arenas_; arena != nullptr; arena = arena->next_) { in TrimMaps() local
135 arena->Release(); in TrimMaps()
142 for (Arena* arena = free_arenas_; arena != nullptr; arena = arena->next_) { in GetBytesAllocated() local
143 total += arena->GetBytesAllocated(); in GetBytesAllocated()
150 for (Arena* arena = first; arena != nullptr; arena = arena->next_) { in FreeArenaChain() local
151 MEMORY_TOOL_MAKE_UNDEFINED(arena->memory_, arena->bytes_allocated_); in FreeArenaChain()
Dgc_visited_arena_pool.h124 for (auto& arena : allocated_arenas_) { in VisitRoots()
125 arena.VisitRoots(visitor); in VisitRoots()
132 for (auto& arena : allocated_arenas_) { in ForEachAllocatedArena()
133 cb(arena); in ForEachAllocatedArena()
151 bool FindAllocatedArena(const TrackedArena* arena) const NO_THREAD_SAFETY_ANALYSIS { in FindAllocatedArena() argument
153 if (arena == &allocated_arena) { in FindAllocatedArena()
Dgc_visited_arena_pool.cc261 for (Arena* arena = first; arena != nullptr; arena = arena->Next()) { in FreeArenaChain() local
262 MEMORY_TOOL_MAKE_UNDEFINED(arena->Begin(), arena->GetBytesAllocated()); in FreeArenaChain()
/art/libartbase/base/
Dmalloc_arena_pool.cc94 Arena* arena = free_arenas_; in ReclaimMemory() local
96 delete arena; in ReclaimMemory()
128 for (Arena* arena = free_arenas_; arena != nullptr; arena = arena->next_) { in GetBytesAllocated() local
129 total += arena->GetBytesAllocated(); in GetBytesAllocated()
136 for (Arena* arena = first; arena != nullptr; arena = arena->next_) { in FreeArenaChain() local
137 MEMORY_TOOL_MAKE_UNDEFINED(arena->memory_, arena->bytes_allocated_); in FreeArenaChain()
Dscoped_arena_allocator.cc112 for (Arena* arena = bottom_arena_; arena != nullptr; arena = arena->next_) { in ApproximatePeakBytes() local
113 sum += arena->bytes_allocated_; in ApproximatePeakBytes()
184 for (Arena* arena = start; arena != nullptr; arena = arena->next_) { in ApproximatePeakBytes() local
185 if (arena == arena_stack_->top_arena_) { in ApproximatePeakBytes()
186 sum += static_cast<size_t>(arena_stack_->top_ptr_ - arena->Begin()); in ApproximatePeakBytes()
189 sum += arena->bytes_allocated_; in ApproximatePeakBytes()
Darena_allocator.cc136 for (const Arena* arena = first; arena != nullptr; arena = arena->next_) { in Dump() local
137 malloc_bytes += arena->Size(); in Dump()
138 lost_bytes += arena->RemainingSpace(); in Dump()
/art/runtime/
Dlinear_alloc-inl.h33 Arena* arena = allocator_.GetHeadArena(); in SetFirstObject() local
34 DCHECK_NE(arena, nullptr); in SetFirstObject()
36 if (UNLIKELY(begin < arena->Begin() || begin >= arena->End())) { in SetFirstObject()
37 arena = arena->Next(); in SetFirstObject()
39 DCHECK(begin >= arena->Begin() && end <= arena->End()); in SetFirstObject()
40 down_cast<TrackedArena*>(arena)->SetFirstObject(static_cast<uint8_t*>(begin), end); in SetFirstObject()
/art/compiler/optimizing/
Dinstruction_simplifier_x86_shared.cc44 ArenaAllocator* arena = instruction->GetBlock()->GetGraph()->GetAllocator(); in TryCombineAndNot() local
45 HX86AndNot* and_not = new (arena) HX86AndNot(type, in TryCombineAndNot()
80 ArenaAllocator* arena = instruction->GetBlock()->GetGraph()->GetAllocator(); in TryGenerateResetLeastSetBit() local
81 HX86MaskOrResetLeastSetBit* lsb = new (arena) HX86MaskOrResetLeastSetBit( in TryGenerateResetLeastSetBit()
113 ArenaAllocator* arena = instruction->GetBlock()->GetGraph()->GetAllocator(); in TryGenerateMaskUptoLeastSetBit() local
114 HX86MaskOrResetLeastSetBit* lsb = new (arena) HX86MaskOrResetLeastSetBit( in TryGenerateMaskUptoLeastSetBit()
Dsuperblock_cloner_test.cc151 ArenaAllocator* arena = GetAllocator(); in TEST_F() local
160 arena, graph_->GetBlocks().size(), false, kArenaAllocSuperblockCloner); in TEST_F()
161 HBasicBlockMap bb_map(std::less<HBasicBlock*>(), arena->Adapter(kArenaAllocSuperblockCloner)); in TEST_F()
162 HInstructionMap hir_map(std::less<HInstruction*>(), arena->Adapter(kArenaAllocSuperblockCloner)); in TEST_F()
232 ArenaAllocator* arena = GetAllocator(); in TEST_F() local
241 arena, graph_->GetBlocks().size(), false, kArenaAllocSuperblockCloner); in TEST_F()
271 ArenaAllocator* arena = GetAllocator(); in TEST_F() local
279 arena, graph_->GetBlocks().size(), false, kArenaAllocSuperblockCloner); in TEST_F()
627 ArenaAllocator* arena = GetAllocator(); in TEST_F() local
637 arena, graph_->GetBlocks().size(), false, kArenaAllocSuperblockCloner); in TEST_F()
[all …]
Dnodes.h1716 HInstruction* Clone(ArenaAllocator* arena) const override { \
1718 return new (arena) H##type(*this->As##type()); \
2571 virtual HInstruction* Clone(ArenaAllocator* arena ATTRIBUTE_UNUSED) const { in FOR_EACH_INSTRUCTION()
/art/runtime/gc/collector/
Dmark_compact.cc2736 [](const TrackedArena& arena) in CompactionPause()
2737 REQUIRES_SHARED(Locks::mutator_lock_) { CHECK(arena.IsPreZygoteForkArena()); }); in CompactionPause()
2746 [this](const TrackedArena& arena) REQUIRES_SHARED(Locks::mutator_lock_) { in CompactionPause()
2750 if (!arena.IsPreZygoteForkArena()) { in CompactionPause()
2751 uint8_t* last_byte = arena.GetLastUsedByte(); in CompactionPause()
2752 CHECK(linear_alloc_arenas_.insert({&arena, last_byte}).second); in CompactionPause()
2755 arena.VisitRoots(updater); in CompactionPause()
3394 const TrackedArena* arena = pair.first; in ProcessLinearAlloc() local
3405 if (arena_pool->AreArenasFreed() && !arena_pool->FindAllocatedArena(arena)) { in ProcessLinearAlloc()
3411 arena_begin = arena->Begin(); in ProcessLinearAlloc()
[all …]
/art/build/boot/hiddenapi/
Dhiddenapi-max-target-o-low-priority.txt7393 Ljava/util/concurrent/Exchanger;->arena:[Ljava/util/concurrent/Exchanger$Node;