/art/runtime/base/ |
D | mem_map_arena_pool.cc | 104 Arena* arena = free_arenas_; in ReclaimMemory() local 106 delete arena; in ReclaimMemory() 134 for (Arena* arena = free_arenas_; arena != nullptr; arena = arena->next_) { in TrimMaps() local 135 arena->Release(); in TrimMaps() 142 for (Arena* arena = free_arenas_; arena != nullptr; arena = arena->next_) { in GetBytesAllocated() local 143 total += arena->GetBytesAllocated(); in GetBytesAllocated() 150 for (Arena* arena = first; arena != nullptr; arena = arena->next_) { in FreeArenaChain() local 151 MEMORY_TOOL_MAKE_UNDEFINED(arena->memory_, arena->bytes_allocated_); in FreeArenaChain()
|
D | gc_visited_arena_pool.h | 124 for (auto& arena : allocated_arenas_) { in VisitRoots() 125 arena.VisitRoots(visitor); in VisitRoots() 132 for (auto& arena : allocated_arenas_) { in ForEachAllocatedArena() 133 cb(arena); in ForEachAllocatedArena() 151 bool FindAllocatedArena(const TrackedArena* arena) const NO_THREAD_SAFETY_ANALYSIS { in FindAllocatedArena() argument 153 if (arena == &allocated_arena) { in FindAllocatedArena()
|
D | gc_visited_arena_pool.cc | 261 for (Arena* arena = first; arena != nullptr; arena = arena->Next()) { in FreeArenaChain() local 262 MEMORY_TOOL_MAKE_UNDEFINED(arena->Begin(), arena->GetBytesAllocated()); in FreeArenaChain()
|
/art/libartbase/base/ |
D | malloc_arena_pool.cc | 94 Arena* arena = free_arenas_; in ReclaimMemory() local 96 delete arena; in ReclaimMemory() 128 for (Arena* arena = free_arenas_; arena != nullptr; arena = arena->next_) { in GetBytesAllocated() local 129 total += arena->GetBytesAllocated(); in GetBytesAllocated() 136 for (Arena* arena = first; arena != nullptr; arena = arena->next_) { in FreeArenaChain() local 137 MEMORY_TOOL_MAKE_UNDEFINED(arena->memory_, arena->bytes_allocated_); in FreeArenaChain()
|
D | scoped_arena_allocator.cc | 112 for (Arena* arena = bottom_arena_; arena != nullptr; arena = arena->next_) { in ApproximatePeakBytes() local 113 sum += arena->bytes_allocated_; in ApproximatePeakBytes() 184 for (Arena* arena = start; arena != nullptr; arena = arena->next_) { in ApproximatePeakBytes() local 185 if (arena == arena_stack_->top_arena_) { in ApproximatePeakBytes() 186 sum += static_cast<size_t>(arena_stack_->top_ptr_ - arena->Begin()); in ApproximatePeakBytes() 189 sum += arena->bytes_allocated_; in ApproximatePeakBytes()
|
D | arena_allocator.cc | 136 for (const Arena* arena = first; arena != nullptr; arena = arena->next_) { in Dump() local 137 malloc_bytes += arena->Size(); in Dump() 138 lost_bytes += arena->RemainingSpace(); in Dump()
|
/art/runtime/ |
D | linear_alloc-inl.h | 33 Arena* arena = allocator_.GetHeadArena(); in SetFirstObject() local 34 DCHECK_NE(arena, nullptr); in SetFirstObject() 36 if (UNLIKELY(begin < arena->Begin() || begin >= arena->End())) { in SetFirstObject() 37 arena = arena->Next(); in SetFirstObject() 39 DCHECK(begin >= arena->Begin() && end <= arena->End()); in SetFirstObject() 40 down_cast<TrackedArena*>(arena)->SetFirstObject(static_cast<uint8_t*>(begin), end); in SetFirstObject()
|
/art/compiler/optimizing/ |
D | instruction_simplifier_x86_shared.cc | 44 ArenaAllocator* arena = instruction->GetBlock()->GetGraph()->GetAllocator(); in TryCombineAndNot() local 45 HX86AndNot* and_not = new (arena) HX86AndNot(type, in TryCombineAndNot() 80 ArenaAllocator* arena = instruction->GetBlock()->GetGraph()->GetAllocator(); in TryGenerateResetLeastSetBit() local 81 HX86MaskOrResetLeastSetBit* lsb = new (arena) HX86MaskOrResetLeastSetBit( in TryGenerateResetLeastSetBit() 113 ArenaAllocator* arena = instruction->GetBlock()->GetGraph()->GetAllocator(); in TryGenerateMaskUptoLeastSetBit() local 114 HX86MaskOrResetLeastSetBit* lsb = new (arena) HX86MaskOrResetLeastSetBit( in TryGenerateMaskUptoLeastSetBit()
|
D | superblock_cloner_test.cc | 151 ArenaAllocator* arena = GetAllocator(); in TEST_F() local 160 arena, graph_->GetBlocks().size(), false, kArenaAllocSuperblockCloner); in TEST_F() 161 HBasicBlockMap bb_map(std::less<HBasicBlock*>(), arena->Adapter(kArenaAllocSuperblockCloner)); in TEST_F() 162 HInstructionMap hir_map(std::less<HInstruction*>(), arena->Adapter(kArenaAllocSuperblockCloner)); in TEST_F() 232 ArenaAllocator* arena = GetAllocator(); in TEST_F() local 241 arena, graph_->GetBlocks().size(), false, kArenaAllocSuperblockCloner); in TEST_F() 271 ArenaAllocator* arena = GetAllocator(); in TEST_F() local 279 arena, graph_->GetBlocks().size(), false, kArenaAllocSuperblockCloner); in TEST_F() 627 ArenaAllocator* arena = GetAllocator(); in TEST_F() local 637 arena, graph_->GetBlocks().size(), false, kArenaAllocSuperblockCloner); in TEST_F() [all …]
|
D | nodes.h | 1716 HInstruction* Clone(ArenaAllocator* arena) const override { \ 1718 return new (arena) H##type(*this->As##type()); \ 2571 virtual HInstruction* Clone(ArenaAllocator* arena ATTRIBUTE_UNUSED) const { in FOR_EACH_INSTRUCTION()
|
/art/runtime/gc/collector/ |
D | mark_compact.cc | 2736 [](const TrackedArena& arena) in CompactionPause() 2737 REQUIRES_SHARED(Locks::mutator_lock_) { CHECK(arena.IsPreZygoteForkArena()); }); in CompactionPause() 2746 [this](const TrackedArena& arena) REQUIRES_SHARED(Locks::mutator_lock_) { in CompactionPause() 2750 if (!arena.IsPreZygoteForkArena()) { in CompactionPause() 2751 uint8_t* last_byte = arena.GetLastUsedByte(); in CompactionPause() 2752 CHECK(linear_alloc_arenas_.insert({&arena, last_byte}).second); in CompactionPause() 2755 arena.VisitRoots(updater); in CompactionPause() 3394 const TrackedArena* arena = pair.first; in ProcessLinearAlloc() local 3405 if (arena_pool->AreArenasFreed() && !arena_pool->FindAllocatedArena(arena)) { in ProcessLinearAlloc() 3411 arena_begin = arena->Begin(); in ProcessLinearAlloc() [all …]
|
/art/build/boot/hiddenapi/ |
D | hiddenapi-max-target-o-low-priority.txt | 7393 Ljava/util/concurrent/Exchanger;->arena:[Ljava/util/concurrent/Exchanger$Node;
|