Home
last modified time | relevance | path

Searched refs:arena_ (Results 1 – 25 of 30) sorted by relevance

12

/art/compiler/optimizing/
Dbuilder.cc76 HLocal* local = new (arena_) HLocal(i); in InitializeLocals()
96 new (arena_) HParameterValue(parameter_index++, Primitive::kPrimNot); in InitializeParameters()
99 entry_block_->AddInstruction(new (arena_) HStoreLocal(local, parameter)); in InitializeParameters()
114 new (arena_) HParameterValue(parameter_index++, Primitive::GetType(shorty[pos - 1])); in InitializeParameters()
119 entry_block_->AddInstruction(new (arena_) HStoreLocal(local, parameter)); in InitializeParameters()
143 T* comparison = new (arena_) T(first, second); in If_22t()
145 HInstruction* ifinst = new (arena_) HIf(comparison); in If_22t()
159 T* comparison = new (arena_) T(value, GetIntConstant(0)); in If_21t()
161 HInstruction* ifinst = new (arena_) HIf(comparison); in If_21t()
181 graph_ = new (arena_) HGraph(arena_); in BuildGraph()
[all …]
Dnodes.cc29 ArenaBitVector visiting(arena_, blocks_.Size(), false); in FindBackEdges()
70 ArenaBitVector visited(arena_, blocks_.Size(), false); in BuildDominatorTree()
87 GrowableArray<size_t> visits(arena_, blocks_.Size()); in BuildDominatorTree()
96 ArenaBitVector visited(arena_, blocks_.Size(), false); in FindCommonDominator()
143 HBasicBlock* new_block = new (arena_) HBasicBlock(this); in SplitCriticalEdge()
145 new_block->AddInstruction(new (arena_) HGoto()); in SplitCriticalEdge()
165 HBasicBlock* new_back_edge = new (arena_) HBasicBlock(this); in SimplifyLoop()
167 new_back_edge->AddInstruction(new (arena_) HGoto()); in SimplifyLoop()
182 HBasicBlock* pre_header = new (arena_) HBasicBlock(this); in SimplifyLoop()
184 pre_header->AddInstruction(new (arena_) HGoto()); in SimplifyLoop()
[all …]
Dbuilder.h38 : arena_(arena), in arena_() function
110 ArenaAllocator* const arena_; variable
Dnodes.h74 : arena_(arena), in HGraph()
83 ArenaAllocator* GetArena() const { return arena_; } in GetArena()
161 ArenaAllocator* const arena_; variable
/art/compiler/dex/
Dssa_transformation.cc97 dfs_order_ = new (arena_) GrowableArray<BasicBlockId>(arena_, GetNumBlocks(), in ComputeDFSOrders()
106 dfs_post_order_ = new (arena_) GrowableArray<BasicBlockId>(arena_, GetNumBlocks(), in ComputeDFSOrders()
152 (arena_->Alloc(sizeof(ArenaBitVector *) * num_registers, in ComputeDefBlockMatrix()
159 new (arena_) ArenaBitVector(arena_, GetNumBlocks(), false, kBitMapBMatrix); in ComputeDefBlockMatrix()
185 new (arena_) GrowableArray<BasicBlockId>(arena_, num_reachable_blocks_, in ComputeDomPostOrderTraversal()
276 bb->dominators = new (arena_) ArenaBitVector(arena_, num_total_blocks, in InitializeDominationInfo()
278 bb->i_dominated = new (arena_) ArenaBitVector(arena_, num_total_blocks, in InitializeDominationInfo()
280 bb->dom_frontier = new (arena_) ArenaBitVector(arena_, num_total_blocks, in InitializeDominationInfo()
392 i_dom_list_ = static_cast<int*>(arena_->Alloc(sizeof(int) * num_reachable_blocks, in ComputeDominators()
562 static_cast<BasicBlockId*>(arena_->Alloc(sizeof(BasicBlockId) * num_uses, in InsertPhiNodeOperands()
Dbackend.h51 explicit Backend(ArenaAllocator* arena) : arena_(arena) {} in Backend()
52 ArenaAllocator* const arena_; variable
Dmir_dataflow.cc933 new (arena_) ArenaBitVector(arena_, cu_->num_dalvik_registers, false, kBitMapUse); in FindLocalLiveIn()
935 new (arena_) ArenaBitVector(arena_, cu_->num_dalvik_registers, false, kBitMapDef); in FindLocalLiveIn()
937 new (arena_) ArenaBitVector(arena_, cu_->num_dalvik_registers, false, kBitMapLiveIn); in FindLocalLiveIn()
1021 … mir->ssa_rep->uses = static_cast<int*>(arena_->Alloc(sizeof(int) * num_uses, kArenaAllocDFInfo)); in AllocateSSAUseData()
1023 …mir->ssa_rep->fp_use = static_cast<bool*>(arena_->Alloc(sizeof(bool) * num_uses, kArenaAllocDFInfo… in AllocateSSAUseData()
1031 mir->ssa_rep->defs = static_cast<int*>(arena_->Alloc(sizeof(int) * num_defs, in AllocateSSADefData()
1033 mir->ssa_rep->fp_def = static_cast<bool*>(arena_->Alloc(sizeof(bool) * num_defs, in AllocateSSADefData()
1080 static_cast<struct SSARepresentation *>(arena_->Alloc(sizeof(SSARepresentation), in DoSSAConversion()
1191 static_cast<int*>(arena_->Alloc(sizeof(int) * cu_->num_dalvik_registers, in DoSSAConversion()
1203 ssa_base_vregs_ = new (arena_) GrowableArray<int>(arena_, num_dalvik_reg + GetDefCount() + 128, in CompilerInitializeSSAConversion()
[all …]
Dmir_graph.cc116 arena_(arena), in MIRGraph()
128 try_block_addr_ = new (arena_) ArenaBitVector(arena_, 0, true /* expandable */); in MIRGraph()
517 new (arena_) GrowableArray<SuccessorBlockInfo*>(arena_, size, kGrowableArraySuccessorBlocks); in ProcessCanSwitch()
523 static_cast<SuccessorBlockInfo*>(arena_->Alloc(sizeof(SuccessorBlockInfo), in ProcessCanSwitch()
570 cur_block->successor_blocks = new (arena_) GrowableArray<SuccessorBlockInfo*>( in ProcessCanThrow()
571 arena_, 2, kGrowableArraySuccessorBlocks); in ProcessCanThrow()
578 (arena_->Alloc(sizeof(SuccessorBlockInfo), kArenaAllocSuccessor)); in ProcessCanThrow()
1340 ret = static_cast<char*>(arena_->Alloc(length, kArenaAllocDFInfo)); in GetDalvikDisassembly()
1454 CallInfo* info = static_cast<CallInfo*>(arena_->Alloc(sizeof(CallInfo), in NewMemCallInfo()
1465 (arena_->Alloc(sizeof(RegLocation) * info->num_arg_words, kArenaAllocMisc)); in NewMemCallInfo()
[all …]
Dmir_optimization.cc257 CompilerTemp *compiler_temp = static_cast<CompilerTemp *>(arena_->Alloc(sizeof(CompilerTemp), in GetNewCompilerTemp()
281 static_cast<CompilerTemp *>(arena_->Alloc(sizeof(CompilerTemp), kArenaAllocRegAlloc)); in GetNewCompilerTemp()
506 static_cast<int*>(arena_->Alloc(sizeof(int) * 3, kArenaAllocDFInfo)); in BasicBlockOpt()
515 static_cast<int*>(arena_->Alloc(sizeof(int) * 1, kArenaAllocDFInfo)); in BasicBlockOpt()
517 static_cast<bool*>(arena_->Alloc(sizeof(bool) * 1, kArenaAllocDFInfo)); in BasicBlockOpt()
521 static_cast<bool*>(arena_->Alloc(sizeof(bool) * mir->ssa_rep->num_uses, in BasicBlockOpt()
1304 static_cast<Checkstats*>(arena_->Alloc(sizeof(Checkstats), kArenaAllocDFInfo)); in DumpCheckStats()
Dmir_graph.h625 return arena_; in GetArena()
629 opcode_count_ = static_cast<int*>(arena_->Alloc(kNumPackedOpcodes * sizeof(int), in EnableOpcodeCounting()
1189 ArenaAllocator* arena_; variable
Dvreg_analysis.cc441 RegLocation* loc = static_cast<RegLocation*>(arena_->Alloc(max_regs * sizeof(*loc), in InitRegLocations()
/art/compiler/utils/
Darena_bit_vector.cc25 explicit ArenaBitVectorAllocator(ArenaAlloc* arena) : arena_(arena) {} in ArenaBitVectorAllocator()
29 return arena_->Alloc(size, kArenaAllocGrowableBitMap); in Alloc()
40 ArenaAlloc* arena_; member in art::ArenaBitVectorAllocator
Dgrowable_array.h85 : arena_(arena), in arena_() function
89 elem_list_ = static_cast<T*>(arena_->Alloc(sizeof(T) * init_length, in arena_()
103 T* new_array = static_cast<T*>(arena_->Alloc(sizeof(T) * target_length, in Resize()
208 ArenaAllocator* const arena_;
/art/compiler/dex/quick/mips/
Dcall_mips.cc71 static_cast<SwitchTable*>(arena_->Alloc(sizeof(SwitchTable), kArenaAllocData)); in GenLargeSparseSwitch()
76 static_cast<LIR**>(arena_->Alloc(elements * sizeof(LIR*), kArenaAllocLIR)); in GenLargeSparseSwitch()
148 static_cast<SwitchTable*>(arena_->Alloc(sizeof(SwitchTable), kArenaAllocData)); in GenLargePackedSwitch()
152 tab_rec->targets = static_cast<LIR**>(arena_->Alloc(size * sizeof(LIR*), in GenLargePackedSwitch()
227 reinterpret_cast<FillArrayData*>(arena_->Alloc(sizeof(FillArrayData), in GenFillArrayData()
343 AddSlowPath(new(arena_)StackOverflowSlowPath(this, branch, spill_count * 4)); in GenEntrySequence()
Dtarget_mips.cc433 reg_pool_ = new (arena_) RegisterPool(this, arena_, core_regs, empty_pool /* core64 */, sp_regs, in CompilerInitializeRegAlloc()
/art/compiler/dex/quick/arm/
Dcall_arm.cc53 static_cast<SwitchTable*>(arena_->Alloc(sizeof(SwitchTable), kArenaAllocData)); in GenLargeSparseSwitch()
57 tab_rec->targets = static_cast<LIR**>(arena_->Alloc(size * sizeof(LIR*), kArenaAllocLIR)); in GenLargeSparseSwitch()
101 static_cast<SwitchTable*>(arena_->Alloc(sizeof(SwitchTable), kArenaAllocData)); in GenLargePackedSwitch()
106 static_cast<LIR**>(arena_->Alloc(size * sizeof(LIR*), kArenaAllocLIR)); in GenLargePackedSwitch()
154 static_cast<FillArrayData*>(arena_->Alloc(sizeof(FillArrayData), kArenaAllocData)); in GenFillArrayData()
436 AddSlowPath(new(arena_)StackOverflowSlowPath(this, branch, true, spill_size)); in GenEntrySequence()
451 AddSlowPath(new(arena_)StackOverflowSlowPath(this, branch, false, frame_size_)); in GenEntrySequence()
Dtarget_arm.cc570 reg_pool_ = new (arena_) RegisterPool(this, arena_, core_regs, empty_pool /* core64 */, sp_regs, in CompilerInitializeRegAlloc()
/art/compiler/dex/quick/arm64/
Dcall_arm64.cc53 static_cast<SwitchTable*>(arena_->Alloc(sizeof(SwitchTable), kArenaAllocData)); in GenLargeSparseSwitch()
57 tab_rec->targets = static_cast<LIR**>(arena_->Alloc(size * sizeof(LIR*), kArenaAllocLIR)); in GenLargeSparseSwitch()
105 static_cast<SwitchTable*>(arena_->Alloc(sizeof(SwitchTable), kArenaAllocData)); in GenLargePackedSwitch()
110 static_cast<LIR**>(arena_->Alloc(size * sizeof(LIR*), kArenaAllocLIR)); in GenLargePackedSwitch()
163 static_cast<FillArrayData*>(arena_->Alloc(sizeof(FillArrayData), kArenaAllocData)); in GenFillArrayData()
398 AddSlowPath(new(arena_)StackOverflowSlowPath(this, branch, frame_size_)); in GenEntrySequence()
Dtarget_arm64.cc598 reg_pool_ = new (arena_) RegisterPool(this, arena_, core_regs, core64_regs, sp_regs, dp_regs, in CompilerInitializeRegAlloc()
/art/compiler/dex/quick/x86/
Dcall_x86.cc70 static_cast<SwitchTable*>(arena_->Alloc(sizeof(SwitchTable), kArenaAllocData)); in GenLargePackedSwitch()
74 tab_rec->targets = static_cast<LIR**>(arena_->Alloc(size * sizeof(LIR*), in GenLargePackedSwitch()
141 static_cast<FillArrayData*>(arena_->Alloc(sizeof(FillArrayData), kArenaAllocData)); in GenFillArrayData()
280 new(arena_)StackOverflowSlowPath(this, branch, in GenEntrySequence()
Dtarget_x86.cc616 reg_pool_ = new (arena_) RegisterPool(this, arena_, core_regs_64, core_regs_64q, sp_regs_64, in CompilerInitializeRegAlloc()
620 reg_pool_ = new (arena_) RegisterPool(this, arena_, core_regs_32, empty_pool, sp_regs_32, in CompilerInitializeRegAlloc()
630 RegisterInfo* info = new (arena_) RegisterInfo(reg, GetRegMaskCommon(reg)); in CompilerInitializeRegAlloc()
2360 LIR* new_value = static_cast<LIR*>(arena_->Alloc(sizeof(LIR), kArenaAllocData)); in AddVectorLiteral()
/art/compiler/dex/quick/
Dgen_common.cc81 AddSlowPath(new (arena_) DivZeroCheckSlowPath(this, branch)); in AddDivZeroCheckSlowPath()
105 AddSlowPath(new (arena_) ArrayBoundsCheckSlowPath(this, branch, index, length)); in GenArrayBoundsCheck()
135 AddSlowPath(new (arena_) ArrayBoundsCheckSlowPath(this, branch, index, length)); in GenArrayBoundsCheck()
154 AddSlowPath(new (arena_) NullCheckSlowPath(this, branch)); in GenNullCheck()
580 AddSlowPath(new (arena_) StaticFieldSlowPath(this, unresolved_branch, uninit_branch, cont, in GenSput()
665 AddSlowPath(new (arena_) StaticFieldSlowPath(this, unresolved_branch, uninit_branch, cont, in GenSget()
872 AddSlowPath(new (arena_) SlowPath(this, branch, cont, type_idx, rl_method, rl_result)); in GenConstClass()
935 AddSlowPath(new (arena_) SlowPath(this, fromfast, cont, r_method, string_idx)); in GenConstString()
1144 AddSlowPath(new (arena_) InitTypeSlowPath(this, slow_path_branch, slow_path_target, in GenInstanceofCallingHelper()
1291 AddSlowPath(new (arena_) SlowPath(this, hop_branch, cont, type_idx, class_reg)); in GenCheckCast()
[all …]
Dmir_to_lir-inl.h46 LIR* insn = static_cast<LIR*>(arena_->Alloc(sizeof(LIR), kArenaAllocLIR)); in RawLIR()
Dlocal_optimizations.cc496 static_cast<LIR*>(arena_->Alloc(sizeof(LIR), kArenaAllocLIR)); in ApplyLoadHoisting()
Dralloc_util.cc1215 (arena_->Alloc(num_regs * sizeof(promotion_map_[0]), kArenaAllocRegAlloc)); in DoPromotion()
1234 static_cast<RefCounts*>(arena_->Alloc(sizeof(RefCounts) * core_reg_count_size, in DoPromotion()
1237 static_cast<RefCounts *>(arena_->Alloc(sizeof(RefCounts) * fp_reg_count_size, in DoPromotion()

12