/art/compiler/optimizing/ |
D | builder.cc | 76 HLocal* local = new (arena_) HLocal(i); in InitializeLocals() 96 new (arena_) HParameterValue(parameter_index++, Primitive::kPrimNot); in InitializeParameters() 99 entry_block_->AddInstruction(new (arena_) HStoreLocal(local, parameter)); in InitializeParameters() 114 new (arena_) HParameterValue(parameter_index++, Primitive::GetType(shorty[pos - 1])); in InitializeParameters() 119 entry_block_->AddInstruction(new (arena_) HStoreLocal(local, parameter)); in InitializeParameters() 143 T* comparison = new (arena_) T(first, second); in If_22t() 145 HInstruction* ifinst = new (arena_) HIf(comparison); in If_22t() 159 T* comparison = new (arena_) T(value, GetIntConstant(0)); in If_21t() 161 HInstruction* ifinst = new (arena_) HIf(comparison); in If_21t() 181 graph_ = new (arena_) HGraph(arena_); in BuildGraph() [all …]
|
D | nodes.cc | 29 ArenaBitVector visiting(arena_, blocks_.Size(), false); in FindBackEdges() 70 ArenaBitVector visited(arena_, blocks_.Size(), false); in BuildDominatorTree() 87 GrowableArray<size_t> visits(arena_, blocks_.Size()); in BuildDominatorTree() 96 ArenaBitVector visited(arena_, blocks_.Size(), false); in FindCommonDominator() 143 HBasicBlock* new_block = new (arena_) HBasicBlock(this); in SplitCriticalEdge() 145 new_block->AddInstruction(new (arena_) HGoto()); in SplitCriticalEdge() 165 HBasicBlock* new_back_edge = new (arena_) HBasicBlock(this); in SimplifyLoop() 167 new_back_edge->AddInstruction(new (arena_) HGoto()); in SimplifyLoop() 182 HBasicBlock* pre_header = new (arena_) HBasicBlock(this); in SimplifyLoop() 184 pre_header->AddInstruction(new (arena_) HGoto()); in SimplifyLoop() [all …]
|
D | builder.h | 38 : arena_(arena), in arena_() function 110 ArenaAllocator* const arena_; variable
|
D | nodes.h | 74 : arena_(arena), in HGraph() 83 ArenaAllocator* GetArena() const { return arena_; } in GetArena() 161 ArenaAllocator* const arena_; variable
|
/art/compiler/dex/ |
D | ssa_transformation.cc | 97 dfs_order_ = new (arena_) GrowableArray<BasicBlockId>(arena_, GetNumBlocks(), in ComputeDFSOrders() 106 dfs_post_order_ = new (arena_) GrowableArray<BasicBlockId>(arena_, GetNumBlocks(), in ComputeDFSOrders() 152 (arena_->Alloc(sizeof(ArenaBitVector *) * num_registers, in ComputeDefBlockMatrix() 159 new (arena_) ArenaBitVector(arena_, GetNumBlocks(), false, kBitMapBMatrix); in ComputeDefBlockMatrix() 185 new (arena_) GrowableArray<BasicBlockId>(arena_, num_reachable_blocks_, in ComputeDomPostOrderTraversal() 276 bb->dominators = new (arena_) ArenaBitVector(arena_, num_total_blocks, in InitializeDominationInfo() 278 bb->i_dominated = new (arena_) ArenaBitVector(arena_, num_total_blocks, in InitializeDominationInfo() 280 bb->dom_frontier = new (arena_) ArenaBitVector(arena_, num_total_blocks, in InitializeDominationInfo() 392 i_dom_list_ = static_cast<int*>(arena_->Alloc(sizeof(int) * num_reachable_blocks, in ComputeDominators() 562 static_cast<BasicBlockId*>(arena_->Alloc(sizeof(BasicBlockId) * num_uses, in InsertPhiNodeOperands()
|
D | backend.h | 51 explicit Backend(ArenaAllocator* arena) : arena_(arena) {} in Backend() 52 ArenaAllocator* const arena_; variable
|
D | mir_dataflow.cc | 933 new (arena_) ArenaBitVector(arena_, cu_->num_dalvik_registers, false, kBitMapUse); in FindLocalLiveIn() 935 new (arena_) ArenaBitVector(arena_, cu_->num_dalvik_registers, false, kBitMapDef); in FindLocalLiveIn() 937 new (arena_) ArenaBitVector(arena_, cu_->num_dalvik_registers, false, kBitMapLiveIn); in FindLocalLiveIn() 1021 … mir->ssa_rep->uses = static_cast<int*>(arena_->Alloc(sizeof(int) * num_uses, kArenaAllocDFInfo)); in AllocateSSAUseData() 1023 …mir->ssa_rep->fp_use = static_cast<bool*>(arena_->Alloc(sizeof(bool) * num_uses, kArenaAllocDFInfo… in AllocateSSAUseData() 1031 mir->ssa_rep->defs = static_cast<int*>(arena_->Alloc(sizeof(int) * num_defs, in AllocateSSADefData() 1033 mir->ssa_rep->fp_def = static_cast<bool*>(arena_->Alloc(sizeof(bool) * num_defs, in AllocateSSADefData() 1080 static_cast<struct SSARepresentation *>(arena_->Alloc(sizeof(SSARepresentation), in DoSSAConversion() 1191 static_cast<int*>(arena_->Alloc(sizeof(int) * cu_->num_dalvik_registers, in DoSSAConversion() 1203 ssa_base_vregs_ = new (arena_) GrowableArray<int>(arena_, num_dalvik_reg + GetDefCount() + 128, in CompilerInitializeSSAConversion() [all …]
|
D | mir_graph.cc | 116 arena_(arena), in MIRGraph() 128 try_block_addr_ = new (arena_) ArenaBitVector(arena_, 0, true /* expandable */); in MIRGraph() 517 new (arena_) GrowableArray<SuccessorBlockInfo*>(arena_, size, kGrowableArraySuccessorBlocks); in ProcessCanSwitch() 523 static_cast<SuccessorBlockInfo*>(arena_->Alloc(sizeof(SuccessorBlockInfo), in ProcessCanSwitch() 570 cur_block->successor_blocks = new (arena_) GrowableArray<SuccessorBlockInfo*>( in ProcessCanThrow() 571 arena_, 2, kGrowableArraySuccessorBlocks); in ProcessCanThrow() 578 (arena_->Alloc(sizeof(SuccessorBlockInfo), kArenaAllocSuccessor)); in ProcessCanThrow() 1340 ret = static_cast<char*>(arena_->Alloc(length, kArenaAllocDFInfo)); in GetDalvikDisassembly() 1454 CallInfo* info = static_cast<CallInfo*>(arena_->Alloc(sizeof(CallInfo), in NewMemCallInfo() 1465 (arena_->Alloc(sizeof(RegLocation) * info->num_arg_words, kArenaAllocMisc)); in NewMemCallInfo() [all …]
|
D | mir_optimization.cc | 257 CompilerTemp *compiler_temp = static_cast<CompilerTemp *>(arena_->Alloc(sizeof(CompilerTemp), in GetNewCompilerTemp() 281 static_cast<CompilerTemp *>(arena_->Alloc(sizeof(CompilerTemp), kArenaAllocRegAlloc)); in GetNewCompilerTemp() 506 static_cast<int*>(arena_->Alloc(sizeof(int) * 3, kArenaAllocDFInfo)); in BasicBlockOpt() 515 static_cast<int*>(arena_->Alloc(sizeof(int) * 1, kArenaAllocDFInfo)); in BasicBlockOpt() 517 static_cast<bool*>(arena_->Alloc(sizeof(bool) * 1, kArenaAllocDFInfo)); in BasicBlockOpt() 521 static_cast<bool*>(arena_->Alloc(sizeof(bool) * mir->ssa_rep->num_uses, in BasicBlockOpt() 1304 static_cast<Checkstats*>(arena_->Alloc(sizeof(Checkstats), kArenaAllocDFInfo)); in DumpCheckStats()
|
D | mir_graph.h | 625 return arena_; in GetArena() 629 opcode_count_ = static_cast<int*>(arena_->Alloc(kNumPackedOpcodes * sizeof(int), in EnableOpcodeCounting() 1189 ArenaAllocator* arena_; variable
|
D | vreg_analysis.cc | 441 RegLocation* loc = static_cast<RegLocation*>(arena_->Alloc(max_regs * sizeof(*loc), in InitRegLocations()
|
/art/compiler/utils/ |
D | arena_bit_vector.cc | 25 explicit ArenaBitVectorAllocator(ArenaAlloc* arena) : arena_(arena) {} in ArenaBitVectorAllocator() 29 return arena_->Alloc(size, kArenaAllocGrowableBitMap); in Alloc() 40 ArenaAlloc* arena_; member in art::ArenaBitVectorAllocator
|
D | growable_array.h | 85 : arena_(arena), in arena_() function 89 elem_list_ = static_cast<T*>(arena_->Alloc(sizeof(T) * init_length, in arena_() 103 T* new_array = static_cast<T*>(arena_->Alloc(sizeof(T) * target_length, in Resize() 208 ArenaAllocator* const arena_;
|
/art/compiler/dex/quick/mips/ |
D | call_mips.cc | 71 static_cast<SwitchTable*>(arena_->Alloc(sizeof(SwitchTable), kArenaAllocData)); in GenLargeSparseSwitch() 76 static_cast<LIR**>(arena_->Alloc(elements * sizeof(LIR*), kArenaAllocLIR)); in GenLargeSparseSwitch() 148 static_cast<SwitchTable*>(arena_->Alloc(sizeof(SwitchTable), kArenaAllocData)); in GenLargePackedSwitch() 152 tab_rec->targets = static_cast<LIR**>(arena_->Alloc(size * sizeof(LIR*), in GenLargePackedSwitch() 227 reinterpret_cast<FillArrayData*>(arena_->Alloc(sizeof(FillArrayData), in GenFillArrayData() 343 AddSlowPath(new(arena_)StackOverflowSlowPath(this, branch, spill_count * 4)); in GenEntrySequence()
|
D | target_mips.cc | 433 reg_pool_ = new (arena_) RegisterPool(this, arena_, core_regs, empty_pool /* core64 */, sp_regs, in CompilerInitializeRegAlloc()
|
/art/compiler/dex/quick/arm/ |
D | call_arm.cc | 53 static_cast<SwitchTable*>(arena_->Alloc(sizeof(SwitchTable), kArenaAllocData)); in GenLargeSparseSwitch() 57 tab_rec->targets = static_cast<LIR**>(arena_->Alloc(size * sizeof(LIR*), kArenaAllocLIR)); in GenLargeSparseSwitch() 101 static_cast<SwitchTable*>(arena_->Alloc(sizeof(SwitchTable), kArenaAllocData)); in GenLargePackedSwitch() 106 static_cast<LIR**>(arena_->Alloc(size * sizeof(LIR*), kArenaAllocLIR)); in GenLargePackedSwitch() 154 static_cast<FillArrayData*>(arena_->Alloc(sizeof(FillArrayData), kArenaAllocData)); in GenFillArrayData() 436 AddSlowPath(new(arena_)StackOverflowSlowPath(this, branch, true, spill_size)); in GenEntrySequence() 451 AddSlowPath(new(arena_)StackOverflowSlowPath(this, branch, false, frame_size_)); in GenEntrySequence()
|
D | target_arm.cc | 570 reg_pool_ = new (arena_) RegisterPool(this, arena_, core_regs, empty_pool /* core64 */, sp_regs, in CompilerInitializeRegAlloc()
|
/art/compiler/dex/quick/arm64/ |
D | call_arm64.cc | 53 static_cast<SwitchTable*>(arena_->Alloc(sizeof(SwitchTable), kArenaAllocData)); in GenLargeSparseSwitch() 57 tab_rec->targets = static_cast<LIR**>(arena_->Alloc(size * sizeof(LIR*), kArenaAllocLIR)); in GenLargeSparseSwitch() 105 static_cast<SwitchTable*>(arena_->Alloc(sizeof(SwitchTable), kArenaAllocData)); in GenLargePackedSwitch() 110 static_cast<LIR**>(arena_->Alloc(size * sizeof(LIR*), kArenaAllocLIR)); in GenLargePackedSwitch() 163 static_cast<FillArrayData*>(arena_->Alloc(sizeof(FillArrayData), kArenaAllocData)); in GenFillArrayData() 398 AddSlowPath(new(arena_)StackOverflowSlowPath(this, branch, frame_size_)); in GenEntrySequence()
|
D | target_arm64.cc | 598 reg_pool_ = new (arena_) RegisterPool(this, arena_, core_regs, core64_regs, sp_regs, dp_regs, in CompilerInitializeRegAlloc()
|
/art/compiler/dex/quick/x86/ |
D | call_x86.cc | 70 static_cast<SwitchTable*>(arena_->Alloc(sizeof(SwitchTable), kArenaAllocData)); in GenLargePackedSwitch() 74 tab_rec->targets = static_cast<LIR**>(arena_->Alloc(size * sizeof(LIR*), in GenLargePackedSwitch() 141 static_cast<FillArrayData*>(arena_->Alloc(sizeof(FillArrayData), kArenaAllocData)); in GenFillArrayData() 280 new(arena_)StackOverflowSlowPath(this, branch, in GenEntrySequence()
|
D | target_x86.cc | 616 reg_pool_ = new (arena_) RegisterPool(this, arena_, core_regs_64, core_regs_64q, sp_regs_64, in CompilerInitializeRegAlloc() 620 reg_pool_ = new (arena_) RegisterPool(this, arena_, core_regs_32, empty_pool, sp_regs_32, in CompilerInitializeRegAlloc() 630 RegisterInfo* info = new (arena_) RegisterInfo(reg, GetRegMaskCommon(reg)); in CompilerInitializeRegAlloc() 2360 LIR* new_value = static_cast<LIR*>(arena_->Alloc(sizeof(LIR), kArenaAllocData)); in AddVectorLiteral()
|
/art/compiler/dex/quick/ |
D | gen_common.cc | 81 AddSlowPath(new (arena_) DivZeroCheckSlowPath(this, branch)); in AddDivZeroCheckSlowPath() 105 AddSlowPath(new (arena_) ArrayBoundsCheckSlowPath(this, branch, index, length)); in GenArrayBoundsCheck() 135 AddSlowPath(new (arena_) ArrayBoundsCheckSlowPath(this, branch, index, length)); in GenArrayBoundsCheck() 154 AddSlowPath(new (arena_) NullCheckSlowPath(this, branch)); in GenNullCheck() 580 AddSlowPath(new (arena_) StaticFieldSlowPath(this, unresolved_branch, uninit_branch, cont, in GenSput() 665 AddSlowPath(new (arena_) StaticFieldSlowPath(this, unresolved_branch, uninit_branch, cont, in GenSget() 872 AddSlowPath(new (arena_) SlowPath(this, branch, cont, type_idx, rl_method, rl_result)); in GenConstClass() 935 AddSlowPath(new (arena_) SlowPath(this, fromfast, cont, r_method, string_idx)); in GenConstString() 1144 AddSlowPath(new (arena_) InitTypeSlowPath(this, slow_path_branch, slow_path_target, in GenInstanceofCallingHelper() 1291 AddSlowPath(new (arena_) SlowPath(this, hop_branch, cont, type_idx, class_reg)); in GenCheckCast() [all …]
|
D | mir_to_lir-inl.h | 46 LIR* insn = static_cast<LIR*>(arena_->Alloc(sizeof(LIR), kArenaAllocLIR)); in RawLIR()
|
D | local_optimizations.cc | 496 static_cast<LIR*>(arena_->Alloc(sizeof(LIR), kArenaAllocLIR)); in ApplyLoadHoisting()
|
D | ralloc_util.cc | 1215 (arena_->Alloc(num_regs * sizeof(promotion_map_[0]), kArenaAllocRegAlloc)); in DoPromotion() 1234 static_cast<RefCounts*>(arena_->Alloc(sizeof(RefCounts) * core_reg_count_size, in DoPromotion() 1237 static_cast<RefCounts *>(arena_->Alloc(sizeof(RefCounts) * fp_reg_count_size, in DoPromotion()
|