/art/compiler/optimizing/ |
D | stack_map_test.cc | 32 const CodeInfo& code_info, in CheckStackMask() argument 35 BitMemoryRegion stack_mask = code_info.GetStackMaskOf(stack_map); in CheckStackMask() 73 CodeInfo code_info(memory.data()); in TEST() local 74 ASSERT_EQ(1u, code_info.GetNumberOfStackMaps()); in TEST() 76 uint32_t number_of_catalog_entries = code_info.GetNumberOfLocationCatalogEntries(); in TEST() 79 StackMap stack_map = code_info.GetStackMapAt(0); in TEST() 80 ASSERT_TRUE(stack_map.Equals(code_info.GetStackMapForDexPc(0))); in TEST() 81 ASSERT_TRUE(stack_map.Equals(code_info.GetStackMapForNativePcOffset(64 * kPcAlign))); in TEST() 84 ASSERT_EQ(0x3u, code_info.GetRegisterMaskOf(stack_map)); in TEST() 86 ASSERT_TRUE(CheckStackMask(code_info, stack_map, sp_mask)); in TEST() [all …]
|
D | stack_map_stream.cc | 70 dchecks_.emplace_back([=](const CodeInfo& code_info) { in BeginMethod() argument 71 DCHECK_EQ(code_info.packed_frame_size_, frame_size_in_bytes / kStackAlignment); in BeginMethod() 72 DCHECK_EQ(code_info.core_spill_mask_, core_spill_mask); in BeginMethod() 73 DCHECK_EQ(code_info.fp_spill_mask_, fp_spill_mask); in BeginMethod() 74 DCHECK_EQ(code_info.number_of_dex_registers_, num_dex_registers); in BeginMethod() 101 dchecks_.emplace_back([=](const CodeInfo& code_info) { in EndMethod() argument 102 CHECK_EQ(code_info.code_size_, code_size); in EndMethod() 156 dchecks_.emplace_back([=](const CodeInfo& code_info) { in BeginStackMapEntry() argument 160 StackMap stack_map = code_info.GetStackMapForNativePcOffset(final_native_pc_offset, in BeginStackMapEntry() 164 StackMap stack_map = code_info.GetCatchStackMapForDexPc( in BeginStackMapEntry() [all …]
|
D | code_generator.cc | 1039 const CodeInfo& code_info, in CheckCovers() argument 1045 DCHECK(code_info.GetOsrStackMapForDexPc(dex_pc).IsValid()); in CheckCovers() 1055 const CodeInfo& code_info, in CheckLoopEntriesCanBeUsedForOsr() argument 1078 CheckCovers(target, graph, code_info, loop_headers, &covered); in CheckLoopEntriesCanBeUsedForOsr() 1088 CheckCovers(target, graph, code_info, loop_headers, &covered); in CheckLoopEntriesCanBeUsedForOsr()
|
D | optimizing_compiler.cc | 1344 info.code_info = nullptr; in JitCompile() 1451 info.code_info = stack_map.size() == 0 ? nullptr : stack_map.data(); in JitCompile()
|
/art/dex2oat/linker/ |
D | code_info_table_deduper.cc | 59 CodeInfo code_info; in Dedupe() local 60 CodeInfo::ForEachHeaderField([&code_info, &header](size_t i, auto member_pointer) { in Dedupe() 61 code_info.*member_pointer = header[i]; in Dedupe() 63 DCHECK(!code_info.HasDedupedBitTables()); // Input `CodeInfo` has no deduped tables. in Dedupe() 67 DCHECK(!code_info.IsBitTableDeduped(i)); in Dedupe() 68 if (LIKELY(code_info.HasBitTable(i))) { in Dedupe() 69 auto& table = code_info.*member_pointer; in Dedupe() 83 if (LIKELY(code_info.HasBitTable(i))) { in Dedupe() 93 code_info.SetBitTableDeduped(i); // Mark as deduped before we write header. in Dedupe() 100 if (code_info.HasDedupedBitTables()) { in Dedupe() [all …]
|
D | code_info_table_deduper_test.cc | 59 CodeInfo code_info(out.data() + deduped); in TEST() local 60 ASSERT_EQ(1u, code_info.GetNumberOfStackMaps()); in TEST() 62 StackMap stack_map = code_info.GetStackMapAt(0); in TEST() 63 ASSERT_TRUE(stack_map.Equals(code_info.GetStackMapForDexPc(0))); in TEST() 64 ASSERT_TRUE(stack_map.Equals(code_info.GetStackMapForNativePcOffset(64 * kPcAlign))); in TEST() 69 DexRegisterMap dex_register_map = code_info.GetDexRegisterMapOf(stack_map); in TEST()
|
D | code_info_table_deduper.h | 43 size_t Dedupe(const uint8_t* code_info);
|
/art/runtime/oat/ |
D | stack_map.cc | 65 CodeInfo code_info(header->GetOptimizedCodeInfoPtr()); in DecodeGcMasksOnly() local 67 copy.stack_maps_ = code_info.stack_maps_; in DecodeGcMasksOnly() 68 copy.register_masks_ = code_info.register_masks_; in DecodeGcMasksOnly() 69 copy.stack_masks_ = code_info.stack_masks_; in DecodeGcMasksOnly() 74 CodeInfo code_info(header->GetOptimizedCodeInfoPtr()); in DecodeInlineInfoOnly() local 76 copy.number_of_dex_registers_ = code_info.number_of_dex_registers_; in DecodeInlineInfoOnly() 77 copy.stack_maps_ = code_info.stack_maps_; in DecodeInlineInfoOnly() 78 copy.inline_infos_ = code_info.inline_infos_; in DecodeInlineInfoOnly() 79 copy.method_infos_ = code_info.method_infos_; in DecodeInlineInfoOnly() 171 CodeInfo code_info(code_info_data, &num_bits, [&](size_t i, auto* table, BitMemoryRegion region) { in CollectSizeStats() local [all …]
|
D | oat_quick_method_header.cc | 48 CodeInfo code_info = CodeInfo::DecodeInlineInfoOnly(this); in ToDexPc() local 49 StackMap stack_map = code_info.GetStackMapForNativePcOffset(sought_offset); in ToDexPc() 73 CodeInfo code_info = CodeInfo::DecodeInlineInfoOnly(this); in ToNativeQuickPc() local 75 StackMap stack_map = code_info.GetStackMapForDexPc(dex_pc); in ToNativeQuickPc() 99 CodeInfo code_info = CodeInfo::DecodeInlineInfoOnly(this); in ToNativeQuickPcForCatchHandlers() local 101 StackMap stack_map = code_info.GetCatchStackMapForDexPc(dex_pc_list); in ToNativeQuickPcForCatchHandlers()
|
D | stack_map.h | 171 const CodeInfo& code_info, 303 CodeInfo code_info = DecodeHeaderOnly(code_info_data); in DecodeFrameInfo() local 304 return QuickMethodFrameInfo(code_info.packed_frame_size_ * kStackAlignment, in DecodeFrameInfo() 305 code_info.core_spill_mask_, in DecodeFrameInfo() 306 code_info.fp_spill_mask_); in DecodeFrameInfo() 310 CodeInfo code_info; in DecodeHeaderOnly() local 313 ForEachHeaderField([&code_info, &header](size_t i, auto member_pointer) { in DecodeHeaderOnly() 314 code_info.*member_pointer = header[i]; in DecodeHeaderOnly() 316 return code_info; in DecodeHeaderOnly() 482 EXPORT static void CollectSizeStats(const uint8_t* code_info, /*out*/ Stats& parent);
|
/art/runtime/ |
D | check_reference_map_visitor.h | 82 CodeInfo code_info(GetCurrentOatQuickMethodHeader()); in CheckOptimizedMethod() 83 StackMap stack_map = code_info.GetStackMapForNativePcOffset(native_pc_offset); in CheckOptimizedMethod() 84 if (search_for_valid_stack_map && !code_info.GetStackMaskOf(stack_map).IsValid()) { in CheckOptimizedMethod() 85 for (StackMap map : code_info.GetStackMaps()) { in CheckOptimizedMethod() 86 if (map.GetDexPc() == dex_pc && code_info.GetStackMaskOf(map).IsValid()) { in CheckOptimizedMethod() 100 DexRegisterMap dex_register_map = code_info.GetDexRegisterMapOf(stack_map); in CheckOptimizedMethod() 102 uint32_t register_mask = code_info.GetRegisterMaskOf(stack_map); in CheckOptimizedMethod() 103 BitMemoryRegion stack_mask = code_info.GetStackMaskOf(stack_map); in CheckOptimizedMethod()
|
D | quick_exception_handler.cc | 302 CodeInfo code_info(handler_method_header_); in SetCatchEnvironmentForOptimizedHandler() local 307 StackMap catch_stack_map = code_info.GetStackMapAt(GetCatchStackMapRow()); in SetCatchEnvironmentForOptimizedHandler() 309 DCHECK_EQ(catch_stack_map.Row(), code_info.GetCatchStackMapForDexPc(dex_pc_list).Row()); in SetCatchEnvironmentForOptimizedHandler() 311 const size_t number_of_registers = stack_visitor->GetNumberOfRegisters(&code_info, catch_depth); in SetCatchEnvironmentForOptimizedHandler() 313 code_info.GetDexRegisterMapOf(catch_stack_map, /* first= */ 0, number_of_registers); in SetCatchEnvironmentForOptimizedHandler() 321 code_info.GetStackMapForNativePcOffset(stack_visitor->GetNativePcOffset()); in SetCatchEnvironmentForOptimizedHandler() 326 code_info.GetDexRegisterMapOf(throw_stack_map, /* first= */ 0, number_of_registers); in SetCatchEnvironmentForOptimizedHandler() 332 : stack_visitor->GetNumberOfRegisters(&code_info, catch_depth - 1); in SetCatchEnvironmentForOptimizedHandler() 588 CodeInfo code_info(method_header); in HandleOptimizingDeoptimization() local 590 StackMap stack_map = code_info.GetStackMapForNativePcOffset(native_pc_offset); in HandleOptimizingDeoptimization() [all …]
|
D | stack.cc | 110 CodeInfo* code_info = GetCurrentInlineInfo(); in GetMethod() local 112 return GetResolvedMethod(*GetCurrentQuickFrame(), *code_info, current_inline_frames_); in GetMethod() 289 size_t StackVisitor::GetNumberOfRegisters(CodeInfo* code_info, int depth) const { in GetNumberOfRegisters() argument 291 ? code_info->GetNumberOfDexRegisters() in GetNumberOfRegisters() 304 CodeInfo code_info(method_header); in GetVRegFromOptimizedCode() local 307 StackMap stack_map = code_info.GetStackMapForNativePcOffset(native_pc_offset); in GetVRegFromOptimizedCode() 311 ? code_info.GetInlineDexRegisterMapOf(stack_map, current_inline_frames_.back()) in GetVRegFromOptimizedCode() 312 : code_info.GetDexRegisterMapOf(stack_map, in GetVRegFromOptimizedCode() 314 GetNumberOfRegisters(&code_info, InlineDepth())); in GetVRegFromOptimizedCode() 326 BitMemoryRegion stack_mask = code_info.GetStackMaskOf(stack_map); in GetVRegFromOptimizedCode() [all …]
|
D | stack.h | 332 size_t GetNumberOfRegisters(CodeInfo* code_info, int depth) const;
|
D | thread.cc | 4244 CodeInfo code_info = kPrecise in VisitQuickFrameWithVregCallback() local 4247 StackMap map = code_info.GetStackMapForNativePcOffset(native_pc_offset); in VisitQuickFrameWithVregCallback() 4250 T vreg_info(m, code_info, map, visitor_); in VisitQuickFrameWithVregCallback() 4253 BitMemoryRegion stack_mask = code_info.GetStackMaskOf(map); in VisitQuickFrameWithVregCallback() 4268 uint32_t register_mask = code_info.GetRegisterMaskOf(map); in VisitQuickFrameWithVregCallback() 4315 [[maybe_unused]] const CodeInfo& code_info, in VisitQuickFrameNonPrecise() 4346 code_info(_code_info), in VisitQuickFramePrecise() 4347 dex_register_map(code_info.GetDexRegisterMapOf(map)), in VisitQuickFramePrecise() 4394 const CodeInfo& code_info; in VisitQuickFramePrecise() member
|
/art/compiler/debug/ |
D | elf_debug_loc_writer.h | 97 const CodeInfo code_info(method_info->code_info); in GetVariableLocations() 99 for (uint32_t s = 0; s < code_info.GetNumberOfStackMaps(); s++) { in GetVariableLocations() 100 StackMap stack_map = code_info.GetStackMapAt(s); in GetVariableLocations() 122 const StackMap stack_map = code_info.GetStackMapAt(stack_map_index); in GetVariableLocations() 183 if (method_info->code_info == nullptr || dex_register_maps.empty()) { in WriteDebugLocEntry()
|
D | elf_debug_line_writer.h | 95 if (mi->code_info != nullptr) { in WriteCompilationUnit() 97 const CodeInfo code_info(mi->code_info); in WriteCompilationUnit() 98 pc2dex_map.reserve(code_info.GetNumberOfStackMaps()); in WriteCompilationUnit() 99 for (StackMap stack_map : code_info.GetStackMaps()) { in WriteCompilationUnit()
|
D | method_debug_info.h | 45 const uint8_t* code_info; member
|
D | elf_debug_info_writer.h | 192 std::unique_ptr<const CodeInfo> code_info; in Write() local 194 if (accessor.HasCodeItem() && mi->code_info != nullptr) { in Write() 195 code_info.reset(new CodeInfo(mi->code_info)); in Write() 196 for (StackMap stack_map : code_info->GetStackMaps()) { in Write() 197 dex_reg_maps.push_back(code_info->GetDexRegisterMapOf(stack_map)); in Write()
|
/art/runtime/entrypoints/jni/ |
D | jni_entrypoints.cc | 167 CodeInfo code_info = CodeInfo::DecodeInlineInfoOnly(current_code); in artCriticalNativeFrameSize() local 168 StackMap stack_map = code_info.GetStackMapForNativePcOffset(native_pc_offset); in artCriticalNativeFrameSize() 170 BitTableRange<InlineInfo> inline_infos = code_info.GetInlineInfosOf(stack_map); in artCriticalNativeFrameSize() 172 inline_infos.empty() ? method : GetResolvedMethod(method, code_info, inline_infos); in artCriticalNativeFrameSize()
|
/art/runtime/entrypoints/ |
D | entrypoint_utils.cc | 224 CodeInfo code_info = CodeInfo::DecodeInlineInfoOnly(current_code); in DoGetCalleeSaveMethodCallerAndDexPc() local 225 StackMap stack_map = code_info.GetStackMapForNativePcOffset(native_pc_offset); in DoGetCalleeSaveMethodCallerAndDexPc() 227 BitTableRange<InlineInfo> inline_infos = code_info.GetInlineInfosOf(stack_map); in DoGetCalleeSaveMethodCallerAndDexPc() 229 caller = GetResolvedMethod(outer_method, code_info, inline_infos); in DoGetCalleeSaveMethodCallerAndDexPc()
|
D | entrypoint_utils-inl.h | 104 const CodeInfo& code_info, in GetResolvedMethod() argument 121 uint32_t method_index = code_info.GetMethodIndexOf(inline_info); in GetResolvedMethod() 136 MethodInfo method_info = code_info.GetMethodInfoOf(inline_info); in GetResolvedMethod()
|
/art/oatdump/ |
D | oatdump.cc | 326 info.code_info = info.is_optimized ? method_header->GetOptimizedCodeInfoPtr() : nullptr; in WalkOatMethod() 1363 CodeInfo code_info(raw_code_info); in DumpVmapData() local 1366 DumpCodeInfo(vios, code_info, oat_method); in DumpVmapData() 1375 const CodeInfo& code_info, in DumpCodeInfo() argument 1377 code_info.Dump(vios, in DumpCodeInfo() 1537 CodeInfo code_info(oat_method.GetVmapTable()); in DumpCode() local 1539 code_info.CollectSizeStats(oat_method.GetVmapTable(), stats_["CodeInfo"]); in DumpCode() 1542 for (const StackMap& it : code_info.GetStackMaps()) { in DumpCode() 1554 stack_map.Dump(vios, code_info, oat_method.GetCodeOffset(), instruction_set_); in DumpCode()
|
/art/runtime/jit/ |
D | jit.cc | 378 CodeInfo code_info(osr_method); in PrepareForOsr() local 381 StackMap stack_map = code_info.GetOsrStackMapForDexPc(dex_pc); in PrepareForOsr() 390 DexRegisterMap vreg_map = code_info.GetDexRegisterMapOf(stack_map); in PrepareForOsr()
|
/art/runtime/entrypoints/quick/ |
D | quick_trampoline_entrypoints.cc | 1085 CodeInfo code_info(current_code); in DumpB74410240DebugData() local 1086 StackMap stack_map = code_info.GetStackMapForNativePcOffset(native_pc_offset); in DumpB74410240DebugData() 1102 BitTableRange<InlineInfo> inline_infos = code_info.GetInlineInfosOf(stack_map); in DumpB74410240DebugData() 1110 uint32_t method_index = code_info.GetMethodIndexOf(inline_info); in DumpB74410240DebugData()
|