1 /*
2 * Copyright (C) 2014 The Android Open Source Project
3 *
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
7 *
8 * http://www.apache.org/licenses/LICENSE-2.0
9 *
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
15 */
16
17 #include "optimizing_compiler.h"
18
19 #include <fstream>
20 #include <memory>
21 #include <stdint.h>
22
23 #ifdef ART_ENABLE_CODEGEN_arm
24 #include "dex_cache_array_fixups_arm.h"
25 #endif
26
27 #ifdef ART_ENABLE_CODEGEN_arm64
28 #include "instruction_simplifier_arm64.h"
29 #endif
30
31 #ifdef ART_ENABLE_CODEGEN_x86
32 #include "pc_relative_fixups_x86.h"
33 #endif
34
35 #include "art_method-inl.h"
36 #include "base/arena_allocator.h"
37 #include "base/arena_containers.h"
38 #include "base/dumpable.h"
39 #include "base/macros.h"
40 #include "base/timing_logger.h"
41 #include "bounds_check_elimination.h"
42 #include "builder.h"
43 #include "code_generator.h"
44 #include "compiled_method.h"
45 #include "compiler.h"
46 #include "constant_folding.h"
47 #include "dead_code_elimination.h"
48 #include "debug/elf_debug_writer.h"
49 #include "debug/method_debug_info.h"
50 #include "dex/quick/dex_file_to_method_inliner_map.h"
51 #include "dex/verification_results.h"
52 #include "dex/verified_method.h"
53 #include "driver/compiler_driver-inl.h"
54 #include "driver/compiler_options.h"
55 #include "driver/dex_compilation_unit.h"
56 #include "elf_writer_quick.h"
57 #include "graph_checker.h"
58 #include "graph_visualizer.h"
59 #include "gvn.h"
60 #include "induction_var_analysis.h"
61 #include "inliner.h"
62 #include "instruction_simplifier.h"
63 #include "instruction_simplifier_arm.h"
64 #include "intrinsics.h"
65 #include "jit/debugger_interface.h"
66 #include "jit/jit.h"
67 #include "jit/jit_code_cache.h"
68 #include "jni/quick/jni_compiler.h"
69 #include "licm.h"
70 #include "load_store_elimination.h"
71 #include "nodes.h"
72 #include "oat_quick_method_header.h"
73 #include "prepare_for_register_allocation.h"
74 #include "reference_type_propagation.h"
75 #include "register_allocator.h"
76 #include "select_generator.h"
77 #include "sharpening.h"
78 #include "side_effects_analysis.h"
79 #include "ssa_builder.h"
80 #include "ssa_liveness_analysis.h"
81 #include "ssa_phi_elimination.h"
82 #include "utils/assembler.h"
83 #include "verifier/method_verifier.h"
84
85 namespace art {
86
87 static constexpr size_t kArenaAllocatorMemoryReportThreshold = 8 * MB;
88
89 /**
90 * Used by the code generator, to allocate the code in a vector.
91 */
92 class CodeVectorAllocator FINAL : public CodeAllocator {
93 public:
CodeVectorAllocator(ArenaAllocator * arena)94 explicit CodeVectorAllocator(ArenaAllocator* arena)
95 : memory_(arena->Adapter(kArenaAllocCodeBuffer)),
96 size_(0) {}
97
Allocate(size_t size)98 virtual uint8_t* Allocate(size_t size) {
99 size_ = size;
100 memory_.resize(size);
101 return &memory_[0];
102 }
103
GetSize() const104 size_t GetSize() const { return size_; }
GetMemory() const105 const ArenaVector<uint8_t>& GetMemory() const { return memory_; }
106
107 private:
108 ArenaVector<uint8_t> memory_;
109 size_t size_;
110
111 DISALLOW_COPY_AND_ASSIGN(CodeVectorAllocator);
112 };
113
114 /**
115 * Filter to apply to the visualizer. Methods whose name contain that filter will
116 * be dumped.
117 */
118 static constexpr const char kStringFilter[] = "";
119
120 class PassScope;
121
122 class PassObserver : public ValueObject {
123 public:
PassObserver(HGraph * graph,CodeGenerator * codegen,std::ostream * visualizer_output,CompilerDriver * compiler_driver)124 PassObserver(HGraph* graph,
125 CodeGenerator* codegen,
126 std::ostream* visualizer_output,
127 CompilerDriver* compiler_driver)
128 : graph_(graph),
129 cached_method_name_(),
130 timing_logger_enabled_(compiler_driver->GetDumpPasses()),
131 timing_logger_(timing_logger_enabled_ ? GetMethodName() : "", true, true),
132 disasm_info_(graph->GetArena()),
133 visualizer_enabled_(!compiler_driver->GetCompilerOptions().GetDumpCfgFileName().empty()),
134 visualizer_(visualizer_output, graph, *codegen),
135 graph_in_bad_state_(false) {
136 if (timing_logger_enabled_ || visualizer_enabled_) {
137 if (!IsVerboseMethod(compiler_driver, GetMethodName())) {
138 timing_logger_enabled_ = visualizer_enabled_ = false;
139 }
140 if (visualizer_enabled_) {
141 visualizer_.PrintHeader(GetMethodName());
142 codegen->SetDisassemblyInformation(&disasm_info_);
143 }
144 }
145 }
146
~PassObserver()147 ~PassObserver() {
148 if (timing_logger_enabled_) {
149 LOG(INFO) << "TIMINGS " << GetMethodName();
150 LOG(INFO) << Dumpable<TimingLogger>(timing_logger_);
151 }
152 }
153
DumpDisassembly() const154 void DumpDisassembly() const {
155 if (visualizer_enabled_) {
156 visualizer_.DumpGraphWithDisassembly();
157 }
158 }
159
SetGraphInBadState()160 void SetGraphInBadState() { graph_in_bad_state_ = true; }
161
GetMethodName()162 const char* GetMethodName() {
163 // PrettyMethod() is expensive, so we delay calling it until we actually have to.
164 if (cached_method_name_.empty()) {
165 cached_method_name_ = PrettyMethod(graph_->GetMethodIdx(), graph_->GetDexFile());
166 }
167 return cached_method_name_.c_str();
168 }
169
170 private:
StartPass(const char * pass_name)171 void StartPass(const char* pass_name) {
172 // Dump graph first, then start timer.
173 if (visualizer_enabled_) {
174 visualizer_.DumpGraph(pass_name, /* is_after_pass */ false, graph_in_bad_state_);
175 }
176 if (timing_logger_enabled_) {
177 timing_logger_.StartTiming(pass_name);
178 }
179 }
180
EndPass(const char * pass_name)181 void EndPass(const char* pass_name) {
182 // Pause timer first, then dump graph.
183 if (timing_logger_enabled_) {
184 timing_logger_.EndTiming();
185 }
186 if (visualizer_enabled_) {
187 visualizer_.DumpGraph(pass_name, /* is_after_pass */ true, graph_in_bad_state_);
188 }
189
190 // Validate the HGraph if running in debug mode.
191 if (kIsDebugBuild) {
192 if (!graph_in_bad_state_) {
193 GraphChecker checker(graph_);
194 checker.Run();
195 if (!checker.IsValid()) {
196 LOG(FATAL) << "Error after " << pass_name << ": " << Dumpable<GraphChecker>(checker);
197 }
198 }
199 }
200 }
201
IsVerboseMethod(CompilerDriver * compiler_driver,const char * method_name)202 static bool IsVerboseMethod(CompilerDriver* compiler_driver, const char* method_name) {
203 // Test an exact match to --verbose-methods. If verbose-methods is set, this overrides an
204 // empty kStringFilter matching all methods.
205 if (compiler_driver->GetCompilerOptions().HasVerboseMethods()) {
206 return compiler_driver->GetCompilerOptions().IsVerboseMethod(method_name);
207 }
208
209 // Test the kStringFilter sub-string. constexpr helper variable to silence unreachable-code
210 // warning when the string is empty.
211 constexpr bool kStringFilterEmpty = arraysize(kStringFilter) <= 1;
212 if (kStringFilterEmpty || strstr(method_name, kStringFilter) != nullptr) {
213 return true;
214 }
215
216 return false;
217 }
218
219 HGraph* const graph_;
220
221 std::string cached_method_name_;
222
223 bool timing_logger_enabled_;
224 TimingLogger timing_logger_;
225
226 DisassemblyInformation disasm_info_;
227
228 bool visualizer_enabled_;
229 HGraphVisualizer visualizer_;
230
231 // Flag to be set by the compiler if the pass failed and the graph is not
232 // expected to validate.
233 bool graph_in_bad_state_;
234
235 friend PassScope;
236
237 DISALLOW_COPY_AND_ASSIGN(PassObserver);
238 };
239
240 class PassScope : public ValueObject {
241 public:
PassScope(const char * pass_name,PassObserver * pass_observer)242 PassScope(const char *pass_name, PassObserver* pass_observer)
243 : pass_name_(pass_name),
244 pass_observer_(pass_observer) {
245 pass_observer_->StartPass(pass_name_);
246 }
247
~PassScope()248 ~PassScope() {
249 pass_observer_->EndPass(pass_name_);
250 }
251
252 private:
253 const char* const pass_name_;
254 PassObserver* const pass_observer_;
255 };
256
257 class OptimizingCompiler FINAL : public Compiler {
258 public:
259 explicit OptimizingCompiler(CompilerDriver* driver);
260 ~OptimizingCompiler();
261
262 bool CanCompileMethod(uint32_t method_idx, const DexFile& dex_file) const OVERRIDE;
263
264 CompiledMethod* Compile(const DexFile::CodeItem* code_item,
265 uint32_t access_flags,
266 InvokeType invoke_type,
267 uint16_t class_def_idx,
268 uint32_t method_idx,
269 jobject class_loader,
270 const DexFile& dex_file,
271 Handle<mirror::DexCache> dex_cache) const OVERRIDE;
272
JniCompile(uint32_t access_flags,uint32_t method_idx,const DexFile & dex_file) const273 CompiledMethod* JniCompile(uint32_t access_flags,
274 uint32_t method_idx,
275 const DexFile& dex_file) const OVERRIDE {
276 return ArtQuickJniCompileMethod(GetCompilerDriver(), access_flags, method_idx, dex_file);
277 }
278
GetEntryPointOf(ArtMethod * method) const279 uintptr_t GetEntryPointOf(ArtMethod* method) const OVERRIDE
280 SHARED_REQUIRES(Locks::mutator_lock_) {
281 return reinterpret_cast<uintptr_t>(method->GetEntryPointFromQuickCompiledCodePtrSize(
282 InstructionSetPointerSize(GetCompilerDriver()->GetInstructionSet())));
283 }
284
285 void Init() OVERRIDE;
286
287 void UnInit() const OVERRIDE;
288
MaybeRecordStat(MethodCompilationStat compilation_stat) const289 void MaybeRecordStat(MethodCompilationStat compilation_stat) const {
290 if (compilation_stats_.get() != nullptr) {
291 compilation_stats_->RecordStat(compilation_stat);
292 }
293 }
294
295 bool JitCompile(Thread* self, jit::JitCodeCache* code_cache, ArtMethod* method, bool osr)
296 OVERRIDE
297 SHARED_REQUIRES(Locks::mutator_lock_);
298
299 private:
300 // Create a 'CompiledMethod' for an optimized graph.
301 CompiledMethod* Emit(ArenaAllocator* arena,
302 CodeVectorAllocator* code_allocator,
303 CodeGenerator* codegen,
304 CompilerDriver* driver,
305 const DexFile::CodeItem* item) const;
306
307 // Try compiling a method and return the code generator used for
308 // compiling it.
309 // This method:
310 // 1) Builds the graph. Returns null if it failed to build it.
311 // 2) Transforms the graph to SSA. Returns null if it failed.
312 // 3) Runs optimizations on the graph, including register allocator.
313 // 4) Generates code with the `code_allocator` provided.
314 CodeGenerator* TryCompile(ArenaAllocator* arena,
315 CodeVectorAllocator* code_allocator,
316 const DexFile::CodeItem* code_item,
317 uint32_t access_flags,
318 InvokeType invoke_type,
319 uint16_t class_def_idx,
320 uint32_t method_idx,
321 jobject class_loader,
322 const DexFile& dex_file,
323 Handle<mirror::DexCache> dex_cache,
324 ArtMethod* method,
325 bool osr) const;
326
327 std::unique_ptr<OptimizingCompilerStats> compilation_stats_;
328
329 std::unique_ptr<std::ostream> visualizer_output_;
330
331 DISALLOW_COPY_AND_ASSIGN(OptimizingCompiler);
332 };
333
334 static const int kMaximumCompilationTimeBeforeWarning = 100; /* ms */
335
OptimizingCompiler(CompilerDriver * driver)336 OptimizingCompiler::OptimizingCompiler(CompilerDriver* driver)
337 : Compiler(driver, kMaximumCompilationTimeBeforeWarning) {}
338
Init()339 void OptimizingCompiler::Init() {
340 // Enable C1visualizer output. Must be done in Init() because the compiler
341 // driver is not fully initialized when passed to the compiler's constructor.
342 CompilerDriver* driver = GetCompilerDriver();
343 const std::string cfg_file_name = driver->GetCompilerOptions().GetDumpCfgFileName();
344 if (!cfg_file_name.empty()) {
345 CHECK_EQ(driver->GetThreadCount(), 1U)
346 << "Graph visualizer requires the compiler to run single-threaded. "
347 << "Invoke the compiler with '-j1'.";
348 std::ios_base::openmode cfg_file_mode =
349 driver->GetCompilerOptions().GetDumpCfgAppend() ? std::ofstream::app : std::ofstream::out;
350 visualizer_output_.reset(new std::ofstream(cfg_file_name, cfg_file_mode));
351 }
352 if (driver->GetDumpStats()) {
353 compilation_stats_.reset(new OptimizingCompilerStats());
354 }
355 }
356
UnInit() const357 void OptimizingCompiler::UnInit() const {
358 }
359
~OptimizingCompiler()360 OptimizingCompiler::~OptimizingCompiler() {
361 if (compilation_stats_.get() != nullptr) {
362 compilation_stats_->Log();
363 }
364 }
365
CanCompileMethod(uint32_t method_idx ATTRIBUTE_UNUSED,const DexFile & dex_file ATTRIBUTE_UNUSED) const366 bool OptimizingCompiler::CanCompileMethod(uint32_t method_idx ATTRIBUTE_UNUSED,
367 const DexFile& dex_file ATTRIBUTE_UNUSED) const {
368 return true;
369 }
370
IsInstructionSetSupported(InstructionSet instruction_set)371 static bool IsInstructionSetSupported(InstructionSet instruction_set) {
372 return (instruction_set == kArm && !kArm32QuickCodeUseSoftFloat)
373 || instruction_set == kArm64
374 || (instruction_set == kThumb2 && !kArm32QuickCodeUseSoftFloat)
375 || instruction_set == kMips
376 || instruction_set == kMips64
377 || instruction_set == kX86
378 || instruction_set == kX86_64;
379 }
380
381 // Read barrier are supported on ARM, ARM64, x86 and x86-64 at the moment.
382 // TODO: Add support for other architectures and remove this function
InstructionSetSupportsReadBarrier(InstructionSet instruction_set)383 static bool InstructionSetSupportsReadBarrier(InstructionSet instruction_set) {
384 return instruction_set == kArm64
385 || instruction_set == kThumb2
386 || instruction_set == kX86
387 || instruction_set == kX86_64;
388 }
389
RunOptimizations(HOptimization * optimizations[],size_t length,PassObserver * pass_observer)390 static void RunOptimizations(HOptimization* optimizations[],
391 size_t length,
392 PassObserver* pass_observer) {
393 for (size_t i = 0; i < length; ++i) {
394 PassScope scope(optimizations[i]->GetPassName(), pass_observer);
395 optimizations[i]->Run();
396 }
397 }
398
MaybeRunInliner(HGraph * graph,CodeGenerator * codegen,CompilerDriver * driver,OptimizingCompilerStats * stats,const DexCompilationUnit & dex_compilation_unit,PassObserver * pass_observer,StackHandleScopeCollection * handles)399 static void MaybeRunInliner(HGraph* graph,
400 CodeGenerator* codegen,
401 CompilerDriver* driver,
402 OptimizingCompilerStats* stats,
403 const DexCompilationUnit& dex_compilation_unit,
404 PassObserver* pass_observer,
405 StackHandleScopeCollection* handles) {
406 const CompilerOptions& compiler_options = driver->GetCompilerOptions();
407 bool should_inline = (compiler_options.GetInlineDepthLimit() > 0)
408 && (compiler_options.GetInlineMaxCodeUnits() > 0);
409 if (!should_inline) {
410 return;
411 }
412 size_t number_of_dex_registers = dex_compilation_unit.GetCodeItem()->registers_size_;
413 HInliner* inliner = new (graph->GetArena()) HInliner(
414 graph,
415 graph,
416 codegen,
417 dex_compilation_unit,
418 dex_compilation_unit,
419 driver,
420 handles,
421 stats,
422 number_of_dex_registers,
423 /* depth */ 0);
424 HOptimization* optimizations[] = { inliner };
425
426 RunOptimizations(optimizations, arraysize(optimizations), pass_observer);
427 }
428
RunArchOptimizations(InstructionSet instruction_set,HGraph * graph,CodeGenerator * codegen,OptimizingCompilerStats * stats,PassObserver * pass_observer)429 static void RunArchOptimizations(InstructionSet instruction_set,
430 HGraph* graph,
431 CodeGenerator* codegen,
432 OptimizingCompilerStats* stats,
433 PassObserver* pass_observer) {
434 ArenaAllocator* arena = graph->GetArena();
435 switch (instruction_set) {
436 #ifdef ART_ENABLE_CODEGEN_arm
437 case kThumb2:
438 case kArm: {
439 arm::DexCacheArrayFixups* fixups = new (arena) arm::DexCacheArrayFixups(graph, stats);
440 arm::InstructionSimplifierArm* simplifier =
441 new (arena) arm::InstructionSimplifierArm(graph, stats);
442 HOptimization* arm_optimizations[] = {
443 simplifier,
444 fixups
445 };
446 RunOptimizations(arm_optimizations, arraysize(arm_optimizations), pass_observer);
447 break;
448 }
449 #endif
450 #ifdef ART_ENABLE_CODEGEN_arm64
451 case kArm64: {
452 arm64::InstructionSimplifierArm64* simplifier =
453 new (arena) arm64::InstructionSimplifierArm64(graph, stats);
454 SideEffectsAnalysis* side_effects = new (arena) SideEffectsAnalysis(graph);
455 GVNOptimization* gvn = new (arena) GVNOptimization(graph, *side_effects, "GVN_after_arch");
456 HOptimization* arm64_optimizations[] = {
457 simplifier,
458 side_effects,
459 gvn
460 };
461 RunOptimizations(arm64_optimizations, arraysize(arm64_optimizations), pass_observer);
462 break;
463 }
464 #endif
465 #ifdef ART_ENABLE_CODEGEN_x86
466 case kX86: {
467 x86::PcRelativeFixups* pc_relative_fixups =
468 new (arena) x86::PcRelativeFixups(graph, codegen, stats);
469 HOptimization* x86_optimizations[] = {
470 pc_relative_fixups
471 };
472 RunOptimizations(x86_optimizations, arraysize(x86_optimizations), pass_observer);
473 break;
474 }
475 #endif
476 default:
477 break;
478 }
479 }
480
481 NO_INLINE // Avoid increasing caller's frame size by large stack-allocated objects.
AllocateRegisters(HGraph * graph,CodeGenerator * codegen,PassObserver * pass_observer)482 static void AllocateRegisters(HGraph* graph,
483 CodeGenerator* codegen,
484 PassObserver* pass_observer) {
485 {
486 PassScope scope(PrepareForRegisterAllocation::kPrepareForRegisterAllocationPassName,
487 pass_observer);
488 PrepareForRegisterAllocation(graph).Run();
489 }
490 SsaLivenessAnalysis liveness(graph, codegen);
491 {
492 PassScope scope(SsaLivenessAnalysis::kLivenessPassName, pass_observer);
493 liveness.Analyze();
494 }
495 {
496 PassScope scope(RegisterAllocator::kRegisterAllocatorPassName, pass_observer);
497 RegisterAllocator(graph->GetArena(), codegen, liveness).AllocateRegisters();
498 }
499 }
500
RunOptimizations(HGraph * graph,CodeGenerator * codegen,CompilerDriver * driver,OptimizingCompilerStats * stats,const DexCompilationUnit & dex_compilation_unit,PassObserver * pass_observer,StackHandleScopeCollection * handles)501 static void RunOptimizations(HGraph* graph,
502 CodeGenerator* codegen,
503 CompilerDriver* driver,
504 OptimizingCompilerStats* stats,
505 const DexCompilationUnit& dex_compilation_unit,
506 PassObserver* pass_observer,
507 StackHandleScopeCollection* handles) {
508 ArenaAllocator* arena = graph->GetArena();
509 HDeadCodeElimination* dce1 = new (arena) HDeadCodeElimination(
510 graph, stats, HDeadCodeElimination::kInitialDeadCodeEliminationPassName);
511 HDeadCodeElimination* dce2 = new (arena) HDeadCodeElimination(
512 graph, stats, HDeadCodeElimination::kFinalDeadCodeEliminationPassName);
513 HConstantFolding* fold1 = new (arena) HConstantFolding(graph);
514 InstructionSimplifier* simplify1 = new (arena) InstructionSimplifier(graph, stats);
515 HSelectGenerator* select_generator = new (arena) HSelectGenerator(graph, stats);
516 HConstantFolding* fold2 = new (arena) HConstantFolding(graph, "constant_folding_after_inlining");
517 HConstantFolding* fold3 = new (arena) HConstantFolding(graph, "constant_folding_after_bce");
518 SideEffectsAnalysis* side_effects = new (arena) SideEffectsAnalysis(graph);
519 GVNOptimization* gvn = new (arena) GVNOptimization(graph, *side_effects);
520 LICM* licm = new (arena) LICM(graph, *side_effects, stats);
521 LoadStoreElimination* lse = new (arena) LoadStoreElimination(graph, *side_effects);
522 HInductionVarAnalysis* induction = new (arena) HInductionVarAnalysis(graph);
523 BoundsCheckElimination* bce = new (arena) BoundsCheckElimination(graph, *side_effects, induction);
524 HSharpening* sharpening = new (arena) HSharpening(graph, codegen, dex_compilation_unit, driver);
525 InstructionSimplifier* simplify2 = new (arena) InstructionSimplifier(
526 graph, stats, "instruction_simplifier_after_bce");
527 InstructionSimplifier* simplify3 = new (arena) InstructionSimplifier(
528 graph, stats, "instruction_simplifier_before_codegen");
529 IntrinsicsRecognizer* intrinsics = new (arena) IntrinsicsRecognizer(graph, driver, stats);
530
531 HOptimization* optimizations1[] = {
532 intrinsics,
533 sharpening,
534 fold1,
535 simplify1,
536 dce1,
537 };
538 RunOptimizations(optimizations1, arraysize(optimizations1), pass_observer);
539
540 MaybeRunInliner(graph, codegen, driver, stats, dex_compilation_unit, pass_observer, handles);
541
542 HOptimization* optimizations2[] = {
543 // SelectGenerator depends on the InstructionSimplifier removing
544 // redundant suspend checks to recognize empty blocks.
545 select_generator,
546 fold2, // TODO: if we don't inline we can also skip fold2.
547 side_effects,
548 gvn,
549 licm,
550 induction,
551 bce,
552 fold3, // evaluates code generated by dynamic bce
553 simplify2,
554 lse,
555 dce2,
556 // The codegen has a few assumptions that only the instruction simplifier
557 // can satisfy. For example, the code generator does not expect to see a
558 // HTypeConversion from a type to the same type.
559 simplify3,
560 };
561 RunOptimizations(optimizations2, arraysize(optimizations2), pass_observer);
562
563 RunArchOptimizations(driver->GetInstructionSet(), graph, codegen, stats, pass_observer);
564 AllocateRegisters(graph, codegen, pass_observer);
565 }
566
EmitAndSortLinkerPatches(CodeGenerator * codegen)567 static ArenaVector<LinkerPatch> EmitAndSortLinkerPatches(CodeGenerator* codegen) {
568 ArenaVector<LinkerPatch> linker_patches(codegen->GetGraph()->GetArena()->Adapter());
569 codegen->EmitLinkerPatches(&linker_patches);
570
571 // Sort patches by literal offset. Required for .oat_patches encoding.
572 std::sort(linker_patches.begin(), linker_patches.end(),
573 [](const LinkerPatch& lhs, const LinkerPatch& rhs) {
574 return lhs.LiteralOffset() < rhs.LiteralOffset();
575 });
576
577 return linker_patches;
578 }
579
Emit(ArenaAllocator * arena,CodeVectorAllocator * code_allocator,CodeGenerator * codegen,CompilerDriver * compiler_driver,const DexFile::CodeItem * code_item) const580 CompiledMethod* OptimizingCompiler::Emit(ArenaAllocator* arena,
581 CodeVectorAllocator* code_allocator,
582 CodeGenerator* codegen,
583 CompilerDriver* compiler_driver,
584 const DexFile::CodeItem* code_item) const {
585 ArenaVector<LinkerPatch> linker_patches = EmitAndSortLinkerPatches(codegen);
586 ArenaVector<uint8_t> stack_map(arena->Adapter(kArenaAllocStackMaps));
587 stack_map.resize(codegen->ComputeStackMapsSize());
588 codegen->BuildStackMaps(MemoryRegion(stack_map.data(), stack_map.size()), *code_item);
589
590 CompiledMethod* compiled_method = CompiledMethod::SwapAllocCompiledMethod(
591 compiler_driver,
592 codegen->GetInstructionSet(),
593 ArrayRef<const uint8_t>(code_allocator->GetMemory()),
594 // Follow Quick's behavior and set the frame size to zero if it is
595 // considered "empty" (see the definition of
596 // art::CodeGenerator::HasEmptyFrame).
597 codegen->HasEmptyFrame() ? 0 : codegen->GetFrameSize(),
598 codegen->GetCoreSpillMask(),
599 codegen->GetFpuSpillMask(),
600 ArrayRef<const SrcMapElem>(),
601 ArrayRef<const uint8_t>(stack_map),
602 ArrayRef<const uint8_t>(*codegen->GetAssembler()->cfi().data()),
603 ArrayRef<const LinkerPatch>(linker_patches));
604
605 return compiled_method;
606 }
607
TryCompile(ArenaAllocator * arena,CodeVectorAllocator * code_allocator,const DexFile::CodeItem * code_item,uint32_t access_flags,InvokeType invoke_type,uint16_t class_def_idx,uint32_t method_idx,jobject class_loader,const DexFile & dex_file,Handle<mirror::DexCache> dex_cache,ArtMethod * method,bool osr) const608 CodeGenerator* OptimizingCompiler::TryCompile(ArenaAllocator* arena,
609 CodeVectorAllocator* code_allocator,
610 const DexFile::CodeItem* code_item,
611 uint32_t access_flags,
612 InvokeType invoke_type,
613 uint16_t class_def_idx,
614 uint32_t method_idx,
615 jobject class_loader,
616 const DexFile& dex_file,
617 Handle<mirror::DexCache> dex_cache,
618 ArtMethod* method,
619 bool osr) const {
620 MaybeRecordStat(MethodCompilationStat::kAttemptCompilation);
621 CompilerDriver* compiler_driver = GetCompilerDriver();
622 InstructionSet instruction_set = compiler_driver->GetInstructionSet();
623
624 // Always use the Thumb-2 assembler: some runtime functionality
625 // (like implicit stack overflow checks) assume Thumb-2.
626 if (instruction_set == kArm) {
627 instruction_set = kThumb2;
628 }
629
630 // Do not attempt to compile on architectures we do not support.
631 if (!IsInstructionSetSupported(instruction_set)) {
632 MaybeRecordStat(MethodCompilationStat::kNotCompiledUnsupportedIsa);
633 return nullptr;
634 }
635
636 // When read barriers are enabled, do not attempt to compile for
637 // instruction sets that have no read barrier support.
638 if (kEmitCompilerReadBarrier && !InstructionSetSupportsReadBarrier(instruction_set)) {
639 return nullptr;
640 }
641
642 if (Compiler::IsPathologicalCase(*code_item, method_idx, dex_file)) {
643 MaybeRecordStat(MethodCompilationStat::kNotCompiledPathological);
644 return nullptr;
645 }
646
647 // Implementation of the space filter: do not compile a code item whose size in
648 // code units is bigger than 128.
649 static constexpr size_t kSpaceFilterOptimizingThreshold = 128;
650 const CompilerOptions& compiler_options = compiler_driver->GetCompilerOptions();
651 if ((compiler_options.GetCompilerFilter() == CompilerFilter::kSpace)
652 && (code_item->insns_size_in_code_units_ > kSpaceFilterOptimizingThreshold)) {
653 MaybeRecordStat(MethodCompilationStat::kNotCompiledSpaceFilter);
654 return nullptr;
655 }
656
657 DexCompilationUnit dex_compilation_unit(
658 class_loader,
659 Runtime::Current()->GetClassLinker(),
660 dex_file,
661 code_item,
662 class_def_idx,
663 method_idx,
664 access_flags,
665 /* verified_method */ nullptr,
666 dex_cache);
667
668 bool requires_barrier = dex_compilation_unit.IsConstructor()
669 && compiler_driver->RequiresConstructorBarrier(Thread::Current(),
670 dex_compilation_unit.GetDexFile(),
671 dex_compilation_unit.GetClassDefIndex());
672
673 HGraph* graph = new (arena) HGraph(
674 arena,
675 dex_file,
676 method_idx,
677 requires_barrier,
678 compiler_driver->GetInstructionSet(),
679 kInvalidInvokeType,
680 compiler_driver->GetCompilerOptions().GetDebuggable(),
681 osr);
682
683 const uint8_t* interpreter_metadata = nullptr;
684 if (method == nullptr) {
685 ScopedObjectAccess soa(Thread::Current());
686 StackHandleScope<1> hs(soa.Self());
687 Handle<mirror::ClassLoader> loader(hs.NewHandle(
688 soa.Decode<mirror::ClassLoader*>(class_loader)));
689 method = compiler_driver->ResolveMethod(
690 soa, dex_cache, loader, &dex_compilation_unit, method_idx, invoke_type);
691 }
692 // For AOT compilation, we may not get a method, for example if its class is erroneous.
693 // JIT should always have a method.
694 DCHECK(Runtime::Current()->IsAotCompiler() || method != nullptr);
695 if (method != nullptr) {
696 graph->SetArtMethod(method);
697 ScopedObjectAccess soa(Thread::Current());
698 interpreter_metadata = method->GetQuickenedInfo();
699 uint16_t type_index = method->GetDeclaringClass()->GetDexTypeIndex();
700
701 // Update the dex cache if the type is not in it yet. Note that under AOT,
702 // the verifier must have set it, but under JIT, there's no guarantee, as we
703 // don't necessarily run the verifier.
704 // The compiler and the compiler driver assume the compiling class is
705 // in the dex cache.
706 if (dex_cache->GetResolvedType(type_index) == nullptr) {
707 dex_cache->SetResolvedType(type_index, method->GetDeclaringClass());
708 }
709 }
710
711 std::unique_ptr<CodeGenerator> codegen(
712 CodeGenerator::Create(graph,
713 instruction_set,
714 *compiler_driver->GetInstructionSetFeatures(),
715 compiler_driver->GetCompilerOptions(),
716 compilation_stats_.get()));
717 if (codegen.get() == nullptr) {
718 MaybeRecordStat(MethodCompilationStat::kNotCompiledNoCodegen);
719 return nullptr;
720 }
721 codegen->GetAssembler()->cfi().SetEnabled(
722 compiler_driver->GetCompilerOptions().GenerateAnyDebugInfo());
723
724 PassObserver pass_observer(graph,
725 codegen.get(),
726 visualizer_output_.get(),
727 compiler_driver);
728
729 VLOG(compiler) << "Building " << pass_observer.GetMethodName();
730
731 {
732 ScopedObjectAccess soa(Thread::Current());
733 StackHandleScopeCollection handles(soa.Self());
734 // Do not hold `mutator_lock_` between optimizations.
735 ScopedThreadSuspension sts(soa.Self(), kNative);
736
737 {
738 PassScope scope(HGraphBuilder::kBuilderPassName, &pass_observer);
739 HGraphBuilder builder(graph,
740 &dex_compilation_unit,
741 &dex_compilation_unit,
742 &dex_file,
743 *code_item,
744 compiler_driver,
745 compilation_stats_.get(),
746 interpreter_metadata,
747 dex_cache,
748 &handles);
749 GraphAnalysisResult result = builder.BuildGraph();
750 if (result != kAnalysisSuccess) {
751 switch (result) {
752 case kAnalysisSkipped:
753 MaybeRecordStat(MethodCompilationStat::kNotCompiledSkipped);
754 break;
755 case kAnalysisInvalidBytecode:
756 MaybeRecordStat(MethodCompilationStat::kNotCompiledInvalidBytecode);
757 break;
758 case kAnalysisFailThrowCatchLoop:
759 MaybeRecordStat(MethodCompilationStat::kNotCompiledThrowCatchLoop);
760 break;
761 case kAnalysisFailAmbiguousArrayOp:
762 MaybeRecordStat(MethodCompilationStat::kNotCompiledAmbiguousArrayOp);
763 break;
764 case kAnalysisSuccess:
765 UNREACHABLE();
766 }
767 pass_observer.SetGraphInBadState();
768 return nullptr;
769 }
770 }
771
772 RunOptimizations(graph,
773 codegen.get(),
774 compiler_driver,
775 compilation_stats_.get(),
776 dex_compilation_unit,
777 &pass_observer,
778 &handles);
779
780 codegen->Compile(code_allocator);
781 pass_observer.DumpDisassembly();
782 }
783
784 return codegen.release();
785 }
786
Compile(const DexFile::CodeItem * code_item,uint32_t access_flags,InvokeType invoke_type,uint16_t class_def_idx,uint32_t method_idx,jobject jclass_loader,const DexFile & dex_file,Handle<mirror::DexCache> dex_cache) const787 CompiledMethod* OptimizingCompiler::Compile(const DexFile::CodeItem* code_item,
788 uint32_t access_flags,
789 InvokeType invoke_type,
790 uint16_t class_def_idx,
791 uint32_t method_idx,
792 jobject jclass_loader,
793 const DexFile& dex_file,
794 Handle<mirror::DexCache> dex_cache) const {
795 CompilerDriver* compiler_driver = GetCompilerDriver();
796 CompiledMethod* method = nullptr;
797 DCHECK(Runtime::Current()->IsAotCompiler());
798 const VerifiedMethod* verified_method = compiler_driver->GetVerifiedMethod(&dex_file, method_idx);
799 DCHECK(!verified_method->HasRuntimeThrow());
800 if (compiler_driver->IsMethodVerifiedWithoutFailures(method_idx, class_def_idx, dex_file)
801 || verifier::MethodVerifier::CanCompilerHandleVerificationFailure(
802 verified_method->GetEncounteredVerificationFailures())) {
803 ArenaAllocator arena(Runtime::Current()->GetArenaPool());
804 CodeVectorAllocator code_allocator(&arena);
805 std::unique_ptr<CodeGenerator> codegen(
806 TryCompile(&arena,
807 &code_allocator,
808 code_item,
809 access_flags,
810 invoke_type,
811 class_def_idx,
812 method_idx,
813 jclass_loader,
814 dex_file,
815 dex_cache,
816 nullptr,
817 /* osr */ false));
818 if (codegen.get() != nullptr) {
819 MaybeRecordStat(MethodCompilationStat::kCompiled);
820 method = Emit(&arena, &code_allocator, codegen.get(), compiler_driver, code_item);
821
822 if (kArenaAllocatorCountAllocations) {
823 if (arena.BytesAllocated() > kArenaAllocatorMemoryReportThreshold) {
824 MemStats mem_stats(arena.GetMemStats());
825 LOG(INFO) << PrettyMethod(method_idx, dex_file) << " " << Dumpable<MemStats>(mem_stats);
826 }
827 }
828 }
829 } else {
830 if (compiler_driver->GetCompilerOptions().VerifyAtRuntime()) {
831 MaybeRecordStat(MethodCompilationStat::kNotCompiledVerifyAtRuntime);
832 } else {
833 MaybeRecordStat(MethodCompilationStat::kNotCompiledVerificationError);
834 }
835 }
836
837 if (kIsDebugBuild &&
838 IsCompilingWithCoreImage() &&
839 IsInstructionSetSupported(compiler_driver->GetInstructionSet()) &&
840 (!kEmitCompilerReadBarrier ||
841 InstructionSetSupportsReadBarrier(compiler_driver->GetInstructionSet()))) {
842 // For testing purposes, we put a special marker on method names
843 // that should be compiled with this compiler (when the the
844 // instruction set is supported -- and has support for read
845 // barriers, if they are enabled). This makes sure we're not
846 // regressing.
847 std::string method_name = PrettyMethod(method_idx, dex_file);
848 bool shouldCompile = method_name.find("$opt$") != std::string::npos;
849 DCHECK((method != nullptr) || !shouldCompile) << "Didn't compile " << method_name;
850 }
851
852 return method;
853 }
854
CreateOptimizingCompiler(CompilerDriver * driver)855 Compiler* CreateOptimizingCompiler(CompilerDriver* driver) {
856 return new OptimizingCompiler(driver);
857 }
858
IsCompilingWithCoreImage()859 bool IsCompilingWithCoreImage() {
860 const std::string& image = Runtime::Current()->GetImageLocation();
861 // TODO: This is under-approximating...
862 if (EndsWith(image, "core.art") || EndsWith(image, "core-optimizing.art")) {
863 return true;
864 }
865 return false;
866 }
867
JitCompile(Thread * self,jit::JitCodeCache * code_cache,ArtMethod * method,bool osr)868 bool OptimizingCompiler::JitCompile(Thread* self,
869 jit::JitCodeCache* code_cache,
870 ArtMethod* method,
871 bool osr) {
872 StackHandleScope<2> hs(self);
873 Handle<mirror::ClassLoader> class_loader(hs.NewHandle(
874 method->GetDeclaringClass()->GetClassLoader()));
875 Handle<mirror::DexCache> dex_cache(hs.NewHandle(method->GetDexCache()));
876 DCHECK(method->IsCompilable());
877
878 jobject jclass_loader = class_loader.ToJObject();
879 const DexFile* dex_file = method->GetDexFile();
880 const uint16_t class_def_idx = method->GetClassDefIndex();
881 const DexFile::CodeItem* code_item = dex_file->GetCodeItem(method->GetCodeItemOffset());
882 const uint32_t method_idx = method->GetDexMethodIndex();
883 const uint32_t access_flags = method->GetAccessFlags();
884 const InvokeType invoke_type = method->GetInvokeType();
885
886 ArenaAllocator arena(Runtime::Current()->GetJitArenaPool());
887 CodeVectorAllocator code_allocator(&arena);
888 std::unique_ptr<CodeGenerator> codegen;
889 {
890 // Go to native so that we don't block GC during compilation.
891 ScopedThreadSuspension sts(self, kNative);
892 codegen.reset(
893 TryCompile(&arena,
894 &code_allocator,
895 code_item,
896 access_flags,
897 invoke_type,
898 class_def_idx,
899 method_idx,
900 jclass_loader,
901 *dex_file,
902 dex_cache,
903 method,
904 osr));
905 if (codegen.get() == nullptr) {
906 return false;
907 }
908
909 if (kArenaAllocatorCountAllocations) {
910 if (arena.BytesAllocated() > kArenaAllocatorMemoryReportThreshold) {
911 MemStats mem_stats(arena.GetMemStats());
912 LOG(INFO) << PrettyMethod(method_idx, *dex_file) << " " << Dumpable<MemStats>(mem_stats);
913 }
914 }
915 }
916
917 size_t stack_map_size = codegen->ComputeStackMapsSize();
918 uint8_t* stack_map_data = code_cache->ReserveData(self, stack_map_size, method);
919 if (stack_map_data == nullptr) {
920 return false;
921 }
922 MaybeRecordStat(MethodCompilationStat::kCompiled);
923 codegen->BuildStackMaps(MemoryRegion(stack_map_data, stack_map_size), *code_item);
924 const void* code = code_cache->CommitCode(
925 self,
926 method,
927 stack_map_data,
928 codegen->HasEmptyFrame() ? 0 : codegen->GetFrameSize(),
929 codegen->GetCoreSpillMask(),
930 codegen->GetFpuSpillMask(),
931 code_allocator.GetMemory().data(),
932 code_allocator.GetSize(),
933 osr);
934
935 if (code == nullptr) {
936 code_cache->ClearData(self, stack_map_data);
937 return false;
938 }
939
940 const CompilerOptions& compiler_options = GetCompilerDriver()->GetCompilerOptions();
941 if (compiler_options.GetGenerateDebugInfo()) {
942 const auto* method_header = reinterpret_cast<const OatQuickMethodHeader*>(code);
943 const uintptr_t code_address = reinterpret_cast<uintptr_t>(method_header->GetCode());
944 debug::MethodDebugInfo info = debug::MethodDebugInfo();
945 info.trampoline_name = nullptr;
946 info.dex_file = dex_file;
947 info.class_def_index = class_def_idx;
948 info.dex_method_index = method_idx;
949 info.access_flags = access_flags;
950 info.code_item = code_item;
951 info.isa = codegen->GetInstructionSet();
952 info.deduped = false;
953 info.is_native_debuggable = compiler_options.GetNativeDebuggable();
954 info.is_optimized = true;
955 info.is_code_address_text_relative = false;
956 info.code_address = code_address;
957 info.code_size = code_allocator.GetSize();
958 info.frame_size_in_bytes = method_header->GetFrameSizeInBytes();
959 info.code_info = stack_map_size == 0 ? nullptr : stack_map_data;
960 info.cfi = ArrayRef<const uint8_t>(*codegen->GetAssembler()->cfi().data());
961 std::vector<uint8_t> elf_file = debug::WriteDebugElfFileForMethods(
962 GetCompilerDriver()->GetInstructionSet(),
963 GetCompilerDriver()->GetInstructionSetFeatures(),
964 ArrayRef<const debug::MethodDebugInfo>(&info, 1));
965 CreateJITCodeEntryForAddress(code_address, std::move(elf_file));
966 }
967
968 Runtime::Current()->GetJit()->AddMemoryUsage(method, arena.BytesUsed());
969
970 return true;
971 }
972
973 } // namespace art
974