• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /*
2  * Copyright (C) 2014 The Android Open Source Project
3  *
4  * Licensed under the Apache License, Version 2.0 (the "License");
5  * you may not use this file except in compliance with the License.
6  * You may obtain a copy of the License at
7  *
8  *      http://www.apache.org/licenses/LICENSE-2.0
9  *
10  * Unless required by applicable law or agreed to in writing, software
11  * distributed under the License is distributed on an "AS IS" BASIS,
12  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13  * See the License for the specific language governing permissions and
14  * limitations under the License.
15  */
16 
17 #ifndef ART_COMPILER_OPTIMIZING_CODE_GENERATOR_ARM_H_
18 #define ART_COMPILER_OPTIMIZING_CODE_GENERATOR_ARM_H_
19 
20 #include "code_generator.h"
21 #include "dex/compiler_enums.h"
22 #include "driver/compiler_options.h"
23 #include "nodes.h"
24 #include "parallel_move_resolver.h"
25 #include "utils/arm/assembler_thumb2.h"
26 #include "utils/string_reference.h"
27 
28 namespace art {
29 namespace arm {
30 
31 class CodeGeneratorARM;
32 
33 // Use a local definition to prevent copying mistakes.
34 static constexpr size_t kArmWordSize = kArmPointerSize;
35 static constexpr size_t kArmBitsPerWord = kArmWordSize * kBitsPerByte;
36 
37 static constexpr Register kParameterCoreRegisters[] = { R1, R2, R3 };
38 static constexpr size_t kParameterCoreRegistersLength = arraysize(kParameterCoreRegisters);
39 static constexpr SRegister kParameterFpuRegisters[] =
40     { S0, S1, S2, S3, S4, S5, S6, S7, S8, S9, S10, S11, S12, S13, S14, S15 };
41 static constexpr size_t kParameterFpuRegistersLength = arraysize(kParameterFpuRegisters);
42 
43 static constexpr Register kArtMethodRegister = R0;
44 
45 static constexpr Register kRuntimeParameterCoreRegisters[] = { R0, R1, R2, R3 };
46 static constexpr size_t kRuntimeParameterCoreRegistersLength =
47     arraysize(kRuntimeParameterCoreRegisters);
48 static constexpr SRegister kRuntimeParameterFpuRegisters[] = { S0, S1, S2, S3 };
49 static constexpr size_t kRuntimeParameterFpuRegistersLength =
50     arraysize(kRuntimeParameterFpuRegisters);
51 
52 class InvokeRuntimeCallingConvention : public CallingConvention<Register, SRegister> {
53  public:
InvokeRuntimeCallingConvention()54   InvokeRuntimeCallingConvention()
55       : CallingConvention(kRuntimeParameterCoreRegisters,
56                           kRuntimeParameterCoreRegistersLength,
57                           kRuntimeParameterFpuRegisters,
58                           kRuntimeParameterFpuRegistersLength,
59                           kArmPointerSize) {}
60 
61  private:
62   DISALLOW_COPY_AND_ASSIGN(InvokeRuntimeCallingConvention);
63 };
64 
FromLowSToD(SRegister reg)65 static constexpr DRegister FromLowSToD(SRegister reg) {
66   return DCHECK_CONSTEXPR(reg % 2 == 0, , D0)
67       static_cast<DRegister>(reg / 2);
68 }
69 
70 
71 class InvokeDexCallingConvention : public CallingConvention<Register, SRegister> {
72  public:
InvokeDexCallingConvention()73   InvokeDexCallingConvention()
74       : CallingConvention(kParameterCoreRegisters,
75                           kParameterCoreRegistersLength,
76                           kParameterFpuRegisters,
77                           kParameterFpuRegistersLength,
78                           kArmPointerSize) {}
79 
80  private:
81   DISALLOW_COPY_AND_ASSIGN(InvokeDexCallingConvention);
82 };
83 
84 class InvokeDexCallingConventionVisitorARM : public InvokeDexCallingConventionVisitor {
85  public:
InvokeDexCallingConventionVisitorARM()86   InvokeDexCallingConventionVisitorARM() {}
~InvokeDexCallingConventionVisitorARM()87   virtual ~InvokeDexCallingConventionVisitorARM() {}
88 
89   Location GetNextLocation(Primitive::Type type) OVERRIDE;
90   Location GetReturnLocation(Primitive::Type type) const OVERRIDE;
91   Location GetMethodLocation() const OVERRIDE;
92 
93  private:
94   InvokeDexCallingConvention calling_convention;
95   uint32_t double_index_ = 0;
96 
97   DISALLOW_COPY_AND_ASSIGN(InvokeDexCallingConventionVisitorARM);
98 };
99 
100 class FieldAccessCallingConventionARM : public FieldAccessCallingConvention {
101  public:
FieldAccessCallingConventionARM()102   FieldAccessCallingConventionARM() {}
103 
GetObjectLocation()104   Location GetObjectLocation() const OVERRIDE {
105     return Location::RegisterLocation(R1);
106   }
GetFieldIndexLocation()107   Location GetFieldIndexLocation() const OVERRIDE {
108     return Location::RegisterLocation(R0);
109   }
GetReturnLocation(Primitive::Type type)110   Location GetReturnLocation(Primitive::Type type) const OVERRIDE {
111     return Primitive::Is64BitType(type)
112         ? Location::RegisterPairLocation(R0, R1)
113         : Location::RegisterLocation(R0);
114   }
GetSetValueLocation(Primitive::Type type,bool is_instance)115   Location GetSetValueLocation(Primitive::Type type, bool is_instance) const OVERRIDE {
116     return Primitive::Is64BitType(type)
117         ? Location::RegisterPairLocation(R2, R3)
118         : (is_instance
119             ? Location::RegisterLocation(R2)
120             : Location::RegisterLocation(R1));
121   }
GetFpuLocation(Primitive::Type type)122   Location GetFpuLocation(Primitive::Type type) const OVERRIDE {
123     return Primitive::Is64BitType(type)
124         ? Location::FpuRegisterPairLocation(S0, S1)
125         : Location::FpuRegisterLocation(S0);
126   }
127 
128  private:
129   DISALLOW_COPY_AND_ASSIGN(FieldAccessCallingConventionARM);
130 };
131 
132 class ParallelMoveResolverARM : public ParallelMoveResolverWithSwap {
133  public:
ParallelMoveResolverARM(ArenaAllocator * allocator,CodeGeneratorARM * codegen)134   ParallelMoveResolverARM(ArenaAllocator* allocator, CodeGeneratorARM* codegen)
135       : ParallelMoveResolverWithSwap(allocator), codegen_(codegen) {}
136 
137   void EmitMove(size_t index) OVERRIDE;
138   void EmitSwap(size_t index) OVERRIDE;
139   void SpillScratch(int reg) OVERRIDE;
140   void RestoreScratch(int reg) OVERRIDE;
141 
142   ArmAssembler* GetAssembler() const;
143 
144  private:
145   void Exchange(Register reg, int mem);
146   void Exchange(int mem1, int mem2);
147 
148   CodeGeneratorARM* const codegen_;
149 
150   DISALLOW_COPY_AND_ASSIGN(ParallelMoveResolverARM);
151 };
152 
153 class LocationsBuilderARM : public HGraphVisitor {
154  public:
LocationsBuilderARM(HGraph * graph,CodeGeneratorARM * codegen)155   LocationsBuilderARM(HGraph* graph, CodeGeneratorARM* codegen)
156       : HGraphVisitor(graph), codegen_(codegen) {}
157 
158 #define DECLARE_VISIT_INSTRUCTION(name, super)     \
159   void Visit##name(H##name* instr) OVERRIDE;
160 
161   FOR_EACH_CONCRETE_INSTRUCTION_COMMON(DECLARE_VISIT_INSTRUCTION)
FOR_EACH_CONCRETE_INSTRUCTION_ARM(DECLARE_VISIT_INSTRUCTION)162   FOR_EACH_CONCRETE_INSTRUCTION_ARM(DECLARE_VISIT_INSTRUCTION)
163   FOR_EACH_CONCRETE_INSTRUCTION_SHARED(DECLARE_VISIT_INSTRUCTION)
164 
165 #undef DECLARE_VISIT_INSTRUCTION
166 
167   void VisitInstruction(HInstruction* instruction) OVERRIDE {
168     LOG(FATAL) << "Unreachable instruction " << instruction->DebugName()
169                << " (id " << instruction->GetId() << ")";
170   }
171 
172  private:
173   void HandleInvoke(HInvoke* invoke);
174   void HandleBitwiseOperation(HBinaryOperation* operation, Opcode opcode);
175   void HandleCondition(HCondition* condition);
176   void HandleIntegerRotate(LocationSummary* locations);
177   void HandleLongRotate(LocationSummary* locations);
178   void HandleShift(HBinaryOperation* operation);
179   void HandleFieldSet(HInstruction* instruction, const FieldInfo& field_info);
180   void HandleFieldGet(HInstruction* instruction, const FieldInfo& field_info);
181 
182   Location ArmEncodableConstantOrRegister(HInstruction* constant, Opcode opcode);
183   bool CanEncodeConstantAsImmediate(HConstant* input_cst, Opcode opcode);
184   bool CanEncodeConstantAsImmediate(uint32_t value, Opcode opcode);
185 
186   CodeGeneratorARM* const codegen_;
187   InvokeDexCallingConventionVisitorARM parameter_visitor_;
188 
189   DISALLOW_COPY_AND_ASSIGN(LocationsBuilderARM);
190 };
191 
192 class InstructionCodeGeneratorARM : public InstructionCodeGenerator {
193  public:
194   InstructionCodeGeneratorARM(HGraph* graph, CodeGeneratorARM* codegen);
195 
196 #define DECLARE_VISIT_INSTRUCTION(name, super)     \
197   void Visit##name(H##name* instr) OVERRIDE;
198 
199   FOR_EACH_CONCRETE_INSTRUCTION_COMMON(DECLARE_VISIT_INSTRUCTION)
FOR_EACH_CONCRETE_INSTRUCTION_ARM(DECLARE_VISIT_INSTRUCTION)200   FOR_EACH_CONCRETE_INSTRUCTION_ARM(DECLARE_VISIT_INSTRUCTION)
201   FOR_EACH_CONCRETE_INSTRUCTION_SHARED(DECLARE_VISIT_INSTRUCTION)
202 
203 #undef DECLARE_VISIT_INSTRUCTION
204 
205   void VisitInstruction(HInstruction* instruction) OVERRIDE {
206     LOG(FATAL) << "Unreachable instruction " << instruction->DebugName()
207                << " (id " << instruction->GetId() << ")";
208   }
209 
GetAssembler()210   ArmAssembler* GetAssembler() const { return assembler_; }
211 
212  private:
213   // Generate code for the given suspend check. If not null, `successor`
214   // is the block to branch to if the suspend check is not needed, and after
215   // the suspend call.
216   void GenerateSuspendCheck(HSuspendCheck* check, HBasicBlock* successor);
217   void GenerateClassInitializationCheck(SlowPathCode* slow_path, Register class_reg);
218   void GenerateAndConst(Register out, Register first, uint32_t value);
219   void GenerateOrrConst(Register out, Register first, uint32_t value);
220   void GenerateEorConst(Register out, Register first, uint32_t value);
221   void HandleBitwiseOperation(HBinaryOperation* operation);
222   void HandleCondition(HCondition* condition);
223   void HandleIntegerRotate(LocationSummary* locations);
224   void HandleLongRotate(LocationSummary* locations);
225   void HandleShift(HBinaryOperation* operation);
226 
227   void GenerateWideAtomicStore(Register addr, uint32_t offset,
228                                Register value_lo, Register value_hi,
229                                Register temp1, Register temp2,
230                                HInstruction* instruction);
231   void GenerateWideAtomicLoad(Register addr, uint32_t offset,
232                               Register out_lo, Register out_hi);
233 
234   void HandleFieldSet(HInstruction* instruction,
235                       const FieldInfo& field_info,
236                       bool value_can_be_null);
237   void HandleFieldGet(HInstruction* instruction, const FieldInfo& field_info);
238 
239   // Generate a heap reference load using one register `out`:
240   //
241   //   out <- *(out + offset)
242   //
243   // while honoring heap poisoning and/or read barriers (if any).
244   //
245   // Location `maybe_temp` is used when generating a read barrier and
246   // shall be a register in that case; it may be an invalid location
247   // otherwise.
248   void GenerateReferenceLoadOneRegister(HInstruction* instruction,
249                                         Location out,
250                                         uint32_t offset,
251                                         Location maybe_temp);
252   // Generate a heap reference load using two different registers
253   // `out` and `obj`:
254   //
255   //   out <- *(obj + offset)
256   //
257   // while honoring heap poisoning and/or read barriers (if any).
258   //
259   // Location `maybe_temp` is used when generating a Baker's (fast
260   // path) read barrier and shall be a register in that case; it may
261   // be an invalid location otherwise.
262   void GenerateReferenceLoadTwoRegisters(HInstruction* instruction,
263                                          Location out,
264                                          Location obj,
265                                          uint32_t offset,
266                                          Location maybe_temp);
267   // Generate a GC root reference load:
268   //
269   //   root <- *(obj + offset)
270   //
271   // while honoring read barriers (if any).
272   void GenerateGcRootFieldLoad(HInstruction* instruction,
273                                Location root,
274                                Register obj,
275                                uint32_t offset);
276   void GenerateTestAndBranch(HInstruction* instruction,
277                              size_t condition_input_index,
278                              Label* true_target,
279                              Label* false_target);
280   void GenerateCompareTestAndBranch(HCondition* condition,
281                                     Label* true_target,
282                                     Label* false_target);
283   void GenerateFPJumps(HCondition* cond, Label* true_label, Label* false_label);
284   void GenerateLongComparesAndJumps(HCondition* cond, Label* true_label, Label* false_label);
285   void DivRemOneOrMinusOne(HBinaryOperation* instruction);
286   void DivRemByPowerOfTwo(HBinaryOperation* instruction);
287   void GenerateDivRemWithAnyConstant(HBinaryOperation* instruction);
288   void GenerateDivRemConstantIntegral(HBinaryOperation* instruction);
289   void HandleGoto(HInstruction* got, HBasicBlock* successor);
290 
291   ArmAssembler* const assembler_;
292   CodeGeneratorARM* const codegen_;
293 
294   DISALLOW_COPY_AND_ASSIGN(InstructionCodeGeneratorARM);
295 };
296 
297 class CodeGeneratorARM : public CodeGenerator {
298  public:
299   CodeGeneratorARM(HGraph* graph,
300                    const ArmInstructionSetFeatures& isa_features,
301                    const CompilerOptions& compiler_options,
302                    OptimizingCompilerStats* stats = nullptr);
~CodeGeneratorARM()303   virtual ~CodeGeneratorARM() {}
304 
305   void GenerateFrameEntry() OVERRIDE;
306   void GenerateFrameExit() OVERRIDE;
307   void Bind(HBasicBlock* block) OVERRIDE;
308   void MoveConstant(Location destination, int32_t value) OVERRIDE;
309   void MoveLocation(Location dst, Location src, Primitive::Type dst_type) OVERRIDE;
310   void AddLocationAsTemp(Location location, LocationSummary* locations) OVERRIDE;
311 
312   size_t SaveCoreRegister(size_t stack_index, uint32_t reg_id) OVERRIDE;
313   size_t RestoreCoreRegister(size_t stack_index, uint32_t reg_id) OVERRIDE;
314   size_t SaveFloatingPointRegister(size_t stack_index, uint32_t reg_id) OVERRIDE;
315   size_t RestoreFloatingPointRegister(size_t stack_index, uint32_t reg_id) OVERRIDE;
316 
GetWordSize()317   size_t GetWordSize() const OVERRIDE {
318     return kArmWordSize;
319   }
320 
GetFloatingPointSpillSlotSize()321   size_t GetFloatingPointSpillSlotSize() const OVERRIDE {
322     // Allocated in S registers, which are word sized.
323     return kArmWordSize;
324   }
325 
GetLocationBuilder()326   HGraphVisitor* GetLocationBuilder() OVERRIDE {
327     return &location_builder_;
328   }
329 
GetInstructionVisitor()330   HGraphVisitor* GetInstructionVisitor() OVERRIDE {
331     return &instruction_visitor_;
332   }
333 
GetAssembler()334   ArmAssembler* GetAssembler() OVERRIDE {
335     return &assembler_;
336   }
337 
GetAssembler()338   const ArmAssembler& GetAssembler() const OVERRIDE {
339     return assembler_;
340   }
341 
GetAddressOf(HBasicBlock * block)342   uintptr_t GetAddressOf(HBasicBlock* block) OVERRIDE {
343     return GetLabelOf(block)->Position();
344   }
345 
346   void SetupBlockedRegisters() const OVERRIDE;
347 
348   void DumpCoreRegister(std::ostream& stream, int reg) const OVERRIDE;
349   void DumpFloatingPointRegister(std::ostream& stream, int reg) const OVERRIDE;
350 
351   // Blocks all register pairs made out of blocked core registers.
352   void UpdateBlockedPairRegisters() const;
353 
GetMoveResolver()354   ParallelMoveResolverARM* GetMoveResolver() OVERRIDE {
355     return &move_resolver_;
356   }
357 
GetInstructionSet()358   InstructionSet GetInstructionSet() const OVERRIDE {
359     return InstructionSet::kThumb2;
360   }
361 
362   // Helper method to move a 32bits value between two locations.
363   void Move32(Location destination, Location source);
364   // Helper method to move a 64bits value between two locations.
365   void Move64(Location destination, Location source);
366 
367   // Generate code to invoke a runtime entry point.
368   void InvokeRuntime(QuickEntrypointEnum entrypoint,
369                      HInstruction* instruction,
370                      uint32_t dex_pc,
371                      SlowPathCode* slow_path) OVERRIDE;
372 
373   void InvokeRuntime(int32_t offset,
374                      HInstruction* instruction,
375                      uint32_t dex_pc,
376                      SlowPathCode* slow_path);
377 
378   // Emit a write barrier.
379   void MarkGCCard(Register temp, Register card, Register object, Register value, bool can_be_null);
380 
381   void GenerateMemoryBarrier(MemBarrierKind kind);
382 
GetLabelOf(HBasicBlock * block)383   Label* GetLabelOf(HBasicBlock* block) const {
384     return CommonGetLabelOf<Label>(block_labels_, block);
385   }
386 
Initialize()387   void Initialize() OVERRIDE {
388     block_labels_ = CommonInitializeLabels<Label>();
389   }
390 
391   void Finalize(CodeAllocator* allocator) OVERRIDE;
392 
GetInstructionSetFeatures()393   const ArmInstructionSetFeatures& GetInstructionSetFeatures() const {
394     return isa_features_;
395   }
396 
NeedsTwoRegisters(Primitive::Type type)397   bool NeedsTwoRegisters(Primitive::Type type) const OVERRIDE {
398     return type == Primitive::kPrimDouble || type == Primitive::kPrimLong;
399   }
400 
401   void ComputeSpillMask() OVERRIDE;
402 
GetFrameEntryLabel()403   Label* GetFrameEntryLabel() { return &frame_entry_label_; }
404 
405   // Check if the desired_string_load_kind is supported. If it is, return it,
406   // otherwise return a fall-back kind that should be used instead.
407   HLoadString::LoadKind GetSupportedLoadStringKind(
408       HLoadString::LoadKind desired_string_load_kind) OVERRIDE;
409 
410   // Check if the desired_dispatch_info is supported. If it is, return it,
411   // otherwise return a fall-back info that should be used instead.
412   HInvokeStaticOrDirect::DispatchInfo GetSupportedInvokeStaticOrDirectDispatch(
413       const HInvokeStaticOrDirect::DispatchInfo& desired_dispatch_info,
414       MethodReference target_method) OVERRIDE;
415 
416   void GenerateStaticOrDirectCall(HInvokeStaticOrDirect* invoke, Location temp) OVERRIDE;
417   void GenerateVirtualCall(HInvokeVirtual* invoke, Location temp) OVERRIDE;
418 
419   void MoveFromReturnRegister(Location trg, Primitive::Type type) OVERRIDE;
420 
421   // The PcRelativePatchInfo is used for PC-relative addressing of dex cache arrays
422   // and boot image strings. The only difference is the interpretation of the offset_or_index.
423   // The PC-relative address is loaded with three instructions, MOVW+MOVT
424   // to load the offset to base_reg and then ADD base_reg, PC. The offset is
425   // calculated from the ADD's effective PC, i.e. PC+4 on Thumb2. Though we
426   // currently emit these 3 instructions together, instruction scheduling could
427   // split this sequence apart, so we keep separate labels for each of them.
428   struct PcRelativePatchInfo {
PcRelativePatchInfoPcRelativePatchInfo429     PcRelativePatchInfo(const DexFile& dex_file, uint32_t off_or_idx)
430         : target_dex_file(dex_file), offset_or_index(off_or_idx) { }
431     PcRelativePatchInfo(PcRelativePatchInfo&& other) = default;
432 
433     const DexFile& target_dex_file;
434     // Either the dex cache array element offset or the string index.
435     uint32_t offset_or_index;
436     Label movw_label;
437     Label movt_label;
438     Label add_pc_label;
439   };
440 
441   PcRelativePatchInfo* NewPcRelativeStringPatch(const DexFile& dex_file, uint32_t string_index);
442   PcRelativePatchInfo* NewPcRelativeDexCacheArrayPatch(const DexFile& dex_file,
443                                                        uint32_t element_offset);
444   Literal* DeduplicateBootImageStringLiteral(const DexFile& dex_file, uint32_t string_index);
445   Literal* DeduplicateBootImageAddressLiteral(uint32_t address);
446   Literal* DeduplicateDexCacheAddressLiteral(uint32_t address);
447 
448   void EmitLinkerPatches(ArenaVector<LinkerPatch>* linker_patches) OVERRIDE;
449 
450   // Fast path implementation of ReadBarrier::Barrier for a heap
451   // reference field load when Baker's read barriers are used.
452   void GenerateFieldLoadWithBakerReadBarrier(HInstruction* instruction,
453                                              Location ref,
454                                              Register obj,
455                                              uint32_t offset,
456                                              Location temp,
457                                              bool needs_null_check);
458   // Fast path implementation of ReadBarrier::Barrier for a heap
459   // reference array load when Baker's read barriers are used.
460   void GenerateArrayLoadWithBakerReadBarrier(HInstruction* instruction,
461                                              Location ref,
462                                              Register obj,
463                                              uint32_t data_offset,
464                                              Location index,
465                                              Location temp,
466                                              bool needs_null_check);
467 
468   // Generate a read barrier for a heap reference within `instruction`
469   // using a slow path.
470   //
471   // A read barrier for an object reference read from the heap is
472   // implemented as a call to the artReadBarrierSlow runtime entry
473   // point, which is passed the values in locations `ref`, `obj`, and
474   // `offset`:
475   //
476   //   mirror::Object* artReadBarrierSlow(mirror::Object* ref,
477   //                                      mirror::Object* obj,
478   //                                      uint32_t offset);
479   //
480   // The `out` location contains the value returned by
481   // artReadBarrierSlow.
482   //
483   // When `index` is provided (i.e. for array accesses), the offset
484   // value passed to artReadBarrierSlow is adjusted to take `index`
485   // into account.
486   void GenerateReadBarrierSlow(HInstruction* instruction,
487                                Location out,
488                                Location ref,
489                                Location obj,
490                                uint32_t offset,
491                                Location index = Location::NoLocation());
492 
493   // If read barriers are enabled, generate a read barrier for a heap
494   // reference using a slow path. If heap poisoning is enabled, also
495   // unpoison the reference in `out`.
496   void MaybeGenerateReadBarrierSlow(HInstruction* instruction,
497                                     Location out,
498                                     Location ref,
499                                     Location obj,
500                                     uint32_t offset,
501                                     Location index = Location::NoLocation());
502 
503   // Generate a read barrier for a GC root within `instruction` using
504   // a slow path.
505   //
506   // A read barrier for an object reference GC root is implemented as
507   // a call to the artReadBarrierForRootSlow runtime entry point,
508   // which is passed the value in location `root`:
509   //
510   //   mirror::Object* artReadBarrierForRootSlow(GcRoot<mirror::Object>* root);
511   //
512   // The `out` location contains the value returned by
513   // artReadBarrierForRootSlow.
514   void GenerateReadBarrierForRootSlow(HInstruction* instruction, Location out, Location root);
515 
516   void GenerateNop();
517 
518   void GenerateImplicitNullCheck(HNullCheck* instruction);
519   void GenerateExplicitNullCheck(HNullCheck* instruction);
520 
521  private:
522   // Factored implementation of GenerateFieldLoadWithBakerReadBarrier
523   // and GenerateArrayLoadWithBakerReadBarrier.
524   void GenerateReferenceLoadWithBakerReadBarrier(HInstruction* instruction,
525                                                  Location ref,
526                                                  Register obj,
527                                                  uint32_t offset,
528                                                  Location index,
529                                                  Location temp,
530                                                  bool needs_null_check);
531 
532   Register GetInvokeStaticOrDirectExtraParameter(HInvokeStaticOrDirect* invoke, Register temp);
533 
534   using Uint32ToLiteralMap = ArenaSafeMap<uint32_t, Literal*>;
535   using MethodToLiteralMap = ArenaSafeMap<MethodReference, Literal*, MethodReferenceComparator>;
536   using BootStringToLiteralMap = ArenaSafeMap<StringReference,
537                                               Literal*,
538                                               StringReferenceValueComparator>;
539 
540   Literal* DeduplicateUint32Literal(uint32_t value, Uint32ToLiteralMap* map);
541   Literal* DeduplicateMethodLiteral(MethodReference target_method, MethodToLiteralMap* map);
542   Literal* DeduplicateMethodAddressLiteral(MethodReference target_method);
543   Literal* DeduplicateMethodCodeLiteral(MethodReference target_method);
544   PcRelativePatchInfo* NewPcRelativePatch(const DexFile& dex_file,
545                                           uint32_t offset_or_index,
546                                           ArenaDeque<PcRelativePatchInfo>* patches);
547 
548   // Labels for each block that will be compiled.
549   Label* block_labels_;  // Indexed by block id.
550   Label frame_entry_label_;
551   LocationsBuilderARM location_builder_;
552   InstructionCodeGeneratorARM instruction_visitor_;
553   ParallelMoveResolverARM move_resolver_;
554   Thumb2Assembler assembler_;
555   const ArmInstructionSetFeatures& isa_features_;
556 
557   // Deduplication map for 32-bit literals, used for non-patchable boot image addresses.
558   Uint32ToLiteralMap uint32_literals_;
559   // Method patch info, map MethodReference to a literal for method address and method code.
560   MethodToLiteralMap method_patches_;
561   MethodToLiteralMap call_patches_;
562   // Relative call patch info.
563   // Using ArenaDeque<> which retains element addresses on push/emplace_back().
564   ArenaDeque<MethodPatchInfo<Label>> relative_call_patches_;
565   // PC-relative patch info for each HArmDexCacheArraysBase.
566   ArenaDeque<PcRelativePatchInfo> pc_relative_dex_cache_patches_;
567   // Deduplication map for boot string literals for kBootImageLinkTimeAddress.
568   BootStringToLiteralMap boot_image_string_patches_;
569   // PC-relative String patch info.
570   ArenaDeque<PcRelativePatchInfo> pc_relative_string_patches_;
571   // Deduplication map for patchable boot image addresses.
572   Uint32ToLiteralMap boot_image_address_patches_;
573 
574   DISALLOW_COPY_AND_ASSIGN(CodeGeneratorARM);
575 };
576 
577 }  // namespace arm
578 }  // namespace art
579 
580 #endif  // ART_COMPILER_OPTIMIZING_CODE_GENERATOR_ARM_H_
581