/* * Copyright (C) 2012 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #ifndef ART_COMPILER_DEX_LOCAL_VALUE_NUMBERING_H_ #define ART_COMPILER_DEX_LOCAL_VALUE_NUMBERING_H_ #include #include "compiler_internals.h" #include "global_value_numbering.h" #include "utils/scoped_arena_allocator.h" #include "utils/scoped_arena_containers.h" namespace art { class DexFile; // Enable/disable tracking values stored in the FILLED_NEW_ARRAY result. static constexpr bool kLocalValueNumberingEnableFilledNewArrayTracking = true; class LocalValueNumbering { private: static constexpr uint16_t kNoValue = GlobalValueNumbering::kNoValue; public: LocalValueNumbering(GlobalValueNumbering* gvn, BasicBlockId id, ScopedArenaAllocator* allocator); BasicBlockId Id() const { return id_; } bool Equals(const LocalValueNumbering& other) const; uint16_t GetSRegValueName(uint16_t s_reg) const { return GetOperandValue(s_reg); } void SetValueNameNullChecked(uint16_t value_name) { null_checked_.insert(value_name); } bool IsValueNullChecked(uint16_t value_name) const { return null_checked_.find(value_name) != null_checked_.end(); } bool IsSregValue(uint16_t s_reg, uint16_t value_name) const { auto it = sreg_value_map_.find(s_reg); if (it != sreg_value_map_.end()) { return it->second == value_name; } else { return gvn_->HasValue(kNoValue, s_reg, kNoValue, kNoValue, value_name); } } enum MergeType { kNormalMerge, kCatchMerge, kReturnMerge, // RETURN or PHI+RETURN. Merge only sreg maps. }; void MergeOne(const LocalValueNumbering& other, MergeType merge_type); void Merge(MergeType merge_type); // Merge gvn_->merge_lvns_. uint16_t GetValueNumber(MIR* mir); // LocalValueNumbering should be allocated on the ArenaStack (or the native stack). static void* operator new(size_t size, ScopedArenaAllocator* allocator) { return allocator->Alloc(sizeof(LocalValueNumbering), kArenaAllocMisc); } // Allow delete-expression to destroy a LocalValueNumbering object without deallocation. static void operator delete(void* ptr) { UNUSED(ptr); } private: // A set of value names. typedef GlobalValueNumbering::ValueNameSet ValueNameSet; // Field types correspond to the ordering of GET/PUT instructions; this order is the same // for IGET, IPUT, SGET, SPUT, AGET and APUT: // op 0 // op_WIDE 1 // op_OBJECT 2 // op_BOOLEAN 3 // op_BYTE 4 // op_CHAR 5 // op_SHORT 6 static constexpr size_t kFieldTypeCount = 7; // Key is s_reg, value is value name. typedef ScopedArenaSafeMap SregValueMap; void SetOperandValueImpl(uint16_t s_reg, uint16_t value, SregValueMap* map) { DCHECK_EQ(map->count(s_reg), 0u) << PrettyMethod(gvn_->cu_->method_idx, *gvn_->cu_->dex_file) << " LVN id: " << id_ << ", s_reg: " << s_reg; map->Put(s_reg, value); } uint16_t GetOperandValueImpl(int s_reg, const SregValueMap* map) const { uint16_t res = kNoValue; auto lb = map->find(s_reg); if (lb != map->end()) { res = lb->second; } else { // Using the original value; s_reg refers to an input reg. res = gvn_->LookupValue(kNoValue, s_reg, kNoValue, kNoValue); } return res; } void SetOperandValue(uint16_t s_reg, uint16_t value) { SetOperandValueImpl(s_reg, value, &sreg_value_map_); }; uint16_t GetOperandValue(int s_reg) const { return GetOperandValueImpl(s_reg, &sreg_value_map_); }; void SetOperandValueWide(uint16_t s_reg, uint16_t value) { SetOperandValueImpl(s_reg, value, &sreg_wide_value_map_); }; uint16_t GetOperandValueWide(int s_reg) const { return GetOperandValueImpl(s_reg, &sreg_wide_value_map_); }; struct RangeCheckKey { uint16_t array; uint16_t index; // NOTE: Can't define this at namespace scope for a private struct. bool operator==(const RangeCheckKey& other) const { return array == other.array && index == other.index; } }; struct RangeCheckKeyComparator { bool operator()(const RangeCheckKey& lhs, const RangeCheckKey& rhs) const { if (lhs.array != rhs.array) { return lhs.array < rhs.array; } return lhs.index < rhs.index; } }; typedef ScopedArenaSet RangeCheckSet; // Maps instance field "location" (derived from base, field_id and type) to value name. typedef ScopedArenaSafeMap IFieldLocToValueMap; // Maps static field id to value name typedef ScopedArenaSafeMap SFieldToValueMap; struct EscapedIFieldClobberKey { uint16_t base; // Or array. uint16_t type; uint16_t field_id; // None (kNoValue) for arrays and unresolved instance field stores. // NOTE: Can't define this at namespace scope for a private struct. bool operator==(const EscapedIFieldClobberKey& other) const { return base == other.base && type == other.type && field_id == other.field_id; } }; struct EscapedIFieldClobberKeyComparator { bool operator()(const EscapedIFieldClobberKey& lhs, const EscapedIFieldClobberKey& rhs) const { // Compare base first. This makes sequential iteration respect the order of base. if (lhs.base != rhs.base) { return lhs.base < rhs.base; } // Compare type second. This makes the type-clobber entries (field_id == kNoValue) last // for given base and type and makes it easy to prune unnecessary entries when merging // escaped_ifield_clobber_set_ from multiple LVNs. if (lhs.type != rhs.type) { return lhs.type < rhs.type; } return lhs.field_id < rhs.field_id; } }; typedef ScopedArenaSet EscapedIFieldClobberSet; struct EscapedArrayClobberKey { uint16_t base; uint16_t type; // NOTE: Can't define this at namespace scope for a private struct. bool operator==(const EscapedArrayClobberKey& other) const { return base == other.base && type == other.type; } }; struct EscapedArrayClobberKeyComparator { bool operator()(const EscapedArrayClobberKey& lhs, const EscapedArrayClobberKey& rhs) const { // Compare base first. This makes sequential iteration respect the order of base. if (lhs.base != rhs.base) { return lhs.base < rhs.base; } return lhs.type < rhs.type; } }; // Clobber set for previously non-aliasing array refs that escaped. typedef ScopedArenaSet EscapedArrayClobberSet; // Known location values for an aliasing set. The set can be tied to one of: // 1. Instance field. The locations are aliasing references used to access the field. // 2. Non-aliasing array reference. The locations are indexes to the array. // 3. Aliasing array type. The locations are (reference, index) pair ids assigned by GVN. // In each case we keep track of the last stored value, if any, and the set of locations // where it was stored. We also keep track of all values known for the current write state // (load_value_map), which can be known either because they have been loaded since the last // store or because they contained the last_stored_value before the store and thus could not // have changed as a result. struct AliasingValues { explicit AliasingValues(LocalValueNumbering* lvn) : memory_version_before_stores(kNoValue), last_stored_value(kNoValue), store_loc_set(std::less(), lvn->null_checked_.get_allocator()), last_load_memory_version(kNoValue), load_value_map(std::less(), lvn->null_checked_.get_allocator()) { } uint16_t memory_version_before_stores; // kNoValue if start version for the field. uint16_t last_stored_value; // Last stored value name, kNoValue if none. ValueNameSet store_loc_set; // Where was last_stored_value stored. // Maps refs (other than stored_to) to currently known values for this field other. On write, // anything that differs from the written value is removed as it may be overwritten. uint16_t last_load_memory_version; // kNoValue if not known. ScopedArenaSafeMap load_value_map; // NOTE: Can't define this at namespace scope for a private struct. bool operator==(const AliasingValues& other) const { return memory_version_before_stores == other.memory_version_before_stores && last_load_memory_version == other.last_load_memory_version && last_stored_value == other.last_stored_value && store_loc_set == other.store_loc_set && load_value_map == other.load_value_map; } }; // Maps instance field id to AliasingValues, locations are object refs. typedef ScopedArenaSafeMap AliasingIFieldValuesMap; // Maps non-aliasing array reference to AliasingValues, locations are array indexes. typedef ScopedArenaSafeMap NonAliasingArrayValuesMap; // Maps aliasing array type to AliasingValues, locations are (array, index) pair ids. typedef ScopedArenaSafeMap AliasingArrayValuesMap; // Helper classes defining versions for updating and merging the AliasingValues maps above. class AliasingIFieldVersions; class NonAliasingArrayVersions; class AliasingArrayVersions; template AliasingValues* GetAliasingValues(Map* map, const typename Map::key_type& key); template void UpdateAliasingValuesLoadVersion(const KeyType& key, AliasingValues* values); template static uint16_t AliasingValuesMergeGet(GlobalValueNumbering* gvn, const LocalValueNumbering* lvn, Map* map, const typename Map::key_type& key, uint16_t location); template uint16_t HandleAliasingValuesGet(Map* map, const typename Map::key_type& key, uint16_t location); template bool HandleAliasingValuesPut(Map* map, const typename Map::key_type& key, uint16_t location, uint16_t value); template void CopyAliasingValuesMap(ScopedArenaSafeMap* dest, const ScopedArenaSafeMap& src); uint16_t MarkNonAliasingNonNull(MIR* mir); bool IsNonAliasing(uint16_t reg) const; bool IsNonAliasingIField(uint16_t reg, uint16_t field_id, uint16_t type) const; bool IsNonAliasingArray(uint16_t reg, uint16_t type) const; void HandleNullCheck(MIR* mir, uint16_t reg); void HandleRangeCheck(MIR* mir, uint16_t array, uint16_t index); void HandlePutObject(MIR* mir); void HandleEscapingRef(uint16_t base); uint16_t HandlePhi(MIR* mir); uint16_t HandleAGet(MIR* mir, uint16_t opcode); void HandleAPut(MIR* mir, uint16_t opcode); uint16_t HandleIGet(MIR* mir, uint16_t opcode); void HandleIPut(MIR* mir, uint16_t opcode); uint16_t HandleSGet(MIR* mir, uint16_t opcode); void HandleSPut(MIR* mir, uint16_t opcode); void RemoveSFieldsForType(uint16_t type); void HandleInvokeOrClInit(MIR* mir); bool SameMemoryVersion(const LocalValueNumbering& other) const; uint16_t NewMemoryVersion(uint16_t* new_version); void MergeMemoryVersions(bool clobbered_catch); void PruneNonAliasingRefsForCatch(); template void IntersectSets(); void CopyLiveSregValues(SregValueMap* dest, const SregValueMap& src); // Intersect maps as sets. The value type must be equality-comparable. template void IntersectSregValueMaps(); // Intersect maps as sets. The value type must be equality-comparable. template static void InPlaceIntersectMaps(Map* work_map, const Map& other_map); template void MergeSets(); void IntersectAliasingValueLocations(AliasingValues* work_values, const AliasingValues* values); void MergeEscapedRefs(const ValueNameSet::value_type& entry, ValueNameSet::iterator hint); void MergeEscapedIFieldTypeClobberSets(const EscapedIFieldClobberSet::value_type& entry, EscapedIFieldClobberSet::iterator hint); void MergeEscapedIFieldClobberSets(const EscapedIFieldClobberSet::value_type& entry, EscapedIFieldClobberSet::iterator hint); void MergeEscapedArrayClobberSets(const EscapedArrayClobberSet::value_type& entry, EscapedArrayClobberSet::iterator hint); void MergeNullChecked(const ValueNameSet::value_type& entry, ValueNameSet::iterator hint); void MergeSFieldValues(const SFieldToValueMap::value_type& entry, SFieldToValueMap::iterator hint); void MergeNonAliasingIFieldValues(const IFieldLocToValueMap::value_type& entry, IFieldLocToValueMap::iterator hint); template void MergeAliasingValues(const typename Map::value_type& entry, typename Map::iterator hint); GlobalValueNumbering* gvn_; // We're using the block id as a 16-bit operand value for some lookups. COMPILE_ASSERT(sizeof(BasicBlockId) == sizeof(uint16_t), BasicBlockId_must_be_16_bit); BasicBlockId id_; SregValueMap sreg_value_map_; SregValueMap sreg_wide_value_map_; SFieldToValueMap sfield_value_map_; IFieldLocToValueMap non_aliasing_ifield_value_map_; AliasingIFieldValuesMap aliasing_ifield_value_map_; NonAliasingArrayValuesMap non_aliasing_array_value_map_; AliasingArrayValuesMap aliasing_array_value_map_; // Data for dealing with memory clobbering and store/load aliasing. uint16_t global_memory_version_; uint16_t unresolved_sfield_version_[kFieldTypeCount]; uint16_t unresolved_ifield_version_[kFieldTypeCount]; // Value names of references to objects that cannot be reached through a different value name. ValueNameSet non_aliasing_refs_; // Previously non-aliasing refs that escaped but can still be used for non-aliasing AGET/IGET. ValueNameSet escaped_refs_; // Blacklists for cases where escaped_refs_ can't be used. EscapedIFieldClobberSet escaped_ifield_clobber_set_; EscapedArrayClobberSet escaped_array_clobber_set_; // Range check and null check elimination. RangeCheckSet range_checked_; ValueNameSet null_checked_; // Reuse one vector for all merges to avoid leaking too much memory on the ArenaStack. ScopedArenaVector merge_names_; // Map to identify when different locations merge the same values. ScopedArenaSafeMap, uint16_t> merge_map_; // New memory version for merge, kNoValue if all memory versions matched. uint16_t merge_new_memory_version_; DISALLOW_COPY_AND_ASSIGN(LocalValueNumbering); }; } // namespace art #endif // ART_COMPILER_DEX_LOCAL_VALUE_NUMBERING_H_