/art/compiler/optimizing/ |
D | code_sinking.cc | 116 if (!discard_blocks.IsBitSet(instruction->GetBlock()->GetBlockId()) && in AddInstruction() 117 !processed_instructions.IsBitSet(instruction->GetId()) && in AddInstruction() 150 !post_dominated.IsBitSet(user->GetBlock()->GetBlockId()); in ShouldFilterUse() 154 !post_dominated.IsBitSet(user->GetBlock()->GetBlockId()); in ShouldFilterUse() 199 if (!post_dominated.IsBitSet(target_block->GetDominator()->GetBlockId())) { in FindIdealPosition() 269 if (!post_dominated.IsBitSet(successor->GetBlockId())) { in SinkCodeToUncommonBranch() 287 if (post_dominated.IsBitSet(i)) { in SinkCodeToUncommonBranch() 297 if (processed_instructions.IsBitSet(instruction->GetId())) { in SinkCodeToUncommonBranch() 308 if (!post_dominated.IsBitSet(user->GetBlock()->GetBlockId()) && in SinkCodeToUncommonBranch() 309 !instructions_that_can_move.IsBitSet(user->GetId())) { in SinkCodeToUncommonBranch() [all …]
|
D | gvn.cc | 90 if (!buckets_owned_.IsBitSet(index)) { in Add() 228 DCHECK(!buckets_owned_.IsBitSet(index)); in CloneBucket() 260 if (!buckets_owned_.IsBitSet(i)) { in DeleteAllImpureWhich() 279 DCHECK(buckets_owned_.IsBitSet(i) || node == nullptr); in DeleteAllImpureWhich() 517 DCHECK(visited_blocks_.IsBitSet(block->GetBlockId())); in WillBeReferencedAgain() 520 if (!visited_blocks_.IsBitSet(dominated_block->GetBlockId())) { in WillBeReferencedAgain() 526 if (!visited_blocks_.IsBitSet(successor->GetBlockId())) { in WillBeReferencedAgain()
|
D | intrinsics.h | 159 bool Get##name() const { return IsBitSet(k##name); } \ 175 bool IsBitSet(uint32_t bit) const { in IsBitSet() function 196 bool Get##name() const { return IsBitSet(k##name); } \
|
D | ssa_phi_elimination.cc | 182 if (!visited_phis_in_cycle.IsBitSet(input->GetId())) { in Run() 233 if (user->IsPhi() && !visited_phis_in_cycle.IsBitSet(user->GetId())) { in Run()
|
D | dead_code_elimination.cc | 37 DCHECK(visited->IsBitSet(block_id)); in MarkReachableBlocks() 77 if (!visited->IsBitSet(successor->GetBlockId())) { in MarkReachableBlocks() 321 if (!live_blocks.IsBitSet(id)) { in RemoveDeadBlocks()
|
D | register_allocator.cc | 131 if (liveness_of_spill_slot->IsBitSet(j)) { in ValidateIntervals() 153 if (liveness_of_register->IsBitSet(j)) { in ValidateIntervals()
|
D | licm.cc | 109 DCHECK(visited->IsBitSet(inner->GetBlockId())); in Run()
|
D | load_store_analysis.h | 260 return aliasing_matrix_.IsBitSet(AliasingMatrixPosition(index1, index2)); in MayAlias() 262 return aliasing_matrix_.IsBitSet(AliasingMatrixPosition(index2, index1)); in MayAlias()
|
D | liveness_test.cc | 41 buffer << vector->IsBitSet(i); in DumpBitVector()
|
D | nodes.cc | 81 if (visiting.IsBitSet(successor_id)) { in FindBackEdges() 84 } else if (!visited->IsBitSet(successor_id)) { in FindBackEdges() 112 if (!visited.IsBitSet(i)) { in RemoveInstructionsAsUsersFromDeadBlocks() 125 if (!visited.IsBitSet(i)) { in RemoveDeadBlocks() 582 if (blocks_.IsBitSet(block->GetBlockId())) { in PopulateRecursive() 606 if (finalized->IsBitSet(block_id)) { in PopulateIrreducibleRecursive() 619 if (blocks_.IsBitSet(pre_header->GetBlockId())) { in PopulateIrreducibleRecursive() 635 if (!is_finalized && blocks_.IsBitSet(predecessor->GetBlockId())) { in PopulateIrreducibleRecursive() 712 return blocks_.IsBitSet(block.GetBlockId()); in Contains() 716 return other.blocks_.IsBitSet(header_->GetBlockId()); in IsIn() [all …]
|
D | stack_map_stream.cc | 525 if (live_dex_registers_mask->IsBitSet(reg)) { in CheckDexRegisterMap() 595 stack_mask.StoreBit(i, stack_map.sp_mask != nullptr && stack_map.sp_mask->IsBitSet(i)); in PrepareStackMasks() 625 DCHECK_EQ(stack_mask.LoadBit(b), entry.sp_mask->IsBitSet(b)); in CheckCodeInfo()
|
D | graph_checker.cc | 318 if (seen_ids_.IsBitSet(instruction->GetId())) { in VisitInstruction() 623 if (!loop_blocks.IsBitSet(back_edge_id)) { in HandleLoop() 692 if (visited->IsBitSet(insn1->GetId())) { in IsConstantEquivalent()
|
D | find_loops_test.cc | 126 ASSERT_TRUE(blocks.IsBitSet(blocks_in_loop[i])); in TestBlock()
|
D | register_allocator_graph_color.cc | 2009 if (taken.IsBitSet(s)) { in ColorSpillSlots() 2038 DCHECK(taken.IsBitSet(s)); in ColorSpillSlots()
|
D | stack_map_test.cc | 39 if (stack_mask.LoadBit(i) != bit_vector.IsBitSet(i)) { in CheckStackMask()
|
D | nodes.h | 6937 if (!blocks_in_loop_.IsBitSet(index_)) { in HBlocksInLoopIterator() 6947 if (blocks_in_loop_.IsBitSet(index_)) { in Advance() 6970 if (!blocks_in_loop_.IsBitSet(blocks_[index_]->GetBlockId())) { in HBlocksInLoopReversePostOrderIterator() 6980 if (blocks_in_loop_.IsBitSet(blocks_[index_]->GetBlockId())) { in Advance()
|
D | instruction_builder.cc | 337 if (native_debuggable && native_debug_info_locations->IsBitSet(dex_pc)) { in Build()
|
/art/runtime/base/ |
D | bit_vector_test.cc | 37 EXPECT_FALSE(bv.IsBitSet(i)); in TEST() 50 EXPECT_TRUE(bv.IsBitSet(0)); in TEST() 52 EXPECT_FALSE(bv.IsBitSet(i)); in TEST() 54 EXPECT_TRUE(bv.IsBitSet(kBits - 1)); in TEST() 166 EXPECT_TRUE(first.IsBitSet(64)); in TEST()
|
D | bit_vector.h | 170 bool IsBitSet(uint32_t idx) const { in IsBitSet() function 173 return (idx < (storage_size_ * kWordBits)) && IsBitSet(storage_, idx); in IsBitSet() 252 static bool IsBitSet(const uint32_t* storage, uint32_t idx) { in IsBitSet() function
|
D | bit_vector.cc | 343 buffer << IsBitSet(i); in DumpHelper()
|
/art/runtime/verifier/ |
D | reg_type_test.cc | 506 EXPECT_TRUE(unresolved_parts.IsBitSet(ref_type_0.GetId())); in TEST_F() 507 EXPECT_TRUE(unresolved_parts.IsBitSet(ref_type_1.GetId())); in TEST_F()
|
/art/runtime/ |
D | oat_file.cc | 1575 if (!BitVector::IsBitSet(bitmap_, method_index)) { in GetOatMethodOffsets()
|
D | stack_map.h | 574 region_.StoreBit(live_bit_mask_offset_in_bits + i, live_dex_registers_mask.IsBitSet(i)); in SetLiveBitMask()
|
/art/compiler/driver/ |
D | compiler_driver.cc | 512 if (driver->GetCurrentDexToDexMethods()->IsBitSet(method_idx)) { in CompileMethod() 706 DCHECK(current_dex_to_dex_methods_->IsBitSet(method_idx)); in CompileOne()
|