/art/compiler/optimizing/ |
D | data_type.h | 40 kInt64, enumerator 65 case Type::kInt64: in SizeShift() 92 case Type::kInt64: in Size() 119 case Type::kInt64: in IsIntegralType() 127 return type == Type::kInt32 || type == Type::kInt64; in IsIntOrLongType() 131 return type == Type::kUint64 || type == Type::kInt64 || type == Type::kFloat64; in Is64BitType() 155 case Type::kInt64: in Kind() 156 return Type::kInt64; in Kind() 180 case Type::kInt64: in MinValueOfIntegralType() 206 case Type::kInt64: in MaxValueOfIntegralType() [all …]
|
D | parallel_move_test.cc | 294 DataType::Type::kInt64, in TYPED_TEST() 306 DataType::Type::kInt64, in TYPED_TEST() 323 DataType::Type::kInt64, in TYPED_TEST() 353 DataType::Type::kInt64, in TYPED_TEST() 374 DataType::Type::kInt64, in TYPED_TEST() 395 DataType::Type::kInt64, in TYPED_TEST() 420 DataType::Type::kInt64, in TYPED_TEST() 425 DataType::Type::kInt64, in TYPED_TEST() 440 DataType::Type::kInt64, in TYPED_TEST() 445 DataType::Type::kInt64, in TYPED_TEST() [all …]
|
D | scheduler_arm.cc | 34 case DataType::Type::kInt64: in HandleBinaryOperationLantencies() 61 case DataType::Type::kInt64: in VisitMul() 77 case DataType::Type::kInt64: in HandleBitwiseOperationLantencies() 108 case DataType::Type::kInt64: { in VisitRor() 142 case DataType::Type::kInt64: in HandleShiftLatencies() 207 DCHECK_EQ(condition->GetLeft()->GetType(), DataType::Type::kInt64); in HandleGenerateLongTestConstant() 273 DCHECK_EQ(condition->GetLeft()->GetType(), DataType::Type::kInt64); in HandleGenerateLongTest() 306 if (type == DataType::Type::kInt64) { in HandleGenerateTest() 320 if (condition->GetLeft()->GetType() == DataType::Type::kInt64) { in CanGenerateTest() 358 DCHECK_EQ(cond->GetLeft()->GetType(), DataType::Type::kInt64); in HandleGenerateEqualLong() [all …]
|
D | data_type-inl.h | 34 case Primitive::kPrimLong: return DataType::Type::kInt64; in DataTypeFromPrimitive() 52 case DataType::Type::kInt64: return Primitive::kPrimLong; in DataTypeToPrimitive() 78 case DataType::Type::kInt64: return 'j'; // Java long (J). in TypeId()
|
D | instruction_simplifier_arm.cc | 93 if (type != DataType::Type::kInt32 && type != DataType::Type::kInt64) { in TryMergeIntoShifterOperand() 129 if (!use->IsAdd() && (!use->IsSub() || use->GetType() != DataType::Type::kInt64)) { in TryMergeIntoShifterOperand() 134 } else if (type == DataType::Type::kInt64 && shift_amount == 1) { in TryMergeIntoShifterOperand() 212 if (type == DataType::Type::kInt64 in VisitArrayGet() 233 if (type == DataType::Type::kInt64 in VisitArraySet()
|
D | code_generator_vector_arm64_neon.cc | 89 case DataType::Type::kInt64: in VisitVecReplicateScalar() 142 case DataType::Type::kInt64: in VisitVecReplicateScalar() 181 case DataType::Type::kInt64: in VisitVecExtractScalar() 204 case DataType::Type::kInt64: in VisitVecExtractScalar() 235 case DataType::Type::kInt64: in CreateVecUnOpLocations() 270 case DataType::Type::kInt64: in VisitVecReduce() 328 case DataType::Type::kInt64: in VisitVecNeg() 367 case DataType::Type::kInt64: in VisitVecAbs() 404 case DataType::Type::kInt64: in VisitVecNot() 423 case DataType::Type::kInt64: in CreateVecBinOpLocations() [all …]
|
D | code_generator_vector_arm64_sve.cc | 91 case DataType::Type::kInt64: in VisitVecReplicateScalar() 143 case DataType::Type::kInt64: in VisitVecReplicateScalar() 179 case DataType::Type::kInt64: in VisitVecExtractScalar() 203 case DataType::Type::kInt64: in VisitVecExtractScalar() 231 case DataType::Type::kInt64: in CreateVecUnOpLocations() 265 case DataType::Type::kInt64: in VisitVecReduce() 324 case DataType::Type::kInt64: in VisitVecNeg() 360 case DataType::Type::kInt64: in VisitVecAbs() 402 case DataType::Type::kInt64: in VisitVecNot() 421 case DataType::Type::kInt64: in CreateVecBinOpLocations() [all …]
|
D | side_effects_test.cc | 32 DataType::Type::kInt64, 173 SideEffects::FieldWriteOfType(DataType::Type::kInt64, /* is_volatile= */ false), in TEST() 176 SideEffects::ArrayWriteOfType(DataType::Type::kInt64), in TEST() 249 SideEffects::ArrayReadOfType(DataType::Type::kInt64).ToString().c_str()); in TEST() 258 s = s.Union(SideEffects::FieldWriteOfType(DataType::Type::kInt64, /* is_volatile= */ false)); in TEST()
|
D | code_generator_vector_x86.cc | 33 case DataType::Type::kInt64: in VisitVecReplicateScalar() 95 case DataType::Type::kInt64: { in VisitVecReplicateScalar() 123 case DataType::Type::kInt64: in VisitVecExtractScalar() 163 case DataType::Type::kInt64: { in VisitVecExtractScalar() 193 case DataType::Type::kInt64: in CreateVecUnOpLocations() 208 if (instruction->GetPackedType() == DataType::Type::kInt64 || in VisitVecReduce() 235 case DataType::Type::kInt64: { in VisitVecReduce() 301 case DataType::Type::kInt64: in VisitVecNeg() 390 case DataType::Type::kInt64: in VisitVecNot() 422 case DataType::Type::kInt64: in CreateVecBinOpLocations() [all …]
|
D | code_generator_vector_x86_64.cc | 39 case DataType::Type::kInt64: in VisitVecReplicateScalar() 90 case DataType::Type::kInt64: in VisitVecReplicateScalar() 120 case DataType::Type::kInt64: in VisitVecExtractScalar() 150 case DataType::Type::kInt64: in VisitVecExtractScalar() 176 case DataType::Type::kInt64: in CreateVecUnOpLocations() 191 if (instruction->GetPackedType() == DataType::Type::kInt64 || in VisitVecReduce() 218 case DataType::Type::kInt64: { in VisitVecReduce() 284 case DataType::Type::kInt64: in VisitVecNeg() 373 case DataType::Type::kInt64: in VisitVecNot() 405 case DataType::Type::kInt64: in CreateVecBinOpLocations() [all …]
|
D | instruction_builder.cc | 1776 bool is_wide = (type == DataType::Type::kInt64) || (type == DataType::Type::kFloat64); in SetupInvokeArguments() 1855 constexpr DataType::Type kInt64 = DataType::Type::kInt64; in BuildSimpleIntrinsic() local 1869 instruction = new (allocator_) HRor(kInt64, /*value=*/ nullptr, /*distance=*/ nullptr); in BuildSimpleIntrinsic() 1877 kInt64, /*first=*/ nullptr, /*second=*/ nullptr, ComparisonBias::kNoBias, dex_pc); in BuildSimpleIntrinsic() 1885 kInt64, /*first=*/ nullptr, graph_->GetLongConstant(0), ComparisonBias::kNoBias, dex_pc); in BuildSimpleIntrinsic() 1944 instruction = new (allocator_) HMin(kInt64, /*left=*/ nullptr, /*right=*/ nullptr, dex_pc); in BuildSimpleIntrinsic() 1956 instruction = new (allocator_) HMax(kInt64, /*left=*/ nullptr, /*right=*/ nullptr, dex_pc); in BuildSimpleIntrinsic() 1968 instruction = new (allocator_) HAbs(kInt64, /*input=*/ nullptr, dex_pc); in BuildSimpleIntrinsic() 2328 DCHECK(type == DataType::Type::kInt32 || type == DataType::Type::kInt64); in BuildCheckedDivRem() 2344 || (type == DataType::Type::kInt64 && second->AsLongConstant()->GetValue() == 0)) { in BuildCheckedDivRem() [all …]
|
D | licm_test.cc | 124 DataType::Type::kInt64, in TEST_F() 151 DataType::Type::kInt64, in TEST_F() 162 DataType::Type::kInt64, in TEST_F()
|
D | data_type_test.cc | 41 CheckConversion<DataType::Type::kInt64, Primitive::kPrimLong>(); in TEST() 70 DataType::Type::kInt64, in TEST()
|
D | intrinsics_arm_vixl.cc | 310 DCHECK((type == DataType::Type::kInt32) || (type == DataType::Type::kInt64)); in GenNumberOfLeadingZeros() 312 if (type == DataType::Type::kInt64) { in GenNumberOfLeadingZeros() 342 GenNumberOfLeadingZeros(invoke, DataType::Type::kInt64, codegen_); in VisitLongNumberOfLeadingZeros() 348 DCHECK((type == DataType::Type::kInt32) || (type == DataType::Type::kInt64)); in GenNumberOfTrailingZeros() 354 if (type == DataType::Type::kInt64) { in GenNumberOfTrailingZeros() 388 GenNumberOfTrailingZeros(invoke, DataType::Type::kInt64, codegen_); in VisitLongNumberOfTrailingZeros() 2016 case DataType::Type::kInt64: in GenerateReverseBytes() 2052 GenerateReverseBytes(assembler, DataType::Type::kInt64, locations->InAt(0), locations->Out()); in VisitLongReverseBytes() 2070 bool is_long = type == DataType::Type::kInt64; in GenBitCount() 2107 GenBitCount(invoke, DataType::Type::kInt64, GetAssembler()); in VisitLongBitCount() [all …]
|
D | nodes_shared.cc | 50 if (result_type == DataType::Type::kInt32 && input_type == DataType::Type::kInt64) { in GetOpInfoFromInstruction()
|
D | code_generator_x86_64.cc | 138 DCHECK_EQ(DataType::Type::kInt64, type_); in EmitNativeCode() 1181 case DataType::Type::kInt64: in GenerateStaticOrDirectCall() 1587 case DataType::Type::kInt64: in SetInForReturnValue() 1847 case DataType::Type::kInt64: in LoadFromMemoryNoReference() 1961 case DataType::Type::kInt64: { in GenerateCompareTest() 2014 case DataType::Type::kInt64: { in GenerateCompareTestAndBranch() 2110 if (type == DataType::Type::kInt64 || DataType::IsFloatingPointType(type)) { in GenerateTestAndBranch() 2304 case DataType::Type::kInt64: in HandleCondition() 2344 case DataType::Type::kInt64: in HandleCondition() 2483 case DataType::Type::kInt64: { in VisitCompare() [all …]
|
D | critical_native_abi_fixup_arm.cc | 47 DataType::Type converted_type = is_double ? DataType::Type::kInt64 : DataType::Type::kInt32; in FixUpArguments()
|
D | code_generator_x86.cc | 1171 case DataType::Type::kInt64: in SetInForReturnValue() 1340 case DataType::Type::kInt64: in GetReturnLocation() 1376 case DataType::Type::kInt64: { in GetNextLocation() 1595 case DataType::Type::kInt64: { in LoadFromMemoryNoBarrier() 1629 if (instr != nullptr && dst_type != DataType::Type::kInt64) { in LoadFromMemoryNoBarrier() 1675 case DataType::Type::kInt64: { in MoveToMemory() 1720 if (dst_type == DataType::Type::kInt64 && !src.IsConstant() && !src.IsFpuRegister()) { in MoveLocation() 1951 case DataType::Type::kInt64: in GenerateCompareTestAndBranch() 1981 cond->InputAt(0)->GetType() != DataType::Type::kInt64 && in AreEflagsSetFrom() 2047 if (type == DataType::Type::kInt64 || DataType::IsFloatingPointType(type)) { in GenerateTestAndBranch() [all …]
|
D | code_generator_arm_vixl.cc | 287 case DataType::Type::kInt64: in GetLoadOperandType() 312 case DataType::Type::kInt64: in GetStoreOperandType() 1191 DCHECK_EQ(instruction->GetType(), DataType::Type::kInt64); in GenerateLongDataProc() 1371 DCHECK_EQ(condition->GetLeft()->GetType(), DataType::Type::kInt64); in GenerateLongTestConstant() 1492 DCHECK_EQ(condition->GetLeft()->GetType(), DataType::Type::kInt64); in GenerateLongTest() 1571 if (type == DataType::Type::kInt64) { in GenerateTest() 1617 DCHECK_EQ(cond->GetLeft()->GetType(), DataType::Type::kInt64); in GenerateEqualLong() 1673 DCHECK_EQ(cond->GetLeft()->GetType(), DataType::Type::kInt64); in GenerateConditionLong() 1742 if (type == DataType::Type::kInt64) { in GenerateConditionIntegralOrNonPrimitive() 1833 if (type == DataType::Type::kInt64) { in CanEncodeConstantAs8BitImmediate() [all …]
|
D | intrinsics_arm64.cc | 267 case DataType::Type::kInt64: in GenerateReverseBytes() 305 GenReverseBytes(invoke->GetLocations(), DataType::Type::kInt64, GetVIXLAssembler()); in VisitLongReverseBytes() 319 DCHECK(type == DataType::Type::kInt32 || type == DataType::Type::kInt64); in GenNumberOfLeadingZeros() 340 GenNumberOfLeadingZeros(invoke->GetLocations(), DataType::Type::kInt64, GetVIXLAssembler()); in VisitLongNumberOfLeadingZeros() 346 DCHECK(type == DataType::Type::kInt32 || type == DataType::Type::kInt64); in GenNumberOfTrailingZeros() 368 GenNumberOfTrailingZeros(invoke->GetLocations(), DataType::Type::kInt64, GetVIXLAssembler()); in VisitLongNumberOfTrailingZeros() 374 DCHECK(type == DataType::Type::kInt32 || type == DataType::Type::kInt64); in GenReverse() 395 GenReverse(invoke->GetLocations(), DataType::Type::kInt64, GetVIXLAssembler()); in VisitLongReverse() 407 VRegister fpr = (type == DataType::Type::kInt64) ? temps.AcquireD() : temps.AcquireS(); in GenBitCount() 420 GenBitCount(invoke, DataType::Type::kInt64, GetVIXLAssembler()); in VisitLongBitCount() [all …]
|
D | intrinsics_x86.cc | 1536 case DataType::Type::kInt64: in GenPeek() 1567 GenPeek(invoke->GetLocations(), DataType::Type::kInt64, GetAssembler()); in VisitMemoryPeekLongNative() 1622 case DataType::Type::kInt64: in GenPoke() 1655 CreateLongIntToVoidLocations(allocator_, DataType::Type::kInt64, invoke); in VisitMemoryPokeLongNative() 1659 GenPoke(invoke->GetLocations(), DataType::Type::kInt64, GetAssembler()); in VisitMemoryPokeLongNative() 1719 case DataType::Type::kInt64: { in GenUnsafeGet() 1773 if (type == DataType::Type::kInt64) { in CreateIntIntIntToIntLocations() 1839 allocator_, invoke, DataType::Type::kInt64, /*is_volatile=*/ false); in VisitJdkUnsafeGetLong() 1842 CreateIntIntIntToIntLocations(allocator_, invoke, DataType::Type::kInt64, /*is_volatile=*/ true); in VisitJdkUnsafeGetLongVolatile() 1845 CreateIntIntIntToIntLocations(allocator_, invoke, DataType::Type::kInt64, /*is_volatile=*/ true); in VisitJdkUnsafeGetLongAcquire() [all …]
|
D | constant_folding_test.cc | 202 DataType::Type::kInt64); in TEST_F() 477 DataType::Type::kInt64); in TEST_F() 541 DataType::Type::kInt64); in TEST_F()
|
D | code_generator.cc | 75 return type == DataType::Type::kInt64; in CheckType() 79 return (DataType::IsIntegralType(type) && type != DataType::Type::kInt64) in CheckType() 83 return (type == DataType::Type::kInt64) || (type == DataType::Type::kFloat64); in CheckType() 86 return DataType::IsIntegralType(type) && (type != DataType::Type::kInt64); in CheckType() 90 return type == DataType::Type::kInt64; in CheckType() 773 (DataType::Is64BitType(field_type) ? DataType::Type::kInt64 in GenerateUnresolvedFieldAccess() 810 case DataType::Type::kInt64: in GenerateUnresolvedFieldAccess() 1452 if (current->GetType() == DataType::Type::kInt64) { in EmitVRegInfo() 1459 if (current->GetType() == DataType::Type::kInt64) { in EmitVRegInfo() 1854 case DataType::Type::kInt64: in ScaleFactorForType()
|
D | intrinsics_x86_64.cc | 200 codegen_->GetInstructionCodegen()->Bswap(invoke->GetLocations()->Out(), DataType::Type::kInt64); in VisitLongReverseBytes() 1734 case DataType::Type::kInt64: in GenPeek() 1764 GenPeek(invoke->GetLocations(), DataType::Type::kInt64, GetAssembler()); in VisitMemoryPeekLongNative() 1812 case DataType::Type::kInt64: in GenPoke() 1849 GenPoke(invoke->GetLocations(), DataType::Type::kInt64, GetAssembler()); in VisitMemoryPokeLongNative() 1908 case DataType::Type::kInt64: in GenUnsafeGet() 2027 GenUnsafeGet(invoke, DataType::Type::kInt64, /*is_volatile=*/ false, codegen_); in VisitJdkUnsafeGetLong() 2030 GenUnsafeGet(invoke, DataType::Type::kInt64, /*is_volatile=*/ true, codegen_); in VisitJdkUnsafeGetLongVolatile() 2033 GenUnsafeGet(invoke, DataType::Type::kInt64, /*is_volatile=*/ true, codegen_); in VisitJdkUnsafeGetLongAcquire() 2115 CreateIntIntIntIntToVoidPlusTempsLocations(allocator_, DataType::Type::kInt64, invoke); in VisitJdkUnsafePutLong() [all …]
|
D | instruction_simplifier.cc | 369 int64_t implicit_mask = (value->GetType() == DataType::Type::kInt64) in VisitShift() 883 DCHECK(type == DataType::Type::kInt32 || type == DataType::Type::kInt64); in NewIntegralAbs() 896 DCHECK(type == DataType::Type::kInt32 || type == DataType::Type::kInt64); in NewIntegralMinMax() 923 to_type == DataType::Type::kInt64); in AreLowerPrecisionArgs() 1187 !(result_type == DataType::Type::kInt64 && input_type == DataType::Type::kFloat32); in IsTypeConversionLossless() 1245 input_type == DataType::Type::kInt64 || in IsTypeConversionForStoringIntoNoWiderFieldOnly() 1246 result_type == DataType::Type::kInt64) { in IsTypeConversionForStoringIntoNoWiderFieldOnly() 1531 input_other->GetType() == DataType::Type::kInt64 && in VisitAnd() 1547 DCHECK_NE(new_and_input->GetType(), DataType::Type::kInt64); in VisitAnd() 1553 new (GetGraph()->GetAllocator()) HTypeConversion(DataType::Type::kInt64, new_and); in VisitAnd() [all …]
|