/external/tensorflow/tensorflow/compiler/xla/service/ |
D | bfloat16_support.cc | 28 case HloOpcode::kCustomCall: in SupportsBF16Operand() 47 case HloOpcode::kCustomCall: in SupportsBF16Output() 66 case HloOpcode::kCustomCall: in SupportsMixedPrecisions()
|
D | hlo_opcode_test.cc | 58 case HloOpcode::kCustomCall: in TEST()
|
D | hlo_matchers.h | 95 : HloMatcher(HloOpcode::kCustomCall, operands), in HloCustomCallMatcher() 373 HloOpcode::kCustomCall, {operands_first, operands_rest...})); in CustomCall() 377 new ::xla::testing::HloMatcher(HloOpcode::kCustomCall, {})); in CustomCall()
|
D | dynamic_padder_test.cc | 47 if (hlo->opcode() != HloOpcode::kCustomCall) { in OpHasDynamismSupport() 186 ASSERT_THAT(slice_to_dynamic->opcode(), HloOpcode::kCustomCall); in TEST_F() 190 ASSERT_THAT(pad_to_static->opcode(), HloOpcode::kCustomCall); in TEST_F() 193 ASSERT_THAT(slice_to_dynamic->opcode(), HloOpcode::kCustomCall); in TEST_F()
|
D | hlo_opcode.h | 80 V(kCustomCall, "custom-call", kHloOpcodeIsVariadic) \
|
D | hlo.proto | 138 // kCustomCall. 231 // Whether the kCustomCall instruction has side-effects, only present for 232 // kCustomCall. 236 // buffers between output and operands for kCustomCall.
|
D | hlo_element_type_converter.cc | 138 if (opcode == HloOpcode::kCustomCall) { in Run()
|
D | bfloat16_conversion_folding.cc | 173 hlo->opcode() == HloOpcode::kCustomCall || // in DefaultAction()
|
D | hlo_dataflow_analysis.cc | 437 CHECK_EQ(custom_call->opcode(), HloOpcode::kCustomCall); in UpdateCustomCallValueSet() 777 case HloOpcode::kCustomCall: in UpdateInstructionValueSet() 1040 case HloOpcode::kCustomCall: { in InitializeInstructionValueSets()
|
D | hlo_instruction.cc | 549 case HloOpcode::kCustomCall: { in CreateFromProto() 1539 case HloOpcode::kCustomCall: in HasSideEffectNoRecurse() 1696 case HloOpcode::kCustomCall: in CloneWithNewOperands() 2234 case HloOpcode::kCustomCall: in IdenticalSlowPath() 2437 case HloOpcode::kCustomCall: in to_apply() 2458 case HloOpcode::kCustomCall: in set_to_apply() 2830 } else if (opcode() == HloOpcode::kCustomCall) { in ExtraAttributesToString() 2999 return opcode() == HloOpcode::kCustomCall && custom_call_target() == target; in IsCustomCall() 3238 case HloOpcode::kCustomCall: in Visit()
|
D | hlo_replication_analysis.cc | 132 if (hlo->opcode() == HloOpcode::kCustomCall && in DetermineHloInstructionIsReplicated()
|
D | call_graph.cc | 69 case HloOpcode::kCustomCall: in GetInstructionCallContext()
|
D | qr_expander.cc | 419 return instruction->opcode() == HloOpcode::kCustomCall && in InstructionMatchesPattern()
|
D | bfloat16_normalization.cc | 501 hlo->opcode() == HloOpcode::kCustomCall || // in DefaultAction()
|
D | instruction_fusion.cc | 156 case HloOpcode::kCustomCall: in IsExpensive()
|
D | hlo_verifier.cc | 51 case HloOpcode::kCustomCall: in IsCallerInstruction() 1086 case HloOpcode::kCustomCall: in CheckMixedPrecisionOperands() 1200 case HloOpcode::kCustomCall: in CheckShape()
|
D | hlo_instructions.cc | 2341 : HloInstruction(HloOpcode::kCustomCall, shape), in HloCustomCallInstruction() 2358 : HloInstruction(HloOpcode::kCustomCall, shape), in HloCustomCallInstruction() 2376 : HloInstruction(HloOpcode::kCustomCall, shape), in HloCustomCallInstruction() 2396 : HloInstruction(HloOpcode::kCustomCall, shape), in HloCustomCallInstruction()
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
D | gpu_layout_assignment.cc | 314 if (instruction->opcode() == HloOpcode::kCustomCall && in PropagateOperandConstraint() 326 if (instruction->opcode() == HloOpcode::kCustomCall && in PropagateOperandConstraint() 341 if (instruction->opcode() == HloOpcode::kCustomCall && in PropagateOperandConstraint() 372 if (instruction->opcode() == HloOpcode::kCustomCall && in PropagateBufferConstraint() 379 if (instruction->opcode() == HloOpcode::kCustomCall && in PropagateBufferConstraint() 386 if (instruction->opcode() == HloOpcode::kCustomCall && in PropagateBufferConstraint()
|
D | thunk.cc | 51 case Thunk::kCustomCall: in ThunkKindToString()
|
D | custom_call_thunk.cc | 34 : Thunk(Thunk::kCustomCall, thunk_info), in CustomCallThunk()
|
D | cudnn_fused_conv_rewriter.cc | 88 >e, Op(&conv_instr).WithOpcode(HloOpcode::kCustomCall), 0); in FindConvWithRelu() 322 >e, Op(&conv_instr).WithOpcode(HloOpcode::kCustomCall), 0), in FindConvWithClampAndConvertToInt8() 443 .WithOpcode(HloOpcode::kCustomCall) in FindConvWithConvertToFloat()
|
D | thunk.h | 55 kCustomCall, enumerator
|
D | ir_emission_utils.cc | 129 return hlo.opcode() == HloOpcode::kCustomCall && in IsCublasGemm() 172 if (hlo.opcode() != HloOpcode::kCustomCall) { in IsCustomCallToDnnBatchNorm() 191 if (hlo.opcode() != HloOpcode::kCustomCall) { in IsCustomCallToDnnConvolution() 204 if (hlo.opcode() != HloOpcode::kCustomCall) { in IsCustomCallToCusolver()
|
D | nvptx_compiler.cc | 195 if (user->opcode() == HloOpcode::kCustomCall && in CanShareBufferHint()
|
/external/tensorflow/tensorflow/compiler/xla/client/ |
D | xla_builder.cc | 143 if (opcode == HloOpcode::kCustomCall && in InstrIsSetBound() 352 case HloOpcode::kCustomCall: in IsConstantVisitor() 1546 return AddInstruction(std::move(instr), HloOpcode::kCustomCall, in DynamicConvInputGrad() 1572 return AddInstruction(std::move(instr), HloOpcode::kCustomCall, in DynamicConvKernelGrad() 1594 return AddInstruction(std::move(instr), HloOpcode::kCustomCall, {lhs, rhs}); in DynamicConvForward() 1955 return AddInstruction(std::move(instr), HloOpcode::kCustomCall, operands); in CustomCallInternal() 2016 return AddInstruction(std::move(instr), HloOpcode::kCustomCall, operands); in CustomCall() 2541 return AddInstruction(std::move(instr), HloOpcode::kCustomCall, args); in ReduceWindow() 3062 return AddInstruction(std::move(instr), HloOpcode::kCustomCall, in SelectAndScatter() 3592 case HloOpcode::kCustomCall: in BuildDynamicInferenceGraph()
|