/external/tensorflow/tensorflow/compiler/xla/service/ |
D | fusion_node_indexing_evaluation_test.cc | 93 HloInstruction* add = sub->mutable_operand(0); in TEST_F() 109 HloInstruction* slice1 = sub->mutable_operand(0); in TEST_F() 110 HloInstruction* slice2 = sub->mutable_operand(1); in TEST_F() 149 HloInstruction* slice2_0 = add2->mutable_operand(0); in TEST_F() 150 HloInstruction* slice2_1 = add2->mutable_operand(1); in TEST_F() 157 HloInstruction* add1 = fusion->mutable_operand(0); in TEST_F() 162 HloInstruction* slice1_0 = fusion->mutable_operand(0); in TEST_F() 168 HloInstruction* slice1_1 = fusion->mutable_operand(0); in TEST_F() 174 HloInstruction* add0 = fusion->mutable_operand(0); in TEST_F() 205 HloInstruction* add0 = fusion->mutable_operand(0); in TEST_F()
|
D | dynamic_index_splitter.cc | 55 dynamic_op, dynamic_op->mutable_operand(1))); in Run() 58 dynamic_op, dynamic_op->mutable_operand(0))); in Run() 66 auto index_operand = dynamic_op->mutable_operand(index_operand_number); in Run() 85 dynamic_op->shape(), dynamic_op->mutable_operand(0), in Run() 86 dynamic_op->mutable_operand(1), absl::MakeSpan(index_array)) in Run() 88 dynamic_op->shape(), dynamic_op->mutable_operand(0), in Run()
|
D | transpose_folding.cc | 113 HloInstruction* new_lhs = dot->mutable_operand(0); in FoldTransposeIntoDot() 114 HloInstruction* new_rhs = dot->mutable_operand(1); in FoldTransposeIntoDot() 128 new_lhs = new_lhs->mutable_operand(0); in FoldTransposeIntoDot() 135 new_rhs = new_rhs->mutable_operand(0); in FoldTransposeIntoDot() 163 HloInstruction& transpose = *convolution.mutable_operand(kLhsIdx); in FoldTransposeIntoConvolution() 165 HloInstruction& transpose_operand = *transpose.mutable_operand(0); in FoldTransposeIntoConvolution() 180 new_lhs = convolution.mutable_operand(kLhsIdx); in FoldTransposeIntoConvolution() 186 HloInstruction& transpose = *convolution.mutable_operand(kRhsIdx); in FoldTransposeIntoConvolution() 188 HloInstruction& transpose_operand = *transpose.mutable_operand(0); in FoldTransposeIntoConvolution() 203 new_rhs = convolution.mutable_operand(kRhsIdx); in FoldTransposeIntoConvolution()
|
D | dynamic_padder.cc | 69 return inst->mutable_operand(2); in ChooseIdentityValue() 93 return inst->mutable_operand(init_value_index); in ChooseIdentityValue() 102 return inst->mutable_operand(1); in ChooseIdentityValue() 114 return inst->mutable_operand(1); in ChooseIdentityValue() 168 HloInstruction* operand = instr->mutable_operand(0); in ReplaceGetSize() 199 HloInstruction* operand = instr->mutable_operand(0); in ReplaceSetSize() 215 HloInstruction* operand = instr->mutable_operand(0); in ReplaceSetBound() 452 operand_shape, reshape->mutable_operand(0), operand_static_dim_size, in RewriteDynamicReshapeSplitInput() 809 HloInstruction* grad = custom_call_conv->mutable_operand(1); in RewriteDynamicConvolutionInputGrad() 810 HloInstruction* kernel = custom_call_conv->mutable_operand(2); in RewriteDynamicConvolutionInputGrad() [all …]
|
D | space_to_batch_converter.cc | 639 if (old_to_new_instrs_.contains(instr->mutable_operand(0))) { in Run() 640 producer = instr->mutable_operand(0); in Run() 641 } else if (old_to_new_instrs_.contains(instr->mutable_operand(1))) { in Run() 642 producer = instr->mutable_operand(1); in Run() 658 if (old_to_new_instrs_.count(instr->mutable_operand(i))) { in Run() 660 BatchToSpace(instr->mutable_operand(i))); in Run() 689 auto old_producer = consumer->mutable_operand(i); in CanPropagate() 760 if (!old_to_new_instrs_.contains(consumer->mutable_operand(0))) { in CanPropagate() 763 auto dim_map_val_op_0 = instr_to_dim_map_[consumer->mutable_operand(0)]; in CanPropagate() 800 auto activations = consumer->mutable_operand(0); in CanPropagate() [all …]
|
D | batchnorm_expander.cc | 172 HloInstruction* operand = batch_norm->mutable_operand(0); in HandleBatchNormTraining() 177 HloInstruction* scale = batch_norm->mutable_operand(1); in HandleBatchNormTraining() 178 HloInstruction* offset = batch_norm->mutable_operand(2); in HandleBatchNormTraining() 299 HloInstruction* operand = batch_norm->mutable_operand(0); in HandleBatchNormInference() 304 HloInstruction* scale = batch_norm->mutable_operand(1); in HandleBatchNormInference() 305 HloInstruction* offset = batch_norm->mutable_operand(2); in HandleBatchNormInference() 306 HloInstruction* mean = batch_norm->mutable_operand(3); in HandleBatchNormInference() 307 HloInstruction* var = batch_norm->mutable_operand(4); in HandleBatchNormInference() 410 HloInstruction* activation = batch_norm->mutable_operand(0); in HandleBatchNormGrad() 413 HloInstruction* scale = batch_norm->mutable_operand(1); in HandleBatchNormGrad() [all …]
|
D | algebraic_simplifier.cc | 210 instr, operand->mutable_operand(0), options); in BitcastingOperandOfReshapeOrCopyChainHelper() 225 instr, instr->mutable_operand(0), options); in BitcastingOperandOfReshapeOrCopyChain() 622 auto lhs = dot->mutable_operand(0); in ScalarMultiplyReduction() 623 auto rhs = dot->mutable_operand(1); in ScalarMultiplyReduction() 769 operand = instruction->mutable_operand(0); in ReplaceWithBitcast() 792 HloInstruction* abs_operand = abs->mutable_operand(0); in HandleAbs() 882 sliced_lhs, rhs->mutable_operand(1))); in HandleAdd() 1167 ReplaceInstructionIfSameShape(bitcast, bitcast->mutable_operand(0)); in HandleBitcast() 1174 ReplaceInstructionIfSameShape(bitcast, bitcast->mutable_operand(0)); in HandleBitcastConvert() 1186 if (ReplaceInstructionIfSameShape(copy, copy->mutable_operand(0))) { in HandleCopy() [all …]
|
D | while_loop_all_reduce_code_motion_test.cc | 104 transformed_while->mutable_operand(0)->mutable_operand(3); in TEST_F() 115 moved_all_reduce->mutable_operand(0)) in TEST_F() 398 transformed_while->mutable_operand(0)->mutable_operand(3); in TEST_F() 569 transformed_while->mutable_operand(0)->mutable_operand(3); in TEST_F() 650 transformed_while->mutable_operand(0)->mutable_operand(3); in TEST_F()
|
D | hlo_query.cc | 82 *matching_operand = instruction->mutable_operand(0); in MatchBinaryInstructionOperand() 83 *other_operand = instruction->mutable_operand(1); in MatchBinaryInstructionOperand() 87 *matching_operand = instruction->mutable_operand(1); in MatchBinaryInstructionOperand() 88 *other_operand = instruction->mutable_operand(0); in MatchBinaryInstructionOperand()
|
D | tuple_simplifier.cc | 42 HloInstruction* operand = tuple->mutable_operand(operand_number); in RemoveWholeTuple() 49 top_tuple = operand->mutable_operand(0); in RemoveWholeTuple() 110 ancestor.first->mutable_operand(ancestor.second[0]))); in Run()
|
D | conditional_to_select.cc | 47 conditional->shape(), {conditional->mutable_operand(1)}, in DoConditionalToSelect() 52 conditional->shape(), {conditional->mutable_operand(2)}, in DoConditionalToSelect() 55 HloInstruction* condition = conditional->mutable_operand(0); in DoConditionalToSelect()
|
D | all_reduce_simplifier.cc | 84 all_reduce, all_reduce->mutable_operand(0))); in Run() 115 all_reduce->mutable_operand(0), multiplier)); in Run() 122 replacement = all_reduce->mutable_operand(0); in Run()
|
D | while_loop_all_reduce_code_motion.cc | 111 instruction = instruction->mutable_operand(0); in IsAllReduceMovable() 121 instruction = instruction->mutable_operand(0); in IsAllReduceMovable() 308 user->mutable_operand(buffer_index); in IsAllReduceMovable() 359 HloInstruction* old_while_init = old_while_instruction->mutable_operand(0); in CreateNewWhileInit() 370 HloInstruction* old_buffer = old_while_init->mutable_operand(tuple_index); in CreateNewWhileInit() 381 new_while_init_elements[i] = old_while_init->mutable_operand(i); in CreateNewWhileInit() 383 tuple_index_to_old_buffer[i] = old_while_init->mutable_operand(i); in CreateNewWhileInit() 594 all_reduce, all_reduce->mutable_operand(0))); in Run()
|
D | convert_operand_folding.cc | 51 auto* operand = instruction->mutable_operand(i); in ExpandInstruction() 54 i, operand->mutable_operand(0))); in ExpandInstruction()
|
D | logistic_expander.cc | 42 HloInstruction* operand = logistic->mutable_operand(0); in ExpandLogisticWithTanh() 58 HloInstruction* operand = logistic->mutable_operand(0); in ExpandLogisticWithExp()
|
D | hlo_domain_remover.cc | 82 operand->ReplaceAllUsesWith(operand->mutable_operand(0))); in Run() 91 computation->set_root_instruction(root->mutable_operand(0)); in Run()
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
D | reduction_splitter.cc | 46 HloInstruction *operand = reduce->mutable_operand(0); in HandleReduce() 87 pre_reduce_shape, reduce->mutable_operand(0), in HandleReduce() 88 reduce->mutable_operand(1), pre_reduce_dims, reduce->to_apply()); in HandleReduce() 105 reduce->mutable_operand(1), final_reduce_dims, reduce->to_apply()); in HandleReduce()
|
D | gpu_conv_rewriter.cc | 73 auto lhs = conv->mutable_operand(0); in ConvertBatchGroupedToFeatureGroupedConvolution() 74 auto rhs = conv->mutable_operand(1); in ConvertBatchGroupedToFeatureGroupedConvolution() 329 HloInstruction* lhs = conv->mutable_operand(0); in MatchBackwardFilter() 343 HloInstruction* reverse_filter = conv->mutable_operand(1); in MatchBackwardInput() 540 rhs = rhs->mutable_operand(0); in MatchBackwardInput() 626 conv->mutable_operand(0), rhs, window, dnums, in CreateCustomCallHelper() 633 conv->mutable_operand(1), window, dnums, in CreateCustomCallHelper() 657 HloInstruction* input_convert = conv->mutable_operand(0); in CreateCustomCallHelper() 658 HloInstruction* kernel_convert = conv->mutable_operand(1); in CreateCustomCallHelper() 669 0, input_convert->mutable_operand(0))); in CreateCustomCallHelper() [all …]
|
D | gemm_rewriter.cc | 52 HloInstruction *lhs = instr->mutable_operand(0); in HandleDot() 53 HloInstruction *rhs = instr->mutable_operand(1); in HandleDot() 115 {existing_gemm->mutable_operand(0), in HandleAdd() 116 existing_gemm->mutable_operand(1), bias}, in HandleAdd()
|
D | cudnn_batchnorm_rewriter.cc | 97 operands[0] = AddConvert(batch_norm->mutable_operand(0), F16); in HandleBatchNormInference() 148 operands[0] = AddConvert(batch_norm->mutable_operand(0), F16); in HandleBatchNormTraining() 238 batch_norm->mutable_operand(3), in HandleBatchNormGrad() 249 HloInstruction* operand_0_convert = batch_norm->mutable_operand(0); in HandleBatchNormGrad() 256 batch_norm->mutable_operand(index)->shape())) { in HandleBatchNormGrad() 257 operands[index] = AddConvert(batch_norm->mutable_operand(index), F16); in HandleBatchNormGrad()
|
D | tree_reduction_rewriter.cc | 76 HloInstruction *input = hlo->mutable_operand(0); in RewriteReduction() 77 HloInstruction *initial_value = hlo->mutable_operand(1); in RewriteReduction() 214 inner_reduce_shape, input, hlo->mutable_operand(1), in RewriteBatchDimensionLargerThanTile() 218 hlo->shape(), inner_reduce, hlo->mutable_operand(1), {0}, in RewriteBatchDimensionLargerThanTile()
|
D | gpu_conv_padding_legalization.cc | 160 conv->mutable_operand(0)); in CanonicalizeForwardConvolution() 163 conv->mutable_operand(1)); in CanonicalizeForwardConvolution() 218 HloInstruction* input = backward_conv->mutable_operand(0); in CanonicalizeBackwardFilterConvolution() 252 HloInstruction* output = backward_conv->mutable_operand(1); in CanonicalizeBackwardFilterConvolution() 328 HloInstruction* output = backward_conv->mutable_operand(0); in CanonicalizeBackwardInputConvolution() 329 HloInstruction* filter = backward_conv->mutable_operand(1); in CanonicalizeBackwardInputConvolution()
|
D | reduction_dimension_grouper.cc | 49 HloInstruction *operand = reduce->mutable_operand(0); in HandleReduce() 93 reduce->shape(), reduce_input_grouped, reduce->mutable_operand(1), in HandleReduce()
|
D | reduction_degenerate_dim_remover.cc | 40 HloInstruction *reduced_op = instr->mutable_operand(0); in HandleReduce() 78 canonical_reduce_shape, input_reshape, instr->mutable_operand(1), in HandleReduce()
|
/external/tensorflow/tensorflow/compiler/xla/tests/ |
D | reduce_hlo_test.cc | 79 module->entry_computation()->root_instruction()->mutable_operand(0); in XLA_TEST_P() 89 reduce_instruction->mutable_operand(0)->mutable_shape(); in XLA_TEST_P()
|