/external/tensorflow/tensorflow/compiler/xla/service/ |
D | hlo_subcomputation_unification_test.cc | 88 EXPECT_NE(x->to_apply(), y->to_apply()); in TEST_F() 91 EXPECT_EQ(x->to_apply(), y->to_apply()); in TEST_F() 117 EXPECT_NE(x->to_apply(), y->to_apply()); in TEST_F() 120 EXPECT_EQ(x->to_apply(), y->to_apply()); in TEST_F() 147 EXPECT_NE(x->to_apply(), y->to_apply()); in TEST_F() 150 EXPECT_NE(x->to_apply(), y->to_apply()); in TEST_F()
|
D | stable_sort_expander_test.cc | 111 root->operand(0)->to_apply()->root_instruction(), /*iota_parameter=*/1); in TEST_F() 157 root->operand(0)->to_apply()->root_instruction(), /*iota_parameter=*/1); in TEST_F() 191 root->operand(0)->operand(0)->to_apply()->root_instruction(), in TEST_F() 255 root->operand(0)->to_apply()->root_instruction(), in TEST_F() 292 root->operand(0)->to_apply()->root_instruction(), in TEST_F() 323 root->operand(0)->to_apply()->root_instruction(), /*iota_parameter=*/1); in TEST_F() 354 root->operand(0)->operand(0)->to_apply()->root_instruction(), in TEST_F()
|
D | hlo_parser_test.cc | 361 ROOT %call = f32[] call(f32[] %constant), to_apply=%Identity.v1 in CreateTestCases() 380 …d, f32[] %constant), window={size=1x1x7x1 stride=1x4x1x1 pad=0_0x0_0x3_3x0_0}, to_apply=%add_F32.v3 in CreateTestCases() 399 …ROOT %reduce-window = f32[] reduce-window(f32[] %constant, f32[] %constant.1), to_apply=%add_F32.v3 in CreateTestCases() 872 …_window_dims={}, scatter_dims_to_operand_dims={0,1,2,3,4}, index_vector_dim=4, to_apply=%add_F32.v3 in CreateTestCases() 1008 ROOT map = f32[4]{0} map(param0, param1), dimensions={0}, to_apply=add_F32.v3 in CreateTestCases() 1027 ROOT reduce = f32[8,16]{1,0} reduce(input, constant), dimensions={2}, to_apply=add_F32.v3 in CreateTestCases() 1053 …(f32[], s32[]) reduce(values, indices, init_value, init_index), dimensions={0}, to_apply=max_argmax in CreateTestCases() 1114 ROOT sorted = f32[1024]{0} sort(x), dimensions={0}, to_apply=compare in CreateTestCases() 1135 ROOT sorted = (f32[1024]{0}, s32[1024]{0}) sort(keys, values), dimensions={0}, to_apply=compare in CreateTestCases() 1153 ROOT sorted = f32[1024,16]{0,1} sort(x), dimensions={0}, to_apply=compare in CreateTestCases() [all …]
|
D | shape_inference_test.cc | 64 ProgramShape to_apply = ShapeUtil::MakeProgramShape({f32_, f32_}, f32_); in ExpectInferredReduceShape() local 66 {&arg, &f32_}, dimensions_to_reduce, to_apply); in ExpectInferredReduceShape() 317 ProgramShape to_apply = ShapeUtil::MakeProgramShape( in TEST_F() local 320 matrix_shape, init_value_shape, window, to_apply); in TEST_F() 567 ProgramShape to_apply = ShapeUtil::MakeProgramShape({f32_}, s32_); in TEST_F() local 568 auto inferred_status = ShapeInference::InferMapShape({&arg}, to_apply, {0}); in TEST_F() 638 ProgramShape to_apply = ShapeUtil::MakeProgramShape({f32_}, f32_); in TEST_F() local 639 auto inferred_status = ShapeInference::InferMapShape({&arg}, to_apply, {0}); in TEST_F() 716 ProgramShape to_apply = ShapeUtil::MakeProgramShape( in TEST_F() local 719 {&f32_arg_shape, &s32_arg_shape, &f32_, &s32_}, {0, 1}, to_apply); in TEST_F() [all …]
|
D | hlo_verifier.cc | 342 HloComputation* compare = sort->to_apply(); in HandleSort() 426 const ProgramShape& to_apply = instruction.to_apply()->ComputeProgramShape(); in SameElementTypesForOperandsAndToApplyParameters() local 428 const Shape& parameter_shape = to_apply.parameters(i); in SameElementTypesForOperandsAndToApplyParameters() 455 reduce->to_apply()->ComputeProgramShape()))); in HandleReduce() 536 CheckParameterCount(call, call->to_apply(), call->operand_count())); in HandleCall() 537 for (int64 i = 0; i < call->to_apply()->num_parameters(); ++i) { in HandleCall() 538 TF_RETURN_IF_ERROR(CheckOperandAndParameter(call, i, call->to_apply(), i)); in HandleCall() 541 return CheckShape(call, call->to_apply()->root_instruction()->shape()); in HandleCall() 614 operand_shapes, map->to_apply()->ComputeProgramShape(), map_dims))); in HandleMap() 628 reduce_window->to_apply()->ComputeProgramShape()))); in HandleReduceWindow() [all …]
|
D | shape_inference.h | 83 absl::Span<const Shape* const> arg_shapes, const ProgramShape& to_apply, 154 const ProgramShape& to_apply); 271 absl::Span<const Shape* const> arg_shapes, const ProgramShape& to_apply);
|
D | hlo_instructions.cc | 573 eq_computations(to_apply(), casted_other.to_apply()) && in IdenticalSlowPath() 583 shape, new_operands, to_apply(), replica_groups(), all_reduce_barrier(), in CloneWithNewOperandsImpl() 760 eq_computations(to_apply(), casted_other.to_apply()); in IdenticalSlowPath() 768 dimensions(), to_apply()); in CloneWithNewOperandsImpl() 814 return eq_computations(to_apply(), other.to_apply()); in IdenticalSlowPath() 821 shape, dimensions(0), new_operands, to_apply(), is_stable()); in CloneWithNewOperandsImpl() 954 return eq_computations(to_apply(), other.to_apply()); in IdenticalSlowPath() 960 return absl::make_unique<HloMapInstruction>(shape, new_operands, to_apply()); in CloneWithNewOperandsImpl() 1982 return eq_computations(to_apply(), casted_other.to_apply()) && in IdenticalSlowPath() 1992 shape, new_operands[0], new_operands[1], window(), to_apply()); in CloneWithNewOperandsImpl() [all …]
|
D | hlo_cost_analysis.cc | 266 ProcessNestedSubcomputation(map->to_apply())); in HandleMap() 279 HloComputation* function = reduce->to_apply(); in HandleReduce() 305 auto function = reduce_window->to_apply(); in HandleReduceWindow() 638 ProcessUnnestedSubcomputation(call->to_apply())); in HandleCall() 727 ProcessNestedSubcomputation(scatter->to_apply())); in HandleScatter()
|
D | shape_inference.cc | 1073 absl::Span<const Shape* const> arg_shapes, const ProgramShape& to_apply, in InferMapShape() argument 1127 if (arg_shapes.size() != to_apply.parameters_size()) { in InferMapShape() 1131 to_apply.parameters_size(), arg_shapes.size()); in InferMapShape() 1135 const Shape& output_shape = to_apply.result(); in InferMapShape() 1142 for (int i = 0; i < to_apply.parameters_size(); ++i) { in InferMapShape() 1143 const Shape& parameter_shape = to_apply.parameters(i); in InferMapShape() 2045 const ProgramShape& to_apply) { in InferReduceShape() argument 2085 TF_RETURN_IF_ERROR(VerifyReducerShape(to_apply, init_values, element_types, in InferReduceShape() 2099 if (ShapeUtil::IsScalar(to_apply.result())) { in InferReduceShape() 2100 return ShapeUtil::MakeShape(to_apply.result().element_type(), in InferReduceShape() [all …]
|
D | stable_sort_expander.cc | 79 auto comparator = sort->to_apply(); in ExpandInstruction() 130 auto comparator = sort->to_apply(); in ExpandInstruction()
|
D | map_inliner.cc | 67 HloComputation* function = map->to_apply(); in HandleMap()
|
D | flatten_call_graph.cc | 47 CHECK_EQ(instruction->to_apply(), computation); in ReplaceCalledComputation()
|
D | sort_simplifier.cc | 55 auto comparator = sort->to_apply(); in RemoveUnusedOperandFromSort()
|
D | bfloat16_normalization_test.cc | 329 EXPECT_EQ(sort->to_apply()->parameter_instruction(1)->shape().element_type(), in TEST_F() 333 auto users = sort->to_apply()->parameter_instruction(1)->users(); in TEST_F()
|
D | hlo_parser.cc | 819 optional<HloComputation*> to_apply; in ParseInstructionRhs() local 824 &to_apply}; in ParseInstructionRhs() 838 shape, operands, *to_apply, replica_groups, barrier ? *barrier : "", in ParseInstructionRhs() 923 optional<HloComputation*> to_apply; in ParseInstructionRhs() local 925 &to_apply}; in ParseInstructionRhs() 932 to_apply.value(), is_stable.value())); in ParseInstructionRhs() 1036 optional<HloComputation*> to_apply; in ParseInstructionRhs() local 1038 &to_apply}; in ParseInstructionRhs() 1043 HloInstruction::CreateCall(shape, operands, *to_apply)); in ParseInstructionRhs() 1180 optional<HloComputation*> to_apply; in ParseInstructionRhs() local [all …]
|
D | tuple_points_to_analysis.cc | 801 auto* param = user->to_apply()->parameter_instruction(operand_indices[0]); in CanShareOperandBufferWithUser() 812 auto* callee_root = user->to_apply()->root_instruction(); in CanShareOperandBufferWithUser()
|
D | hlo_module_test.cc | 287 HloComputation* reduction = root->to_apply(); in TEST_F()
|
D | hlo_ordering.cc | 226 call->to_apply())) { in UseIsBeforeValueDefinition()
|
D | scatter_expander.cc | 303 scatter->to_apply())); in ScatterLoopBody()
|
D | hlo_dataflow_analysis.cc | 366 GetInstructionValueSet(call->to_apply()->root_instruction()); in UpdateCallValueSet() 1132 auto* callee_root = user->to_apply()->root_instruction(); in CanShareOperandBufferWithUser()
|
D | hlo_instruction.cc | 1450 clone = CreateCall(shape, new_operands, to_apply()); in CloneWithNewOperands() 1784 return eq_computations(to_apply(), other.to_apply()); in IdenticalSlowPath() 1988 HloComputation* HloInstruction::to_apply() const { in to_apply() function in xla::HloInstruction 2313 StrCat("to_apply=", PrintName(to_apply()->name(), options))); in ExtraAttributesToString() 2360 StrCat("to_apply=\n", to_apply()->ToString(new_options))); in ExtraAttributesToString()
|
/external/tensorflow/tensorflow/compiler/xla/service/cpu/ |
D | parallel_task_assignment.cc | 197 changed |= AssignParallelTasksHelper(module, instruction->to_apply(), in AssignParallelTasksHelper() 224 auto* new_root = call->to_apply()->root_instruction(); in AssignParallelTasksHelper()
|
D | ir_emitter.cc | 576 auto less_than_function = FindOrDie(emitted_functions_, sort->to_apply()); in HandleSort() 577 CHECK(absl::c_binary_search(thread_local_computations_, sort->to_apply())); in HandleSort() 637 return EmitThreadLocalCall(*map_instr.to_apply(), elemental_operands, name); in EmitElementalMap() 720 *reduce_window->to_apply(), {Load(accumulator_address), input_value}, in EmitElementalReduceWindow() 1864 *reduce->to_apply(), {Load(accumulator_addr), input_element}, in EmitElementalReduce() 1880 HloComputation* function = reduce->to_apply(); in HandleReduce() 2203 HloComputation* computation = call->to_apply(); in HandleCall()
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
D | elemental_ir_emitter.cc | 337 TF_RET_CHECK(hlo->to_apply()->num_parameters() > 0); in MakeElementGenerator() 344 return compute_nested_(*hlo->to_apply(), operand_elements); in MakeElementGenerator() 434 compute_nested_(*hlo->to_apply(), {Load(accum_ptr), input_value})); in MakeElementGenerator() 479 compute_nested_(*hlo->to_apply(), in MakeElementGenerator()
|
D | ir_emitter.cc | 672 HloComputation* function = reduce->to_apply(); in HandleReduce() 740 return EmitCallToNestedComputation(*call->to_apply(), operand_addresses, in HandleCall()
|