Home
last modified time | relevance | path

Searched refs:input_layout (Results 1 – 25 of 58) sorted by relevance

123

/external/webrtc/audio/utility/
Dchannel_mixing_matrix_unittest.cc28 for (ChannelLayout input_layout = CHANNEL_LAYOUT_MONO; in TEST() local
29 input_layout <= CHANNEL_LAYOUT_MAX; in TEST()
30 input_layout = static_cast<ChannelLayout>(input_layout + 1)) { in TEST()
37 if (input_layout == CHANNEL_LAYOUT_BITSTREAM || in TEST()
38 input_layout == CHANNEL_LAYOUT_DISCRETE || in TEST()
39 input_layout == CHANNEL_LAYOUT_STEREO_AND_KEYBOARD_MIC || in TEST()
48 ss << "Input Layout: " << input_layout in TEST()
52 input_layout, ChannelLayoutToChannelCount(input_layout), in TEST()
54 const int input_channels = ChannelLayoutToChannelCount(input_layout); in TEST()
87 ChannelLayout input_layout = CHANNEL_LAYOUT_STEREO; in TEST() local
[all …]
Dchannel_mixer_unittest.cc147 for (ChannelLayout input_layout = CHANNEL_LAYOUT_MONO; in TEST_F() local
148 input_layout <= CHANNEL_LAYOUT_MAX; in TEST_F()
149 input_layout = static_cast<ChannelLayout>(input_layout + 1)) { in TEST_F()
156 if (input_layout == CHANNEL_LAYOUT_BITSTREAM || in TEST_F()
157 input_layout == CHANNEL_LAYOUT_DISCRETE || in TEST_F()
158 input_layout == CHANNEL_LAYOUT_STEREO_AND_KEYBOARD_MIC || in TEST_F()
167 ss << "Input Layout: " << input_layout in TEST_F()
170 ChannelMixer mixer(input_layout, output_layout); in TEST_F()
174 ChannelLayoutToChannelCount(input_layout)); in TEST_F()
185 for (ChannelLayout input_layout = CHANNEL_LAYOUT_MONO; in TEST_F() local
[all …]
Dchannel_mixer.cc20 ChannelMixer::ChannelMixer(ChannelLayout input_layout, in ChannelMixer() argument
22 : input_layout_(input_layout), in ChannelMixer()
24 input_channels_(ChannelLayoutToChannelCount(input_layout)), in ChannelMixer()
/external/tensorflow/tensorflow/dtensor/mlir/ir/
Dtf_dtensor.cc89 const tensorflow::dtensor::Layout input_layout = op.input_layout(); in verify() local
92 if (input_layout.rank() != output_layout.rank()) in verify()
95 << input_layout.rank() << " and " << output_layout.rank(); in verify()
97 for (int32_t i = 0; i < input_layout.rank(); ++i) { in verify()
98 if (input_layout.sharding_spec(i) != output_layout.sharding_spec(i) && in verify()
105 << input_layout.sharding_spec(i); in verify()
113 if (input_type.getRank() != input_layout.rank()) in verify()
115 << "input layout rank " << input_layout.rank() in verify()
129 input_layout.GlobalShapeFromLocalShape(input_type.getShape())); in verify()
145 const tensorflow::dtensor::Layout input_layout = op.input_layout(); in verify() local
[all …]
/external/tensorflow/tensorflow/dtensor/mlir/expansions/
Dtop_k_spmd_expander.cc31 StatusOr<Layout> GetSuggestedLayout(const Layout& input_layout) { in GetSuggestedLayout() argument
32 std::vector<ShardingSpec> layout_specs(input_layout.rank()); in GetSuggestedLayout()
34 for (int i = 0; i < input_layout.rank() - 1; ++i) { in GetSuggestedLayout()
35 layout_specs[i].set_sharding_spec(input_layout.sharding_spec(i)); in GetSuggestedLayout()
37 layout_specs[input_layout.rank() - 1].set_sharding_spec( in GetSuggestedLayout()
40 return Layout::GetLayout(layout_specs, input_layout.mesh()); in GetSuggestedLayout()
46 TF_ASSIGN_OR_RETURN(auto input_layout, ExtractLayoutFromOperand(input)); in ExpandOp()
48 if (!input_layout) in ExpandOp()
60 if (!input_layout->IsLastDimReplicated()) { in ExpandOp()
61 TF_ASSIGN_OR_RETURN(Layout new_layout, GetSuggestedLayout(*input_layout)); in ExpandOp()
[all …]
Dargmax_spmd_expander.cc41 const Layout& input_layout) { in ComputeResultLayout() argument
54 *output_layout_proto.mutable_mesh_config() = input_layout.mesh().ToProto(); in ComputeResultLayout()
59 input_layout.sharding_spec(i)); in ComputeResultLayout()
69 TF_ASSIGN_OR_RETURN(auto input_layout, in ExpandOp()
72 if (!input_layout || !output_layout) in ExpandOp()
88 input_layout->mesh().ToProto(); in ExpandOp()
99 input_layout->sharding_spec(i)); in ExpandOp()
103 if (!Layout::IsUnshardedDimension(input_layout->sharding_spec(axis))) { in ExpandOp()
106 builder, input, *input_layout, in ExpandOp()
126 const Layout& input_layout = input_layouts.lookup(0); in ComputeLayoutForward() local
[all …]
Dbias_add_spmd_expander.cc62 TF_ASSIGN_OR_RETURN(Layout input_layout, in ExpandOp()
73 input_layout.num_shards_for_dim(input_layout.dim(c_dim_idx)); in ExpandOp()
83 Layout::GetLayout(input_new_specs, input_layout.mesh())); in ExpandOp()
85 output, EmitRelayout(input.get(), input_layout, new_input_layout)); in ExpandOp()
87 input_layout = new_input_layout; in ExpandOp()
92 if (bias_layout.sharding_spec(0) != input_layout.sharding_spec(c_dim_idx)) { in ExpandOp()
96 input_layout.sharding_spec_strs()[c_dim_idx]}; in ExpandOp()
110 EmitRelayout(new_local_op->getOpResult(0), input_layout, in ExpandOp()
122 Layout input_layout = input_layouts.lookup(0); in ComputeLayoutForward() local
125 int c_dim_idx = get_c_dimension_idx(input_layout, data_format); in ComputeLayoutForward()
[all …]
Dmeta_spmd_expander.cc185 TF_ASSIGN_OR_RETURN(const absl::optional<Layout> input_layout, in ExpandOp()
187 if (!input_layout) { in ExpandOp()
199 if (input_layout->num_shards_for_dim(input_layout->dim(axis)) != 1) { in ExpandOp()
207 new_layout_specs[input_index] = input_layout->dim(input_index); in ExpandOp()
212 Layout::GetLayout(std::move(new_layout_specs), input_layout->mesh())); in ExpandOp()
215 EmitRelayout(unpack.getOperand(), *input_layout, new_input_layout)); in ExpandOp()
276 TF_ASSIGN_OR_RETURN(auto input_layout, ExtractLayoutFromOperand(pad_input)); in ExpandOp()
277 assert(input_layout && op_layout); in ExpandOp()
279 if (op_layout != input_layout) in ExpandOp()
294 const Layout input_layout = input_layouts.lookup(0); in ComputeLayoutForward() local
[all …]
Dsqueeze_spmd_expander.cc54 const Layout& input_layout = input_layouts.lookup(0); in ComputeLayoutForward() local
59 layout_specs.reserve(input_layout.rank()); in ComputeLayoutForward()
60 for (int64 i = 0; i < input_layout.rank(); ++i) { in ComputeLayoutForward()
63 layout_specs.push_back(input_layout.dim(i)); in ComputeLayoutForward()
67 layout_specs.push_back(input_layout.dim(i)); in ComputeLayoutForward()
73 Layout::GetLayout(layout_specs, input_layout.mesh())); in ComputeLayoutForward()
110 TF_ASSIGN_OR_RETURN(const Layout input_layout, in ComputeLayoutBackward()
112 return llvm::DenseMap<int, Layout>({{0, input_layout}}); in ComputeLayoutBackward()
Dbroadcast_to_spmd_expander.cc48 TF_ASSIGN_OR_RETURN(const Layout input_layout, in ExpandOp()
64 const int broadcasted_dimensions = output_layout.rank() - input_layout.rank(); in ExpandOp()
67 for (int i = 0; i < input_layout.rank(); ++i) { in ExpandOp()
72 input_layout.sharding_spec(i) != output_layout_dim) { in ExpandOp()
87 Layout::ReplicatedOnMesh(mesh, input_layout.rank()); in ExpandOp()
91 EmitAllGather(builder, input_data, input_layout, in ExpandOp()
149 Layout input_layout = input_layouts.lookup(0); in ComputeLayoutForward() local
157 input_layout.sharding_spec(i - broadcasted_dimensions)); in ComputeLayoutForward()
Dconv_spmd_expander.cc41 Status VerifyConvLayout(const Layout& input_layout, const Layout& filter_layout, in VerifyConvLayout() argument
54 if (input_layout.sharding_spec(channel_dim) != Layout::kUnshardedDim) in VerifyConvLayout()
58 if (input_layout.IsBatchParallel()) in VerifyConvLayout()
141 TF_ASSIGN_OR_RETURN(const Layout input_layout, in HandleConv()
148 TF_RETURN_IF_ERROR(VerifyConvLayout(input_layout, filter_layout, conv_op)); in HandleConv()
150 if (input_layout.IsBatchParallel()) in HandleConv()
158 const Mesh& mesh = input_layout.mesh(); in HandleConv()
162 input_layout.sharding_spec_strs(); in HandleConv()
168 const auto input_num_shards = input_layout.num_shards(); in HandleConv()
242 input_sharding_spec[curr_input_dim], input_layout, in HandleConv()
[all …]
Dexpanddims_spmd_expander.cc87 auto input_layout = input_layouts.lookup(0); in ComputeLayoutForward() local
89 if (dim < 0) dim += input_layout.rank() + 1; in ComputeLayoutForward()
92 for (int i = 0; i <= input_layout.rank(); ++i) { in ComputeLayoutForward()
94 if (i < input_layout.rank()) in ComputeLayoutForward()
95 layout_sharding.push_back(input_layout.sharding_spec(i)); in ComputeLayoutForward()
Dslice_spmd_expander.cc451 TF_ASSIGN_OR_RETURN(auto input_layout, in ExpandOp()
455 if (!output_layout || !input_layout) in ExpandOp()
481 *input_layout, &global_shape)); in ExpandOp()
486 EmitRelayout(op->getOperand(0), *input_layout, in ExpandOp()
551 const Layout& input_layout = input_layouts.lookup(0); in ComputeLayoutForward() local
554 VerifySliceLayout(slice_op, slice_op.input(), input_layout)); in ComputeLayoutForward()
583 TF_ASSIGN_OR_RETURN(Layout input_layout, ExtractRequiredLayoutFromOperand( in ExpandOp()
594 GetStridedSliceIntermediateLayout(strided_slice_op, input_layout, in ExpandOp()
598 EmitRelayout(strided_slice_op.input(), input_layout, in ExpandOp()
658 const Layout& input_layout = input_layouts.lookup(0); in ComputeLayoutForward() local
[all …]
Dreduce_spmd_expander.cc167 const Layout& input_layout) { in ComputeResultLayout() argument
170 return Layout::ReplicatedOnMesh(input_layout.mesh(), /*rank=*/0); in ComputeResultLayout()
177 return input_layout.GetLayoutWithReducedDims(reduced_dims_set, keep_dims); in ComputeResultLayout()
185 TF_ASSIGN_OR_RETURN(auto input_layout, in ExpandOp()
188 if (!input_layout || !requested_output_layout) in ExpandOp()
192 if (input_layout->mesh().is_tpu_mesh()) { in ExpandOp()
205 ComputeResultLayout(op, input_layout.value())); in ExpandOp()
208 ReducedMeshDimensions(*input_layout, output_layout); in ExpandOp()
Ddtensor_op_spmd_expander.cc181 TF_ASSIGN_OR_RETURN(const Layout input_layout, in ExpandOp()
192 if (input_layout == output_layout) { in ExpandOp()
205 EmitRelayout(relayout.input(), input_layout, output_layout); in ExpandOp()
278 TF_ASSIGN_OR_RETURN(const Layout input_layout, in ExpandOp()
284 if (IsOneToOneHostMeshTransfer(input_layout, dtensor_send.target_layout())) { in ExpandOp()
285 return LowerDTensorSendToXlaOp(input_layout, dtensor_send.input(), in ExpandOp()
295 Layout::ReplicatedOnMesh(input_layout.mesh(), rank); in ExpandOp()
300 EmitAllGather(builder, dtensor_send.input(), input_layout, in ExpandOp()
349 if (SendRecvOpUsesXla(input_layout.mesh(), recv_mesh)) { in ExpandOp()
353 LowerDTensorSendToXlaOp(input_layout, send_input, dtensor_send, in ExpandOp()
[all …]
Dtrivial_spmd_expander.cc51 TF_ASSIGN_OR_RETURN(auto input_layout, ExtractLayoutFromOperand(operand)); in ExpandOp()
52 if (!input_layout.has_value()) in ExpandOp()
57 if (!input_layout->IsFullyReplicated()) in ExpandOp()
/external/iamf_tools/iamf/cli/user_metadata_builder/
Daudio_element_metadata_builder.cc32 IamfInputLayout input_layout) { in LookupNumSubstreamsFromInputLayout() argument
53 return LookupInMap(*kInputLayoutToNumSubstreams, input_layout, in LookupNumSubstreamsFromInputLayout()
58 IamfInputLayout input_layout) { in LookupCoupledSubstreamCountFromInputLayout() argument
75 return LookupInMap(*kInputLayoutToCoupledSubstreamCount, input_layout, in LookupCoupledSubstreamCountFromInputLayout()
80 LookupLoudspeakerLayoutFromInputLayout(IamfInputLayout input_layout) { in LookupLoudspeakerLayoutFromInputLayout() argument
100 return LookupInMap(*KInputLayoutToLoudspeakerLayout, input_layout, in LookupLoudspeakerLayoutFromInputLayout()
105 LookupExpandedLoudspeakerLayoutFromInputLayout(IamfInputLayout input_layout) { in LookupExpandedLoudspeakerLayoutFromInputLayout() argument
117 return LookupInMap(*KInputLayoutToExpandedLoudspeakerLayout, input_layout, in LookupExpandedLoudspeakerLayoutFromInputLayout()
122 LookupAudioElementTypeFromInputLayout(IamfInputLayout input_layout) { in LookupAudioElementTypeFromInputLayout() argument
145 return LookupInMap(*KInputLayoutToAudioElementType, input_layout, in LookupAudioElementTypeFromInputLayout()
[all …]
Daudio_frame_metadata_builder.cc33 LookupLabelsFromInputLayout(IamfInputLayout input_layout) { in LookupLabelsFromInputLayout() argument
82 return LookupInMap(*kIamfInputLayoutToProtoLabels, input_layout, in LookupLabelsFromInputLayout()
91 IamfInputLayout input_layout, in PopulateAudioFrameMetadata() argument
103 const auto& labels = LookupLabelsFromInputLayout(input_layout); in PopulateAudioFrameMetadata()
/external/tensorflow/tensorflow/dtensor/cc/
Ddtensor_meta_ops.cc63 TF_ASSIGN_OR_RETURN(Layout input_layout, in __anon2ae02bb80102()
67 if (c->Rank(in) != input_layout.rank() || in __anon2ae02bb80102()
71 c->Rank(in), " input layout rank ", input_layout.rank(), in __anon2ae02bb80102()
82 input_layout.sharding_spec(i) == output_layout.sharding_spec(i)) { in __anon2ae02bb80102()
85 input_layout.sharding_spec(i))) { in __anon2ae02bb80102()
94 input_layout.sharding_spec(i), " and output sharding spec ", in __anon2ae02bb80102()
120 TF_ASSIGN_OR_RETURN(Layout input_layout, in __anon2ae02bb80202()
124 if (c->Rank(in) != input_layout.rank() || in __anon2ae02bb80202()
128 c->Rank(in), " input layout rank ", input_layout.rank(), in __anon2ae02bb80202()
133 const std::vector<int32> input_sharding = input_layout.num_shards(); in __anon2ae02bb80202()
[all …]
/external/tensorflow/tensorflow/compiler/xla/service/gpu/
Dstream_executor_util.cc70 std::vector<int64_t> input_layout; in StreamExecutorConvLayoutsToXlaLayouts() local
73 input_layout.push_back(dnums.input_batch_dimension()); in StreamExecutorConvLayoutsToXlaLayouts()
74 input_layout.push_back(dnums.input_feature_dimension()); in StreamExecutorConvLayoutsToXlaLayouts()
75 input_layout.insert(input_layout.end(), in StreamExecutorConvLayoutsToXlaLayouts()
81 input_layout.push_back(dnums.input_batch_dimension()); in StreamExecutorConvLayoutsToXlaLayouts()
82 input_layout.push_back(dnums.input_feature_dimension()); in StreamExecutorConvLayoutsToXlaLayouts()
83 input_layout.insert(input_layout.end(), in StreamExecutorConvLayoutsToXlaLayouts()
86 input_layout.push_back(FindMissingDnum(input_layout)); in StreamExecutorConvLayoutsToXlaLayouts()
89 input_layout.push_back(dnums.input_batch_dimension()); in StreamExecutorConvLayoutsToXlaLayouts()
90 input_layout.insert(input_layout.end(), in StreamExecutorConvLayoutsToXlaLayouts()
[all …]
/external/iamf_tools/iamf/cli/adm_to_user_metadata/iamf/
Dmix_presentation_handler.cc137 const auto& input_layout = LookupInputLayoutFromAudioPackFormatId( in SubMixAudioElementMetadataBuilder() local
139 if (!input_layout.ok()) { in SubMixAudioElementMetadataBuilder()
140 return input_layout.status(); in SubMixAudioElementMetadataBuilder()
168 IamfInputLayout input_layout, const LoudnessMetadata& loudness_metadata, in MixPresentationLayoutHandler() argument
175 if (input_layout == IamfInputLayout::kBinaural) { in MixPresentationLayoutHandler()
182 const auto& sound_system = LookupSoundSystemFromInputLayout(input_layout); in MixPresentationLayoutHandler()
192 bool IsChannelBasedAndNotStereo(IamfInputLayout input_layout) { in IsChannelBasedAndNotStereo() argument
193 switch (input_layout) { in IsChannelBasedAndNotStereo()
212 << static_cast<int>(input_layout); in IsChannelBasedAndNotStereo()
Diamf.cc249 const auto input_layout = LookupInputLayoutFromAudioPackFormatId( in Create() local
251 if (!input_layout.ok()) { in Create()
252 return input_layout.status(); in Create()
255 input_layouts.push_back(*input_layout); in Create()
/external/libopus/dnn/torch/lpcnet/models/
Dlpcnet.py47 self.input_layout = config['input_layout']
76 …self.gru_a_input_dim = len(self.input_layout['signals']) * self.signal_embedding_dim + self…
237 num_input_signals = len(self.input_layout['signals'])
238 pitch_corr_position = self.input_layout['features']['pitch_corr'][0]
276 if 'prediction' in self.input_layout['signals']:
277 … input_signals[0, 0, self.input_layout['signals']['prediction']] = lin2ulawq(pred)
297 if 'last_signal' in self.input_layout['signals']:
298 … input_signals[0, 0, self.input_layout['signals']['last_signal']] = lin2ulawq(sig)
300 if 'last_error' in self.input_layout['signals']:
301 … input_signals[0, 0, self.input_layout['signals']['last_error']] = lin2ulawq(exc)
/external/pytorch/torch/distributed/tensor/parallel/
Dstyle.py407 input_layout: Optional[Placement],
410 if input_layout is not None:
417 dt_inp = DTensor.from_local(input, mesh, (input_layout,), run_check=False)
419 if desired_layout is not None and input_layout != desired_layout:
436 …for inp, input_layout, desired_layout in zip(inputs, self.input_layouts, self.desired_input_layout…
437 … prepared_inputs.append(self._prepare_input_arg(inp, device_mesh, input_layout, desired_layout))
445 input_layout = self.input_kwarg_layouts.get(kwarg_key)
448 …g_inputs[kwarg_key] = self._prepare_input_arg(kwarg_val, device_mesh, input_layout, desired_input_…
/external/tensorflow/tensorflow/lite/delegates/gpu/common/tasks/
Dconv_weights_converter.h33 Layout input_layout);
58 Layout input_layout);

123