/external/webrtc/audio/utility/ |
D | channel_mixing_matrix_unittest.cc | 28 for (ChannelLayout input_layout = CHANNEL_LAYOUT_MONO; in TEST() local 29 input_layout <= CHANNEL_LAYOUT_MAX; in TEST() 30 input_layout = static_cast<ChannelLayout>(input_layout + 1)) { in TEST() 37 if (input_layout == CHANNEL_LAYOUT_BITSTREAM || in TEST() 38 input_layout == CHANNEL_LAYOUT_DISCRETE || in TEST() 39 input_layout == CHANNEL_LAYOUT_STEREO_AND_KEYBOARD_MIC || in TEST() 48 ss << "Input Layout: " << input_layout in TEST() 52 input_layout, ChannelLayoutToChannelCount(input_layout), in TEST() 54 const int input_channels = ChannelLayoutToChannelCount(input_layout); in TEST() 87 ChannelLayout input_layout = CHANNEL_LAYOUT_STEREO; in TEST() local [all …]
|
D | channel_mixer_unittest.cc | 147 for (ChannelLayout input_layout = CHANNEL_LAYOUT_MONO; in TEST_F() local 148 input_layout <= CHANNEL_LAYOUT_MAX; in TEST_F() 149 input_layout = static_cast<ChannelLayout>(input_layout + 1)) { in TEST_F() 156 if (input_layout == CHANNEL_LAYOUT_BITSTREAM || in TEST_F() 157 input_layout == CHANNEL_LAYOUT_DISCRETE || in TEST_F() 158 input_layout == CHANNEL_LAYOUT_STEREO_AND_KEYBOARD_MIC || in TEST_F() 167 ss << "Input Layout: " << input_layout in TEST_F() 170 ChannelMixer mixer(input_layout, output_layout); in TEST_F() 174 ChannelLayoutToChannelCount(input_layout)); in TEST_F() 185 for (ChannelLayout input_layout = CHANNEL_LAYOUT_MONO; in TEST_F() local [all …]
|
D | channel_mixer.cc | 20 ChannelMixer::ChannelMixer(ChannelLayout input_layout, in ChannelMixer() argument 22 : input_layout_(input_layout), in ChannelMixer() 24 input_channels_(ChannelLayoutToChannelCount(input_layout)), in ChannelMixer()
|
/external/tensorflow/tensorflow/dtensor/mlir/ir/ |
D | tf_dtensor.cc | 89 const tensorflow::dtensor::Layout input_layout = op.input_layout(); in verify() local 92 if (input_layout.rank() != output_layout.rank()) in verify() 95 << input_layout.rank() << " and " << output_layout.rank(); in verify() 97 for (int32_t i = 0; i < input_layout.rank(); ++i) { in verify() 98 if (input_layout.sharding_spec(i) != output_layout.sharding_spec(i) && in verify() 105 << input_layout.sharding_spec(i); in verify() 113 if (input_type.getRank() != input_layout.rank()) in verify() 115 << "input layout rank " << input_layout.rank() in verify() 129 input_layout.GlobalShapeFromLocalShape(input_type.getShape())); in verify() 145 const tensorflow::dtensor::Layout input_layout = op.input_layout(); in verify() local [all …]
|
/external/tensorflow/tensorflow/dtensor/mlir/expansions/ |
D | top_k_spmd_expander.cc | 31 StatusOr<Layout> GetSuggestedLayout(const Layout& input_layout) { in GetSuggestedLayout() argument 32 std::vector<ShardingSpec> layout_specs(input_layout.rank()); in GetSuggestedLayout() 34 for (int i = 0; i < input_layout.rank() - 1; ++i) { in GetSuggestedLayout() 35 layout_specs[i].set_sharding_spec(input_layout.sharding_spec(i)); in GetSuggestedLayout() 37 layout_specs[input_layout.rank() - 1].set_sharding_spec( in GetSuggestedLayout() 40 return Layout::GetLayout(layout_specs, input_layout.mesh()); in GetSuggestedLayout() 46 TF_ASSIGN_OR_RETURN(auto input_layout, ExtractLayoutFromOperand(input)); in ExpandOp() 48 if (!input_layout) in ExpandOp() 60 if (!input_layout->IsLastDimReplicated()) { in ExpandOp() 61 TF_ASSIGN_OR_RETURN(Layout new_layout, GetSuggestedLayout(*input_layout)); in ExpandOp() [all …]
|
D | argmax_spmd_expander.cc | 41 const Layout& input_layout) { in ComputeResultLayout() argument 54 *output_layout_proto.mutable_mesh_config() = input_layout.mesh().ToProto(); in ComputeResultLayout() 59 input_layout.sharding_spec(i)); in ComputeResultLayout() 69 TF_ASSIGN_OR_RETURN(auto input_layout, in ExpandOp() 72 if (!input_layout || !output_layout) in ExpandOp() 88 input_layout->mesh().ToProto(); in ExpandOp() 99 input_layout->sharding_spec(i)); in ExpandOp() 103 if (!Layout::IsUnshardedDimension(input_layout->sharding_spec(axis))) { in ExpandOp() 106 builder, input, *input_layout, in ExpandOp() 126 const Layout& input_layout = input_layouts.lookup(0); in ComputeLayoutForward() local [all …]
|
D | bias_add_spmd_expander.cc | 62 TF_ASSIGN_OR_RETURN(Layout input_layout, in ExpandOp() 73 input_layout.num_shards_for_dim(input_layout.dim(c_dim_idx)); in ExpandOp() 83 Layout::GetLayout(input_new_specs, input_layout.mesh())); in ExpandOp() 85 output, EmitRelayout(input.get(), input_layout, new_input_layout)); in ExpandOp() 87 input_layout = new_input_layout; in ExpandOp() 92 if (bias_layout.sharding_spec(0) != input_layout.sharding_spec(c_dim_idx)) { in ExpandOp() 96 input_layout.sharding_spec_strs()[c_dim_idx]}; in ExpandOp() 110 EmitRelayout(new_local_op->getOpResult(0), input_layout, in ExpandOp() 122 Layout input_layout = input_layouts.lookup(0); in ComputeLayoutForward() local 125 int c_dim_idx = get_c_dimension_idx(input_layout, data_format); in ComputeLayoutForward() [all …]
|
D | meta_spmd_expander.cc | 185 TF_ASSIGN_OR_RETURN(const absl::optional<Layout> input_layout, in ExpandOp() 187 if (!input_layout) { in ExpandOp() 199 if (input_layout->num_shards_for_dim(input_layout->dim(axis)) != 1) { in ExpandOp() 207 new_layout_specs[input_index] = input_layout->dim(input_index); in ExpandOp() 212 Layout::GetLayout(std::move(new_layout_specs), input_layout->mesh())); in ExpandOp() 215 EmitRelayout(unpack.getOperand(), *input_layout, new_input_layout)); in ExpandOp() 276 TF_ASSIGN_OR_RETURN(auto input_layout, ExtractLayoutFromOperand(pad_input)); in ExpandOp() 277 assert(input_layout && op_layout); in ExpandOp() 279 if (op_layout != input_layout) in ExpandOp() 294 const Layout input_layout = input_layouts.lookup(0); in ComputeLayoutForward() local [all …]
|
D | squeeze_spmd_expander.cc | 54 const Layout& input_layout = input_layouts.lookup(0); in ComputeLayoutForward() local 59 layout_specs.reserve(input_layout.rank()); in ComputeLayoutForward() 60 for (int64 i = 0; i < input_layout.rank(); ++i) { in ComputeLayoutForward() 63 layout_specs.push_back(input_layout.dim(i)); in ComputeLayoutForward() 67 layout_specs.push_back(input_layout.dim(i)); in ComputeLayoutForward() 73 Layout::GetLayout(layout_specs, input_layout.mesh())); in ComputeLayoutForward() 110 TF_ASSIGN_OR_RETURN(const Layout input_layout, in ComputeLayoutBackward() 112 return llvm::DenseMap<int, Layout>({{0, input_layout}}); in ComputeLayoutBackward()
|
D | broadcast_to_spmd_expander.cc | 48 TF_ASSIGN_OR_RETURN(const Layout input_layout, in ExpandOp() 64 const int broadcasted_dimensions = output_layout.rank() - input_layout.rank(); in ExpandOp() 67 for (int i = 0; i < input_layout.rank(); ++i) { in ExpandOp() 72 input_layout.sharding_spec(i) != output_layout_dim) { in ExpandOp() 87 Layout::ReplicatedOnMesh(mesh, input_layout.rank()); in ExpandOp() 91 EmitAllGather(builder, input_data, input_layout, in ExpandOp() 149 Layout input_layout = input_layouts.lookup(0); in ComputeLayoutForward() local 157 input_layout.sharding_spec(i - broadcasted_dimensions)); in ComputeLayoutForward()
|
D | conv_spmd_expander.cc | 41 Status VerifyConvLayout(const Layout& input_layout, const Layout& filter_layout, in VerifyConvLayout() argument 54 if (input_layout.sharding_spec(channel_dim) != Layout::kUnshardedDim) in VerifyConvLayout() 58 if (input_layout.IsBatchParallel()) in VerifyConvLayout() 141 TF_ASSIGN_OR_RETURN(const Layout input_layout, in HandleConv() 148 TF_RETURN_IF_ERROR(VerifyConvLayout(input_layout, filter_layout, conv_op)); in HandleConv() 150 if (input_layout.IsBatchParallel()) in HandleConv() 158 const Mesh& mesh = input_layout.mesh(); in HandleConv() 162 input_layout.sharding_spec_strs(); in HandleConv() 168 const auto input_num_shards = input_layout.num_shards(); in HandleConv() 242 input_sharding_spec[curr_input_dim], input_layout, in HandleConv() [all …]
|
D | expanddims_spmd_expander.cc | 87 auto input_layout = input_layouts.lookup(0); in ComputeLayoutForward() local 89 if (dim < 0) dim += input_layout.rank() + 1; in ComputeLayoutForward() 92 for (int i = 0; i <= input_layout.rank(); ++i) { in ComputeLayoutForward() 94 if (i < input_layout.rank()) in ComputeLayoutForward() 95 layout_sharding.push_back(input_layout.sharding_spec(i)); in ComputeLayoutForward()
|
D | slice_spmd_expander.cc | 451 TF_ASSIGN_OR_RETURN(auto input_layout, in ExpandOp() 455 if (!output_layout || !input_layout) in ExpandOp() 481 *input_layout, &global_shape)); in ExpandOp() 486 EmitRelayout(op->getOperand(0), *input_layout, in ExpandOp() 551 const Layout& input_layout = input_layouts.lookup(0); in ComputeLayoutForward() local 554 VerifySliceLayout(slice_op, slice_op.input(), input_layout)); in ComputeLayoutForward() 583 TF_ASSIGN_OR_RETURN(Layout input_layout, ExtractRequiredLayoutFromOperand( in ExpandOp() 594 GetStridedSliceIntermediateLayout(strided_slice_op, input_layout, in ExpandOp() 598 EmitRelayout(strided_slice_op.input(), input_layout, in ExpandOp() 658 const Layout& input_layout = input_layouts.lookup(0); in ComputeLayoutForward() local [all …]
|
D | reduce_spmd_expander.cc | 167 const Layout& input_layout) { in ComputeResultLayout() argument 170 return Layout::ReplicatedOnMesh(input_layout.mesh(), /*rank=*/0); in ComputeResultLayout() 177 return input_layout.GetLayoutWithReducedDims(reduced_dims_set, keep_dims); in ComputeResultLayout() 185 TF_ASSIGN_OR_RETURN(auto input_layout, in ExpandOp() 188 if (!input_layout || !requested_output_layout) in ExpandOp() 192 if (input_layout->mesh().is_tpu_mesh()) { in ExpandOp() 205 ComputeResultLayout(op, input_layout.value())); in ExpandOp() 208 ReducedMeshDimensions(*input_layout, output_layout); in ExpandOp()
|
D | dtensor_op_spmd_expander.cc | 181 TF_ASSIGN_OR_RETURN(const Layout input_layout, in ExpandOp() 192 if (input_layout == output_layout) { in ExpandOp() 205 EmitRelayout(relayout.input(), input_layout, output_layout); in ExpandOp() 278 TF_ASSIGN_OR_RETURN(const Layout input_layout, in ExpandOp() 284 if (IsOneToOneHostMeshTransfer(input_layout, dtensor_send.target_layout())) { in ExpandOp() 285 return LowerDTensorSendToXlaOp(input_layout, dtensor_send.input(), in ExpandOp() 295 Layout::ReplicatedOnMesh(input_layout.mesh(), rank); in ExpandOp() 300 EmitAllGather(builder, dtensor_send.input(), input_layout, in ExpandOp() 349 if (SendRecvOpUsesXla(input_layout.mesh(), recv_mesh)) { in ExpandOp() 353 LowerDTensorSendToXlaOp(input_layout, send_input, dtensor_send, in ExpandOp() [all …]
|
D | trivial_spmd_expander.cc | 51 TF_ASSIGN_OR_RETURN(auto input_layout, ExtractLayoutFromOperand(operand)); in ExpandOp() 52 if (!input_layout.has_value()) in ExpandOp() 57 if (!input_layout->IsFullyReplicated()) in ExpandOp()
|
/external/iamf_tools/iamf/cli/user_metadata_builder/ |
D | audio_element_metadata_builder.cc | 32 IamfInputLayout input_layout) { in LookupNumSubstreamsFromInputLayout() argument 53 return LookupInMap(*kInputLayoutToNumSubstreams, input_layout, in LookupNumSubstreamsFromInputLayout() 58 IamfInputLayout input_layout) { in LookupCoupledSubstreamCountFromInputLayout() argument 75 return LookupInMap(*kInputLayoutToCoupledSubstreamCount, input_layout, in LookupCoupledSubstreamCountFromInputLayout() 80 LookupLoudspeakerLayoutFromInputLayout(IamfInputLayout input_layout) { in LookupLoudspeakerLayoutFromInputLayout() argument 100 return LookupInMap(*KInputLayoutToLoudspeakerLayout, input_layout, in LookupLoudspeakerLayoutFromInputLayout() 105 LookupExpandedLoudspeakerLayoutFromInputLayout(IamfInputLayout input_layout) { in LookupExpandedLoudspeakerLayoutFromInputLayout() argument 117 return LookupInMap(*KInputLayoutToExpandedLoudspeakerLayout, input_layout, in LookupExpandedLoudspeakerLayoutFromInputLayout() 122 LookupAudioElementTypeFromInputLayout(IamfInputLayout input_layout) { in LookupAudioElementTypeFromInputLayout() argument 145 return LookupInMap(*KInputLayoutToAudioElementType, input_layout, in LookupAudioElementTypeFromInputLayout() [all …]
|
D | audio_frame_metadata_builder.cc | 33 LookupLabelsFromInputLayout(IamfInputLayout input_layout) { in LookupLabelsFromInputLayout() argument 82 return LookupInMap(*kIamfInputLayoutToProtoLabels, input_layout, in LookupLabelsFromInputLayout() 91 IamfInputLayout input_layout, in PopulateAudioFrameMetadata() argument 103 const auto& labels = LookupLabelsFromInputLayout(input_layout); in PopulateAudioFrameMetadata()
|
/external/tensorflow/tensorflow/dtensor/cc/ |
D | dtensor_meta_ops.cc | 63 TF_ASSIGN_OR_RETURN(Layout input_layout, in __anon2ae02bb80102() 67 if (c->Rank(in) != input_layout.rank() || in __anon2ae02bb80102() 71 c->Rank(in), " input layout rank ", input_layout.rank(), in __anon2ae02bb80102() 82 input_layout.sharding_spec(i) == output_layout.sharding_spec(i)) { in __anon2ae02bb80102() 85 input_layout.sharding_spec(i))) { in __anon2ae02bb80102() 94 input_layout.sharding_spec(i), " and output sharding spec ", in __anon2ae02bb80102() 120 TF_ASSIGN_OR_RETURN(Layout input_layout, in __anon2ae02bb80202() 124 if (c->Rank(in) != input_layout.rank() || in __anon2ae02bb80202() 128 c->Rank(in), " input layout rank ", input_layout.rank(), in __anon2ae02bb80202() 133 const std::vector<int32> input_sharding = input_layout.num_shards(); in __anon2ae02bb80202() [all …]
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
D | stream_executor_util.cc | 70 std::vector<int64_t> input_layout; in StreamExecutorConvLayoutsToXlaLayouts() local 73 input_layout.push_back(dnums.input_batch_dimension()); in StreamExecutorConvLayoutsToXlaLayouts() 74 input_layout.push_back(dnums.input_feature_dimension()); in StreamExecutorConvLayoutsToXlaLayouts() 75 input_layout.insert(input_layout.end(), in StreamExecutorConvLayoutsToXlaLayouts() 81 input_layout.push_back(dnums.input_batch_dimension()); in StreamExecutorConvLayoutsToXlaLayouts() 82 input_layout.push_back(dnums.input_feature_dimension()); in StreamExecutorConvLayoutsToXlaLayouts() 83 input_layout.insert(input_layout.end(), in StreamExecutorConvLayoutsToXlaLayouts() 86 input_layout.push_back(FindMissingDnum(input_layout)); in StreamExecutorConvLayoutsToXlaLayouts() 89 input_layout.push_back(dnums.input_batch_dimension()); in StreamExecutorConvLayoutsToXlaLayouts() 90 input_layout.insert(input_layout.end(), in StreamExecutorConvLayoutsToXlaLayouts() [all …]
|
/external/iamf_tools/iamf/cli/adm_to_user_metadata/iamf/ |
D | mix_presentation_handler.cc | 137 const auto& input_layout = LookupInputLayoutFromAudioPackFormatId( in SubMixAudioElementMetadataBuilder() local 139 if (!input_layout.ok()) { in SubMixAudioElementMetadataBuilder() 140 return input_layout.status(); in SubMixAudioElementMetadataBuilder() 168 IamfInputLayout input_layout, const LoudnessMetadata& loudness_metadata, in MixPresentationLayoutHandler() argument 175 if (input_layout == IamfInputLayout::kBinaural) { in MixPresentationLayoutHandler() 182 const auto& sound_system = LookupSoundSystemFromInputLayout(input_layout); in MixPresentationLayoutHandler() 192 bool IsChannelBasedAndNotStereo(IamfInputLayout input_layout) { in IsChannelBasedAndNotStereo() argument 193 switch (input_layout) { in IsChannelBasedAndNotStereo() 212 << static_cast<int>(input_layout); in IsChannelBasedAndNotStereo()
|
D | iamf.cc | 249 const auto input_layout = LookupInputLayoutFromAudioPackFormatId( in Create() local 251 if (!input_layout.ok()) { in Create() 252 return input_layout.status(); in Create() 255 input_layouts.push_back(*input_layout); in Create()
|
/external/libopus/dnn/torch/lpcnet/models/ |
D | lpcnet.py | 47 self.input_layout = config['input_layout'] 76 …self.gru_a_input_dim = len(self.input_layout['signals']) * self.signal_embedding_dim + self… 237 num_input_signals = len(self.input_layout['signals']) 238 pitch_corr_position = self.input_layout['features']['pitch_corr'][0] 276 if 'prediction' in self.input_layout['signals']: 277 … input_signals[0, 0, self.input_layout['signals']['prediction']] = lin2ulawq(pred) 297 if 'last_signal' in self.input_layout['signals']: 298 … input_signals[0, 0, self.input_layout['signals']['last_signal']] = lin2ulawq(sig) 300 if 'last_error' in self.input_layout['signals']: 301 … input_signals[0, 0, self.input_layout['signals']['last_error']] = lin2ulawq(exc)
|
/external/pytorch/torch/distributed/tensor/parallel/ |
D | style.py | 407 input_layout: Optional[Placement], 410 if input_layout is not None: 417 dt_inp = DTensor.from_local(input, mesh, (input_layout,), run_check=False) 419 if desired_layout is not None and input_layout != desired_layout: 436 …for inp, input_layout, desired_layout in zip(inputs, self.input_layouts, self.desired_input_layout… 437 … prepared_inputs.append(self._prepare_input_arg(inp, device_mesh, input_layout, desired_layout)) 445 input_layout = self.input_kwarg_layouts.get(kwarg_key) 448 …g_inputs[kwarg_key] = self._prepare_input_arg(kwarg_val, device_mesh, input_layout, desired_input_…
|
/external/tensorflow/tensorflow/lite/delegates/gpu/common/tasks/ |
D | conv_weights_converter.h | 33 Layout input_layout); 58 Layout input_layout);
|