/external/webrtc/webrtc/modules/audio_conference_mixer/source/ |
D | audio_frame_manipulator.cc | 42 void CalculateEnergy(AudioFrame& audioFrame) in CalculateEnergy() argument 44 audioFrame.energy_ = 0; in CalculateEnergy() 45 for(size_t position = 0; position < audioFrame.samples_per_channel_; in CalculateEnergy() 49 audioFrame.energy_ += audioFrame.data_[position] * in CalculateEnergy() 50 audioFrame.data_[position]; in CalculateEnergy() 54 void RampIn(AudioFrame& audioFrame) in RampIn() argument 56 assert(rampSize <= audioFrame.samples_per_channel_); in RampIn() 59 audioFrame.data_[i] = static_cast<int16_t>(rampArray[i] * in RampIn() 60 audioFrame.data_[i]); in RampIn() 64 void RampOut(AudioFrame& audioFrame) in RampOut() argument [all …]
|
D | audio_conference_mixer_impl.cc | 24 AudioFrame* audioFrame; member 548 AudioFrame* audioFrame = NULL; in UpdateToMix() local 549 if(_audioFramePool->PopMemory(audioFrame) == -1) { in UpdateToMix() 555 audioFrame->sample_rate_hz_ = _outputFrequency; in UpdateToMix() 557 if((*participant)->GetAudioFrame(_id, audioFrame) != 0) { in UpdateToMix() 560 _audioFramePool->PushMemory(audioFrame); in UpdateToMix() 565 audioFrame->ntp_time_ms_ = -1; in UpdateToMix() 572 if (audioFrame->vad_activity_ == AudioFrame::kVadUnknown) { in UpdateToMix() 577 if(audioFrame->vad_activity_ == AudioFrame::kVadActive) { in UpdateToMix() 579 RampIn(*audioFrame); in UpdateToMix() [all …]
|
D | audio_frame_manipulator.h | 18 void CalculateEnergy(AudioFrame& audioFrame); 21 void RampIn(AudioFrame& audioFrame); 22 void RampOut(AudioFrame& audioFrame);
|
/external/webrtc/webrtc/modules/audio_coding/test/ |
D | SpatialAudio.cc | 153 AudioFrame audioFrame; in EncodeDecode() local 161 _inFile.Read10MsData(audioFrame); in EncodeDecode() 162 for (size_t n = 0; n < audioFrame.samples_per_channel_; n++) { in EncodeDecode() 163 audioFrame.data_[n] = (int16_t) floor( in EncodeDecode() 164 audioFrame.data_[n] * leftPanning + 0.5); in EncodeDecode() 166 CHECK_ERROR(_acmLeft->Add10MsData(audioFrame)); in EncodeDecode() 168 for (size_t n = 0; n < audioFrame.samples_per_channel_; n++) { in EncodeDecode() 169 audioFrame.data_[n] = (int16_t) floor( in EncodeDecode() 170 audioFrame.data_[n] * rightToLeftRatio + 0.5); in EncodeDecode() 172 CHECK_ERROR(_acmRight->Add10MsData(audioFrame)); in EncodeDecode() [all …]
|
D | TwoWayCommunication.cc | 251 AudioFrame audioFrame; in Perform() local 262 EXPECT_GT(_inFileA.Read10MsData(audioFrame), 0); in Perform() 263 EXPECT_GE(_acmA->Add10MsData(audioFrame), 0); in Perform() 264 EXPECT_GE(_acmRefA->Add10MsData(audioFrame), 0); in Perform() 266 EXPECT_GT(_inFileB.Read10MsData(audioFrame), 0); in Perform() 268 EXPECT_GE(_acmB->Add10MsData(audioFrame), 0); in Perform() 269 EXPECT_GE(_acmRefB->Add10MsData(audioFrame), 0); in Perform() 270 EXPECT_EQ(0, _acmA->PlayoutData10Ms(outFreqHzA, &audioFrame)); in Perform() 271 _outFileA.Write10MsData(audioFrame); in Perform() 272 EXPECT_EQ(0, _acmRefA->PlayoutData10Ms(outFreqHzA, &audioFrame)); in Perform() [all …]
|
D | iSACTest.cc | 198 AudioFrame audioFrame; in Run10ms() local 199 EXPECT_GT(_inFileA.Read10MsData(audioFrame), 0); in Run10ms() 200 EXPECT_GE(_acmA->Add10MsData(audioFrame), 0); in Run10ms() 201 EXPECT_GE(_acmB->Add10MsData(audioFrame), 0); in Run10ms() 202 EXPECT_EQ(0, _acmA->PlayoutData10Ms(32000, &audioFrame)); in Run10ms() 203 _outFileA.Write10MsData(audioFrame); in Run10ms() 204 EXPECT_EQ(0, _acmB->PlayoutData10Ms(32000, &audioFrame)); in Run10ms() 205 _outFileB.Write10MsData(audioFrame); in Run10ms()
|
D | APITest.cc | 325 AudioFrame audioFrame; in PullAudioRunA() local 326 if (_acmA->PlayoutData10Ms(_outFreqHzA, &audioFrame) < 0) { in PullAudioRunA() 337 _outFileA.Write10MsData(audioFrame); in PullAudioRunA() 345 AudioFrame audioFrame; in PullAudioRunB() local 346 if (_acmB->PlayoutData10Ms(_outFreqHzB, &audioFrame) < 0) { in PullAudioRunB() 358 _outFileB.Write10MsData(audioFrame); in PullAudioRunB() 366 AudioFrame audioFrame; in PushAudioRunA() local 367 _inFileA.Read10MsData(audioFrame); in PushAudioRunA() 368 if (_acmA->Add10MsData(audioFrame) < 0) { in PushAudioRunA() 383 AudioFrame audioFrame; in PushAudioRunB() local [all …]
|
D | EncodeDecodeTest.cc | 210 AudioFrame audioFrame; in PlayoutData() local 212 int32_t ok =_acm->PlayoutData10Ms(_frequency, &audioFrame); in PlayoutData() 220 _pcmFile.Write10MsData(audioFrame.data_, in PlayoutData() 221 audioFrame.samples_per_channel_ * audioFrame.num_channels_); in PlayoutData()
|
D | TestRedFec.cc | 454 AudioFrame audioFrame; in Run() local 462 EXPECT_GT(_inFileA.Read10MsData(audioFrame), 0); in Run() 463 EXPECT_GE(_acmA->Add10MsData(audioFrame), 0); in Run() 464 EXPECT_EQ(0, _acmB->PlayoutData10Ms(outFreqHzB, &audioFrame)); in Run() 465 _outFileB.Write10MsData(audioFrame.data_, audioFrame.samples_per_channel_); in Run()
|
/external/webrtc/webrtc/modules/utility/source/ |
D | coder.cc | 83 AudioFrame audioFrame; in Encode() local 84 audioFrame.CopyFrom(audio); in Encode() 85 audioFrame.timestamp_ = _encodeTimestamp; in Encode() 86 _encodeTimestamp += static_cast<uint32_t>(audioFrame.samples_per_channel_); in Encode() 91 if(_acm->Add10MsData((AudioFrame&)audioFrame) == -1) in Encode()
|
/external/webrtc/webrtc/voice_engine/ |
D | level_indicator.cc | 48 void AudioLevel::ComputeLevel(const AudioFrame& audioFrame) in ComputeLevel() argument 54 audioFrame.data_, in ComputeLevel() 55 audioFrame.samples_per_channel_*audioFrame.num_channels_); in ComputeLevel()
|
D | channel.cc | 523 int32_t Channel::GetAudioFrame(int32_t id, AudioFrame* audioFrame) in GetAudioFrame() argument 531 if (audio_coding_->PlayoutData10Ms(audioFrame->sample_rate_hz_, in GetAudioFrame() 532 audioFrame) == -1) in GetAudioFrame() 546 UpdateRxVadDetection(*audioFrame); in GetAudioFrame() 550 audioFrame->id_ = VoEChannelId(audioFrame->id_); in GetAudioFrame() 552 _outputSpeechType = audioFrame->speech_type_; in GetAudioFrame() 557 int err = rx_audioproc_->ProcessStream(audioFrame); in GetAudioFrame() 572 &audioFrame->data_[0], in GetAudioFrame() 573 audioFrame->samples_per_channel_, audioFrame->sample_rate_hz_, in GetAudioFrame() 574 audioFrame->num_channels_, audioFrame->timestamp_); in GetAudioFrame() [all …]
|
D | level_indicator.h | 38 void ComputeLevel(const AudioFrame& audioFrame);
|
D | channel.h | 308 int UpdateRxVadDetection(AudioFrame& audioFrame); 401 int32_t GetAudioFrame(int32_t id, AudioFrame* audioFrame) override; 447 uint32_t Demultiplex(const AudioFrame& audioFrame); 483 int32_t MixAudioWithFile(AudioFrame& audioFrame, int mixingFrequency);
|
D | output_mixer.h | 67 AudioFrame* audioFrame);
|
/external/webrtc/webrtc/modules/audio_conference_mixer/include/ |
D | audio_conference_mixer_defines.h | 29 AudioFrame* audioFrame) = 0;
|
/external/webrtc/webrtc/modules/audio_processing/ |
D | audio_buffer.h | 109 void DeinterleaveFrom(AudioFrame* audioFrame);
|