/external/armnn/python/pyarmnn/examples/tests/ |
D | test_mfcc.py | 41 def audio_data(test_data_folder, file, audio_cap_params): function 49 def test_audio_file(audio_data, test_data_folder, file, audio_cap_params): argument 114 def test_mfcc_compute_first_frame(audio_data, mfcc_test_params, test_out, file, audio_cap_params): argument 175 def test_feat_extraction_full_sized_input(audio_data, argument
|
/external/armnn/python/pyarmnn/examples/common/ |
D | mfcc.py | 78 def spectrum_calc(self, audio_data): argument 86 def mfcc_compute(self, audio_data): argument 201 def _get_features(self, features, mfcc_instance, audio_data): argument 214 def extract_features(self, audio_data): argument
|
D | utils.py | 45 def prepare_input_data(audio_data, input_data_type, input_quant_scale, input_quant_offset, mfcc_pre… argument
|
/external/webrtc/audio/ |
D | audio_transport_impl.cc | 109 const void* audio_data, in RecordedDataIsAvailable() 128 const void* audio_data, in RecordedDataIsAvailable() 244 void* audio_data, in PullRenderData()
|
D | audio_state_unittest.cc | 146 std::vector<int16_t> audio_data(samples_per_channel * num_channels, 0); in Create10msTestData() local 223 auto audio_data = Create10msTestData(kSampleRate, kNumChannels); in TEST_P() local 282 auto audio_data = Create10msTestData(kSampleRate, kNumChannels); in TEST_P() local 322 auto audio_data = Create10msTestData(kSampleRate, kNumChannels); in TEST_P() local
|
/external/tensorflow/tensorflow/lite/experimental/microfrontend/lib/ |
D | frontend_memmap_main.c | 32 int16_t* audio_data = malloc(audio_file_size * sizeof(int16_t)); in main() local
|
D | frontend_main.c | 42 int16_t* audio_data = malloc(audio_file_size * sizeof(int16_t)); in main() local
|
/external/armnn/python/pyarmnn/examples/keyword_spotting/ |
D | run_audio_classification.py | 70 def recognise_speech(audio_data, network, preprocessor, threshold): argument
|
/external/webrtc/api/ |
D | media_stream_interface.h | 201 virtual void OnData(const void* audio_data, in OnData() 213 virtual void OnData(const void* audio_data, in OnData()
|
/external/webrtc/modules/audio_device/android/ |
D | aaudio_player.cc | 151 aaudio_data_callback_result_t AAudioPlayer::OnDataCallback(void* audio_data, in OnDataCallback()
|
D | aaudio_recorder.cc | 150 void* audio_data, in OnDataCallback()
|
D | aaudio_wrapper.cc | 102 void* audio_data, in DataCallback() 270 void AAudioWrapper::ClearInputStream(void* audio_data, int32_t num_frames) { in ClearInputStream()
|
/external/webrtc/sdk/android/src/jni/audio_device/ |
D | aaudio_wrapper.cc | 103 void* audio_data, in DataCallback() 270 void AAudioWrapper::ClearInputStream(void* audio_data, int32_t num_frames) { in ClearInputStream()
|
D | aaudio_recorder.cc | 160 void* audio_data, in OnDataCallback()
|
D | aaudio_player.cc | 167 aaudio_data_callback_result_t AAudioPlayer::OnDataCallback(void* audio_data, in OnDataCallback()
|
/external/armnn/python/pyarmnn/examples/speech_recognition/ |
D | wav2letter_mfcc.py | 20 def spectrum_calc(self, audio_data): argument
|
/external/webrtc/modules/audio_processing/agc2/ |
D | clipping_predictor_unittest.cc | 62 std::vector<float> audio_data(num_channels * kSamplesPerChannel, 0.0f); in AnalyzeNonZeroCrestFactorAudio() local 121 std::vector<float> audio_data(num_channels * kSamplesPerChannel, 0.f); in AnalyzeZeroCrestFactorAudio() local
|
/external/googleapis/google/assistant/embedded/v1alpha1/ |
D | embedded_assistant.proto | 172 bytes audio_data = 1; field
|
/external/tensorflow/tensorflow/lite/experimental/microfrontend/ops/ |
D | audio_microfrontend_op.cc | 206 auto audio_data = in Compute() local
|
/external/webrtc/modules/audio_device/win/ |
D | core_audio_output_win.cc | 318 uint8_t* audio_data; in OnDataCallback() local
|
D | core_audio_input_win.cc | 297 uint8_t* audio_data; in OnDataCallback() local
|
/external/webrtc/pc/test/ |
D | fake_audio_capture_module_unittest.cc | 67 void* audio_data, in PullRenderData()
|
/external/webrtc/pc/ |
D | rtp_sender.h | 308 void OnData(const void* audio_data, in OnData()
|
/external/tensorflow/tensorflow/examples/speech_commands/ |
D | test_streaming_accuracy.cc | 217 std::vector<float> audio_data; in main() local
|
/external/tensorflow/tensorflow/lite/experimental/microfrontend/ |
D | audio_microfrontend.cc | 125 const int16_t* audio_data = GetTensorData<int16_t>(input); in GenerateFeatures() local
|