Home
last modified time | relevance | path

Searched refs:stereo (Results 1 – 25 of 50) sorted by relevance

12

/external/speex/libspeex/
Dstereo.c77 SpeexStereoState *stereo = speex_alloc(sizeof(SpeexStereoState)); in speex_stereo_state_init() local
78 speex_stereo_state_reset(stereo); in speex_stereo_state_init()
79 return stereo; in speex_stereo_state_init()
84 RealSpeexStereoState *stereo = (RealSpeexStereoState*)_stereo; in speex_stereo_state_reset() local
86 stereo->balance = 65536; in speex_stereo_state_reset()
87 stereo->e_ratio = 16384; in speex_stereo_state_reset()
88 stereo->smooth_left = 16384; in speex_stereo_state_reset()
89 stereo->smooth_right = 16384; in speex_stereo_state_reset()
90 stereo->reserved1 = 0xdeadbeef; in speex_stereo_state_reset()
91 stereo->reserved2 = 0; in speex_stereo_state_reset()
[all …]
/external/speex/include/speex/
Dspeex_stereo.h66 void speex_stereo_state_reset(SpeexStereoState *stereo);
69 void speex_stereo_state_destroy(SpeexStereoState *stereo);
78 void speex_decode_stereo(float *data, int frame_size, SpeexStereoState *stereo);
81 void speex_decode_stereo_int(spx_int16_t *data, int frame_size, SpeexStereoState *stereo);
/external/libxaac/decoder/
Dixheaacd_sbrdecoder.c254 FLAG stereo = 0; in ixheaacd_applysbr() local
366 stereo = 0; in ixheaacd_applysbr()
369 stereo = 1; in ixheaacd_applysbr()
414 ptr_header_data[k], it_bit_buff, stereo, ptr_sbr_dflt_header); in ixheaacd_applysbr()
454 ptr_header_data[k], it_bit_buff, stereo, ptr_sbr_dflt_header); in ixheaacd_applysbr()
507 if (stereo) { in ixheaacd_applysbr()
547 stereo = (num_channels == 2) ? 1 : 0; in ixheaacd_applysbr()
549 ptr_header_data[0]->channel_mode = stereo ? SBR_STEREO : SBR_MONO; in ixheaacd_applysbr()
553 if (!(stereo || dual_mono)) { in ixheaacd_applysbr()
572 (stereo || dual_mono) ? ptr_frame_data[1] : NULL, in ixheaacd_applysbr()
[all …]
/external/adhd/cras/src/tests/
Daudio_area_unittest.cc13 static const int8_t stereo[CRAS_CH_MAX] = { variable
37 fmt.channel_layout[i] = stereo[i]; in TEST()
68 fmt.channel_layout[i] = stereo[i]; in TEST()
104 fmt.channel_layout[i] = stereo[i]; in TEST()
137 fmt.channel_layout[i] = stereo[i]; in TEST()
175 dst_fmt.channel_layout[i] = stereo[i]; in TEST()
218 fmt.channel_layout[i] = stereo[i]; in TEST()
250 fmt.channel_layout[i] = stereo[i]; in TEST()
/external/webrtc/talk/media/base/
Daudioframe.h41 AudioFrame(int16_t* audio, size_t audio_length, int sample_freq, bool stereo) in AudioFrame() argument
45 stereo_(stereo) {} in AudioFrame()
/external/aac/libSBRdec/src/
Dsbrdecoder.cpp1148 int stereo; in sbrDecoder_Parse() local
1224 stereo = (hSbrElement->elementID == ID_CPE) ? 1 : 0; in sbrDecoder_Parse()
1229 if (stereo) { in sbrDecoder_Parse()
1237 if (stereo) { in sbrDecoder_Parse()
1341 if (stereo && newSbrInfo.pvc_mode > 0) { in sbrDecoder_Parse()
1425 if (!stereo && (self->hParametricStereoDec != NULL)) { in sbrDecoder_Parse()
1432 hSbrHeader, hFrameDataLeft, (stereo) ? hFrameDataRight : NULL, in sbrDecoder_Parse()
1435 (stereo) ? NULL : self->hParametricStereoDec, self->flags, in sbrDecoder_Parse()
1512 if (stereo) { in sbrDecoder_Parse()
1521 if (!stereo) { in sbrDecoder_Parse()
[all …]
/external/webrtc/webrtc/voice_engine/test/auto_test/fakes/
Dfake_media_process.h23 bool stereo) { in Process() argument
25 if (!stereo) { in Process()
/external/deqp/framework/platform/win32/
DtcuWGL.hpp103 bool stereo; member in tcu::wgl::PixelFormatInfo
143 , stereo (false) in PixelFormatInfo()
/external/libopus/celt/
Dbands.c647 static int compute_qn(int N, int b, int offset, int pulse_cap, int stereo) in compute_qn() argument
653 if (stereo && N==2) in compute_qn()
703 int stereo, int *fill) in compute_theta() argument
730 offset = (pulse_cap>>1) - (stereo&&N==2 ? QTHETA_OFFSET_TWOPHASE : QTHETA_OFFSET); in compute_theta()
731 qn = compute_qn(N, *b, offset, pulse_cap, stereo); in compute_theta()
732 if (stereo && i>=intensity) in compute_theta()
740 itheta = stereo_itheta(X, Y, stereo, N, ctx->arch); in compute_theta()
747 if (!stereo || ctx->theta_round == 0) in compute_theta()
750 if (!stereo && ctx->avoid_split_noise && itheta > 0 && itheta < qn) in compute_theta()
777 if (stereo && N>2) in compute_theta()
[all …]
Drate.c257 int stereo; in interp_bits2pulses() local
266 stereo = C>1; in interp_bits2pulses()
478 ebits[j] = bits[j] >> stereo >> BITRES; in interp_bits2pulses()
505 extra_fine = IMIN(excess>>(stereo+BITRES),MAX_FINE_BITS-ebits[j]); in interp_bits2pulses()
523 ebits[j] = bits[j] >> stereo >> BITRES; in interp_bits2pulses()
Dvq.h81 int stereo_itheta(const celt_norm *X, const celt_norm *Y, int stereo, int N, int arch);
/external/webrtc/webrtc/modules/audio_device/
Daudio_device_impl.cc1031 bool stereo(false); in StereoRecording() local
1033 if (_ptrAudioDevice->StereoRecording(stereo) == -1) in StereoRecording()
1038 *enabled = stereo; in StereoRecording()
1061 bool stereo(false); in SetRecordingChannel() local
1063 if (_ptrAudioDevice->StereoRecording(stereo) == -1) in SetRecordingChannel()
1161 bool stereo(false); in StereoPlayout() local
1163 if (_ptrAudioDevice->StereoPlayout(stereo) == -1) in StereoPlayout()
1168 *enabled = stereo; in StereoPlayout()
/external/webrtc/data/voice_engine/stereo_rtp_files/
DREADME.txt3 This sends the stereo rtp file to port 1236.
/external/python/cpython2/Doc/library/
Daudioop.rst213 Convert a stereo fragment to a mono fragment. The left channel is multiplied by
220 Generate a stereo fragment from a mono fragment. Each pair of samples in the
221 stereo fragment are computed from the mono sample, whereby left channel samples
232 between mono and stereo fragments, i.e. all samples are treated equal. If this
233 is a problem the stereo fragment should be split into two mono fragments first
Daifc.rst30 sampled. The number of channels indicate if the audio is mono, stereo, or
37 channels (stereo) and has a frame rate of 44,100 frames/second. This gives a
61 Return the number of audio channels (1 for mono, 2 for stereo).
Dwave.rst14 It does not support compression/decompression, but it does support mono/stereo.
70 Returns number of audio channels (``1`` for mono, ``2`` for stereo).
/external/python/cpython3/Doc/library/
Daudioop.rst221 Convert a stereo fragment to a mono fragment. The left channel is multiplied by
228 Generate a stereo fragment from a mono fragment. Each pair of samples in the
229 stereo fragment are computed from the mono sample, whereby left channel samples
240 between mono and stereo fragments, i.e. all samples are treated equal. If this
241 is a problem the stereo fragment should be split into two mono fragments first
Daifc.rst29 sampled. The number of channels indicate if the audio is mono, stereo, or
36 channels (stereo) and has a frame rate of 44,100 frames/second. This gives a
65 Return the number of audio channels (1 for mono, 2 for stereo).
/external/autotest/client/site_tests/audio_Microphone/
Dcontrol10 Mono and stereo capturing should work fine @ 44.1KHz and 48KHz.
/external/webrtc/resources/audio_coding/
DREAD.ME4 teststereo32kHz.pcm - stereo speech file samples at 32 kHz
/external/libopus/tests/
Drun_vectors.sh104 echo Testing stereo
/external/sonic/debian/
Dchangelog.vinux3 * Converted to mostly integer arithmetic, and added stereo support.
/external/perfetto/src/ftrace_reader/test/data/android_flounder_lte_LRX16F_3.10.40/events/display/display_mode/
Dformat24 …h: H=%d V=%d back_porch: H=%d V=%d active: H=%d V=%d front_porch: H=%d V=%d pclk=%ld stereo mode=%d
/external/tensorflow/tensorflow/core/api_def/base_api/
Dapi_def_DecodeWav.pbtxt47 number of samples. For example, a ten-sample-long stereo WAV file should give an
/external/webrtc/webrtc/modules/audio_processing/test/
Daudio_processing_unittest.cc105 void MixStereoToMono(const float* stereo, float* mono, in MixStereoToMono() argument
108 mono[i] = (stereo[i * 2] + stereo[i * 2 + 1]) / 2; in MixStereoToMono()
111 void MixStereoToMono(const int16_t* stereo, int16_t* mono, in MixStereoToMono() argument
114 mono[i] = (stereo[i * 2] + stereo[i * 2 + 1]) >> 1; in MixStereoToMono()
117 void CopyLeftToRightChannel(int16_t* stereo, size_t samples_per_channel) { in CopyLeftToRightChannel() argument
119 stereo[i * 2 + 1] = stereo[i * 2]; in CopyLeftToRightChannel()
123 void VerifyChannelsAreEqual(int16_t* stereo, size_t samples_per_channel) { in VerifyChannelsAreEqual() argument
125 EXPECT_EQ(stereo[i * 2 + 1], stereo[i * 2]); in VerifyChannelsAreEqual()

12