Home
last modified time | relevance | path

Searched refs:SHR32 (Results 1 – 25 of 38) sorted by relevance

12

/external/speex/libspeex/
Dlpc_bfin.h56 ac0 = ADD32(ac0,SHR32(MULT16_16(x[j],x[j]),8)); in _spx_autocorr()
115 d = ADD32(d,SHR32(MULT16_16(x[j],x[j]), shift)); in _spx_autocorr()
124 d = ADD32(d,SHR32(MULT16_16(x[j],x[j-i]), shift)); in _spx_autocorr()
128 ac[i] = SHR32(ac32[i], ac_shift); in _spx_autocorr()
Dpreprocess.c98 if (SHR32(a,7) >= b) in DIV32_16_Q8()
104 a = SHR32(a,8); in DIV32_16_Q8()
105 b = SHR32(b,8); in DIV32_16_Q8()
109 a = SHR32(a,4); in DIV32_16_Q8()
110 b = SHR32(b,4); in DIV32_16_Q8()
114 a = SHR32(a,4); in DIV32_16_Q8()
115 b = SHR32(b,4); in DIV32_16_Q8()
124 if (SHR32(a,15) >= b) in DIV32_16_Q15()
130 a = SHR32(a,8); in DIV32_16_Q15()
131 b = SHR32(b,8); in DIV32_16_Q15()
[all …]
Dfilters_bfin.h460 ytmp[i] = EXTRACT16(SHR32(yi,LPC_SHIFT));
462 y[i] = EXTRACT16(SHR32(yi2,LPC_SHIFT));
478 x[i] = EXTRACT16(SHR32(_x[i],SIG_SHIFT));
490 y[i] = EXTRACT16(SHR32(_y[i],SIG_SHIFT));
501 y[i] = EXTRACT16(SHR32(_y[i],SIG_SHIFT));
Dlpc.c141 ac0 = ADD32(ac0,SHR32(MULT16_16(x[j],x[j]),8)); in _spx_autocorr()
162 d = ADD32(d,SHR32(MULT16_16(x[j],x[j-i]), shift)); in _spx_autocorr()
165 ac[i] = SHR32(d, ac_shift); in _spx_autocorr()
Dstereo.c164 e_left += SHR32(MULT16_16(data[2*i],data[2*i]),8); in speex_encode_stereo_int()
165 e_right += SHR32(MULT16_16(data[2*i+1],data[2*i+1]),8); in speex_encode_stereo_int()
172 e_tot += SHR32(MULT16_16(data[i],data[i]),8); in speex_encode_stereo_int()
235 e_left = SHR32(MULT16_16(spx_sqrt(balance), e_right), 8); in speex_decode_stereo()
262 e_left = SHR32(MULT16_16(spx_sqrt(balance), e_right), 8); in speex_decode_stereo_int()
Dfixed_debug.h132 static inline int SHR32(long long a, int shift) in SHR32() function
164 #define PSHR32(a,shift) (SHR32(ADD32((a),((EXTEND32(1)<<((shift))>>1))),shift))
165 #define VSHR32(a, shift) (((shift)>0) ? SHR32(a, shift) : SHL32(a, -(shift)))
269 #define MAC16_16_Q11(c,a,b) (EXTRACT16(ADD16((c),EXTRACT16(SHR32(MULT16_16((a),(b)),11)))))
270 #define MAC16_16_Q13(c,a,b) (EXTRACT16(ADD16((c),EXTRACT16(SHR32(MULT16_16((a),(b)),13)))))
271 #define MAC16_16_P13(c,a,b) (EXTRACT16(ADD32((c),SHR32(ADD32(4096,MULT16_16((a),(b))),13))))
Dvq.c86 dist=SUB32(SHR32(E[i],1),dist); in vq_nbest()
128 dist = ADD32(dist,SHR32(E[i],1)); in vq_nbest_sign()
Dfixed_generic.h47 #define SHR32(a,shift) ((a) >> (shift)) macro
50 #define PSHR32(a,shift) (SHR32((a)+((EXTEND32(1)<<((shift))>>1)),shift))
51 #define VSHR32(a, shift) (((shift)>0) ? SHR32(a, shift) : SHL32(a, -(shift)))
Dmath_approx.h126 res = MULT16_16(EXTRACT16(SHR32(*seed,16)),std); in speex_rand()
127 return EXTRACT16(PSHR32(SUB32(res, SHR32(res, 3)),14)); in speex_rand()
299 x = DIV32_16(SHL32(EXTEND32(32767),29-e), EXTRACT16(SHR32(x, e-14))); in spx_atan()
Dfilters.c118 y[i] = SHL32(MULT16_32_Q14(EXTRACT16(SHR32(x[i],7)),scale),7); in signal_mul()
134 } else if (scale > SHR32(EXTEND32(SIG_SCALING), 2)) { in signal_div()
206 tmp = EXTRACT16(SHR32(x[i],sig_shift)); in compute_rms()
208 tmp = EXTRACT16(SHR32(x[i+1],sig_shift)); in compute_rms()
210 tmp = EXTRACT16(SHR32(x[i+2],sig_shift)); in compute_rms()
212 tmp = EXTRACT16(SHR32(x[i+3],sig_shift)); in compute_rms()
214 sum = ADD32(sum,SHR32(sum2,6)); in compute_rms()
243 sum = ADD32(sum,SHR32(sum2,6)); in compute_rms16()
262 sum = ADD32(sum,SHR32(sum2,6)); in compute_rms16()
292 y[i] = EXTRACT16(SHR32(x[i], sig_shift)); in normalize16()
Dmdf.c228 sum = ADD32(sum,SHR32(part,6)); in mdf_inner_prod()
363 … tmp += MULT16_16(EXTRACT16(SHR32(W[p*N*M + i*N+j],18)), EXTRACT16(SHR32(W[p*N*M + i*N+j],18))); in mdf_adjust_prop()
488 spx_word16_t decay = SHR32(spx_exp(NEG16(DIV32_16(QCONST16(2.4,11),M))),1); in speex_echo_state_init_mc()
1030 } else if (SHR32(Sff, 2) > ADD32(Sdd, SHR32(MULT16_16(N, 10000),6))) in speex_echo_cancellation()
1046 See = MAX32(See, SHR32(MULT16_16(N, 100),6)); in speex_echo_cancellation()
1107 tmp32 = ADD32(SHR32(Sxx,13), ADD32(tmp32, SHL32(tmp32,1))); in speex_echo_cancellation()
1117 if (tmp32 > SHR32(See,1)) in speex_echo_cancellation()
1118 tmp32 = SHR32(See,1); in speex_echo_cancellation()
1145 if (r>SHR32(e,1)) in speex_echo_cancellation()
1146 r = SHR32(e,1); in speex_echo_cancellation()
[all …]
Dltp.c71 sum = ADD32(sum,SHR32(part,6)); in inner_prod()
129 sum1 = ADD32(sum1,SHR32(part1,6));
130 sum2 = ADD32(sum2,SHR32(part2,6));
131 sum3 = ADD32(sum3,SHR32(part3,6));
132 sum4 = ADD32(sum4,SHR32(part4,6));
235 …energy[i-start+1] = SUB32(ADD32(energy[i-start],SHR32(MULT16_16(sw[-i-1],sw[-i-1]),6)), SHR32(MULT… in open_loop_nbest_pitch()
298 …g = DIV32(SHL32(EXTEND32(corr16[i-start]),cshift), 10+SHR32(MULT16_16(spx_sqrt(e0),spx_sqrt(SHL32(… in open_loop_nbest_pitch()
Dkiss_fftr.c237 f2k.r = SHR32(SUB32(EXTEND32(st->tmpbuf[k].r), EXTEND32(st->tmpbuf[ncfft-k].r)),1); in kiss_fftr2()
243 …twr = SHR32(SUB32(MULT16_16(f2k.r,st->super_twiddles[k].r),MULT16_16(f2k.i,st->super_twiddles[k].i… in kiss_fftr2()
244 …twi = SHR32(ADD32(MULT16_16(f2k.i,st->super_twiddles[k].r),MULT16_16(f2k.r,st->super_twiddles[k].i… in kiss_fftr2()
/external/libopus/celt/
Dfixed_debug.h45 …DD32(SHL32(MULT16_16(SHR32((a),16),SHR((b),16)),1), SHR32(MULT16_16SU(SHR32((a),16),((b)&0x0000fff…
48 #define MULT16_32_Q16(a,b) ADD32(MULT16_16((a),SHR32((b),16)), SHR32(MULT16_16SU((a),((b)&0x0000fff…
59 #define SHR(a,b) SHR32(a,b)
182 static OPUS_INLINE int SHR32(opus_int64 a, int shift) in SHR32() function
226 #define PSHR32(a,shift) (celt_mips--,SHR32(ADD32((a),(((opus_val32)(1)<<((shift))>>1))),shift))
227 #define VSHR32(a, shift) (((shift)>0) ? SHR32(a, shift) : SHL32(a, -(shift)))
231 #define HALF32(x) (SHR32(x,1))
Dpitch.c69 Syy = ADD32(Syy, SHR32(MULT16_16(y[j],y[j]), yshift)); in find_best_pitch()
100 Syy += SHR32(MULT16_16(y[i+len],y[i+len]),yshift) - SHR32(MULT16_16(y[i],y[i]),yshift); in find_best_pitch()
173 x_lp[i] = SHR32(HALF32(HALF32(x[0][(2*i-1)]+x[0][(2*i+1)])+x[0][2*i]), shift); in pitch_downsample()
174 x_lp[0] = SHR32(HALF32(HALF32(x[0][1])+x[0][0]), shift); in pitch_downsample()
178 x_lp[i] += SHR32(HALF32(HALF32(x[1][(2*i-1)]+x[1][(2*i+1)])+x[1][2*i]), shift); in pitch_downsample()
179 x_lp[0] += SHR32(HALF32(HALF32(x[1][1])+x[1][0]), shift); in pitch_downsample()
187 ac[0] += SHR32(ac[0],13); in pitch_downsample()
369 sum += SHR32(MULT16_16(x_lp[j],y[i+j]), shift); in pitch_search()
513 pg = SHR32(frac_div32(best_xy,best_yy+1),16); in remove_doubling()
Dcelt_lpc.c61 rr += SHR32(ac[i + 1],3); in _celt_lpc()
64 lpc[i] = SHR32(r,3); in _celt_lpc()
77 if (error<SHR32(ac[0],10)) in _celt_lpc()
261 if (n&1) ac0 += SHR32(MULT16_16(xptr[0],xptr[0]),9); in _celt_autocorr()
264 ac0 += SHR32(MULT16_16(xptr[i],xptr[i]),9); in _celt_autocorr()
265 ac0 += SHR32(MULT16_16(xptr[i+1],xptr[i+1]),9); in _celt_autocorr()
302 ac[i] = SHR32(ac[i], shift2); in _celt_autocorr()
Dfixed_generic.h72 #define SHR32(a,shift) ((a) >> (shift)) macro
77 #define PSHR32(a,shift) (SHR32((a)+((EXTEND32(1)<<((shift))>>1)),shift))
79 #define VSHR32(a, shift) (((shift)>0) ? SHR32(a, shift) : SHL32(a, -(shift)))
93 #define HALF32(x) (SHR32(x,1))
Dvq.c51 Xptr[stride] = EXTRACT16(SHR32(MULT16_16(c,x2) + MULT16_16(s,x1), 15)); in exp_rotation1()
52 *Xptr++ = EXTRACT16(SHR32(MULT16_16(c,x1) - MULT16_16(s,x2), 15)); in exp_rotation1()
60 Xptr[stride] = EXTRACT16(SHR32(MULT16_16(c,x2) + MULT16_16(s,x1), 15)); in exp_rotation1()
61 *Xptr-- = EXTRACT16(SHR32(MULT16_16(c,x1) - MULT16_16(s,x2), 15)); in exp_rotation1()
272 Rxy = EXTRACT16(SHR32(ADD32(xy, EXTEND32(X[j])),rshift)); in alg_quant()
Dcelt_decoder.c489 ac[0] += SHR32(ac[0],13); in celt_decode_lost()
536 E1 += SHR32(MULT16_16(e, e), shift); in celt_decode_lost()
538 E2 += SHR32(MULT16_16(e, e), shift); in celt_decode_lost()
541 decay = celt_sqrt(frac_div32(SHR32(E1, 1), E2)); in celt_decode_lost()
573 S1 += SHR32(MULT16_16(tmp, tmp), 8); in celt_decode_lost()
597 S2 += SHR32(MULT16_16(tmp, tmp), 8); in celt_decode_lost()
601 if (!(S1 > SHR32(S2,2))) in celt_decode_lost()
612 opus_val16 ratio = celt_sqrt(frac_div32(SHR32(S1,1)+1,S2+1)); in celt_decode_lost()
D_kiss_fft_guts.h69 do{ (m).r = SHR32(SUB32(S_MUL((a).r,(b).r) , S_MUL((a).i,(b).i)),2); \
70 (m).i = SHR32(ADD32(S_MUL((a).r,(b).i) , S_MUL((a).i,(b).r)),2); }while(0)
Dcelt_encoder.c265 x = SHR32(in[i+c*len],SIG_SHIFT); in transient_analysis()
269 mem1 = x - SHR32(y,1); in transient_analysis()
274 tmp[i] = EXTRACT16(SHR32(y,2)); in transient_analysis()
341 norm = SHL32(EXTEND32(len2),6+14)/ADD32(EPSILON,SHR32(mean,1)); in transient_analysis()
776 sum = ADD16(sum, EXTRACT16(SHR32(partial, 18))); in alloc_trim_analysis()
787 minXC = MIN16(minXC, ABS16(EXTRACT16(SHR32(partial, 18)))); in alloc_trim_analysis()
978 boost = (int)SHR32(EXTEND32(follower[i]),DB_SHIFT); in dynalloc_analysis()
981 boost = (int)SHR32(EXTEND32(follower[i])*8,DB_SHIFT); in dynalloc_analysis()
984 boost = (int)SHR32(EXTEND32(follower[i])*width/6,DB_SHIFT); in dynalloc_analysis()
1175SHR32(MULT16_16(stereo_saving-QCONST16(0.1f,8),(coded_stereo_dof<<BITRES)),8)); in compute_vbr()
[all …]
Dquant_bands.c153 return MIN32(200,SHR32(dist,2*DB_SHIFT-6)); in loss_distortion()
387 …offset = SUB16(SHR32(SHL32(EXTEND32(q2),DB_SHIFT)+QCONST16(.5f,DB_SHIFT),fine_quant[i]),QCONST16(.… in quant_fine_energy()
506 …offset = SUB16(SHR32(SHL32(EXTEND32(q2),DB_SHIFT)+QCONST16(.5f,DB_SHIFT),fine_quant[i]),QCONST16(.… in unquant_fine_energy()
/external/libopus/src/
Dmlp.c53 xx = EXTRACT16(SHR32(_x, 8));
55 i = SHR32(ADD32(1024,MULT16_16(25, xx)),11);
57 xx -= EXTRACT16(SHR32(MULT16_16(20972,i),8));
Dopus_encoder.c420 g = SHR32(MAC16_16(MULT16_16(w,g2), in stereo_fade()
452 g = SHR32(MAC16_16(MULT16_16(w,g2), in gain_fade()
461 g = SHR32(MAC16_16(MULT16_16(w,g2), in gain_fade()
871 pxx = SHR32(MULT16_16(x,x),2); in compute_stereo_width()
872 pxy = SHR32(MULT16_16(x,y),2); in compute_stereo_width()
873 pyy = SHR32(MULT16_16(y,y),2); in compute_stereo_width()
876 pxx += SHR32(MULT16_16(x,x),2); in compute_stereo_width()
877 pxy += SHR32(MULT16_16(x,y),2); in compute_stereo_width()
878 pyy += SHR32(MULT16_16(y,y),2); in compute_stereo_width()
881 pxx += SHR32(MULT16_16(x,x),2); in compute_stereo_width()
[all …]
/external/libopus/celt/arm/
Dkiss_fft_armv5e.h86 (m).r = SHR32(SUB32(mr1__, mr2__), 1); \
87 (m).i = SHR32(mi__, 1); \

12