/third_party/cmsis/CMSIS/DSP/Include/dsp/ |
D | matrix_utils.h | 83 f16x8_t tmpa,tmpb; \ 86 tmpa=vldrhq_z_f16(&data[i*_numCols + _w],p0);\ 90 vstrhq_p(&data[j*_numCols + _w], tmpa, p0); \ 105 f16x8_t tmpa; \ 107 tmpa = vldrhq_z_f16(&data[i*_numCols + _w],p0);\ 108 tmpa = vmulq_n_f16(tmpa,(_Float16)v); \ 109 vstrhq_p(&data[i*_numCols + _w], tmpa, p0); \ 125 f16x8_t tmpa,tmpb; \ 127 tmpa = vldrhq_z_f16(&dataA[i*_numCols + _w],p0);\ 129 tmpa = vfmaq_n_f16(tmpa,tmpb,v); \ [all …]
|
/third_party/wpa_supplicant/wpa_supplicant-2.9_standard/src/tls/ |
D | libtommath.c | 250 register mp_digit u, *tmpa, *tmpb, *tmpc; in s_mp_add() local 256 tmpa = a->dp; in s_mp_add() 268 *tmpc = *tmpa++ + *tmpb++ + u; in s_mp_add() 326 register mp_digit u, *tmpa, *tmpb, *tmpc; in s_mp_sub() local 330 tmpa = a->dp; in s_mp_sub() 338 *tmpc = *tmpa++ - *tmpb++ - u; in s_mp_sub() 354 *tmpc = *tmpa++ - u; in s_mp_sub() 930 mp_digit *tmpa, *tmpb; in mp_cmp_mag() local 942 tmpa = a->dp + (a->used - 1); in mp_cmp_mag() 948 for (n = 0; n < a->used; ++n, --tmpa, --tmpb) { in mp_cmp_mag() [all …]
|
/third_party/wpa_supplicant/wpa_supplicant-2.9/src/tls/ |
D | libtommath.c | 250 register mp_digit u, *tmpa, *tmpb, *tmpc; in s_mp_add() local 256 tmpa = a->dp; in s_mp_add() 268 *tmpc = *tmpa++ + *tmpb++ + u; in s_mp_add() 326 register mp_digit u, *tmpa, *tmpb, *tmpc; in s_mp_sub() local 330 tmpa = a->dp; in s_mp_sub() 338 *tmpc = *tmpa++ - *tmpb++ - u; in s_mp_sub() 354 *tmpc = *tmpa++ - u; in s_mp_sub() 930 mp_digit *tmpa, *tmpb; in mp_cmp_mag() local 942 tmpa = a->dp + (a->used - 1); in mp_cmp_mag() 948 for (n = 0; n < a->used; ++n, --tmpa, --tmpb) { in mp_cmp_mag() [all …]
|
/third_party/cmsis/CMSIS/DSP/Source/FilteringFunctions/ |
D | arm_levinson_durbin_f32.c | 161 uint64_t tmpa,tmpb; in arm_levinson_durbin_f32() local 165 tmpa = vgetq_lane_u64((uint64x2_t)vecA,0); in arm_levinson_durbin_f32() 168 vecRevA = (f32x4_t) vsetq_lane_u64(tmpa,(uint64x2_t)vecRevA,1); in arm_levinson_durbin_f32()
|
D | arm_levinson_durbin_f16.c | 158 uint64_t tmpa,tmpb; in arm_levinson_durbin_f16() local 162 tmpa = vgetq_lane_u64((uint64x2_t)vecA,0); in arm_levinson_durbin_f16() 165 vecRevA = (f16x8_t) vsetq_lane_u64(tmpa,(uint64x2_t)vecRevA,1); in arm_levinson_durbin_f16()
|
D | arm_levinson_durbin_q31.c | 230 uint64_t tmpa,tmpb; in arm_levinson_durbin_q31() local 234 tmpa = vgetq_lane_u64((uint64x2_t)vecA,0); in arm_levinson_durbin_q31() 237 vecRevA = (q31x4_t) vsetq_lane_u64(tmpa,(uint64x2_t)vecRevA,1); in arm_levinson_durbin_q31()
|
/third_party/openh264/test/encoder/ |
D | EncUT_Sample.cpp | 16 int32_t tmpa, tmpb; \ 33 tmpa = ref (pDec + 128, iLineSizeDec, pEnc, iLineSizeEnc, &iBestMode_c, iLambda, pDst); \ 35 ASSERT_EQ (tmpa, tmpb); \ 43 int32_t tmpa, tmpb; \ 62 …tmpa = WelsSampleSatdIntra4x4Combined3_c (pDec + 128, iLineSizeDec, pEnc, iLineSizeEnc, pDst, &iBe… 66 ASSERT_EQ (tmpa, tmpb); \ 74 int32_t tmpa, tmpb; \ 97 tmpa = ref (pDecCb + 128, iLineSizeDec, pEncCb, iLineSizeEnc, &iBestMode_c, iLambda, \ 101 ASSERT_EQ (tmpa, tmpb); \
|
/third_party/mksh/ |
D | edit.c | 2189 unsigned int tmpa, tmpb; in x_transpose() local 2218 if (utf_mbtowc(&tmpa, xcp) == (size_t)-1) { in x_transpose() 2227 utf_wctomb(xcp, tmpa); in x_transpose() 2236 if (utf_mbtowc(&tmpa, xcp) == (size_t)-1) { in x_transpose() 2245 utf_wctomb(xcp, tmpa); in x_transpose()
|