/third_party/cmsis/CMSIS/DSP/Source/FilteringFunctions/ |
D | arm_biquad_cascade_df2T_f64.c | 148 float64_t acc1; /* Accumulator */ in arm_biquad_cascade_df2T_f64() local 184 acc1 = b0 * Xn1 + d1; in arm_biquad_cascade_df2T_f64() 187 d1 += a1 * acc1; in arm_biquad_cascade_df2T_f64() 190 d2 += a2 * acc1; in arm_biquad_cascade_df2T_f64() 192 *pOut++ = acc1; in arm_biquad_cascade_df2T_f64() 198 acc1 = b0 * Xn1 + d1; in arm_biquad_cascade_df2T_f64() 201 d1 += a1 * acc1; in arm_biquad_cascade_df2T_f64() 204 d2 += a2 * acc1; in arm_biquad_cascade_df2T_f64() 206 *pOut++ = acc1; in arm_biquad_cascade_df2T_f64() 211 acc1 = b0 * Xn1 + d1; in arm_biquad_cascade_df2T_f64() [all …]
|
D | arm_biquad_cascade_df2T_f16.c | 59 float16_t acc0, acc1; in arm_biquad_cascade_df2T_f16() local 134 acc1 = vgetq_lane(state, 1); in arm_biquad_cascade_df2T_f16() 135 state = vfmaq(state, a1Coeffs, acc1); in arm_biquad_cascade_df2T_f16() 146 *pOut++ = acc1; in arm_biquad_cascade_df2T_f16() 200 _Float16 acc1; /* Accumulator */ in arm_biquad_cascade_df2T_f16() local 235 acc1 = b0 * Xn1 + d1; in arm_biquad_cascade_df2T_f16() 238 d1 += a1 * acc1; in arm_biquad_cascade_df2T_f16() 241 d2 += a2 * acc1; in arm_biquad_cascade_df2T_f16() 243 *pOut++ = acc1; in arm_biquad_cascade_df2T_f16() 248 acc1 = b0 * Xn1 + d1; in arm_biquad_cascade_df2T_f16() [all …]
|
D | arm_biquad_cascade_df2T_f32.c | 59 float32_t acc0, acc1; in arm_biquad_cascade_df2T_f32() local 129 acc1 = vgetq_lane(state, 1); in arm_biquad_cascade_df2T_f32() 130 state = vfmaq(state, a1Coeffs, acc1); in arm_biquad_cascade_df2T_f32() 141 *pOut++ = acc1; in arm_biquad_cascade_df2T_f32() 195 float32_t acc1; /* accumulator */ in arm_biquad_cascade_df2T_f32() local 315 acc1 = (b0 * Xn1) + d1; in arm_biquad_cascade_df2T_f32() 318 *pOut++ = acc1; in arm_biquad_cascade_df2T_f32() 322 d1 = ((b1 * Xn1) + (a1 * acc1)) + d2; in arm_biquad_cascade_df2T_f32() 325 d2 = (b2 * Xn1) + (a2 * acc1); in arm_biquad_cascade_df2T_f32() 357 float32_t acc1; /* Accumulator */ in arm_biquad_cascade_df2T_f32() local [all …]
|
D | arm_fir_q31.c | 235 q63_t acc0=0, acc1=0, acc2=0, acc3=0; in arm_fir_q31_1_4_mve() local 257 acc1 = vrmlaldavhq(vecIn0, vecCoeffs); in arm_fir_q31_1_4_mve() 266 acc1 = asrl(acc1, 23); in arm_fir_q31_1_4_mve() 271 *pOutput++ = (q31_t) acc1; in arm_fir_q31_1_4_mve() 298 acc1 = vrmlaldavhq(vecIn0, vecCoeffs); in arm_fir_q31_1_4_mve() 304 acc1 = asrl(acc1, 23); in arm_fir_q31_1_4_mve() 308 *pOutput++ = (q31_t) acc1; in arm_fir_q31_1_4_mve() 326 acc1 = vrmlaldavhq(vecIn0, vecCoeffs); in arm_fir_q31_1_4_mve() 329 acc1 = asrl(acc1, 23); in arm_fir_q31_1_4_mve() 332 *pOutput++ = (q31_t) acc1; in arm_fir_q31_1_4_mve() [all …]
|
D | arm_fir_interpolate_q31.c | 116 q63_t acc1 = 0LL; in arm_fir_interpolate_q31() local 129 acc1 = vrmlaldavhaq(acc1, vecState, vecCoef); in arm_fir_interpolate_q31() 152 acc1 = vrmlaldavhaq(acc1, vecState, vecCoef); in arm_fir_interpolate_q31() 162 acc1 = asrl(acc1, 31 - 8); in arm_fir_interpolate_q31() 167 *pDst++ = (q31_t) acc1; in arm_fir_interpolate_q31() 180 q63_t acc1 = 0LL; in arm_fir_interpolate_q31() local 192 acc1 = vrmlaldavhaq(acc1, vecState, vecCoef); in arm_fir_interpolate_q31() 212 acc1 = vrmlaldavhaq(acc1, vecState, vecCoef); in arm_fir_interpolate_q31() 219 acc1 = asrl(acc1, 31 - 8); in arm_fir_interpolate_q31() 223 *pDst++ = (q31_t) acc1; in arm_fir_interpolate_q31() [all …]
|
D | arm_fir_interpolate_q15.c | 119 q63_t acc1 = 0LL; in arm_fir_interpolate_q15() local 132 acc1 = vmlaldavaq(acc1, vecState, vecCoef); in arm_fir_interpolate_q15() 155 acc1 = vmlaldavaq(acc1, vecState, vecCoef); in arm_fir_interpolate_q15() 165 acc1 = asrl(acc1, 15); in arm_fir_interpolate_q15() 170 *pDst++ = (q15_t) __SSAT(acc1, 16); in arm_fir_interpolate_q15() 183 q63_t acc1 = 0LL; in arm_fir_interpolate_q15() local 195 acc1 = vmlaldavaq(acc1, vecState, vecCoef); in arm_fir_interpolate_q15() 215 acc1 = vmlaldavaq(acc1, vecState, vecCoef); in arm_fir_interpolate_q15() 222 acc1 = asrl(acc1, 15); in arm_fir_interpolate_q15() 226 *pDst++ = (q15_t) __SSAT(acc1, 16);; in arm_fir_interpolate_q15() [all …]
|
D | arm_conv_q15.c | 117 int64_t acc1 = 0LL; in arm_conv_q15() local 122 MVE_INTR_CONV_DUAL_INC_Y_INC_SIZE_Q15(acc0, acc1, pX, pY, count); in arm_conv_q15() 124 *pDst++ = (q15_t) acc1; in arm_conv_q15() 144 int64_t acc1 = 0LL; in arm_conv_q15() local 155 MVE_INTR_CONV_QUAD_INC_X_FIXED_SIZE_Q15(acc0, acc1, acc2, acc3, pX, pY, count); in arm_conv_q15() 157 *pDst++ = (q15_t) acc1; in arm_conv_q15() 167 int64_t acc1 = 0LL; in arm_conv_q15() local 176 MVE_INTR_CONV_DUAL_INC_X_FIXED_SIZE_Q15(acc0, acc1, pX, pY, count); in arm_conv_q15() 178 *pDst++ = (q15_t) acc1; in arm_conv_q15() 199 int64_t acc1 = 0LL; in arm_conv_q15() local [all …]
|
D | arm_correlate_q15.c | 137 int64_t acc1 = 0LL; in arm_correlate_q15() local 146 MVE_INTR_CORR_DUAL_DEC_Y_INC_SIZE_Q15(acc0, acc1, pX, pY, count); in arm_correlate_q15() 150 *pDst = (q15_t) acc1; in arm_correlate_q15() 171 int64_t acc1 = 0LL; in arm_correlate_q15() local 182 MVE_INTR_CORR_QUAD_INC_X_FIXED_SIZE_Q15(acc0, acc1, acc2, acc3, pX, pY, srcBLen); in arm_correlate_q15() 186 *pDst = (q15_t) acc1; in arm_correlate_q15() 198 int64_t acc1 = 0LL; in arm_correlate_q15() local 207 MVE_INTR_CORR_DUAL_INC_X_FIXED_SIZE_Q15(acc0, acc1, pX, pY, srcBLen); in arm_correlate_q15() 211 *pDst = (q15_t) acc1; in arm_correlate_q15() 234 int64_t acc1 = 0LL; in arm_correlate_q15() local [all …]
|
/third_party/cmsis/CMSIS/DSP/PrivateInclude/ |
D | arm_vec_filtering.h | 38 #define MVE_INTR_CORR_QUAD_INC_X_FIXED_SIZE_F32(acc0, acc1, acc2, acc3, pX, pY, count)\ argument 87 acc1 = vecAddAcrossF32Mve(acc1Vec); \ 127 #define MVE_INTR_CORR_DUAL_INC_X_DEC_SIZE_F32(acc0, acc1, pX, pY, count)\ argument 168 acc1 = vecAddAcrossF32Mve(acc1Vec); \ 171 #define MVE_INTR_CORR_DUAL_INC_X_FIXED_SIZE_F32(acc0, acc1, pX, pY, count)\ argument 210 acc1 = vecAddAcrossF32Mve(acc1Vec); \ 213 #define MVE_INTR_CORR_DUAL_DEC_Y_INC_SIZE_F32(acc0, acc1, pX, pY, count)\ argument 253 acc1 = vecAddAcrossF32Mve(acc1Vec); \ 256 #define MVE_INTR_CONV_DUAL_INC_X_DEC_SIZE_F32(acc0, acc1, pX, pY, count) … argument 289 … acc1 = vecAddAcrossF32Mve(acc1Vec); \ [all …]
|
/third_party/boost/libs/accumulators/test/ |
D | variance.cpp | 34 accumulator_set<int, stats<tag::variance(lazy)> > acc1; in test_stat() local 36 acc1(1); in test_stat() 37 acc1(2); in test_stat() 38 acc1(3); in test_stat() 39 acc1(4); in test_stat() 40 acc1(5); in test_stat() 42 BOOST_CHECK_EQUAL(5u, count(acc1)); in test_stat() 43 BOOST_CHECK_CLOSE(3., mean(acc1), 1e-5); in test_stat() 44 BOOST_CHECK_CLOSE(11., accumulators::moment<2>(acc1), 1e-5); in test_stat() 45 BOOST_CHECK_CLOSE(2., variance(acc1), 1e-5); in test_stat() [all …]
|
D | weighted_variance.cpp | 23 accumulator_set<int, stats<tag::weighted_variance(lazy)>, int> acc1; in test_stat() local 25 acc1(1, weight = 2); // 2 in test_stat() 26 acc1(2, weight = 3); // 6 in test_stat() 27 acc1(3, weight = 1); // 3 in test_stat() 28 acc1(4, weight = 4); // 16 in test_stat() 29 acc1(5, weight = 1); // 5 in test_stat() 33 BOOST_CHECK_EQUAL(5u, count(acc1)); in test_stat() 34 BOOST_CHECK_CLOSE(2.9090909, weighted_mean(acc1), 1e-5); in test_stat() 35 BOOST_CHECK_CLOSE(10.1818182, accumulators::weighted_moment<2>(acc1), 1e-5); in test_stat() 36 BOOST_CHECK_CLOSE(1.7190083, weighted_variance(acc1), 1e-5); in test_stat()
|
D | tail_variate_means.cpp | 38 accumulator_t1 acc1( right_tail_cache_size = c ); in test_stat() local 55 acc1(100., covariate1 = cov1); in test_stat() 56 acc1( 50., covariate1 = cov2); in test_stat() 57 acc1(200., covariate1 = cov3); in test_stat() 58 acc1( 80., covariate1 = cov4); in test_stat() 59 acc1( 20., covariate1 = cov5); in test_stat() 80 …BOOST_CHECK_EQUAL( *(relative_tail_variate_means(acc1, quantile_probability = 0.7).begin() ), … in test_stat() 81 …BOOST_CHECK_EQUAL( *(relative_tail_variate_means(acc1, quantile_probability = 0.7).begin() + 1), … in test_stat() 82 …BOOST_CHECK_EQUAL( *(relative_tail_variate_means(acc1, quantile_probability = 0.7).begin() + 2), … in test_stat() 83 …BOOST_CHECK_EQUAL( *(relative_tail_variate_means(acc1, quantile_probability = 0.7).begin() + 3), … in test_stat() [all …]
|
D | p_square_quantile.cpp | 39 accumulator_t acc1(quantile_probability = 0.01 ); in test_stat() local 52 acc1(sample); in test_stat() 63 BOOST_CHECK_CLOSE( p_square_quantile(acc1), 0.01 , 7*epsilon ); in test_stat() 85 accumulator_t acc1(quantile_probability = 0.75 ); in test_persistency() local 91 acc1(sample); in test_persistency() 95 BOOST_CHECK_CLOSE(p_square_quantile(acc1), 0.75 , epsilon); in test_persistency() 98 acc1.serialize(oa, 0); in test_persistency() 101 accumulator_t acc1(quantile_probability = 0.75); in test_persistency() local 104 acc1.serialize(ia, 0); in test_persistency() 106 BOOST_CHECK_CLOSE(p_square_quantile(acc1), 0.75 , epsilon); in test_persistency()
|
/third_party/openssl/crypto/ec/asm/ |
D | ecp_nistz256-armv8.pl | 47 $acc0,$acc1,$acc2,$acc3,$acc4,$acc5) = 225 ldp $acc0,$acc1,[$ap] 248 ldp $acc0,$acc1,[$ap] 269 ldp $acc0,$acc1,[$ap] 274 mov $t1,$acc1 294 ldp $acc0,$acc1,[$ap] 299 mov $t1,$acc1 303 mov $a1,$acc1 331 ldp $acc0,$acc1,[$ap] 354 mov $acc1,xzr [all …]
|
D | ecp_nistz256-ppc64.pl | 42 my ($rp,$ap,$bp,$bi,$acc0,$acc1,$acc2,$acc3,$poly1,$poly3, 210 ld $acc1,8($ap) 248 ld $acc1,8($ap) 283 ld $acc1,8($ap) 288 mr $t1,$acc1 323 ld $acc1,8($ap) 329 mr $t1,$acc1 330 std $acc1,72($sp) 375 ld $acc1,8($ap) 411 li $acc1,0 [all …]
|
D | ecp_nistz256-x86_64.pl | 490 my ($acc0,$acc1,$acc2,$acc3,$acc4,$acc5,$acc6,$acc7)=map("%r$_",(8..15)); 538 mov %rdx, $acc1 541 add %rax, $acc1 573 add $t0, $acc1 575 add %rax, $acc1 594 add %rax, $acc1 614 mov $acc1, $t0 615 imulq %r15, $acc1 623 mov $acc1, %rax 629 mov $acc1, $t1 [all …]
|
/third_party/cmsis/CMSIS/DSP/Source/MatrixFunctions/ |
D | arm_mat_cmplx_mult_f32.c | 82 f32x4_t acc0, acc1; in arm_mat_cmplx_mult_f32_2x2_mve() local 100 acc1 = vcmulq(vecA, vecB); in arm_mat_cmplx_mult_f32_2x2_mve() 101 acc1 = vcmlaq_rot90(acc1, vecA, vecB); in arm_mat_cmplx_mult_f32_2x2_mve() 105 pOut[1 * CMPLX_DIM * MATRIX_DIM2 + 0] = acc1[0] + acc1[2]; in arm_mat_cmplx_mult_f32_2x2_mve() 106 pOut[1 * CMPLX_DIM * MATRIX_DIM2 + 1] = acc1[1] + acc1[3]; in arm_mat_cmplx_mult_f32_2x2_mve() 121 acc1 = vcmulq(vecA, vecB); in arm_mat_cmplx_mult_f32_2x2_mve() 122 acc1 = vcmlaq_rot90(acc1, vecA, vecB); in arm_mat_cmplx_mult_f32_2x2_mve() 126 pOut[1 * CMPLX_DIM * MATRIX_DIM2 + 0] = acc1[0] + acc1[2]; in arm_mat_cmplx_mult_f32_2x2_mve() 127 pOut[1 * CMPLX_DIM * MATRIX_DIM2 + 1] = acc1[1] + acc1[3]; in arm_mat_cmplx_mult_f32_2x2_mve() 147 f32x4_t acc0, acc1, acc2; in arm_mat_cmplx_mult_f32_3x3_mve() local [all …]
|
D | arm_mat_cmplx_mult_q31.c | 78 q63_t acc0, acc1, acc2, acc3; in arm_mat_cmplx_mult_q31_2x2_mve() local 93 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_2x2_mve() 100 pOut[0 * CMPLX_DIM * MATRIX_DIM2 + 1] = (q31_t) asrl(acc1, 31); in arm_mat_cmplx_mult_q31_2x2_mve() 111 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_2x2_mve() 120 pOut[0 * CMPLX_DIM * MATRIX_DIM2 + 1] = (q31_t) asrl(acc1, 31); in arm_mat_cmplx_mult_q31_2x2_mve() 141 q63_t acc0, acc1, acc2, acc3; in arm_mat_cmplx_mult_q31_3x3_mve() local 167 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve() 175 acc1 = vmlaldavaxq_s32(acc1, vecA, vecB1); in arm_mat_cmplx_mult_q31_3x3_mve() 182 pOut[0 * CMPLX_DIM * MATRIX_DIM3 + 1] = (q31_t) asrl(acc1, 31); in arm_mat_cmplx_mult_q31_3x3_mve() 188 acc1 = vmlaldavxq_s32(vecA, vecB); in arm_mat_cmplx_mult_q31_3x3_mve() [all …]
|
D | arm_mat_mult_q31.c | 78 q63_t acc0, acc1; in arm_mat_mult_q31_2x2_mve() local 96 acc1 = vrmlaldavhq(vecA1, vecB); in arm_mat_mult_q31_2x2_mve() 99 acc1 = asrl(acc1, 23); in arm_mat_mult_q31_2x2_mve() 102 pOut[1 * MATRIX_DIM2] = (q31_t) acc1; in arm_mat_mult_q31_2x2_mve() 111 acc1 = vrmlaldavhq(vecA1, vecB); in arm_mat_mult_q31_2x2_mve() 114 acc1 = asrl(acc1, 23); in arm_mat_mult_q31_2x2_mve() 117 pOut[1 * MATRIX_DIM2] = (q31_t) acc1; in arm_mat_mult_q31_2x2_mve() 138 q63_t acc0, acc1, acc2; in arm_mat_mult_q31_3x3_mve() local 153 acc1 = vrmlaldavhq(vecA, vecB); in arm_mat_mult_q31_3x3_mve() 158 acc1 = asrl(acc1, 23); in arm_mat_mult_q31_3x3_mve() [all …]
|
D | arm_mat_mult_q15.c | 79 q63_t acc0, acc1; in arm_mat_mult_q15_2x2_mve() local 93 acc1 = vmlaldavq(vecA1, vecB); in arm_mat_mult_q15_2x2_mve() 96 acc1 = asrl(acc1, 15); in arm_mat_mult_q15_2x2_mve() 99 pOut[1 * MATRIX_DIM2] = (q15_t) __SSAT(acc1, 16); in arm_mat_mult_q15_2x2_mve() 108 acc1 = vmlaldavq(vecA1, vecB); in arm_mat_mult_q15_2x2_mve() 111 acc1 = asrl(acc1, 15); in arm_mat_mult_q15_2x2_mve() 114 pOut[1 * MATRIX_DIM2] = (q15_t) __SSAT(acc1, 16); in arm_mat_mult_q15_2x2_mve() 136 q63_t acc0, acc1, acc2; in arm_mat_mult_q15_3x3_mve() local 152 acc1 = vmlaldavq(vecA1, vecB); in arm_mat_mult_q15_3x3_mve() 156 acc1 = asrl(acc1, 15); in arm_mat_mult_q15_3x3_mve() [all …]
|
D | arm_mat_cmplx_mult_f16.c | 76 f16x8_t acc0, acc1; in arm_mat_cmplx_mult_f16_2x2_mve() local 107 acc1 = vcmulq(vecA1, vecB); in arm_mat_cmplx_mult_f16_2x2_mve() 108 acc1 = vcmlaq_rot90(acc1, vecA1, vecB); in arm_mat_cmplx_mult_f16_2x2_mve() 124 vecTmp = (f16x8_t) vrev64q_s32((int32x4_t) acc1); in arm_mat_cmplx_mult_f16_2x2_mve() 125 vecTmp = vaddq(vecTmp, acc1); in arm_mat_cmplx_mult_f16_2x2_mve() 151 f16x8_t acc0, acc1, acc2; in arm_mat_cmplx_mult_f16_3x3_mve() local 176 acc1 = vcmulq(vecA1, vecB); in arm_mat_cmplx_mult_f16_3x3_mve() 177 acc1 = vcmlaq_rot90(acc1, vecA1, vecB); in arm_mat_cmplx_mult_f16_3x3_mve() 183 mve_cmplx_sum_intra_vec_f16(acc1, &pOut[1 * CMPLX_DIM * MATRIX_DIM]); in arm_mat_cmplx_mult_f16_3x3_mve() 196 acc1 = vcmulq(vecA1, vecB); in arm_mat_cmplx_mult_f16_3x3_mve() [all …]
|
/third_party/openssl/crypto/bn/asm/ |
D | armv8-mont.pl | 279 my ($acc0,$acc1,$acc2,$acc3,$acc4,$acc5,$acc6,$acc7)=map("x$_",(19..26)); 329 mov $acc1,xzr 375 adds $acc1,$acc1,$t0 // t[1]+lo(a[1]*a[0]) 389 stp $acc0,$acc1,[$tp],#8*2 // t[0..1] 419 adc $acc1,xzr,xzr // t[9] 430 adc $acc1,$acc1,$t1 441 adcs $acc1,$acc1,$t2 451 adcs $acc1,$acc1,$t2 460 adcs $acc1,$acc1,$t2 468 adcs $acc1,$acc1,$t1 [all …]
|
/third_party/cmsis/CMSIS/DSP/Source/ComplexMathFunctions/ |
D | arm_cmplx_mag_squared_q31.c | 65 q31_t acc0, acc1; /* Accumulators */ in arm_cmplx_mag_squared_q31() local 97 acc1 = (q31_t) (((q63_t) imag * imag) >> 33); in arm_cmplx_mag_squared_q31() 100 *pDst++ = acc0 + acc1; in arm_cmplx_mag_squared_q31() 115 q31_t acc0, acc1; /* Accumulators */ in arm_cmplx_mag_squared_q31() local 129 acc1 = (q31_t) (((q63_t) imag * imag) >> 33); in arm_cmplx_mag_squared_q31() 131 *pDst++ = acc0 + acc1; in arm_cmplx_mag_squared_q31() 136 acc1 = (q31_t) (((q63_t) imag * imag) >> 33); in arm_cmplx_mag_squared_q31() 137 *pDst++ = acc0 + acc1; in arm_cmplx_mag_squared_q31() 142 acc1 = (q31_t) (((q63_t) imag * imag) >> 33); in arm_cmplx_mag_squared_q31() 143 *pDst++ = acc0 + acc1; in arm_cmplx_mag_squared_q31() [all …]
|
D | arm_cmplx_mag_q31.c | 68 q31_t acc0, acc1; /* Accumulators */ in arm_cmplx_mag_q31() local 111 acc1 = (q31_t) (((q63_t) imag * imag) >> 33); in arm_cmplx_mag_q31() 114 arm_sqrt_q31(acc0 + acc1, pDst++); in arm_cmplx_mag_q31() 129 q31_t acc0, acc1; /* Accumulators */ in arm_cmplx_mag_q31() local 143 acc1 = (q31_t) (((q63_t) imag * imag) >> 33); in arm_cmplx_mag_q31() 146 arm_sqrt_q31(acc0 + acc1, pDst++); in arm_cmplx_mag_q31() 151 acc1 = (q31_t) (((q63_t) imag * imag) >> 33); in arm_cmplx_mag_q31() 152 arm_sqrt_q31(acc0 + acc1, pDst++); in arm_cmplx_mag_q31() 157 acc1 = (q31_t) (((q63_t) imag * imag) >> 33); in arm_cmplx_mag_q31() 158 arm_sqrt_q31(acc0 + acc1, pDst++); in arm_cmplx_mag_q31() [all …]
|
/third_party/openssl/crypto/aes/asm/ |
D | aes-sparcv9.pl | 48 $acc1="%o0"; 213 srl $s1,13,$acc1 ! 223 and $acc1,2040,$acc1 227 ldx [$tbl+$acc1],$acc1 275 srlx $acc1,8,$acc1 280 xor $acc1,$t0,$t0 309 srl $t1,13,$acc1 314 and $acc1,2040,$acc1 318 ldx [$tbl+$acc1],$acc1 358 srlx $acc1,8,$acc1 [all …]
|