/external/libvpx/libvpx/vp9/common/x86/ |
D | vp9_highbd_iht16x16_add_sse4.c | 58 s10[2], s11[2], s12[2], s13[2], s14[2], s15[2]; in highbd_iadst16_4col_sse4_1() local 72 highbd_iadst_butterfly_sse4_1(io[1], io[14], cospi_29_64, cospi_3_64, s14, in highbd_iadst16_4col_sse4_1() 87 x6[0] = _mm_add_epi64(s6[0], s14[0]); in highbd_iadst16_4col_sse4_1() 88 x6[1] = _mm_add_epi64(s6[1], s14[1]); in highbd_iadst16_4col_sse4_1() 103 x14[0] = _mm_sub_epi64(s6[0], s14[0]); in highbd_iadst16_4col_sse4_1() 104 x14[1] = _mm_sub_epi64(s6[1], s14[1]); in highbd_iadst16_4col_sse4_1() 181 s14); in highbd_iadst16_4col_sse4_1() 187 x10[0] = _mm_add_epi64(s10[0], s14[0]); in highbd_iadst16_4col_sse4_1() 188 x10[1] = _mm_add_epi64(s10[1], s14[1]); in highbd_iadst16_4col_sse4_1() 195 x14[0] = _mm_sub_epi64(s10[0], s14[0]); in highbd_iadst16_4col_sse4_1() [all …]
|
/external/compiler-rt/lib/builtins/arm/ |
D | subsf3vfp.S | 22 vmov s14, r0 // move first param from r0 into float register 24 vsub.f32 s14, s14, s15 25 vmov r0, s14 // move result back to r0
|
D | addsf3vfp.S | 21 vmov s14, r0 // move first param from r0 into float register 23 vadd.f32 s14, s14, s15 24 vmov r0, s14 // move result back to r0
|
D | divsf3vfp.S | 21 vmov s14, r0 // move first param from r0 into float register 23 vdiv.f32 s13, s14, s15
|
D | mulsf3vfp.S | 21 vmov s14, r0 // move first param from r0 into float register 23 vmul.f32 s13, s14, s15
|
D | gesf2vfp.S | 22 vmov s14, r0 // move from GPR 0 to float register 24 vcmp.f32 s14, s15
|
D | unordsf2vfp.S | 22 vmov s14, r0 // move from GPR 0 to float register 24 vcmp.f32 s14, s15
|
D | nesf2vfp.S | 22 vmov s14, r0 // move from GPR 0 to float register 24 vcmp.f32 s14, s15
|
D | lesf2vfp.S | 22 vmov s14, r0 // move from GPR 0 to float register 24 vcmp.f32 s14, s15
|
D | eqsf2vfp.S | 22 vmov s14, r0 // move from GPR 0 to float register 24 vcmp.f32 s14, s15
|
D | ltsf2vfp.S | 22 vmov s14, r0 // move from GPR 0 to float register 24 vcmp.f32 s14, s15
|
D | gtsf2vfp.S | 22 vmov s14, r0 // move from GPR 0 to float register 24 vcmp.f32 s14, s15
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/MC/AArch64/ |
D | neon-scalar-shift-imm.s | 73 sqshl s14, s17, #22 86 uqshl s14, s19, #18 99 sqshlu s16, s14, #25 136 uqshrn h10, s14, #5 169 sqshrun h20, s14, #3
|
/external/capstone/suite/MC/AArch64/ |
D | neon-scalar-shift-imm.s.cs | 13 0x2e,0x76,0x36,0x5f = sqshl s14, s17, #22 17 0x6e,0x76,0x32,0x7f = uqshl s14, s19, #18 21 0xd0,0x65,0x39,0x7f = sqshlu s16, s14, #25 29 0xca,0x95,0x1b,0x7f = uqshrn h10, s14, #5 38 0xd4,0x85,0x1d,0x7f = sqshrun h20, s14, #3
|
/external/llvm/test/MC/AArch64/ |
D | neon-scalar-shift-imm.s | 73 sqshl s14, s17, #22 86 uqshl s14, s19, #18 99 sqshlu s16, s14, #25 136 uqshrn h10, s14, #5 169 sqshrun h20, s14, #3
|
/external/libvpx/libvpx/vp9/common/arm/neon/ |
D | vp9_iht16x16_add_neon.c | 40 int32x4_t s8[2], s9[2], s10[2], s11[2], s12[2], s13[2], s14[2], s15[2]; in vpx_iadst16x16_256_add_half1d() local 142 iadst_butterfly_lane_2_3_neon(x[14], x[15], c_25_7_29_3, s14, s15); in vpx_iadst16x16_256_add_half1d() 150 x[6] = add_dct_const_round_shift_low_8(s6, s14); in vpx_iadst16x16_256_add_half1d() 158 x[14] = sub_dct_const_round_shift_low_8(s6, s14); in vpx_iadst16x16_256_add_half1d() 173 iadst_butterfly_lane_3_2_neon(x[15], x[14], c_4_28_20_12, s15, s14); in vpx_iadst16x16_256_add_half1d() 185 x[10] = add_dct_const_round_shift_low_8(s10, s14); in vpx_iadst16x16_256_add_half1d() 189 x[14] = sub_dct_const_round_shift_low_8(s10, s14); in vpx_iadst16x16_256_add_half1d() 204 iadst_butterfly_lane_3_2_neon(x[15], x[14], c_16_n16_8_24, s15, s14); in vpx_iadst16x16_256_add_half1d() 218 x[12] = add_dct_const_round_shift_low_8(s12, s14); in vpx_iadst16x16_256_add_half1d() 220 x[14] = sub_dct_const_round_shift_low_8(s12, s14); in vpx_iadst16x16_256_add_half1d()
|
D | vp9_highbd_iht16x16_add_neon.c | 170 int64x2x2_t s8[2], s9[2], s10[2], s11[2], s12[2], s13[2], s14[2], s15[2]; in vpx_highbd_iadst16_neon() local 254 highbd_iadst_butterfly(x[14], x[15], vget_high_s32(c_25_7_29_3), 0, 1, s14, in vpx_highbd_iadst16_neon() 263 x[6] = highbd_add_dct_const_round_shift_low_8(s6, s14); in vpx_highbd_iadst16_neon() 271 x[14] = highbd_sub_dct_const_round_shift_low_8(s6, s14); in vpx_highbd_iadst16_neon() 289 s14); in vpx_highbd_iadst16_neon() 301 x[10] = highbd_add_dct_const_round_shift_low_8(s10, s14); in vpx_highbd_iadst16_neon() 305 x[14] = highbd_sub_dct_const_round_shift_low_8(s10, s14); in vpx_highbd_iadst16_neon() 324 s14); in vpx_highbd_iadst16_neon() 338 x[12] = highbd_add_dct_const_round_shift_low_8(s12, s14); in vpx_highbd_iadst16_neon() 340 x[14] = highbd_sub_dct_const_round_shift_low_8(s12, s14); in vpx_highbd_iadst16_neon()
|
/external/libvpx/libvpx/vpx_dsp/mips/ |
D | itrans16_dspr2.c | 1061 int s0, s1, s2, s3, s4, s5, s6, s7, s8, s9, s10, s11, s12, s13, s14, s15; in iadst16_dspr2() local 1103 s14 = x14 * cospi_29_64 + x15 * cospi_3_64; in iadst16_dspr2() 1112 x6 = dct_const_round_shift(s6 + s14); in iadst16_dspr2() 1120 x14 = dct_const_round_shift(s6 - s14); in iadst16_dspr2() 1138 s14 = -x14 * cospi_12_64 + x15 * cospi_20_64; in iadst16_dspr2() 1151 x10 = dct_const_round_shift(s10 + s14); in iadst16_dspr2() 1155 x14 = dct_const_round_shift(s10 - s14); in iadst16_dspr2() 1173 s14 = -x14 * cospi_24_64 + x15 * cospi_8_64; in iadst16_dspr2() 1188 x12 = dct_const_round_shift(s12 + s14); in iadst16_dspr2() 1190 x14 = dct_const_round_shift(s12 - s14); in iadst16_dspr2() [all …]
|
/external/boringssl/src/ssl/test/runner/ed25519/internal/edwards25519/ |
D | edwards25519.go | 1079 s14 := a3*b11 + a4*b10 + a5*b9 + a6*b8 + a7*b7 + a8*b6 + a9*b5 + a10*b4 + a11*b3 1111 carry[14] = (s14 + (1 << 20)) >> 21 1113 s14 -= carry[14] << 21 1146 s14 += carry[13] 1164 s14 -= s23 * 997805 1173 s14 += s22 * 136657 1182 s14 -= s21 * 683901 1221 carry[14] = (s14 + (1 << 20)) >> 21 1223 s14 -= carry[14] << 21 1238 s14 += carry[13] [all …]
|
/external/boringssl/src/third_party/fiat/ |
D | curve25519.c | 1080 int64_t s14 = 2097151 & (load_4(s + 36) >> 6); in x25519_sc_reduce() local 1111 s14 -= s23 * 997805; in x25519_sc_reduce() 1120 s14 += s22 * 136657; in x25519_sc_reduce() 1129 s14 -= s21 * 683901; in x25519_sc_reduce() 1168 carry14 = (s14 + (1 << 20)) >> 21; in x25519_sc_reduce() 1170 s14 -= int64_lshift21(carry14); in x25519_sc_reduce() 1185 s14 += carry13; in x25519_sc_reduce() 1215 s2 += s14 * 666643; in x25519_sc_reduce() 1216 s3 += s14 * 470296; in x25519_sc_reduce() 1217 s4 += s14 * 654183; in x25519_sc_reduce() [all …]
|
/external/libvpx/libvpx/vp9/encoder/ |
D | vp9_dct.c | 348 tran_high_t s9, s10, s11, s12, s13, s14, s15; in fadst16() local 382 s14 = x14 * cospi_29_64 + x15 * cospi_3_64; in fadst16() 391 x6 = fdct_round_shift(s6 + s14); in fadst16() 399 x14 = fdct_round_shift(s6 - s14); in fadst16() 417 s14 = -x14 * cospi_12_64 + x15 * cospi_20_64; in fadst16() 430 x10 = fdct_round_shift(s10 + s14); in fadst16() 434 x14 = fdct_round_shift(s10 - s14); in fadst16() 452 s14 = -x14 * cospi_24_64 + x15 * cospi_8_64; in fadst16() 467 x12 = fdct_round_shift(s12 + s14); in fadst16() 469 x14 = fdct_round_shift(s12 - s14); in fadst16() [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/MC/ARM/ |
D | thumb-fp-armv8.s | 79 vselvs.f32 s21, s16, s14 80 @ CHECK: vselvs.f32 s21, s16, s14 @ encoding: [0x58,0xfe,0x07,0xaa] 110 vrintxvs.f32 s10, s14 111 @ CHECK: vrintxvs.f32 s10, s14 @ encoding: [0xb7,0xee,0x47,0x5a]
|
D | fp-armv8.s | 76 vselvs.f32 s21, s16, s14 77 @ CHECK: vselvs.f32 s21, s16, s14 @ encoding: [0x07,0xaa,0x58,0xfe] 104 vrintxvs.f32 s10, s14 105 @ CHECK: vrintxvs.f32 s10, s14 @ encoding: [0x47,0x5a,0xb7,0x6e]
|
/external/llvm/test/MC/ARM/ |
D | fp-armv8.s | 76 vselvs.f32 s21, s16, s14 77 @ CHECK: vselvs.f32 s21, s16, s14 @ encoding: [0x07,0xaa,0x58,0xfe] 104 vrintxvs.f32 s10, s14 105 @ CHECK: vrintxvs.f32 s10, s14 @ encoding: [0x47,0x5a,0xb7,0x6e]
|
D | thumb-fp-armv8.s | 79 vselvs.f32 s21, s16, s14 80 @ CHECK: vselvs.f32 s21, s16, s14 @ encoding: [0x58,0xfe,0x07,0xaa] 110 vrintxvs.f32 s10, s14 111 @ CHECK: vrintxvs.f32 s10, s14 @ encoding: [0xb7,0xee,0x47,0x5a]
|