Home
last modified time | relevance | path

Searched refs:s14 (Results 1 – 25 of 112) sorted by relevance

12345

/external/libvpx/libvpx/vp9/common/x86/
Dvp9_highbd_iht16x16_add_sse4.c58 s10[2], s11[2], s12[2], s13[2], s14[2], s15[2]; in highbd_iadst16_4col_sse4_1() local
72 highbd_iadst_butterfly_sse4_1(io[1], io[14], cospi_29_64, cospi_3_64, s14, in highbd_iadst16_4col_sse4_1()
87 x6[0] = _mm_add_epi64(s6[0], s14[0]); in highbd_iadst16_4col_sse4_1()
88 x6[1] = _mm_add_epi64(s6[1], s14[1]); in highbd_iadst16_4col_sse4_1()
103 x14[0] = _mm_sub_epi64(s6[0], s14[0]); in highbd_iadst16_4col_sse4_1()
104 x14[1] = _mm_sub_epi64(s6[1], s14[1]); in highbd_iadst16_4col_sse4_1()
181 s14); in highbd_iadst16_4col_sse4_1()
187 x10[0] = _mm_add_epi64(s10[0], s14[0]); in highbd_iadst16_4col_sse4_1()
188 x10[1] = _mm_add_epi64(s10[1], s14[1]); in highbd_iadst16_4col_sse4_1()
195 x14[0] = _mm_sub_epi64(s10[0], s14[0]); in highbd_iadst16_4col_sse4_1()
[all …]
/external/compiler-rt/lib/builtins/arm/
Dsubsf3vfp.S22 vmov s14, r0 // move first param from r0 into float register
24 vsub.f32 s14, s14, s15
25 vmov r0, s14 // move result back to r0
Daddsf3vfp.S21 vmov s14, r0 // move first param from r0 into float register
23 vadd.f32 s14, s14, s15
24 vmov r0, s14 // move result back to r0
Ddivsf3vfp.S21 vmov s14, r0 // move first param from r0 into float register
23 vdiv.f32 s13, s14, s15
Dmulsf3vfp.S21 vmov s14, r0 // move first param from r0 into float register
23 vmul.f32 s13, s14, s15
Dgesf2vfp.S22 vmov s14, r0 // move from GPR 0 to float register
24 vcmp.f32 s14, s15
Dunordsf2vfp.S22 vmov s14, r0 // move from GPR 0 to float register
24 vcmp.f32 s14, s15
Dnesf2vfp.S22 vmov s14, r0 // move from GPR 0 to float register
24 vcmp.f32 s14, s15
Dlesf2vfp.S22 vmov s14, r0 // move from GPR 0 to float register
24 vcmp.f32 s14, s15
Deqsf2vfp.S22 vmov s14, r0 // move from GPR 0 to float register
24 vcmp.f32 s14, s15
Dltsf2vfp.S22 vmov s14, r0 // move from GPR 0 to float register
24 vcmp.f32 s14, s15
Dgtsf2vfp.S22 vmov s14, r0 // move from GPR 0 to float register
24 vcmp.f32 s14, s15
/external/swiftshader/third_party/llvm-7.0/llvm/test/MC/AArch64/
Dneon-scalar-shift-imm.s73 sqshl s14, s17, #22
86 uqshl s14, s19, #18
99 sqshlu s16, s14, #25
136 uqshrn h10, s14, #5
169 sqshrun h20, s14, #3
/external/capstone/suite/MC/AArch64/
Dneon-scalar-shift-imm.s.cs13 0x2e,0x76,0x36,0x5f = sqshl s14, s17, #22
17 0x6e,0x76,0x32,0x7f = uqshl s14, s19, #18
21 0xd0,0x65,0x39,0x7f = sqshlu s16, s14, #25
29 0xca,0x95,0x1b,0x7f = uqshrn h10, s14, #5
38 0xd4,0x85,0x1d,0x7f = sqshrun h20, s14, #3
/external/llvm/test/MC/AArch64/
Dneon-scalar-shift-imm.s73 sqshl s14, s17, #22
86 uqshl s14, s19, #18
99 sqshlu s16, s14, #25
136 uqshrn h10, s14, #5
169 sqshrun h20, s14, #3
/external/libvpx/libvpx/vp9/common/arm/neon/
Dvp9_iht16x16_add_neon.c40 int32x4_t s8[2], s9[2], s10[2], s11[2], s12[2], s13[2], s14[2], s15[2]; in vpx_iadst16x16_256_add_half1d() local
142 iadst_butterfly_lane_2_3_neon(x[14], x[15], c_25_7_29_3, s14, s15); in vpx_iadst16x16_256_add_half1d()
150 x[6] = add_dct_const_round_shift_low_8(s6, s14); in vpx_iadst16x16_256_add_half1d()
158 x[14] = sub_dct_const_round_shift_low_8(s6, s14); in vpx_iadst16x16_256_add_half1d()
173 iadst_butterfly_lane_3_2_neon(x[15], x[14], c_4_28_20_12, s15, s14); in vpx_iadst16x16_256_add_half1d()
185 x[10] = add_dct_const_round_shift_low_8(s10, s14); in vpx_iadst16x16_256_add_half1d()
189 x[14] = sub_dct_const_round_shift_low_8(s10, s14); in vpx_iadst16x16_256_add_half1d()
204 iadst_butterfly_lane_3_2_neon(x[15], x[14], c_16_n16_8_24, s15, s14); in vpx_iadst16x16_256_add_half1d()
218 x[12] = add_dct_const_round_shift_low_8(s12, s14); in vpx_iadst16x16_256_add_half1d()
220 x[14] = sub_dct_const_round_shift_low_8(s12, s14); in vpx_iadst16x16_256_add_half1d()
Dvp9_highbd_iht16x16_add_neon.c170 int64x2x2_t s8[2], s9[2], s10[2], s11[2], s12[2], s13[2], s14[2], s15[2]; in vpx_highbd_iadst16_neon() local
254 highbd_iadst_butterfly(x[14], x[15], vget_high_s32(c_25_7_29_3), 0, 1, s14, in vpx_highbd_iadst16_neon()
263 x[6] = highbd_add_dct_const_round_shift_low_8(s6, s14); in vpx_highbd_iadst16_neon()
271 x[14] = highbd_sub_dct_const_round_shift_low_8(s6, s14); in vpx_highbd_iadst16_neon()
289 s14); in vpx_highbd_iadst16_neon()
301 x[10] = highbd_add_dct_const_round_shift_low_8(s10, s14); in vpx_highbd_iadst16_neon()
305 x[14] = highbd_sub_dct_const_round_shift_low_8(s10, s14); in vpx_highbd_iadst16_neon()
324 s14); in vpx_highbd_iadst16_neon()
338 x[12] = highbd_add_dct_const_round_shift_low_8(s12, s14); in vpx_highbd_iadst16_neon()
340 x[14] = highbd_sub_dct_const_round_shift_low_8(s12, s14); in vpx_highbd_iadst16_neon()
/external/libvpx/libvpx/vpx_dsp/mips/
Ditrans16_dspr2.c1061 int s0, s1, s2, s3, s4, s5, s6, s7, s8, s9, s10, s11, s12, s13, s14, s15; in iadst16_dspr2() local
1103 s14 = x14 * cospi_29_64 + x15 * cospi_3_64; in iadst16_dspr2()
1112 x6 = dct_const_round_shift(s6 + s14); in iadst16_dspr2()
1120 x14 = dct_const_round_shift(s6 - s14); in iadst16_dspr2()
1138 s14 = -x14 * cospi_12_64 + x15 * cospi_20_64; in iadst16_dspr2()
1151 x10 = dct_const_round_shift(s10 + s14); in iadst16_dspr2()
1155 x14 = dct_const_round_shift(s10 - s14); in iadst16_dspr2()
1173 s14 = -x14 * cospi_24_64 + x15 * cospi_8_64; in iadst16_dspr2()
1188 x12 = dct_const_round_shift(s12 + s14); in iadst16_dspr2()
1190 x14 = dct_const_round_shift(s12 - s14); in iadst16_dspr2()
[all …]
/external/boringssl/src/ssl/test/runner/ed25519/internal/edwards25519/
Dedwards25519.go1079 s14 := a3*b11 + a4*b10 + a5*b9 + a6*b8 + a7*b7 + a8*b6 + a9*b5 + a10*b4 + a11*b3
1111 carry[14] = (s14 + (1 << 20)) >> 21
1113 s14 -= carry[14] << 21
1146 s14 += carry[13]
1164 s14 -= s23 * 997805
1173 s14 += s22 * 136657
1182 s14 -= s21 * 683901
1221 carry[14] = (s14 + (1 << 20)) >> 21
1223 s14 -= carry[14] << 21
1238 s14 += carry[13]
[all …]
/external/boringssl/src/third_party/fiat/
Dcurve25519.c1080 int64_t s14 = 2097151 & (load_4(s + 36) >> 6); in x25519_sc_reduce() local
1111 s14 -= s23 * 997805; in x25519_sc_reduce()
1120 s14 += s22 * 136657; in x25519_sc_reduce()
1129 s14 -= s21 * 683901; in x25519_sc_reduce()
1168 carry14 = (s14 + (1 << 20)) >> 21; in x25519_sc_reduce()
1170 s14 -= int64_lshift21(carry14); in x25519_sc_reduce()
1185 s14 += carry13; in x25519_sc_reduce()
1215 s2 += s14 * 666643; in x25519_sc_reduce()
1216 s3 += s14 * 470296; in x25519_sc_reduce()
1217 s4 += s14 * 654183; in x25519_sc_reduce()
[all …]
/external/libvpx/libvpx/vp9/encoder/
Dvp9_dct.c348 tran_high_t s9, s10, s11, s12, s13, s14, s15; in fadst16() local
382 s14 = x14 * cospi_29_64 + x15 * cospi_3_64; in fadst16()
391 x6 = fdct_round_shift(s6 + s14); in fadst16()
399 x14 = fdct_round_shift(s6 - s14); in fadst16()
417 s14 = -x14 * cospi_12_64 + x15 * cospi_20_64; in fadst16()
430 x10 = fdct_round_shift(s10 + s14); in fadst16()
434 x14 = fdct_round_shift(s10 - s14); in fadst16()
452 s14 = -x14 * cospi_24_64 + x15 * cospi_8_64; in fadst16()
467 x12 = fdct_round_shift(s12 + s14); in fadst16()
469 x14 = fdct_round_shift(s12 - s14); in fadst16()
[all …]
/external/swiftshader/third_party/llvm-7.0/llvm/test/MC/ARM/
Dthumb-fp-armv8.s79 vselvs.f32 s21, s16, s14
80 @ CHECK: vselvs.f32 s21, s16, s14 @ encoding: [0x58,0xfe,0x07,0xaa]
110 vrintxvs.f32 s10, s14
111 @ CHECK: vrintxvs.f32 s10, s14 @ encoding: [0xb7,0xee,0x47,0x5a]
Dfp-armv8.s76 vselvs.f32 s21, s16, s14
77 @ CHECK: vselvs.f32 s21, s16, s14 @ encoding: [0x07,0xaa,0x58,0xfe]
104 vrintxvs.f32 s10, s14
105 @ CHECK: vrintxvs.f32 s10, s14 @ encoding: [0x47,0x5a,0xb7,0x6e]
/external/llvm/test/MC/ARM/
Dfp-armv8.s76 vselvs.f32 s21, s16, s14
77 @ CHECK: vselvs.f32 s21, s16, s14 @ encoding: [0x07,0xaa,0x58,0xfe]
104 vrintxvs.f32 s10, s14
105 @ CHECK: vrintxvs.f32 s10, s14 @ encoding: [0x47,0x5a,0xb7,0x6e]
Dthumb-fp-armv8.s79 vselvs.f32 s21, s16, s14
80 @ CHECK: vselvs.f32 s21, s16, s14 @ encoding: [0x58,0xfe,0x07,0xaa]
110 vrintxvs.f32 s10, s14
111 @ CHECK: vrintxvs.f32 s10, s14 @ encoding: [0xb7,0xee,0x47,0x5a]

12345