Home
last modified time | relevance | path

Searched refs:s14 (Results 1 – 25 of 51) sorted by relevance

123

/external/compiler-rt/lib/builtins/arm/
Dsubsf3vfp.S22 vmov s14, r0 // move first param from r0 into float register
24 vsub.f32 s14, s14, s15
25 vmov r0, s14 // move result back to r0
Daddsf3vfp.S21 vmov s14, r0 // move first param from r0 into float register
23 vadd.f32 s14, s14, s15
24 vmov r0, s14 // move result back to r0
Dmulsf3vfp.S21 vmov s14, r0 // move first param from r0 into float register
23 vmul.f32 s13, s14, s15
Ddivsf3vfp.S21 vmov s14, r0 // move first param from r0 into float register
23 vdiv.f32 s13, s14, s15
Dlesf2vfp.S22 vmov s14, r0 // move from GPR 0 to float register
24 vcmp.f32 s14, s15
Dgtsf2vfp.S22 vmov s14, r0 // move from GPR 0 to float register
24 vcmp.f32 s14, s15
Deqsf2vfp.S22 vmov s14, r0 // move from GPR 0 to float register
24 vcmp.f32 s14, s15
Dgesf2vfp.S22 vmov s14, r0 // move from GPR 0 to float register
24 vcmp.f32 s14, s15
Dunordsf2vfp.S22 vmov s14, r0 // move from GPR 0 to float register
24 vcmp.f32 s14, s15
Dltsf2vfp.S22 vmov s14, r0 // move from GPR 0 to float register
24 vcmp.f32 s14, s15
Dnesf2vfp.S22 vmov s14, r0 // move from GPR 0 to float register
24 vcmp.f32 s14, s15
/external/llvm/test/MC/AArch64/
Dneon-scalar-shift-imm.s73 sqshl s14, s17, #22
86 uqshl s14, s19, #18
99 sqshlu s16, s14, #25
136 uqshrn h10, s14, #5
169 sqshrun h20, s14, #3
Dneon-scalar-recip.s34 frecpe s19, s14
Dneon-scalar-mul.s51 sqdmlsl s14, h12, h25
/external/libvpx/libvpx/vpx_dsp/mips/
Ditrans16_dspr2.c1056 int s0, s1, s2, s3, s4, s5, s6, s7, s8, s9, s10, s11, s12, s13, s14, s15; in iadst16_dspr2() local
1099 s14 = x14 * cospi_29_64 + x15 * cospi_3_64; in iadst16_dspr2()
1108 x6 = dct_const_round_shift(s6 + s14); in iadst16_dspr2()
1116 x14 = dct_const_round_shift(s6 - s14); in iadst16_dspr2()
1134 s14 = - x14 * cospi_12_64 + x15 * cospi_20_64; in iadst16_dspr2()
1147 x10 = dct_const_round_shift(s10 + s14); in iadst16_dspr2()
1151 x14 = dct_const_round_shift(s10 - s14); in iadst16_dspr2()
1169 s14 = - x14 * cospi_24_64 + x15 * cospi_8_64; in iadst16_dspr2()
1184 x12 = dct_const_round_shift(s12 + s14); in iadst16_dspr2()
1186 x14 = dct_const_round_shift(s12 - s14); in iadst16_dspr2()
[all …]
/external/libvpx/libvpx/vp9/encoder/
Dvp9_dct.c348 tran_high_t s9, s10, s11, s12, s13, s14, s15; in fadst16() local
382 s14 = x14 * cospi_29_64 + x15 * cospi_3_64; in fadst16()
391 x6 = fdct_round_shift(s6 + s14); in fadst16()
399 x14 = fdct_round_shift(s6 - s14); in fadst16()
417 s14 = - x14 * cospi_12_64 + x15 * cospi_20_64; in fadst16()
430 x10 = fdct_round_shift(s10 + s14); in fadst16()
434 x14 = fdct_round_shift(s10 - s14); in fadst16()
452 s14 = - x14 * cospi_24_64 + x15 * cospi_8_64; in fadst16()
467 x12 = fdct_round_shift(s12 + s14); in fadst16()
469 x14 = fdct_round_shift(s12 - s14); in fadst16()
[all …]
/external/llvm/test/MC/ARM/
Dthumb-fp-armv8.s79 vselvs.f32 s21, s16, s14
80 @ CHECK: vselvs.f32 s21, s16, s14 @ encoding: [0x58,0xfe,0x07,0xaa]
110 vrintxvs.f32 s10, s14
111 @ CHECK: vrintxvs.f32 s10, s14 @ encoding: [0xb7,0xee,0x47,0x5a]
Dfp-armv8.s76 vselvs.f32 s21, s16, s14
77 @ CHECK: vselvs.f32 s21, s16, s14 @ encoding: [0x07,0xaa,0x58,0xfe]
104 vrintxvs.f32 s10, s14
105 @ CHECK: vrintxvs.f32 s10, s14 @ encoding: [0x47,0x5a,0xb7,0x6e]
/external/boringssl/src/crypto/curve25519/
Dcurve25519.c3793 int64_t s14 = 2097151 & (load_4(s + 36) >> 6); in sc_reduce() local
3824 s14 -= s23 * 997805; in sc_reduce()
3833 s14 += s22 * 136657; in sc_reduce()
3842 s14 -= s21 * 683901; in sc_reduce()
3881 carry14 = (s14 + (1 << 20)) >> 21; in sc_reduce()
3883 s14 -= carry14 << 21; in sc_reduce()
3898 s14 += carry13; in sc_reduce()
3928 s2 += s14 * 666643; in sc_reduce()
3929 s3 += s14 * 470296; in sc_reduce()
3930 s4 += s14 * 654183; in sc_reduce()
[all …]
/external/libvpx/libvpx/vpx_dsp/
Dinv_txfm.c581 tran_high_t s9, s10, s11, s12, s13, s14, s15; in iadst16_c() local
624 s14 = x14 * cospi_29_64 + x15 * cospi_3_64; in iadst16_c()
633 x6 = WRAPLOW(dct_const_round_shift(s6 + s14), 8); in iadst16_c()
641 x14 = WRAPLOW(dct_const_round_shift(s6 - s14), 8); in iadst16_c()
659 s14 = - x14 * cospi_12_64 + x15 * cospi_20_64; in iadst16_c()
672 x10 = WRAPLOW(dct_const_round_shift(s10 + s14), 8); in iadst16_c()
676 x14 = WRAPLOW(dct_const_round_shift(s10 - s14), 8); in iadst16_c()
694 s14 = - x14 * cospi_24_64 + x15 * cospi_8_64; in iadst16_c()
709 x12 = WRAPLOW(dct_const_round_shift(s12 + s14), 8); in iadst16_c()
711 x14 = WRAPLOW(dct_const_round_shift(s12 - s14), 8); in iadst16_c()
[all …]
/external/llvm/test/CodeGen/AArch64/
Dremat-float0.ll15 …2},~{s3},~{s4},~{s5},~{s6},~{s7},~{s8},~{s9},~{s10},~{s11},~{s12},~{s13},~{s14},~{s15},~{s16},~{s1…
/external/llvm/test/MC/Disassembler/ARM/
Dfp-armv8.txt96 # CHECK: vselvs.f32 s21, s16, s14
131 # CHECK: vrintxvs.f32 s10, s14
Dthumb-fp-armv8.txt100 # CHECK: vselvs.f32 s21, s16, s14
139 # CHECK: vrintxvs.f32 s10, s14
/external/clang/test/CodeGen/
Dexprs.c151 void f14(struct s14 *a) { in f14()
/external/llvm/test/CodeGen/AMDGPU/
Dllvm.SI.imageload.ll79 %s14 = add i32 %s10, %e15
80 %s15 = add i32 %s14, %e16

123