/external/compiler-rt/lib/builtins/arm/ |
D | subsf3vfp.S | 22 vmov s14, r0 // move first param from r0 into float register 24 vsub.f32 s14, s14, s15 25 vmov r0, s14 // move result back to r0
|
D | addsf3vfp.S | 21 vmov s14, r0 // move first param from r0 into float register 23 vadd.f32 s14, s14, s15 24 vmov r0, s14 // move result back to r0
|
D | mulsf3vfp.S | 21 vmov s14, r0 // move first param from r0 into float register 23 vmul.f32 s13, s14, s15
|
D | divsf3vfp.S | 21 vmov s14, r0 // move first param from r0 into float register 23 vdiv.f32 s13, s14, s15
|
D | lesf2vfp.S | 22 vmov s14, r0 // move from GPR 0 to float register 24 vcmp.f32 s14, s15
|
D | gtsf2vfp.S | 22 vmov s14, r0 // move from GPR 0 to float register 24 vcmp.f32 s14, s15
|
D | eqsf2vfp.S | 22 vmov s14, r0 // move from GPR 0 to float register 24 vcmp.f32 s14, s15
|
D | gesf2vfp.S | 22 vmov s14, r0 // move from GPR 0 to float register 24 vcmp.f32 s14, s15
|
D | unordsf2vfp.S | 22 vmov s14, r0 // move from GPR 0 to float register 24 vcmp.f32 s14, s15
|
D | ltsf2vfp.S | 22 vmov s14, r0 // move from GPR 0 to float register 24 vcmp.f32 s14, s15
|
D | nesf2vfp.S | 22 vmov s14, r0 // move from GPR 0 to float register 24 vcmp.f32 s14, s15
|
/external/llvm/test/MC/AArch64/ |
D | neon-scalar-shift-imm.s | 73 sqshl s14, s17, #22 86 uqshl s14, s19, #18 99 sqshlu s16, s14, #25 136 uqshrn h10, s14, #5 169 sqshrun h20, s14, #3
|
D | neon-scalar-recip.s | 34 frecpe s19, s14
|
D | neon-scalar-mul.s | 51 sqdmlsl s14, h12, h25
|
/external/libvpx/libvpx/vpx_dsp/mips/ |
D | itrans16_dspr2.c | 1056 int s0, s1, s2, s3, s4, s5, s6, s7, s8, s9, s10, s11, s12, s13, s14, s15; in iadst16_dspr2() local 1099 s14 = x14 * cospi_29_64 + x15 * cospi_3_64; in iadst16_dspr2() 1108 x6 = dct_const_round_shift(s6 + s14); in iadst16_dspr2() 1116 x14 = dct_const_round_shift(s6 - s14); in iadst16_dspr2() 1134 s14 = - x14 * cospi_12_64 + x15 * cospi_20_64; in iadst16_dspr2() 1147 x10 = dct_const_round_shift(s10 + s14); in iadst16_dspr2() 1151 x14 = dct_const_round_shift(s10 - s14); in iadst16_dspr2() 1169 s14 = - x14 * cospi_24_64 + x15 * cospi_8_64; in iadst16_dspr2() 1184 x12 = dct_const_round_shift(s12 + s14); in iadst16_dspr2() 1186 x14 = dct_const_round_shift(s12 - s14); in iadst16_dspr2() [all …]
|
/external/libvpx/libvpx/vp9/encoder/ |
D | vp9_dct.c | 348 tran_high_t s9, s10, s11, s12, s13, s14, s15; in fadst16() local 382 s14 = x14 * cospi_29_64 + x15 * cospi_3_64; in fadst16() 391 x6 = fdct_round_shift(s6 + s14); in fadst16() 399 x14 = fdct_round_shift(s6 - s14); in fadst16() 417 s14 = - x14 * cospi_12_64 + x15 * cospi_20_64; in fadst16() 430 x10 = fdct_round_shift(s10 + s14); in fadst16() 434 x14 = fdct_round_shift(s10 - s14); in fadst16() 452 s14 = - x14 * cospi_24_64 + x15 * cospi_8_64; in fadst16() 467 x12 = fdct_round_shift(s12 + s14); in fadst16() 469 x14 = fdct_round_shift(s12 - s14); in fadst16() [all …]
|
/external/llvm/test/MC/ARM/ |
D | thumb-fp-armv8.s | 79 vselvs.f32 s21, s16, s14 80 @ CHECK: vselvs.f32 s21, s16, s14 @ encoding: [0x58,0xfe,0x07,0xaa] 110 vrintxvs.f32 s10, s14 111 @ CHECK: vrintxvs.f32 s10, s14 @ encoding: [0xb7,0xee,0x47,0x5a]
|
D | fp-armv8.s | 76 vselvs.f32 s21, s16, s14 77 @ CHECK: vselvs.f32 s21, s16, s14 @ encoding: [0x07,0xaa,0x58,0xfe] 104 vrintxvs.f32 s10, s14 105 @ CHECK: vrintxvs.f32 s10, s14 @ encoding: [0x47,0x5a,0xb7,0x6e]
|
/external/boringssl/src/crypto/curve25519/ |
D | curve25519.c | 3793 int64_t s14 = 2097151 & (load_4(s + 36) >> 6); in sc_reduce() local 3824 s14 -= s23 * 997805; in sc_reduce() 3833 s14 += s22 * 136657; in sc_reduce() 3842 s14 -= s21 * 683901; in sc_reduce() 3881 carry14 = (s14 + (1 << 20)) >> 21; in sc_reduce() 3883 s14 -= carry14 << 21; in sc_reduce() 3898 s14 += carry13; in sc_reduce() 3928 s2 += s14 * 666643; in sc_reduce() 3929 s3 += s14 * 470296; in sc_reduce() 3930 s4 += s14 * 654183; in sc_reduce() [all …]
|
/external/libvpx/libvpx/vpx_dsp/ |
D | inv_txfm.c | 581 tran_high_t s9, s10, s11, s12, s13, s14, s15; in iadst16_c() local 624 s14 = x14 * cospi_29_64 + x15 * cospi_3_64; in iadst16_c() 633 x6 = WRAPLOW(dct_const_round_shift(s6 + s14), 8); in iadst16_c() 641 x14 = WRAPLOW(dct_const_round_shift(s6 - s14), 8); in iadst16_c() 659 s14 = - x14 * cospi_12_64 + x15 * cospi_20_64; in iadst16_c() 672 x10 = WRAPLOW(dct_const_round_shift(s10 + s14), 8); in iadst16_c() 676 x14 = WRAPLOW(dct_const_round_shift(s10 - s14), 8); in iadst16_c() 694 s14 = - x14 * cospi_24_64 + x15 * cospi_8_64; in iadst16_c() 709 x12 = WRAPLOW(dct_const_round_shift(s12 + s14), 8); in iadst16_c() 711 x14 = WRAPLOW(dct_const_round_shift(s12 - s14), 8); in iadst16_c() [all …]
|
/external/llvm/test/CodeGen/AArch64/ |
D | remat-float0.ll | 15 …2},~{s3},~{s4},~{s5},~{s6},~{s7},~{s8},~{s9},~{s10},~{s11},~{s12},~{s13},~{s14},~{s15},~{s16},~{s1…
|
/external/llvm/test/MC/Disassembler/ARM/ |
D | fp-armv8.txt | 96 # CHECK: vselvs.f32 s21, s16, s14 131 # CHECK: vrintxvs.f32 s10, s14
|
D | thumb-fp-armv8.txt | 100 # CHECK: vselvs.f32 s21, s16, s14 139 # CHECK: vrintxvs.f32 s10, s14
|
/external/clang/test/CodeGen/ |
D | exprs.c | 151 void f14(struct s14 *a) { in f14()
|
/external/llvm/test/CodeGen/AMDGPU/ |
D | llvm.SI.imageload.ll | 79 %s14 = add i32 %s10, %e15 80 %s15 = add i32 %s14, %e16
|