/external/boringssl/linux-arm/crypto/aes/ |
D | aesv8-armx32.S | 36 veor q0,q0,q0 49 vext.8 q9,q0,q3,#12 51 .byte 0x00,0x43,0xf0,0xf3 @ aese q10,q0 55 vext.8 q9,q0,q9,#12 57 vext.8 q9,q0,q9,#12 68 vext.8 q9,q0,q3,#12 70 .byte 0x00,0x43,0xf0,0xf3 @ aese q10,q0 73 vext.8 q9,q0,q9,#12 75 vext.8 q9,q0,q9,#12 83 vext.8 q9,q0,q3,#12 [all …]
|
D | bsaes-armv7.S | 1140 vld1.8 {q0,q1}, [r0]! @ load input 1157 veor q0, q0, q14 @ ^= IV 1166 vst1.8 {q0,q1}, [r1]! @ write output 1182 vld1.8 {q0}, [r0]! @ load input 1211 veor q0, q0, q14 @ ^= IV 1220 vst1.8 {q0,q1}, [r1]! @ write output 1234 veor q0, q0, q14 @ ^= IV 1243 vst1.8 {q0,q1}, [r1]! @ write output 1255 veor q0, q0, q14 @ ^= IV 1261 vst1.8 {q0,q1}, [r1]! @ write output [all …]
|
/external/valgrind/none/tests/arm/ |
D | neon128.c | 355 TESTINSN_imm("vmov.i32 q0", q0, 0x7); in main() 365 TESTINSN_imm("vmov.f32 q0", q0, 0.328125); in main() 366 TESTINSN_imm("vmov.f32 q0", q0, -0.328125); in main() 369 TESTINSN_imm("vmvn.i32 q0", q0, 0x7); in main() 381 TESTINSN_imm("vorr.i32 q0", q0, 0x7); in main() 389 TESTINSN_imm("vbic.i32 q0", q0, 0x7); in main() 397 TESTINSN_un("vmvn q0, q1", q0, q1, i32, 24); in main() 399 TESTINSN_un("vmvn q0, q14", q0, q14, i32, 24); in main() 402 TESTINSN_un("vmov q0, q1", q0, q1, i32, 24); in main() 404 TESTINSN_un("vmov q0, q14", q0, q14, i32, 24); in main() [all …]
|
/external/llvm/test/MC/ARM/ |
D | thumb-neon-crypto.s | 3 aesd.8 q0, q1 4 @ CHECK: aesd.8 q0, q1 @ encoding: [0xb0,0xff,0x42,0x03] 5 aese.8 q0, q1 6 @ CHECK: aese.8 q0, q1 @ encoding: [0xb0,0xff,0x02,0x03] 7 aesimc.8 q0, q1 8 @ CHECK: aesimc.8 q0, q1 @ encoding: [0xb0,0xff,0xc2,0x03] 9 aesmc.8 q0, q1 10 @ CHECK: aesmc.8 q0, q1 @ encoding: [0xb0,0xff,0x82,0x03] 12 sha1h.32 q0, q1 13 @ CHECK: sha1h.32 q0, q1 @ encoding: [0xb9,0xff,0xc2,0x02] [all …]
|
D | neon-crypto.s | 4 aesd.8 q0, q1 5 aese.8 q0, q1 6 aesimc.8 q0, q1 7 aesmc.8 q0, q1 8 @ CHECK: aesd.8 q0, q1 @ encoding: [0x42,0x03,0xb0,0xf3] 9 @ CHECK: aese.8 q0, q1 @ encoding: [0x02,0x03,0xb0,0xf3] 10 @ CHECK: aesimc.8 q0, q1 @ encoding: [0xc2,0x03,0xb0,0xf3] 11 @ CHECK: aesmc.8 q0, q1 @ encoding: [0x82,0x03,0xb0,0xf3] 17 sha1h.32 q0, q1 18 sha1su1.32 q0, q1 [all …]
|
D | fullfp16-neon.s | 5 vadd.f16 q0, q1, q2 7 @ ARM: vadd.f16 q0, q1, q2 @ encoding: [0x44,0x0d,0x12,0xf2] 9 @ THUMB: vadd.f16 q0, q1, q2 @ encoding: [0x12,0xef,0x44,0x0d] 12 vsub.f16 q0, q1, q2 14 @ ARM: vsub.f16 q0, q1, q2 @ encoding: [0x44,0x0d,0x32,0xf2] 16 @ THUMB: vsub.f16 q0, q1, q2 @ encoding: [0x32,0xef,0x44,0x0d] 19 vmul.f16 q0, q1, q2 21 @ ARM: vmul.f16 q0, q1, q2 @ encoding: [0x54,0x0d,0x12,0xf3] 23 @ THUMB: vmul.f16 q0, q1, q2 @ encoding: [0x12,0xff,0x54,0x0d] 33 vmla.f16 q0, q1, q2 [all …]
|
D | directive-arch_extension-crypto.s | 19 vmull.p64 q0, d0, d1 22 aesd.8 q0, q1 24 aese.8 q0, q1 26 aesimc.8 q0, q1 28 aesmc.8 q0, q1 31 sha1h.32 q0, q1 33 sha1su1.32 q0, q1 35 sha256su0.32 q0, q1 38 sha1c.32 q0, q1, q2 40 sha1m.32 q0, q1, q2 [all …]
|
D | fullfp16-neon-neg.s | 7 vadd.f16 q0, q1, q2 12 vsub.f16 q0, q1, q2 17 vmul.f16 q0, q1, q2 27 vmla.f16 q0, q1, q2 37 vmls.f16 q0, q1, q2 47 vfma.f16 q0, q1, q2 52 vfms.f16 q0, q1, q2 107 vacge.f16 q0, q1, q2 112 vacgt.f16 q0, q1, q2 117 vacle.f16 q0, q1, q2 [all …]
|
/external/libvpx/libvpx/vpx_dsp/arm/ |
D | intrapred_neon_asm.asm | 76 vld1.8 {q0}, [r2] 77 vst1.8 {q0}, [r0], r1 78 vst1.8 {q0}, [r0], r1 79 vst1.8 {q0}, [r0], r1 80 vst1.8 {q0}, [r0], r1 81 vst1.8 {q0}, [r0], r1 82 vst1.8 {q0}, [r0], r1 83 vst1.8 {q0}, [r0], r1 84 vst1.8 {q0}, [r0], r1 85 vst1.8 {q0}, [r0], r1 [all …]
|
D | idct16x16_1_add_neon.asm | 47 vdup.s16 q0, r0 ; duplicate a1 61 vaddw.u8 q9, q0, d2 ; dest[x] + a1 62 vaddw.u8 q10, q0, d3 ; dest[x] + a1 63 vaddw.u8 q11, q0, d4 ; dest[x] + a1 64 vaddw.u8 q12, q0, d5 ; dest[x] + a1 74 vaddw.u8 q9, q0, d6 ; dest[x] + a1 75 vaddw.u8 q10, q0, d7 ; dest[x] + a1 76 vaddw.u8 q11, q0, d16 ; dest[x] + a1 77 vaddw.u8 q12, q0, d17 ; dest[x] + a1 97 vaddw.u8 q9, q0, d2 ; dest[x] + a1 [all …]
|
/external/libavc/common/arm/ |
D | ih264_deblk_luma_a9.s | 62 @* Pointer to the src sample q0 106 vld1.8 {q0}, [r0], r1 @q0 values are loaded into q0 109 vabd.u8 q11, q3, q0 @Q11 = ABS(p0 - q0) 111 vabd.u8 q12, q1, q0 @Q12 = ABS(q1 - q0) 119 vabd.u8 q15, q2, q0 @Q15 = Aq = ABS(q2 - q0) 122 vcge.u8 q9, q11, q10 @Q9 = ( ABS(p0 - q0) >= Alpha ) 123 vcge.u8 q12, q12, q8 @Q12=( ABS(q1 - q0) >= Beta ) 128 vorr q9, q9, q12 @Q9 = ( ABS(p0 - q0) >= Alpha ) | ( ABS(q1 - q0) >= Beta ) 130 vsubl.u8 q12, d0, d6 @Q15,Q12 = (q0 - p0) 131 …vorr q9, q9, q13 @Q9 = ( ABS(p0 - q0) >= Alpha ) | ( ABS(q1 - q0) >= Beta ) | (… [all …]
|
D | ih264_intra_pred_luma_16x16_a9q.s | 113 vld1.8 {q0}, [r0] 115 vst1.8 {q0}, [r1], r3 116 vst1.8 {q0}, [r1], r3 117 vst1.8 {q0}, [r1], r3 118 vst1.8 {q0}, [r1], r3 119 vst1.8 {q0}, [r1], r3 120 vst1.8 {q0}, [r1], r3 121 vst1.8 {q0}, [r1], r3 122 vst1.8 {q0}, [r1], r3 123 vst1.8 {q0}, [r1], r3 [all …]
|
D | ih264_deblk_chroma_a9.s | 69 @* Pointer to the src sample q0 102 vaddl.u8 q5, d7, d1 @Q4,Q5 = q0 + p1 108 vabd.u8 q11, q2, q0 @Q11 = ABS(p0 - q0) 109 vabd.u8 q12, q1, q0 @Q12 = ABS(q1 - q0) 115 vcge.u8 q9, q11, q10 @Q9 = ( ABS(p0 - q0) >= Alpha ) 116 vcge.u8 q12, q12, q8 @Q12= ( ABS(q1 - q0) >= Beta ) 120 vorr q9, q9, q12 @Q9 = ( ABS(p0 - q0) >= Alpha ) | ( ABS(q1 - q0) >= Beta ) 123 …vorr q9, q9, q13 @Q9 = ( ABS(p0 - q0) >= Alpha ) | ( ABS(q1 - q0) >= Beta ) | (… 125 vbit q4, q0, q9 @ 145 @* Pointer to the src sample q0 [all …]
|
/external/boringssl/linux-arm/crypto/modes/ |
D | ghashv8-armx32.S | 30 .byte 0xa8,0x0e,0xa8,0xf2 @ pmull q0,q12,q12 35 vext.8 q9,q0,q2,#8 @ Karatsuba post-processing 36 veor q10,q0,q2 39 .byte 0x26,0x4e,0xe0,0xf2 @ pmull q10,q0,q11 @ 1st phase 43 veor q0,q1,q10 45 vext.8 q10,q0,q0,#8 @ 2nd phase 46 .byte 0x26,0x0e,0xa0,0xf2 @ pmull q0,q0,q11 48 veor q14,q0,q10 70 .byte 0x86,0x0e,0xa8,0xf2 @ pmull q0,q12,q3 @ H.lo·Xi.lo 75 vext.8 q9,q0,q2,#8 @ Karatsuba post-processing [all …]
|
/external/opencv3/modules/features2d/src/ |
D | fast_score.cpp | 129 __m128i q0 = _mm_set1_epi16(-1000), q1 = _mm_set1_epi16(1000); in cornerScore() local 155 q0 = _mm_max_epi16(q0, _mm_min_epi16(a, v0)); in cornerScore() 158 q0 = _mm_max_epi16(q0, _mm_min_epi16(a, v0)); in cornerScore() 161 q0 = _mm_max_epi16(q0, _mm_sub_epi16(_mm_setzero_si128(), q1)); in cornerScore() 162 q0 = _mm_max_epi16(q0, _mm_unpackhi_epi64(q0, q0)); in cornerScore() 163 q0 = _mm_max_epi16(q0, _mm_srli_si128(q0, 4)); in cornerScore() 164 q0 = _mm_max_epi16(q0, _mm_srli_si128(q0, 2)); in cornerScore() 165 threshold = (short)_mm_cvtsi128_si32(q0) - 1; in cornerScore() 223 __m128i q0 = _mm_set1_epi16(-1000), q1 = _mm_set1_epi16(1000); in cornerScore() local 243 q0 = _mm_max_epi16(q0, _mm_min_epi16(a, v0)); in cornerScore() [all …]
|
/external/opencv3/3rdparty/libwebp/dsp/ |
D | dec_sse2.c | 281 #define GET_NOTHEV(p1, p0, q0, q1, hev_thresh, not_hev) { \ argument 284 const __m128i t_2 = MM_ABS(q1, q0); \ 294 #define GET_BASE_DELTA(p1, p0, q0, q1, o) { \ argument 295 const __m128i qp0 = _mm_subs_epi8(q0, p0); /* q0 - p0 */ \ 302 #define DO_SIMPLE_FILTER(p0, q0, fl) { \ argument 310 q0 = _mm_subs_epi8(q0, v4); /* q0 -= v4 */ \ 328 static void NeedsFilter(const __m128i* p1, const __m128i* p0, const __m128i* q0, in NeedsFilter() argument 335 *mask = MM_ABS(*p0, *q0); // abs(p0 - q0) in NeedsFilter() 348 static WEBP_INLINE void DoFilter2(const __m128i* p1, __m128i* p0, __m128i* q0, in DoFilter2() argument 355 NeedsFilter(p1, p0, q0, q1, thresh, &mask); in DoFilter2() [all …]
|
/external/llvm/test/MC/Disassembler/ARM/ |
D | neon-crypto.txt | 4 # CHECK: aesd.8 q0, q1 6 # CHECK: aese.8 q0, q1 8 # CHECK: aesimc.8 q0, q1 10 # CHECK: aesmc.8 q0, q1 13 # CHECK: sha1h.32 q0, q1 15 # CHECK: sha1su1.32 q0, q1 17 # CHECK: sha256su0.32 q0, q1 20 # CHECK: sha1c.32 q0, q1, q2 22 # CHECK: sha1m.32 q0, q1, q2 24 # CHECK: sha1p.32 q0, q1, q2 [all …]
|
D | thumb-neon-crypto.txt | 4 # CHECK: aesd.8 q0, q1 6 # CHECK: aese.8 q0, q1 8 # CHECK: aesimc.8 q0, q1 10 # CHECK: aesmc.8 q0, q1 13 # CHECK: sha1h.32 q0, q1 15 # CHECK: sha1su1.32 q0, q1 17 # CHECK: sha256su0.32 q0, q1 20 # CHECK: sha1c.32 q0, q1, q2 22 # CHECK: sha1m.32 q0, q1, q2 24 # CHECK: sha1p.32 q0, q1, q2 [all …]
|
D | fullfp16-neon-arm.txt | 5 # CHECK: vadd.f16 q0, q1, q2 10 # CHECK: vsub.f16 q0, q1, q2 15 # CHECK: vmul.f16 q0, q1, q2 25 # CHECK: vmla.f16 q0, q1, q2 35 # CHECK: vmls.f16 q0, q1, q2 45 # CHECK: vfma.f16 q0, q1, q2 50 # CHECK: vfms.f16 q0, q1, q2 95 # CHECK: vacge.f16 q0, q1, q2 100 # CHECK: vacgt.f16 q0, q1, q2 105 # CHECK: vabd.f16 q0, q1, q2 [all …]
|
D | fullfp16-neon-thumb.txt | 5 # CHECK: vadd.f16 q0, q1, q2 10 # CHECK: vsub.f16 q0, q1, q2 15 # CHECK: vmul.f16 q0, q1, q2 25 # CHECK: vmla.f16 q0, q1, q2 35 # CHECK: vmls.f16 q0, q1, q2 45 # CHECK: vfma.f16 q0, q1, q2 50 # CHECK: vfms.f16 q0, q1, q2 95 # CHECK: vacge.f16 q0, q1, q2 100 # CHECK: vacgt.f16 q0, q1, q2 105 # CHECK: vabd.f16 q0, q1, q2 [all …]
|
/external/libvpx/libvpx/vpx_dsp/ |
D | loopfilter.c | 39 uint8_t q0, uint8_t q1, in filter_mask() argument 45 mask |= (abs(q1 - q0) > limit) * -1; in filter_mask() 48 mask |= (abs(p0 - q0) * 2 + abs(p1 - q1) / 2 > blimit) * -1; in filter_mask() 55 uint8_t q0, uint8_t q1, in flat_mask4() argument 59 mask |= (abs(q1 - q0) > thresh) * -1; in flat_mask4() 61 mask |= (abs(q2 - q0) > thresh) * -1; in flat_mask4() 63 mask |= (abs(q3 - q0) > thresh) * -1; in flat_mask4() 70 uint8_t p0, uint8_t q0, in flat_mask5() argument 73 int8_t mask = ~flat_mask4(thresh, p3, p2, p1, p0, q0, q1, q2, q3); in flat_mask5() 75 mask |= (abs(q4 - q0) > thresh) * -1; in flat_mask5() [all …]
|
/external/llvm/test/CodeGen/ARM/ |
D | thumb-big-stack.ll | 145 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 147 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 149 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 151 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 153 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 155 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 157 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 159 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 161 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 163 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… [all …]
|
/external/libvpx/libvpx/vpx_dsp/mips/ |
D | loopfilter_4_msa.c | 20 v16u8 p3, p2, p1, p0, q3, q2, q1, q0, p1_out, p0_out, q0_out, q1_out; in vpx_lpf_horizontal_4_msa() local 25 LD_UB8((src - 4 * pitch), pitch, p3, p2, p1, p0, q0, q1, q2, q3); in vpx_lpf_horizontal_4_msa() 31 LPF_MASK_HEV(p3, p2, p1, p0, q0, q1, q2, q3, limit, b_limit, thresh, in vpx_lpf_horizontal_4_msa() 33 VP9_LPF_FILTER4_8W(p1, p0, q0, q1, mask, hev, p1_out, p0_out, q0_out, q1_out); in vpx_lpf_horizontal_4_msa() 50 v16u8 p3, p2, p1, p0, q3, q2, q1, q0; in vpx_lpf_horizontal_4_dual_msa() local 53 LD_UB8((src - 4 * pitch), pitch, p3, p2, p1, p0, q0, q1, q2, q3); in vpx_lpf_horizontal_4_dual_msa() 67 LPF_MASK_HEV(p3, p2, p1, p0, q0, q1, q2, q3, limit0, b_limit0, thresh0, in vpx_lpf_horizontal_4_dual_msa() 69 VP9_LPF_FILTER4_4W(p1, p0, q0, q1, mask, hev, p1, p0, q0, q1); in vpx_lpf_horizontal_4_dual_msa() 71 ST_UB4(p1, p0, q0, q1, (src - 2 * pitch), pitch); in vpx_lpf_horizontal_4_dual_msa() 80 v16u8 p3, p2, p1, p0, q3, q2, q1, q0; in vpx_lpf_vertical_4_msa() local [all …]
|
/external/llvm/test/MC/Hexagon/ |
D | v60-vcmp.s | 14 #CHECK: 1c80e720 { q0 &= vcmp.gt(v7.ub{{ *}},{{ *}}v0.ub) } 15 q0&=vcmp.gt(v7.ub,v0.ub) 32 #CHECK: 1c85e900 { q0 &= vcmp.eq(v9.b{{ *}},{{ *}}v5.b) } 33 q0&=vcmp.eq(v9.b,v5.b) 35 #CHECK: 1c9cfca8 { q0 ^= vcmp.gt(v28.uw{{ *}},{{ *}}v28.uw) } 36 q0^=vcmp.gt(v28.uw,v28.uw) 38 #CHECK: 1c81faa0 { q0 ^= vcmp.gt(v26.ub{{ *}},{{ *}}v1.ub) } 39 q0^=vcmp.gt(v26.ub,v1.ub) 41 #CHECK: 1c96f0a4 { q0 ^= vcmp.gt(v16.uh{{ *}},{{ *}}v22.uh) } 42 q0^=vcmp.gt(v16.uh,v22.uh) [all …]
|
/external/apache-commons-math/src/main/java/org/apache/commons/math/geometry/ |
D | Rotation.java | 104 private final double q0; field in Rotation 133 public Rotation(double q0, double q1, double q2, double q3, in Rotation() argument 138 double inv = 1.0 / FastMath.sqrt(q0 * q0 + q1 * q1 + q2 * q2 + q3 * q3); in Rotation() 139 q0 *= inv; in Rotation() 145 this.q0 = q0; in Rotation() 183 q0 = FastMath.cos (halfAngle); in Rotation() 258 q0 = 0.5 * FastMath.sqrt(s + 1.0); in Rotation() 259 double inv = 0.25 / q0; in Rotation() 269 q0 = inv * (ort[1][2] - ort[2][1]); in Rotation() 278 q0 = inv * (ort[2][0] - ort[0][2]); in Rotation() [all …]
|