/external/capstone/suite/MC/ARM/ |
D | neon-bitwise-encoding.s.cs | 23 0x56,0x81,0x0e,0xf3 = veor q4, q7, q3 24 0x56,0x81,0x0e,0xf3 = veor q4, q7, q3 25 0x56,0x81,0x0e,0xf3 = veor q4, q7, q3 26 0x56,0x81,0x0e,0xf3 = veor q4, q7, q3 27 0x56,0x81,0x0e,0xf3 = veor q4, q7, q3 28 0x56,0x81,0x0e,0xf3 = veor q4, q7, q3 29 0x56,0x81,0x0e,0xf3 = veor q4, q7, q3 30 0x56,0x81,0x0e,0xf3 = veor q4, q7, q3 31 0x56,0x81,0x0e,0xf3 = veor q4, q7, q3 32 0x56,0x81,0x0e,0xf3 = veor q4, q7, q3 [all …]
|
D | neon-shiftaccum-encoding.s.cs | 6 0x54,0xe1,0x88,0xf2 = vsra.s8 q7, q2, #8 14 0x5e,0x21,0x88,0xf3 = vsra.u8 q1, q7, #8 15 0x5e,0x41,0x9a,0xf3 = vsra.u16 q2, q7, #6 30 0x5e,0xe1,0x88,0xf3 = vsra.u8 q7, q7, #8 31 0x5e,0xe1,0x9a,0xf3 = vsra.u16 q7, q7, #6 47 0x5e,0xc3,0x90,0xf3 = vrsra.u16 q6, q7, #16 48 0x70,0xe3,0xa0,0xf3 = vrsra.u32 q7, q8, #32 63 0x5e,0xe3,0x90,0xf3 = vrsra.u16 q7, q7, #16 71 0x5e,0x45,0x9f,0xf3 = vsli.16 q2, q7, #15 80 0x58,0xe4,0xa0,0xf3 = vsri.32 q7, q4, #32 [all …]
|
D | neont2-shiftaccum-encoding.s.cs | 6 0x88,0xef,0x54,0xe1 = vsra.s8 q7, q2, #8 14 0x88,0xff,0x5e,0x21 = vsra.u8 q1, q7, #8 15 0x9a,0xff,0x5e,0x41 = vsra.u16 q2, q7, #6 30 0x88,0xff,0x5e,0xe1 = vsra.u8 q7, q7, #8 31 0x9a,0xff,0x5e,0xe1 = vsra.u16 q7, q7, #6 47 0x90,0xff,0x5e,0xc3 = vrsra.u16 q6, q7, #16 48 0xa0,0xff,0x70,0xe3 = vrsra.u32 q7, q8, #32 63 0x90,0xff,0x5e,0xe3 = vrsra.u16 q7, q7, #16 71 0x9f,0xff,0x5e,0x45 = vsli.16 q2, q7, #15 80 0xa0,0xff,0x58,0xe4 = vsri.32 q7, q4, #32 [all …]
|
D | neon-shift-encoding.s.cs | 55 0x5e,0x41,0x91,0xf2 = vsra.s16 q2, q7, #15 65 0xde,0xe1,0x81,0xf2 = vsra.s64 q7, q7, #63 71 0x5e,0x41,0x91,0xf3 = vsra.u16 q2, q7, #15 81 0xde,0xe1,0x81,0xf3 = vsra.u64 q7, q7, #63 87 0x5e,0x44,0x91,0xf3 = vsri.16 q2, q7, #15 97 0xde,0xe4,0x81,0xf3 = vsri.64 q7, q7, #63 103 0x5e,0x45,0x9f,0xf3 = vsli.16 q2, q7, #15 113 0xde,0xe5,0xbf,0xf3 = vsli.64 q7, q7, #63 205 0xce,0xe5,0x24,0xf3 = vrshl.u32 q7, q7, q10 221 0x5e,0xe0,0xa0,0xf3 = vshr.u32 q7, q7, #32 [all …]
|
/external/llvm/test/MC/ARM/ |
D | neon-bitwise-encoding.s | 110 veor q4, q7, q3 111 veor.8 q4, q7, q3 112 veor.16 q4, q7, q3 113 veor.32 q4, q7, q3 114 veor.64 q4, q7, q3 116 veor.i8 q4, q7, q3 117 veor.i16 q4, q7, q3 118 veor.i32 q4, q7, q3 119 veor.i64 q4, q7, q3 121 veor.s8 q4, q7, q3 [all …]
|
D | neont2-shiftaccum-encoding.s | 9 vsra.s8 q7, q2, #8 17 vsra.u8 q1, q7, #8 18 vsra.u16 q2, q7, #6 35 vsra.u8 q7, #8 36 vsra.u16 q7, #6 44 @ CHECK: vsra.s8 q7, q2, #8 @ encoding: [0x88,0xef,0x54,0xe1] 52 @ CHECK: vsra.u8 q1, q7, #8 @ encoding: [0x88,0xff,0x5e,0x21] 53 @ CHECK: vsra.u16 q2, q7, #6 @ encoding: [0x9a,0xff,0x5e,0x41] 69 @ CHECK: vsra.u8 q7, q7, #8 @ encoding: [0x88,0xff,0x5e,0xe1] 70 @ CHECK: vsra.u16 q7, q7, #6 @ encoding: [0x9a,0xff,0x5e,0xe1] [all …]
|
D | neon-shiftaccum-encoding.s | 7 vsra.s8 q7, q2, #8 15 vsra.u8 q1, q7, #8 16 vsra.u16 q2, q7, #6 33 vsra.u8 q7, #8 34 vsra.u16 q7, #6 42 @ CHECK: vsra.s8 q7, q2, #8 @ encoding: [0x54,0xe1,0x88,0xf2] 50 @ CHECK: vsra.u8 q1, q7, #8 @ encoding: [0x5e,0x21,0x88,0xf3] 51 @ CHECK: vsra.u16 q2, q7, #6 @ encoding: [0x5e,0x41,0x9a,0xf3] 67 @ CHECK: vsra.u8 q7, q7, #8 @ encoding: [0x5e,0xe1,0x88,0xf3] 68 @ CHECK: vsra.u16 q7, q7, #6 @ encoding: [0x5e,0xe1,0x9a,0xf3] [all …]
|
D | neont2-minmax-encoding.s | 23 vmax.s32 q7, q8, q9 26 vmax.u32 q6, q7, q8 34 vmax.u32 q7, q8 53 @ CHECK: vmax.s32 q7, q8, q9 @ encoding: [0x20,0xef,0xe2,0xe6] 56 @ CHECK: vmax.u32 q6, q7, q8 @ encoding: [0x2e,0xff,0x60,0xc6] 63 @ CHECK: vmax.u32 q7, q7, q8 @ encoding: [0x2e,0xff,0x60,0xe6] 85 vmin.s32 q7, q8, q9 88 vmin.u32 q6, q7, q8 96 vmin.u32 q7, q8 115 @ CHECK: vmin.s32 q7, q8, q9 @ encoding: [0x20,0xef,0xf2,0xe6] [all …]
|
D | neon-minmax-encoding.s | 21 vmax.s32 q7, q8, q9 24 vmax.u32 q6, q7, q8 32 vmax.u32 q7, q8 51 @ CHECK: vmax.s32 q7, q8, q9 @ encoding: [0xe2,0xe6,0x20,0xf2] 54 @ CHECK: vmax.u32 q6, q7, q8 @ encoding: [0x60,0xc6,0x2e,0xf3] 61 @ CHECK: vmax.u32 q7, q7, q8 @ encoding: [0x60,0xe6,0x2e,0xf3] 83 vmin.s32 q7, q8, q9 86 vmin.u32 q6, q7, q8 94 vmin.u32 q7, q8 113 @ CHECK: vmin.s32 q7, q8, q9 @ encoding: [0xf2,0xe6,0x20,0xf2] [all …]
|
/external/llvm/test/CodeGen/ARM/ |
D | thumb-big-stack.ll | 145 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 147 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 149 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 151 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 153 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 155 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 157 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 159 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 161 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… 163 …tail call void asm sideeffect "", "~{q0}{q1}{q2}{q3}{q4}{q5}{q6}{q7}{q8}{q9}{q10}{q11}{q12}{q13}{q… [all …]
|
/external/libavc/common/arm/ |
D | ih264_inter_pred_luma_horz_qpel_vert_hpel_a9q.s | 329 vaddl.u8 q7, d4, d6 332 vmla.u16 q6, q7, q13 334 vaddl.u8 q7, d1, d11 336 vmla.u16 q7, q9, q13 340 vmls.u16 q7, q11, q12 343 vext.16 q11, q6, q7, #5 347 vst1.32 {q7}, [r9], r7 @ store row 0 to temp buffer: col 1 349 vext.16 q8, q6, q7, #2 351 vext.16 q9, q6, q7, #3 352 vext.16 q10, q6, q7, #4 [all …]
|
D | ih264_inter_pred_filters_luma_horz_a9q.s | 131 vaddl.u8 q7, d28, d5 @// a0 + a5 (column1,row1) 139 vmlal.u8 q7, d28, d1 @// a0 + a5 + 20a2 (column1,row1) 147 vmlal.u8 q7, d28, d1 @// a0 + a5 + 20a2 + 20a3 (column1,row1) 155 vmlsl.u8 q7, d28, d0 @// a0 + a5 + 20a2 + 20a3 - 5a1 (column1,row1) 163 vmlsl.u8 q7, d28, d0 @// a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 (column1,row1) 169 …vqrshrun.s16 d23, q7, #5 @// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column1,r… 188 vaddl.u8 q7, d28, d5 @// a0 + a5 (column1,row1) 190 vmlal.u8 q7, d25, d1 @// a0 + a5 + 20a2 (column1,row1) 191 vmlal.u8 q7, d24, d1 @// a0 + a5 + 20a2 + 20a3 (column1,row1) 192 vmlsl.u8 q7, d23, d0 @// a0 + a5 + 20a2 + 20a3 - 5a1 (column1,row1) [all …]
|
D | ih264_inter_pred_luma_horz_qpel_a9q.s | 138 vaddl.u8 q7, d28, d5 @// a0 + a5 (column1,row1) 146 vmlal.u8 q7, d28, d1 @// a0 + a5 + 20a2 (column1,row1) 154 vmlal.u8 q7, d28, d1 @// a0 + a5 + 20a2 + 20a3 (column1,row1) 162 vmlsl.u8 q7, d28, d0 @// a0 + a5 + 20a2 + 20a3 - 5a1 (column1,row1) 170 vmlsl.u8 q7, d28, d0 @// a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 (column1,row1) 177 …vqrshrun.s16 d18, q7, #5 @// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column1,r… 200 vaddl.u8 q7, d28, d5 @// a0 + a5 (column1,row1) 202 vmlal.u8 q7, d25, d1 @// a0 + a5 + 20a2 (column1,row1) 203 vmlal.u8 q7, d24, d1 @// a0 + a5 + 20a2 + 20a3 (column1,row1) 204 vmlsl.u8 q7, d23, d0 @// a0 + a5 + 20a2 + 20a3 - 5a1 (column1,row1) [all …]
|
D | ih264_inter_pred_filters_luma_vert_a9q.s | 129 vaddl.u8 q7, d0, d10 @ temp = src[0_0] + src[5_0] 131 vmla.u16 q7, q6, q11 @ temp += temp1 * 20 138 vmls.u16 q7, q8, q12 @ temp -= temp2 * 5 145 vqrshrun.s16 d30, q7, #5 @ dst[0_0] = CLIP_U8((temp +16) >> 5) 146 vaddl.u8 q7, d3, d1 148 vmla.u16 q7, q6, q11 157 vmls.u16 q7, q13, q12 166 vqrshrun.s16 d31, q7, #5 168 vaddl.u8 q7, d6, d4 170 vmla.u16 q7, q6, q11 [all …]
|
/external/boringssl/linux-arm/crypto/fipsmodule/ |
D | bsaes-armv7.S | 1117 veor q7, q7, q6 @ fix up round 0 key 1118 vstmia sp, {q7} 1133 veor q7, q7, q6 @ fix up round 0 key 1134 vstmia r4, {q7} 1157 vld1.8 {q6,q7}, [r0] 1173 veor q7, q7, q12 1180 vst1.8 {q7}, [r1]! 1227 veor q7, q7, q12 1233 vst1.8 {q7}, [r1]! 1250 veor q7, q7, q12 [all …]
|
/external/boringssl/ios-arm/crypto/fipsmodule/ |
D | bsaes-armv7.S | 1124 veor q7, q7, q6 @ fix up round 0 key 1125 vstmia sp, {q7} 1140 veor q7, q7, q6 @ fix up round 0 key 1141 vstmia r4, {q7} 1164 vld1.8 {q6,q7}, [r0] 1180 veor q7, q7, q12 1187 vst1.8 {q7}, [r1]! 1234 veor q7, q7, q12 1240 vst1.8 {q7}, [r1]! 1257 veor q7, q7, q12 [all …]
|
/external/boringssl/src/crypto/curve25519/asm/ |
D | x25519-asm-arm.S | 30 vpush {q4,q5,q6,q7} 101 vshr.u64 q7,q7,#29 113 vand q7,q7,q3 140 vadd.i64 q7,q7,q12 142 vadd.i64 q15,q7,q0 156 vsub.i64 q7,q7,q12 167 vsub.i64 q7,q8,q12 184 vshl.i64 q7,q8,#26 189 vsub.i64 q3,q6,q7 250 veor q10,q7,q9 [all …]
|
/external/libhevc/common/arm/ |
D | ihevc_inter_pred_luma_vert_w16inp_w16out.s | 210 vmull.s16 q7,d4,d23 211 vmlal.s16 q7,d3,d22 212 vmlal.s16 q7,d5,d24 213 vmlal.s16 q7,d6,d25 215 vmlal.s16 q7,d7,d26 217 vmlal.s16 q7,d16,d27 219 vmlal.s16 q7,d17,d28 221 vmlal.s16 q7,d18,d29 250 vsub.s32 q7, q7, q15 269 vshrn.s32 d14, q7, #6 [all …]
|
D | ihevc_inter_pred_filters_luma_vert_w16inp.s | 199 vmull.s16 q7,d4,d23 200 vmlal.s16 q7,d3,d22 201 vmlal.s16 q7,d5,d24 202 vmlal.s16 q7,d6,d25 204 vmlal.s16 q7,d7,d26 206 vmlal.s16 q7,d16,d27 208 vmlal.s16 q7,d17,d28 210 vmlal.s16 q7,d18,d29 237 vqshrn.s32 d14, q7, #6 255 vqrshrun.s16 d14,q7,#6 [all …]
|
/external/libvpx/libvpx/vp8/common/arm/neon/ |
D | vp8_loopfilter_neon.c | 22 uint8x16_t q7, // q0 in vp8_loop_filter_neon() argument 39 q14u8 = vabdq_u8(q8, q7); in vp8_loop_filter_neon() 48 q9 = vabdq_u8(q6, q7); in vp8_loop_filter_neon() 64 q7 = veorq_u8(q7, q10); in vp8_loop_filter_neon() 73 q2s16 = vsubl_s8(vget_low_s8(vreinterpretq_s8_u8(q7)), in vp8_loop_filter_neon() 75 q11s16 = vsubl_s8(vget_high_s8(vreinterpretq_s8_u8(q7)), in vp8_loop_filter_neon() 109 q10s8 = vqsubq_s8(vreinterpretq_s8_u8(q7), q1s8); in vp8_loop_filter_neon() 130 uint8x16_t q5, q6, q7, q8, q9, q10; in vp8_loop_filter_horizontal_edge_y_neon() local 145 q7 = vld1q_u8(src); in vp8_loop_filter_horizontal_edge_y_neon() 153 vp8_loop_filter_neon(qblimit, qlimit, qthresh, q3, q4, q5, q6, q7, q8, q9, in vp8_loop_filter_horizontal_edge_y_neon() [all …]
|
D | mbloopfilter_neon.c | 21 uint8x16_t q7, // q0 in vp8_mbloop_filter_neon() argument 41 q14u8 = vabdq_u8(q8, q7); in vp8_mbloop_filter_neon() 50 q12u8 = vabdq_u8(q6, q7); in vp8_mbloop_filter_neon() 67 q7 = veorq_u8(q7, q0u8); in vp8_mbloop_filter_neon() 78 q2s16 = vsubl_s8(vget_low_s8(vreinterpretq_s8_u8(q7)), in vp8_mbloop_filter_neon() 80 q13s16 = vsubl_s8(vget_high_s8(vreinterpretq_s8_u8(q7)), in vp8_mbloop_filter_neon() 108 q7s8 = vqsubq_s8(vreinterpretq_s8_u8(q7), q2s8); in vp8_mbloop_filter_neon() 158 uint8x16_t q5, q6, q7, q8, q9, q10; in vp8_mbloop_filter_horizontal_edge_y_neon() local 174 q7 = vld1q_u8(src); in vp8_mbloop_filter_horizontal_edge_y_neon() 182 vp8_mbloop_filter_neon(qblimit, qlimit, qthresh, q3, q4, q5, q6, q7, q8, q9, in vp8_mbloop_filter_horizontal_edge_y_neon() [all …]
|
D | idct_dequant_full_2x_neon.c | 21 int16x8_t q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; in idct_dequant_full_2x_neon() local 87 q7 = vqdmulhq_n_s16(q5, sinpi8sqrt2); in idct_dequant_full_2x_neon() 101 q3 = vqaddq_s16(q7, q4); in idct_dequant_full_2x_neon() 106 q7 = vqsubq_s16(q10, q3); in idct_dequant_full_2x_neon() 109 q2tmp1 = vtrnq_s32(vreinterpretq_s32_s16(q5), vreinterpretq_s32_s16(q7)); in idct_dequant_full_2x_neon() 136 q7 = vqsubq_s16(q2, q9); in idct_dequant_full_2x_neon() 141 q7 = vrshrq_n_s16(q7, 3); in idct_dequant_full_2x_neon() 144 q2tmp1 = vtrnq_s32(vreinterpretq_s32_s16(q5), vreinterpretq_s32_s16(q7)); in idct_dequant_full_2x_neon() 156 q7 = vreinterpretq_s16_u16( in idct_dequant_full_2x_neon() 162 d31 = vreinterpret_s32_u8(vqmovun_s16(q7)); in idct_dequant_full_2x_neon()
|
/external/valgrind/none/tests/arm/ |
D | neon128.c | 359 TESTINSN_imm("vmov.i16 q7", q7, 0x700); in main() 373 TESTINSN_imm("vmvn.i16 q7", q7, 0x700); in main() 451 TESTINSN_bin("vorr q7, q3, q0", q7, q3, i8, 0x24, q0, i16, 0xff); in main() 457 TESTINSN_bin("vorn q7, q3, q0", q7, q3, i8, 0x24, q0, i16, 0xff); in main() 467 TESTINSN_bin("veor q7, q3, q0", q7, q3, i8, 0x24, q0, i16, 0xff); in main() 477 TESTINSN_bin("vbsl q7, q3, q0", q7, q3, i8, 0x24, q0, i16, 0xff); in main() 487 TESTINSN_bin("vbit q7, q3, q0", q7, q3, i8, 0x24, q0, i16, 0xff); in main() 497 TESTINSN_bin("vbif q7, q3, q0", q7, q3, i8, 0x24, q0, i16, 0xff); in main() 589 TESTINSN_bin("vrhadd.s8 q5, q7, q5", q5, q7, i32, (1 << 31) + 1, q5, i32, (1 << 31) + 2); in main() 592 TESTINSN_bin("vrhadd.s8 q5, q7, q5", q5, q7, i32, (1 << 31) + 1, q5, i32, (1 << 31) + 3); in main() [all …]
|
D | v8crypto_a.stdout.exp | 255 sha1h.32 q6, q7 62fd870459d3b6088d5e5d5e6ab23ddf b9d47da8de6340ab234dd474bfd8d4c6 0000000000000… 256 sha1h.32 q6, q7 d8308a161f694382213cafb53a36aff3 86ff9089fca3cac10660484e4f7344dc 0000000000000… 257 sha1h.32 q6, q7 be27c1a5174d78728950fadd603a4d91 7b86d0bd834973a67642acd16da757d4 0000000000000… 258 sha1h.32 q6, q7 f3400e1021de3339a4fd1e35bc20f51a 77c81da655b51bba5253e15cf9d4ed0e 0000000000000… 259 sha1h.32 q6, q7 59de50b71e7c553654a0fa1d2d4888ed 5c2557a35145a25b79f3c74fd35be6ea 0000000000000… 260 sha1h.32 q6, q7 ce5f66faeb88beca779c70f69511e66a 08fe5d15565be8ebcc833e0bdb9b22c9 0000000000000… 261 sha1h.32 q6, q7 3423313a6b614e53ee4e5e1fd2dbeff2 5db2115b4557cdca2b6225eff1f5800b 0000000000000… 262 sha1h.32 q6, q7 698a92d77d68e5349918a6f9c50683e4 39a151d5fe98305776f15e5cf5c8e20f 0000000000000… 263 sha1h.32 q6, q7 4ef5672f00fb63ca585926e34ef381a1 7d2cfee4617ef2f28d8fc7b1c6742635 0000000000000… 264 sha1h.32 q6, q7 c3c391a4d67da7770a72bf3d4d01cb88 09b2f9e74e69f4fb509d414e46592ddd 0000000000000… [all …]
|
/external/libxaac/decoder/armv7/ |
D | ixheaacd_sbr_imdct_using_fft.s | 186 VADD.I32 q2, q3, q7 190 VSUB.I32 q6, q3, q7 196 VSUB.S32 q7, q9, q6 219 VPUSH {q7} 230 VADD.I32 q7, q14, q12 259 VADD.S32 q11, q7, q15 260 VSUB.S32 q2, q7, q15 261 VADD.S32 q7, q14, q10 283 VADD.S32 q4, q8, q7 287 VSUB.S32 q11, q8, q7 [all …]
|