/external/swiftshader/third_party/LLVM/test/MC/ARM/ |
D | neon-shiftaccum-encoding.s | 11 @ CHECK: vsra.s8 q8, q9, #8 @ encoding: [0x72,0x01,0xc8,0xf2] 12 vsra.s8 q8, q9, #8 13 @ CHECK: vsra.s16 q8, q9, #16 @ encoding: [0x72,0x01,0xd0,0xf2] 14 vsra.s16 q8, q9, #16 15 @ CHECK: vsra.s32 q8, q9, #32 @ encoding: [0x72,0x01,0xe0,0xf2] 16 vsra.s32 q8, q9, #32 17 @ CHECK: vsra.s64 q8, q9, #64 @ encoding: [0xf2,0x01,0xc0,0xf2] 18 vsra.s64 q8, q9, #64 27 @ CHECK: vsra.u8 q8, q9, #8 @ encoding: [0x72,0x01,0xc8,0xf3] 28 vsra.u8 q8, q9, #8 [all …]
|
D | neont2-shiftaccum-encoding.s | 13 @ CHECK: vsra.s8 q8, q9, #8 @ encoding: [0xc8,0xef,0x72,0x01] 14 vsra.s8 q8, q9, #8 15 @ CHECK: vsra.s16 q8, q9, #16 @ encoding: [0xd0,0xef,0x72,0x01] 16 vsra.s16 q8, q9, #16 17 @ CHECK: vsra.s32 q8, q9, #32 @ encoding: [0xe0,0xef,0x72,0x01] 18 vsra.s32 q8, q9, #32 19 @ CHECK: vsra.s64 q8, q9, #64 @ encoding: [0xc0,0xef,0xf2,0x01] 20 vsra.s64 q8, q9, #64 29 @ CHECK: vsra.u8 q8, q9, #8 @ encoding: [0xc8,0xff,0x72,0x01] 30 vsra.u8 q8, q9, #8 [all …]
|
D | neont2-add-encoding.s | 15 @ CHECK: vadd.f32 q8, q8, q9 @ encoding: [0x40,0xef,0xe2,0x0d] 16 vadd.f32 q8, q8, q9 56 @ CHECK: vhadd.s8 q8, q8, q9 @ encoding: [0x40,0xef,0xe2,0x00] 57 vhadd.s8 q8, q8, q9 58 @ CHECK: vhadd.s16 q8, q8, q9 @ encoding: [0x50,0xef,0xe2,0x00] 59 vhadd.s16 q8, q8, q9 60 @ CHECK: vhadd.s32 q8, q8, q9 @ encoding: [0x60,0xef,0xe2,0x00] 61 vhadd.s32 q8, q8, q9 62 @ CHECK: vhadd.u8 q8, q8, q9 @ encoding: [0x40,0xff,0xe2,0x00] 63 vhadd.u8 q8, q8, q9 [all …]
|
D | neon-add-encoding.s | 14 @ CHECK: vadd.f32 q8, q8, q9 @ encoding: [0xe2,0x0d,0x40,0xf2] 15 vadd.f32 q8, q8, q9 55 @ CHECK: vhadd.s8 q8, q8, q9 @ encoding: [0xe2,0x00,0x40,0xf2] 56 vhadd.s8 q8, q8, q9 57 @ CHECK: vhadd.s16 q8, q8, q9 @ encoding: [0xe2,0x00,0x50,0xf2] 58 vhadd.s16 q8, q8, q9 59 @ CHECK: vhadd.s32 q8, q8, q9 @ encoding: [0xe2,0x00,0x60,0xf2] 60 vhadd.s32 q8, q8, q9 61 @ CHECK: vhadd.u8 q8, q8, q9 @ encoding: [0xe2,0x00,0x40,0xf3] 62 vhadd.u8 q8, q8, q9 [all …]
|
D | neon-cmp-encoding.s | 7 vceq.i8 q8, q8, q9 8 vceq.i16 q8, q8, q9 9 vceq.i32 q8, q8, q9 10 vceq.f32 q8, q8, q9 16 @ CHECK: vceq.i8 q8, q8, q9 @ encoding: [0xf2,0x08,0x40,0xf3] 17 @ CHECK: vceq.i16 q8, q8, q9 @ encoding: [0xf2,0x08,0x50,0xf3] 18 @ CHECK: vceq.i32 q8, q8, q9 @ encoding: [0xf2,0x08,0x60,0xf3] 19 @ CHECK: vceq.f32 q8, q8, q9 @ encoding: [0xe2,0x0e,0x40,0xf2] 28 vcge.s8 q8, q8, q9 29 vcge.s16 q8, q8, q9 [all …]
|
D | neon-sub-encoding.s | 13 @ CHECK: vsub.i8 q8, q8, q9 @ encoding: [0xe2,0x08,0x40,0xf3] 14 vsub.i8 q8, q8, q9 15 @ CHECK: vsub.i16 q8, q8, q9 @ encoding: [0xe2,0x08,0x50,0xf3] 16 vsub.i16 q8, q8, q9 17 @ CHECK: vsub.i32 q8, q8, q9 @ encoding: [0xe2,0x08,0x60,0xf3] 18 vsub.i32 q8, q8, q9 19 @ CHECK: vsub.i64 q8, q8, q9 @ encoding: [0xe2,0x08,0x70,0xf3] 20 vsub.i64 q8, q8, q9 21 @ CHECK: vsub.f32 q8, q8, q9 @ encoding: [0xe2,0x0d,0x60,0xf2] 22 vsub.f32 q8, q8, q9 [all …]
|
D | neon-shuffle-encoding.s | 7 @ CHECK: vext.8 q8, q9, q8, #3 @ encoding: [0xe0,0x03,0xf2,0xf2] 8 vext.8 q8, q9, q8, #3 9 @ CHECK: vext.8 q8, q9, q8, #7 @ encoding: [0xe0,0x07,0xf2,0xf2] 10 vext.8 q8, q9, q8, #7 13 @ CHECK: vext.32 q8, q9, q8, #3 @ encoding: [0xe0,0x0c,0xf2,0xf2] 14 vext.32 q8, q9, q8, #3 21 @ CHECK: vtrn.8 q9, q8 @ encoding: [0xe0,0x20,0xf2,0xf3] 22 vtrn.8 q9, q8 23 @ CHECK: vtrn.16 q9, q8 @ encoding: [0xe0,0x20,0xf6,0xf3] 24 vtrn.16 q9, q8 [all …]
|
D | neont2-shuffle-encoding.s | 9 @ CHECK: vext.8 q8, q9, q8, #3 @ encoding: [0xf2,0xef,0xe0,0x03] 10 vext.8 q8, q9, q8, #3 11 @ CHECK: vext.8 q8, q9, q8, #7 @ encoding: [0xf2,0xef,0xe0,0x07] 12 vext.8 q8, q9, q8, #7 15 @ CHECK: vext.32 q8, q9, q8, #3 @ encoding: [0xf2,0xef,0xe0,0x0c] 16 vext.32 q8, q9, q8, #3 23 @ CHECK: vtrn.8 q9, q8 @ encoding: [0xf2,0xff,0xe0,0x20] 24 vtrn.8 q9, q8 25 @ CHECK: vtrn.16 q9, q8 @ encoding: [0xf6,0xff,0xe0,0x20] 26 vtrn.16 q9, q8 [all …]
|
D | neont2-sub-encoding.s | 7 @ CHECK: vext.8 q8, q9, q8, #3 @ encoding: [0xf2,0xef,0xe0,0x03] 8 vext.8 q8, q9, q8, #3 9 @ CHECK: vext.8 q8, q9, q8, #7 @ encoding: [0xf2,0xef,0xe0,0x07] 10 vext.8 q8, q9, q8, #7 13 @ CHECK: vext.32 q8, q9, q8, #3 @ encoding: [0xf2,0xef,0xe0,0x0c] 14 vext.32 q8, q9, q8, #3 21 @ CHECK: vtrn.8 q9, q8 @ encoding: [0xf2,0xff,0xe0,0x20] 22 vtrn.8 q9, q8 23 @ CHECK: vtrn.16 q9, q8 @ encoding: [0xf6,0xff,0xe0,0x20] 24 vtrn.16 q9, q8 [all …]
|
D | neon-minmax-encoding.s | 17 @ CHECK: vmin.s8 q8, q8, q9 @ encoding: [0xf2,0x06,0x40,0xf2] 18 vmin.s8 q8, q8, q9 19 @ CHECK: vmin.s16 q8, q8, q9 @ encoding: [0xf2,0x06,0x50,0xf2] 20 vmin.s16 q8, q8, q9 21 @ CHECK: vmin.s32 q8, q8, q9 @ encoding: [0xf2,0x06,0x60,0xf2] 22 vmin.s32 q8, q8, q9 23 @ CHECK: vmin.u8 q8, q8, q9 @ encoding: [0xf2,0x06,0x40,0xf3] 24 vmin.u8 q8, q8, q9 25 @ CHECK: vmin.u16 q8, q8, q9 @ encoding: [0xf2,0x06,0x50,0xf3] 26 vmin.u16 q8, q8, q9 [all …]
|
D | neont2-minmax-encoding.s | 19 @ CHECK: vmin.s8 q8, q8, q9 @ encoding: [0x40,0xef,0xf2,0x06] 20 vmin.s8 q8, q8, q9 21 @ CHECK: vmin.s16 q8, q8, q9 @ encoding: [0x50,0xef,0xf2,0x06] 22 vmin.s16 q8, q8, q9 23 @ CHECK: vmin.s32 q8, q8, q9 @ encoding: [0x60,0xef,0xf2,0x06] 24 vmin.s32 q8, q8, q9 25 @ CHECK: vmin.u8 q8, q8, q9 @ encoding: [0x40,0xff,0xf2,0x06] 26 vmin.u8 q8, q8, q9 27 @ CHECK: vmin.u16 q8, q8, q9 @ encoding: [0x50,0xff,0xf2,0x06] 28 vmin.u16 q8, q8, q9 [all …]
|
D | neon-absdiff-encoding.s | 17 @ CHECK: vabd.s8 q8, q8, q9 @ encoding: [0xe2,0x07,0x40,0xf2] 18 vabd.s8 q8, q8, q9 19 @ CHECK: vabd.s16 q8, q8, q9 @ encoding: [0xe2,0x07,0x50,0xf2] 20 vabd.s16 q8, q8, q9 21 @ CHECK: vabd.s32 q8, q8, q9 @ encoding: [0xe2,0x07,0x60,0xf2] 22 vabd.s32 q8, q8, q9 23 @ CHECK: vabd.u8 q8, q8, q9 @ encoding: [0xe2,0x07,0x40,0xf3] 24 vabd.u8 q8, q8, q9 25 @ CHECK: vabd.u16 q8, q8, q9 @ encoding: [0xe2,0x07,0x50,0xf3] 26 vabd.u16 q8, q8, q9 [all …]
|
/external/llvm/test/MC/ARM/ |
D | neon-add-encoding.s | 14 @ CHECK: vadd.f32 q8, q8, q9 @ encoding: [0xe2,0x0d,0x40,0xf2] 15 vadd.f32 q8, q8, q9 55 @ CHECK: vhadd.s8 q8, q8, q9 @ encoding: [0xe2,0x00,0x40,0xf2] 56 vhadd.s8 q8, q8, q9 57 @ CHECK: vhadd.s16 q8, q8, q9 @ encoding: [0xe2,0x00,0x50,0xf2] 58 vhadd.s16 q8, q8, q9 59 @ CHECK: vhadd.s32 q8, q8, q9 @ encoding: [0xe2,0x00,0x60,0xf2] 60 vhadd.s32 q8, q8, q9 61 @ CHECK: vhadd.u8 q8, q8, q9 @ encoding: [0xe2,0x00,0x40,0xf3] 62 vhadd.u8 q8, q8, q9 [all …]
|
D | neont2-add-encoding.s | 15 @ CHECK: vadd.f32 q8, q8, q9 @ encoding: [0x40,0xef,0xe2,0x0d] 16 vadd.f32 q8, q8, q9 56 @ CHECK: vhadd.s8 q8, q8, q9 @ encoding: [0x40,0xef,0xe2,0x00] 57 vhadd.s8 q8, q8, q9 58 @ CHECK: vhadd.s16 q8, q8, q9 @ encoding: [0x50,0xef,0xe2,0x00] 59 vhadd.s16 q8, q8, q9 60 @ CHECK: vhadd.s32 q8, q8, q9 @ encoding: [0x60,0xef,0xe2,0x00] 61 vhadd.s32 q8, q8, q9 62 @ CHECK: vhadd.u8 q8, q8, q9 @ encoding: [0x40,0xff,0xe2,0x00] 63 vhadd.u8 q8, q8, q9 [all …]
|
D | neont2-sub-encoding.s | 7 @ CHECK: vext.8 q8, q9, q8, #3 @ encoding: [0xf2,0xef,0xe0,0x03] 8 vext.8 q8, q9, q8, #3 9 @ CHECK: vext.8 q8, q9, q8, #7 @ encoding: [0xf2,0xef,0xe0,0x07] 10 vext.8 q8, q9, q8, #7 13 @ CHECK: vext.32 q8, q9, q8, #3 @ encoding: [0xf2,0xef,0xe0,0x0c] 14 vext.32 q8, q9, q8, #3 21 @ CHECK: vtrn.8 q9, q8 @ encoding: [0xf2,0xff,0xe0,0x20] 22 vtrn.8 q9, q8 23 @ CHECK: vtrn.16 q9, q8 @ encoding: [0xf6,0xff,0xe0,0x20] 24 vtrn.16 q9, q8 [all …]
|
D | neont2-shuffle-encoding.s | 9 @ CHECK: vext.8 q8, q9, q8, #3 @ encoding: [0xf2,0xef,0xe0,0x03] 10 vext.8 q8, q9, q8, #3 11 @ CHECK: vext.8 q8, q9, q8, #7 @ encoding: [0xf2,0xef,0xe0,0x07] 12 vext.8 q8, q9, q8, #7 15 @ CHECK: vext.32 q8, q9, q8, #3 @ encoding: [0xf2,0xef,0xe0,0x0c] 16 vext.32 q8, q9, q8, #3 23 @ CHECK: vtrn.8 q9, q8 @ encoding: [0xf2,0xff,0xe0,0x20] 24 vtrn.8 q9, q8 25 @ CHECK: vtrn.16 q9, q8 @ encoding: [0xf6,0xff,0xe0,0x20] 26 vtrn.16 q9, q8 [all …]
|
D | neon-cmp-encoding.s | 7 vceq.i8 q8, q8, q9 8 vceq.i16 q8, q8, q9 9 vceq.i32 q8, q8, q9 10 vceq.f32 q8, q8, q9 16 @ CHECK: vceq.i8 q8, q8, q9 @ encoding: [0xf2,0x08,0x40,0xf3] 17 @ CHECK: vceq.i16 q8, q8, q9 @ encoding: [0xf2,0x08,0x50,0xf3] 18 @ CHECK: vceq.i32 q8, q8, q9 @ encoding: [0xf2,0x08,0x60,0xf3] 19 @ CHECK: vceq.f32 q8, q8, q9 @ encoding: [0xe2,0x0e,0x40,0xf2] 28 vcge.s8 q8, q8, q9 29 vcge.s16 q8, q8, q9 [all …]
|
D | neon-sub-encoding.s | 8 vsub.i8 q8, q8, q9 9 vsub.i16 q8, q8, q9 10 vsub.i32 q8, q8, q9 11 vsub.i64 q8, q8, q9 12 vsub.f32 q8, q8, q9 20 vsub.i16 q2, q9 30 @ CHECK: vsub.i8 q8, q8, q9 @ encoding: [0xe2,0x08,0x40,0xf3] 31 @ CHECK: vsub.i16 q8, q8, q9 @ encoding: [0xe2,0x08,0x50,0xf3] 32 @ CHECK: vsub.i32 q8, q8, q9 @ encoding: [0xe2,0x08,0x60,0xf3] 33 @ CHECK: vsub.i64 q8, q8, q9 @ encoding: [0xe2,0x08,0x70,0xf3] [all …]
|
D | neon-shuffle-encoding.s | 5 vext.8 q8, q9, q8, #3 6 vext.8 q8, q9, q8, #7 8 vext.32 q8, q9, q8, #3 9 vext.64 q8, q9, q8, #1 14 vext.8 q9, q4, #7 22 @ CHECK: vext.8 q8, q9, q8, #3 @ encoding: [0xe0,0x03,0xf2,0xf2] 23 @ CHECK: vext.8 q8, q9, q8, #7 @ encoding: [0xe0,0x07,0xf2,0xf2] 25 @ CHECK: vext.32 q8, q9, q8, #3 @ encoding: [0xe0,0x0c,0xf2,0xf2] 26 @ CHECK: vext.64 q8, q9, q8, #1 @ encoding: [0xe0,0x08,0xf2,0xf2] 31 @ CHECK: vext.8 q9, q9, q4, #7 @ encoding: [0xc8,0x27,0xf2,0xf2] [all …]
|
/external/capstone/suite/MC/ARM/ |
D | neon-add-encoding.s.cs | 7 0xe2,0x0d,0x40,0xf2 = vadd.f32 q8, q8, q9 26 0xe2,0x00,0x40,0xf2 = vhadd.s8 q8, q8, q9 27 0xe2,0x00,0x50,0xf2 = vhadd.s16 q8, q8, q9 28 0xe2,0x00,0x60,0xf2 = vhadd.s32 q8, q8, q9 29 0xe2,0x00,0x40,0xf3 = vhadd.u8 q8, q8, q9 30 0xe2,0x00,0x50,0xf3 = vhadd.u16 q8, q8, q9 31 0xe2,0x00,0x60,0xf3 = vhadd.u32 q8, q8, q9 41 0x62,0x80,0x08,0xf3 = vhadd.u8 q4, q4, q9 50 0xe2,0x01,0x40,0xf2 = vrhadd.s8 q8, q8, q9 51 0xe2,0x01,0x50,0xf2 = vrhadd.s16 q8, q8, q9 [all …]
|
D | neon-cmp-encoding.s.cs | 6 0xf2,0x08,0x40,0xf3 = vceq.i8 q8, q8, q9 7 0xf2,0x08,0x50,0xf3 = vceq.i16 q8, q8, q9 8 0xf2,0x08,0x60,0xf3 = vceq.i32 q8, q8, q9 9 0xe2,0x0e,0x40,0xf2 = vceq.f32 q8, q8, q9 17 0xf2,0x03,0x40,0xf2 = vcge.s8 q8, q8, q9 18 0xf2,0x03,0x50,0xf2 = vcge.s16 q8, q8, q9 19 0xf2,0x03,0x60,0xf2 = vcge.s32 q8, q8, q9 20 0xf2,0x03,0x40,0xf3 = vcge.u8 q8, q8, q9 21 0xf2,0x03,0x50,0xf3 = vcge.u16 q8, q8, q9 22 0xf2,0x03,0x60,0xf3 = vcge.u32 q8, q8, q9 [all …]
|
D | neont2-add-encoding.s.cs | 7 0x40,0xef,0xe2,0x0d = vadd.f32 q8, q8, q9 26 0x40,0xef,0xe2,0x00 = vhadd.s8 q8, q8, q9 27 0x50,0xef,0xe2,0x00 = vhadd.s16 q8, q8, q9 28 0x60,0xef,0xe2,0x00 = vhadd.s32 q8, q8, q9 29 0x40,0xff,0xe2,0x00 = vhadd.u8 q8, q8, q9 30 0x50,0xff,0xe2,0x00 = vhadd.u16 q8, q8, q9 31 0x60,0xff,0xe2,0x00 = vhadd.u32 q8, q8, q9 38 0x40,0xef,0xe2,0x01 = vrhadd.s8 q8, q8, q9 39 0x50,0xef,0xe2,0x01 = vrhadd.s16 q8, q8, q9 40 0x60,0xef,0xe2,0x01 = vrhadd.s32 q8, q8, q9 [all …]
|
/external/boringssl/ios-arm/crypto/fipsmodule/ |
D | aesv8-armx32.S | 52 vext.8 q9,q0,q3,#12 57 veor q3,q3,q9 58 vext.8 q9,q0,q9,#12 59 veor q3,q3,q9 60 vext.8 q9,q0,q9,#12 62 veor q3,q3,q9 71 vext.8 q9,q0,q3,#12 75 veor q3,q3,q9 76 vext.8 q9,q0,q9,#12 77 veor q3,q3,q9 [all …]
|
D | ghashv8-armx32.S | 14 vld1.64 {q9},[r1] @ load input H 17 vext.8 q3,q9,q9,#8 19 vdup.32 q9,d18[1] 22 vshr.s32 q9,q9,#31 @ broadcast carry bit 26 vand q8,q8,q9 38 vext.8 q9,q0,q2,#8 @ Karatsuba post-processing 40 veor q1,q1,q9 53 vext.8 q9,q14,q14,#8 @ Karatsuba pre-processing 54 veor q9,q9,q14 55 vext.8 q13,q8,q9,#8 @ pack Karatsuba pre-processed [all …]
|
/external/boringssl/linux-arm/crypto/fipsmodule/ |
D | ghashv8-armx32.S | 13 vld1.64 {q9},[r1] @ load input H 16 vext.8 q3,q9,q9,#8 18 vdup.32 q9,d18[1] 21 vshr.s32 q9,q9,#31 @ broadcast carry bit 25 vand q8,q8,q9 37 vext.8 q9,q0,q2,#8 @ Karatsuba post-processing 39 veor q1,q1,q9 52 vext.8 q9,q14,q14,#8 @ Karatsuba pre-processing 53 veor q9,q9,q14 54 vext.8 q13,q8,q9,#8 @ pack Karatsuba pre-processed [all …]
|