/external/llvm/test/MC/ARM/ |
D | vpush-vpop.s | 7 vpush {s8, s9, s10, s11, s12} 9 vpop {s8, s9, s10, s11, s12} 11 vpush.s8 {d8, d9, d10, d11, d12} 12 vpush.16 {s8, s9, s10, s11, s12} 14 vpop.64 {s8, s9, s10, s11, s12} 17 @ CHECK-THUMB: vpush {s8, s9, s10, s11, s12} @ encoding: [0x2d,0xed,0x05,0x4a] 19 @ CHECK-THUMB: vpop {s8, s9, s10, s11, s12} @ encoding: [0xbd,0xec,0x05,0x4a] 22 @ CHECK-ARM: vpush {s8, s9, s10, s11, s12} @ encoding: [0x05,0x4a,0x2d,0xed] 24 @ CHECK-ARM: vpop {s8, s9, s10, s11, s12} @ encoding: [0x05,0x4a,0xbd,0xec] 27 @ CHECK-THUMB: vpush {s8, s9, s10, s11, s12} @ encoding: [0x2d,0xed,0x05,0x4a] [all …]
|
D | neon-abs-encoding.s | 3 @ CHECK: vabs.s8 d16, d16 @ encoding: [0x20,0x03,0xf1,0xf3] 4 vabs.s8 d16, d16 11 @ CHECK: vabs.s8 q8, q8 @ encoding: [0x60,0x03,0xf1,0xf3] 12 vabs.s8 q8, q8 20 @ CHECK: vqabs.s8 d16, d16 @ encoding: [0x20,0x07,0xf0,0xf3] 21 vqabs.s8 d16, d16 26 @ CHECK: vqabs.s8 q8, q8 @ encoding: [0x60,0x07,0xf0,0xf3] 27 vqabs.s8 q8, q8
|
D | neon-neg-encoding.s | 3 @ CHECK: vneg.s8 d16, d16 @ encoding: [0xa0,0x03,0xf1,0xf3] 4 vneg.s8 d16, d16 11 @ CHECK: vneg.s8 q8, q8 @ encoding: [0xe0,0x03,0xf1,0xf3] 12 vneg.s8 q8, q8 19 @ CHECK: vqneg.s8 d16, d16 @ encoding: [0xa0,0x07,0xf0,0xf3] 20 vqneg.s8 d16, d16 25 @ CHECK: vqneg.s8 q8, q8 @ encoding: [0xe0,0x07,0xf0,0xf3] 26 vqneg.s8 q8, q8
|
D | neont2-neg-encoding.s | 5 @ CHECK: vneg.s8 d16, d16 @ encoding: [0xf1,0xff,0xa0,0x03] 6 vneg.s8 d16, d16 13 @ CHECK: vneg.s8 q8, q8 @ encoding: [0xf1,0xff,0xe0,0x03] 14 vneg.s8 q8, q8 21 @ CHECK: vqneg.s8 d16, d16 @ encoding: [0xf0,0xff,0xa0,0x07] 22 vqneg.s8 d16, d16 27 @ CHECK: vqneg.s8 q8, q8 @ encoding: [0xf0,0xff,0xe0,0x07] 28 vqneg.s8 q8, q8
|
D | neont2-abs-encoding.s | 5 @ CHECK: vabs.s8 d16, d16 @ encoding: [0xf1,0xff,0x20,0x03] 6 vabs.s8 d16, d16 13 @ CHECK: vabs.s8 q8, q8 @ encoding: [0xf1,0xff,0x60,0x03] 14 vabs.s8 q8, q8 22 @ CHECK: vqabs.s8 d16, d16 @ encoding: [0xf0,0xff,0x20,0x07] 23 vqabs.s8 d16, d16 28 @ CHECK: vqabs.s8 q8, q8 @ encoding: [0xf0,0xff,0x60,0x07] 29 vqabs.s8 q8, q8
|
D | neont2-minmax-encoding.s | 5 vmax.s8 d1, d2, d3 13 vmax.s8 d2, d3 21 vmax.s8 q1, q2, q3 29 vmax.s8 q2, q3 37 @ CHECK: vmax.s8 d1, d2, d3 @ encoding: [0x02,0xef,0x03,0x16] 44 @ CHECK: vmax.s8 d2, d2, d3 @ encoding: [0x02,0xef,0x03,0x26] 51 @ CHECK: vmax.s8 q1, q2, q3 @ encoding: [0x04,0xef,0x46,0x26] 58 @ CHECK: vmax.s8 q2, q2, q3 @ encoding: [0x04,0xef,0x46,0x46] 67 vmin.s8 d1, d2, d3 75 vmin.s8 d2, d3 [all …]
|
D | neon-minmax-encoding.s | 3 vmax.s8 d1, d2, d3 11 vmax.s8 d2, d3 19 vmax.s8 q1, q2, q3 27 vmax.s8 q2, q3 35 @ CHECK: vmax.s8 d1, d2, d3 @ encoding: [0x03,0x16,0x02,0xf2] 42 @ CHECK: vmax.s8 d2, d2, d3 @ encoding: [0x03,0x26,0x02,0xf2] 49 @ CHECK: vmax.s8 q1, q2, q3 @ encoding: [0x46,0x26,0x04,0xf2] 56 @ CHECK: vmax.s8 q2, q2, q3 @ encoding: [0x46,0x46,0x04,0xf2] 65 vmin.s8 d1, d2, d3 73 vmin.s8 d2, d3 [all …]
|
D | neon-add-encoding.s | 17 @ CHECK: vaddl.s8 q8, d17, d16 @ encoding: [0xa0,0x00,0xc1,0xf2] 18 vaddl.s8 q8, d17, d16 30 @ CHECK: vaddw.s8 q8, q8, d18 @ encoding: [0xa2,0x01,0xc0,0xf2] 31 vaddw.s8 q8, q8, d18 43 @ CHECK: vhadd.s8 d16, d16, d17 @ encoding: [0xa1,0x00,0x40,0xf2] 44 vhadd.s8 d16, d16, d17 55 @ CHECK: vhadd.s8 q8, q8, q9 @ encoding: [0xe2,0x00,0x40,0xf2] 56 vhadd.s8 q8, q8, q9 69 vhadd.s8 d11, d24 75 vhadd.s8 q1, q12 [all …]
|
D | neon-cmp-encoding.s | 21 vcge.s8 d16, d16, d17 28 vcge.s8 q8, q8, q9 38 @ CHECK: vcge.s8 d16, d16, d17 @ encoding: [0xb1,0x03,0x40,0xf2] 45 @ CHECK: vcge.s8 q8, q8, q9 @ encoding: [0xf2,0x03,0x40,0xf2] 55 vcgt.s8 d16, d16, d17 62 vcgt.s8 q8, q8, q9 72 @ CHECK: vcgt.s8 d16, d16, d17 @ encoding: [0xa1,0x03,0x40,0xf2] 79 @ CHECK: vcgt.s8 q8, q8, q9 @ encoding: [0xe2,0x03,0x40,0xf2] 104 vcge.s8 d16, d16, #0 105 vcle.s8 d16, d16, #0 [all …]
|
D | neon-absdiff-encoding.s | 3 @ CHECK: vabd.s8 d16, d16, d17 @ encoding: [0xa1,0x07,0x40,0xf2] 4 vabd.s8 d16, d16, d17 17 @ CHECK: vabd.s8 q8, q8, q9 @ encoding: [0xe2,0x07,0x40,0xf2] 18 vabd.s8 q8, q8, q9 32 @ CHECK: vabdl.s8 q8, d16, d17 @ encoding: [0xa1,0x07,0xc0,0xf2] 33 vabdl.s8 q8, d16, d17 45 @ CHECK: vaba.s8 d16, d18, d17 @ encoding: [0xb1,0x07,0x42,0xf2] 46 vaba.s8 d16, d18, d17 57 @ CHECK: vaba.s8 q9, q8, q10 @ encoding: [0xf4,0x27,0x40,0xf2] 58 vaba.s8 q9, q8, q10 [all …]
|
D | neont2-absdiff-encoding.s | 5 vabd.s8 d16, d16, d17 12 vabd.s8 q8, q8, q9 20 @ CHECK: vabd.s8 d16, d16, d17 @ encoding: [0x40,0xef,0xa1,0x07] 27 @ CHECK: vabd.s8 q8, q8, q9 @ encoding: [0x40,0xef,0xe2,0x07] 36 vabdl.s8 q8, d16, d17 43 @ CHECK: vabdl.s8 q8, d16, d17 @ encoding: [0xc0,0xef,0xa1,0x07] 51 vaba.s8 d16, d18, d17 57 vaba.s8 q9, q8, q10 64 @ CHECK: vaba.s8 d16, d18, d17 @ encoding: [0x42,0xef,0xb1,0x07] 70 @ CHECK: vaba.s8 q9, q8, q10 @ encoding: [0x40,0xef,0xf4,0x27] [all …]
|
D | neont2-pairwise-encoding.s | 15 vpaddl.s8 d7, d10 21 vpaddl.s8 q4, q7 28 @ CHECK: vpaddl.s8 d7, d10 @ encoding: [0xb0,0xff,0x0a,0x72] 34 @ CHECK: vpaddl.s8 q4, q7 @ encoding: [0xb0,0xff,0x4e,0x82] 42 vpadal.s8 d16, d4 48 vpadal.s8 q4, q10 55 @ CHECK: vpadal.s8 d16, d4 @ encoding: [0xf0,0xff,0x04,0x06] 61 @ CHECK: vpadal.s8 q4, q10 @ encoding: [0xb0,0xff,0x64,0x86] 69 vpmin.s8 d16, d29, d10 77 @ CHECK: vpmin.s8 d16, d29, d10 @ encoding: [0x4d,0xef,0x9a,0x0a] [all …]
|
D | neon-shift-encoding.s | 47 vshr.s8 d16, d16, #7 51 vshr.s8 q8, q8, #7 64 @ CHECK: vshr.s8 d16, d16, #7 @ encoding: [0x30,0x00,0xc9,0xf2] 68 @ CHECK: vshr.s8 q8, q8, #7 @ encoding: [0x70,0x00,0xc9,0xf2] 82 vshr.s8 d16, #7 86 vshr.s8 q8, #7 99 @ CHECK: vshr.s8 d16, d16, #7 @ encoding: [0x30,0x00,0xc9,0xf2] 103 @ CHECK: vshr.s8 q8, q8, #7 @ encoding: [0x70,0x00,0xc9,0xf2] 109 vsra.s8 d16, d6, #7 113 vsra.s8 q1, q8, #7 [all …]
|
D | neont2-add-encoding.s | 18 @ CHECK: vaddl.s8 q8, d17, d16 @ encoding: [0xc1,0xef,0xa0,0x00] 19 vaddl.s8 q8, d17, d16 31 @ CHECK: vaddw.s8 q8, q8, d18 @ encoding: [0xc0,0xef,0xa2,0x01] 32 vaddw.s8 q8, q8, d18 44 @ CHECK: vhadd.s8 d16, d16, d17 @ encoding: [0x40,0xef,0xa1,0x00] 45 vhadd.s8 d16, d16, d17 56 @ CHECK: vhadd.s8 q8, q8, q9 @ encoding: [0x40,0xef,0xe2,0x00] 57 vhadd.s8 q8, q8, q9 69 @ CHECK: vrhadd.s8 d16, d16, d17 @ encoding: [0x40,0xef,0xa1,0x01] 70 vrhadd.s8 d16, d16, d17 [all …]
|
D | neon-pairwise-encoding.s | 21 @ CHECK: vpaddl.s8 d16, d16 @ encoding: [0x20,0x02,0xf0,0xf3] 22 vpaddl.s8 d16, d16 33 @ CHECK: vpaddl.s8 q8, q8 @ encoding: [0x60,0x02,0xf0,0xf3] 34 vpaddl.s8 q8, q8 45 @ CHECK: vpadal.s8 d16, d17 @ encoding: [0x21,0x06,0xf0,0xf3] 46 vpadal.s8 d16, d17 57 @ CHECK: vpadal.s8 q9, q8 @ encoding: [0x60,0x26,0xf0,0xf3] 58 vpadal.s8 q9, q8 69 @ CHECK: vpmin.s8 d16, d16, d17 @ encoding: [0xb1,0x0a,0x40,0xf2] 70 vpmin.s8 d16, d16, d17 [all …]
|
D | neon-satshift-encoding.s | 3 @ CHECK: vqshl.s8 d16, d16, d17 @ encoding: [0xb0,0x04,0x41,0xf2] 4 vqshl.s8 d16, d16, d17 19 @ CHECK: vqshl.s8 q8, q8, q9 @ encoding: [0xf0,0x04,0x42,0xf2] 20 vqshl.s8 q8, q8, q9 35 @ CHECK: vqshl.s8 d16, d16, #7 @ encoding: [0x30,0x07,0xcf,0xf2] 36 vqshl.s8 d16, d16, #7 51 @ CHECK: vqshlu.s8 d16, d16, #7 @ encoding: [0x30,0x06,0xcf,0xf3] 52 vqshlu.s8 d16, d16, #7 59 @ CHECK: vqshl.s8 q8, q8, #7 @ encoding: [0x70,0x07,0xcf,0xf2] 60 vqshl.s8 q8, q8, #7 [all …]
|
D | neont2-satshift-encoding.s | 5 @ CHECK: vqshl.s8 d16, d16, d17 @ encoding: [0x41,0xef,0xb0,0x04] 6 vqshl.s8 d16, d16, d17 21 @ CHECK: vqshl.s8 q8, q8, q9 @ encoding: [0x42,0xef,0xf0,0x04] 22 vqshl.s8 q8, q8, q9 37 @ CHECK: vqshl.s8 d16, d16, #7 @ encoding: [0xcf,0xef,0x30,0x07] 38 vqshl.s8 d16, d16, #7 53 @ CHECK: vqshlu.s8 d16, d16, #7 @ encoding: [0xcf,0xff,0x30,0x06] 54 vqshlu.s8 d16, d16, #7 61 @ CHECK: vqshl.s8 q8, q8, #7 @ encoding: [0xcf,0xef,0x70,0x07] 62 vqshl.s8 q8, q8, #7 [all …]
|
D | neon-sub-encoding.s | 49 @ CHECK: vsubl.s8 q8, d17, d16 @ encoding: [0xa0,0x02,0xc1,0xf2] 50 vsubl.s8 q8, d17, d16 61 @ CHECK: vsubw.s8 q8, q8, d18 @ encoding: [0xa2,0x03,0xc0,0xf2] 62 vsubw.s8 q8, q8, d18 73 @ CHECK: vhsub.s8 d16, d16, d17 @ encoding: [0xa1,0x02,0x40,0xf2] 74 vhsub.s8 d16, d16, d17 85 @ CHECK: vhsub.s8 q8, q8, q9 @ encoding: [0xe2,0x02,0x40,0xf2] 86 vhsub.s8 q8, q8, q9 91 @ CHECK: vqsub.s8 d16, d16, d17 @ encoding: [0xb1,0x02,0x40,0xf2] 92 vqsub.s8 d16, d16, d17 [all …]
|
D | neont2-shiftaccum-encoding.s | 5 vsra.s8 d17, d16, #8 9 vsra.s8 q7, q2, #8 23 vsra.s8 d16, #8 27 vsra.s8 q2, #8 40 @ CHECK: vsra.s8 d17, d16, #8 @ encoding: [0xc8,0xef,0x30,0x11] 44 @ CHECK: vsra.s8 q7, q2, #8 @ encoding: [0x88,0xef,0x54,0xe1] 57 @ CHECK: vsra.s8 d16, d16, #8 @ encoding: [0xc8,0xef,0x30,0x01] 61 @ CHECK: vsra.s8 q2, q2, #8 @ encoding: [0x88,0xef,0x54,0x41] 75 vrsra.s8 d5, d26, #8 83 vrsra.s8 q1, q2, #8 [all …]
|
D | neon-shiftaccum-encoding.s | 3 vsra.s8 d17, d16, #8 7 vsra.s8 q7, q2, #8 21 vsra.s8 d16, #8 25 vsra.s8 q2, #8 38 @ CHECK: vsra.s8 d17, d16, #8 @ encoding: [0x30,0x11,0xc8,0xf2] 42 @ CHECK: vsra.s8 q7, q2, #8 @ encoding: [0x54,0xe1,0x88,0xf2] 55 @ CHECK: vsra.s8 d16, d16, #8 @ encoding: [0x30,0x01,0xc8,0xf2] 59 @ CHECK: vsra.s8 q2, q2, #8 @ encoding: [0x54,0x41,0x88,0xf2] 72 vrsra.s8 d5, d26, #8 80 vrsra.s8 q1, q2, #8 [all …]
|
/external/grub/stage2/ |
D | jfs.h | 98 typedef char s8; typedef 219 s8 s_fscklog; /* 1: which fsck service log is most recent 283 s8 next; /* 1: */ 284 s8 cnt; /* 1: */ 296 s8 next; /* 1: */ 308 s8 next; /* 1: */ 343 s8 nextindex; /* 1: next free entry in stbl */ 344 s8 freecnt; /* 1: free count */ 345 s8 freelist; /* 1: freelist header */ 349 s8 stbl[8]; /* 8: sorted entry index table */ [all …]
|
/external/libvpx/libvpx/vp8/common/arm/neon/ |
D | loopfiltersimplehorizontaledge_neon.asm | 49 vsubl.s8 q2, d14, d12 ; ( qs0 - ps0) 50 vsubl.s8 q3, d15, d13 52 vqsub.s8 q4, q5, q8 ; q4: vp8_filter = vp8_signed_char_clamp(ps1-qs1) 60 vaddw.s8 q2, q2, d8 ; vp8_filter + 3 * ( qs0 - ps0) 61 vaddw.s8 q3, q3, d9 68 vqadd.s8 q2, q14, q10 ; Filter2 = vp8_signed_char_clamp(vp8_filter+3) 69 vqadd.s8 q3, q14, q9 ; Filter1 = vp8_signed_char_clamp(vp8_filter+4) 70 vshr.s8 q2, q2, #3 ; Filter2 >>= 3 71 vshr.s8 q4, q3, #3 ; Filter1 >>= 3 76 vqadd.s8 q11, q6, q2 ; u = vp8_signed_char_clamp(ps0 + Filter2) [all …]
|
D | loopfiltersimpleverticaledge_neon.asm | 69 vsubl.s8 q2, d8, d10 ; ( qs0 - ps0) 70 vsubl.s8 q13, d9, d11 72 vqsub.s8 q14, q3, q6 ; vp8_filter = vp8_signed_char_clamp(ps1-qs1) 80 vaddw.s8 q2, q2, d28 ; vp8_filter + 3 * ( qs0 - ps0) 81 vaddw.s8 q13, q13, d29 91 vqadd.s8 q2, q14, q11 ; Filter2 = vp8_signed_char_clamp(vp8_filter+3) 92 vqadd.s8 q3, q14, q12 ; Filter1 = vp8_signed_char_clamp(vp8_filter+4) 93 vshr.s8 q2, q2, #3 ; Filter2 >>= 3 94 vshr.s8 q14, q3, #3 ; Filter1 >>= 3 97 vqadd.s8 q11, q5, q2 ; u = vp8_signed_char_clamp(ps0 + Filter2) [all …]
|
/external/compiler-rt/lib/asan/ |
D | asan_poisoning.cc | 42 s8 offset; // in [0, SHADOW_GRANULARITY) 43 s8 value; // = *chunk; 79 s8 value = beg.value; in __asan_poison_memory_region() 121 s8 value = beg.value; in __asan_unpoison_memory_region() 230 s8 end_offset = (s8)(size - aligned_size); in PoisonAlignedStackMemory() 231 s8* shadow_end = (s8*)MemToShadow(addr + aligned_size); in PoisonAlignedStackMemory() 232 s8 end_value = *shadow_end; in PoisonAlignedStackMemory() 237 *shadow_end = (s8)kAsanStackUseAfterScopeMagic; in PoisonAlignedStackMemory()
|
/external/clang/test/CodeGenCXX/ |
D | 2010-06-22-ZeroBitfield.cpp | 4 struct s8 { s8_0 a; s8_1 b; }; struct 5 s8 f8() { return s8(); } in f8()
|