/external/llvm-project/llvm/test/CodeGen/AArch64/ |
D | arm64-popcnt.ll | 9 ; CHECK-NEXT: fmov d0, x8 76 ; CHECK-NONEON-NEXT: lsr x8, x0, #1 77 ; CHECK-NONEON-NEXT: and x8, x8, #0x5555555555555555 78 ; CHECK-NONEON-NEXT: sub x8, x0, x8 79 ; CHECK-NONEON-NEXT: and x9, x8, #0x3333333333333333 80 ; CHECK-NONEON-NEXT: lsr x8, x8, #2 81 ; CHECK-NONEON-NEXT: and x8, x8, #0x3333333333333333 82 ; CHECK-NONEON-NEXT: add x8, x9, x8 83 ; CHECK-NONEON-NEXT: add x8, x8, x8, lsr #4 84 ; CHECK-NONEON-NEXT: and x8, x8, #0xf0f0f0f0f0f0f0f [all …]
|
D | i128_volatile_load_store.ll | 9 ; CHECK-NEXT: adrp x8, x 10 ; CHECK-NEXT: add x8, x8, :lo12:x 11 ; CHECK-NEXT: ldp x8, x9, [x8] 14 ; CHECK-NEXT: stp x8, x9, [x10] 24 ; CHECK-NEXT: adrp x8, x 25 ; CHECK-NEXT: add x8, x8, :lo12:x 26 ; CHECK-NEXT: ldp x8, x9, [x8, #504] 29 ; CHECK-NEXT: stp x8, x9, [x10, #504] 39 ; CHECK-NEXT: adrp x8, x 40 ; CHECK-NEXT: add x8, x8, :lo12:x [all …]
|
D | sve-insert-vector.ll | 12 ; CHECK-NEXT: cntd x8 13 ; CHECK-NEXT: sub x8, x8, #1 // =1 14 ; CHECK-NEXT: cmp x8, #0 // =0 15 ; CHECK-NEXT: csel x8, x8, xzr, lo 17 ; CHECK-NEXT: lsl x8, x8, #3 20 ; CHECK-NEXT: str q1, [x9, x8] 34 ; CHECK-NEXT: cntd x8 35 ; CHECK-NEXT: sub x8, x8, #1 // =1 36 ; CHECK-NEXT: cmp x8, #1 // =1 37 ; CHECK-NEXT: csinc x8, x8, xzr, lo [all …]
|
D | expand-movi-renamable.mir | 56 # CHECK-NEXT: renamable $x8 = MOVZXi 4, 0 57 # CHECK-NEXT: STRXui killed renamable $x8, $sp, 0 58 # CHECK-NEXT: $x8 = MOVZXi 45926, 0 59 # CHECK-NEXT: $x8 = MOVKXi $x8, 102, 16 60 # CHECK-NEXT: STRXui killed renamable $x8, $sp, 1 61 # CHECK-NEXT: renamable $x8 = MOVNXi 25166, 0 62 # CHECK-NEXT: renamable $x8 = MOVKXi $x8, 55686, 16 63 # CHECK-NEXT: renamable $x8 = MOVKXi $x8, 63160, 32 64 # CHECK-NEXT: STRXui killed renamable $x8, $sp, 2 65 # CHECK-NEXT: renamable $x8 = MOVZXi 30600, 0 [all …]
|
D | arm64-long-shift.ll | 7 ; CHECK-NEXT: neg x8, x2 8 ; CHECK-NEXT: lsr x8, x0, x8 10 ; CHECK-NEXT: csel x8, xzr, x8, eq 12 ; CHECK-NEXT: orr x8, x8, x9 16 ; CHECK-NEXT: csel x1, x9, x8, ge 26 ; CHECK-NEXT: lsl x8, x1, x2 31 ; CHECK-NEXT: orr x1, x8, x9 42 ; CHECK-NEXT: neg x8, x2 43 ; CHECK-NEXT: lsl x8, x1, x8 45 ; CHECK-NEXT: csel x8, xzr, x8, eq [all …]
|
D | addg_subg.mir | 3 # CHECK: renamable $x8 = IRG $sp, $xzr 5 # CHECK: renamable $x0 = ADDG $x8, 0, 0 6 # CHECK: renamable $x0 = ADDG $x8, 5, 0 7 # CHECK: renamable $x0 = ADDG $x8, 63, 0 8 # CHECK: $[[R:x[0-9]+]] = ADDXri $x8, 16, 0 11 # CHECK: renamable $x0 = SUBG $x8, 5, 0 12 # CHECK: renamable $x0 = SUBG $x8, 63, 0 13 # CHECK: $[[R:x[0-9]+]] = SUBXri $x8, 16, 0 25 renamable $x8 = IRGstack $sp, $xzr 27 renamable $x0 = TAGPstack %stack.0, 0, killed renamable $x8, 0 [all …]
|
D | mingw-refptr.ll | 13 ; CHECK: adrp x8, .refptr.var 14 ; CHECK: ldr x8, [x8, .refptr.var] 15 ; CHECK: ldr w0, [x8] 24 ; CHECK: adrp x8, dsolocalvar 25 ; CHECK: ldr w0, [x8, dsolocalvar] 34 ; CHECK: adrp x8, localvar 35 ; CHECK: ldr w0, [x8, localvar] 44 ; CHECK: adrp x8, localcommon 45 ; CHECK: ldr w0, [x8, localcommon] 54 ; CHECK: adrp x8, __imp_extvar [all …]
|
D | stp-opt-with-renaming.mir | 13 # CHECK: $x10, renamable $x8 = LDPXi renamable $x0, 0 :: (load 8) 16 # CHECK-NEXT: renamable $x8 = ADDXrr $x8, $x8 17 # CHECK-NEXT: STPXi renamable $x8, killed $x10, renamable $x0, 10 :: (store 8, align 4) 26 - { reg: '$x8' } 34 renamable $x9, renamable $x8 = LDPXi renamable $x0, 0 :: (load 8) 38 renamable $x8 = ADDXrr $x8, $x8 39 STRXui renamable $x8, renamable $x0, 10 :: (store 8, align 4) 48 # CHECK: $x10, renamable $x8 = LDPXi renamable $x9, 0 :: (load 8) 51 # CHECK-NEXT: renamable $x8 = ADDXrr $x8, $x8 52 # CHECK-NEXT: STPXi renamable $x8, killed $x10, renamable $x0, 10 :: (store 8, align 4) [all …]
|
D | fold-global-offsets.ll | 9 ; CHECK: adrp x8, x1+16 10 ; CHECK: ldr x0, [x8, :lo12:x1+16] 17 ; CHECK: adrp x8, x1 18 ; CHECK: add x8, x8, :lo12:x1 19 ; CHECK: ldr x0, [x8, #24] 26 ; CHECK: adrp x8, x1+1 27 ; CHECK: add x8, x8, :lo12:x1+1 28 ; CHECK: ldr x0, [x8] 35 ; CHECK: adrp x8, x2+8 36 ; CHECK: add x8, x8, :lo12:x2+8 [all …]
|
D | urem-seteq-vec-nonsplat.ll | 8 ; CHECK-NEXT: adrp x8, .LCPI0_0 9 ; CHECK-NEXT: ldr q1, [x8, :lo12:.LCPI0_0] 10 ; CHECK-NEXT: adrp x8, .LCPI0_1 11 ; CHECK-NEXT: ldr q2, [x8, :lo12:.LCPI0_1] 12 ; CHECK-NEXT: adrp x8, .LCPI0_2 13 ; CHECK-NEXT: ldr q3, [x8, :lo12:.LCPI0_2] 15 ; CHECK-NEXT: adrp x8, .LCPI0_3 19 ; CHECK-NEXT: ldr q2, [x8, :lo12:.LCPI0_3] 40 ; CHECK-NEXT: adrp x8, .LCPI1_0 42 ; CHECK-NEXT: ldr q1, [x8, :lo12:.LCPI1_0] [all …]
|
D | sdivpow2.ll | 44 ; CHECK-NEXT: add x8, x0, #7 // =7 46 ; CHECK-NEXT: csel x8, x8, x0, lt 47 ; CHECK-NEXT: asr x0, x8, #3 56 ; CHECK-NEXT: add x8, x0, #7 // =7 58 ; CHECK-NEXT: csel x8, x8, x0, lt 59 ; CHECK-NEXT: neg x0, x8, asr #3 68 ; CHECK-NEXT: add x8, x0, #63 // =63 70 ; CHECK-NEXT: csel x8, x8, x0, lt 71 ; CHECK-NEXT: asr x0, x8, #6 80 ; CHECK-NEXT: mov x8, #281474976710655 [all …]
|
D | srem-seteq-vec-nonsplat.ll | 8 ; CHECK-NEXT: adrp x8, .LCPI0_0 9 ; CHECK-NEXT: ldr q1, [x8, :lo12:.LCPI0_0] 10 ; CHECK-NEXT: adrp x8, .LCPI0_1 11 ; CHECK-NEXT: ldr q2, [x8, :lo12:.LCPI0_1] 12 ; CHECK-NEXT: adrp x8, .LCPI0_2 16 ; CHECK-NEXT: ldr q3, [x8, :lo12:.LCPI0_2] 17 ; CHECK-NEXT: adrp x8, .LCPI0_3 20 ; CHECK-NEXT: ldr q2, [x8, :lo12:.LCPI0_3] 85 ; CHECK-NEXT: adrp x8, .LCPI3_0 86 ; CHECK-NEXT: ldr q1, [x8, :lo12:.LCPI3_0] [all …]
|
/external/libhevc/common/arm64/ |
D | ihevc_intra_pred_luma_mode_18_34.s | 130 mov x8,x0 134 ld1 {v0.8b},[x8],x6 136 ld1 {v1.8b},[x8],x6 138 ld1 {v2.8b},[x8],x6 139 ld1 {v3.8b},[x8],x6 141 ld1 {v4.8b},[x8],x6 142 ld1 {v5.8b},[x8],x6 143 ld1 {v6.8b},[x8],x6 145 ld1 {v7.8b},[x8],x6 154 csel x8, x0, x8,ne [all …]
|
D | ihevc_sao_edge_offset_class3.s | 92 MOV x17,x8 // ht 103 SUB x10,x8,#1 //ht-1 160 SUB x11,x8,#1 //ht - 1 210 MOV x12,x8 //Move ht 244 CMP x8,#4 //Compare ht with 4 255 csel x8, x20, x8,NE 265 CMP x8,#0 269 csel x8, x20, x8,EQ 271 csel x8, x3, x8,NE 275 ADD x8,x8,#1 //pu1_src - src_strd + 1 [all …]
|
D | ihevc_sao_edge_offset_class2_chroma.s | 75 ldr x8,[sp,#0] 93 mov x22,x8 // *pi1_sao_offset_u 0x38 117 SUB x10,x8,#1 //ht-1 218 SUB x11,x8,#1 //ht - 1 307 MOV x12,x8 //Move ht 341 CMP x8,#4 //Compare ht with 4 352 csel x8, x20, x8,NE 372 csel x8, x20, x8,EQ 375 csel x8, x3, x8,NE //pu1_src_top_cpy 377 SUB x8,x8,#2 //pu1_src - src_strd - 2 [all …]
|
/external/llvm-project/lld/test/ELF/ |
D | aarch64-tlsld-ldst.s | 10 _start: mrs x8, TPIDR_EL0 12 add x8, x8, :tprel_hi12:var0 13 ldr q20, [x8, :tprel_lo12_nc:var0] 15 add x8, x8, :tprel_hi12:var1 16 ldr x0, [x8, :tprel_lo12_nc:var1] 18 add x8, x8, :tprel_hi12:var2 19 ldr w0, [x8, :tprel_lo12_nc:var2] 21 add x8, x8, :tprel_hi12:var3 22 ldrh w0, [x8, :tprel_lo12_nc:var3] 24 add x8, x8, :tprel_hi12:var4 [all …]
|
/external/tcpdump/tests/ |
D | telnet-iac-check-oobr.out | 2 …x8 0xa 0x7 0x1d 0xa8 0xd4 0x59 0x3c 0x88 0xa8 0xff 0xfe 0x24 0xff 0xfa 0x18 0x1 SE, SB 0x7b 0xf5 0…
|
/external/llvm-project/lldb/tools/lldb-vscode/syntaxes/ |
D | arm64.disasm | 7 liblog.so[0x6024] <+16>: adrp x8, 15 8 liblog.so[0x6028] <+20>: ldr x8, [x8, #0x230] 9 liblog.so[0x602c] <+24>: ldr w8, [x8] 20 liblog.so[0x6058] <+68>: ldr x8, [x22] 21 liblog.so[0x605c] <+72>: cmp x22, x8 23 liblog.so[0x6064] <+80>: ldr x8, [x22, #0x18] 24 liblog.so[0x6068] <+84>: cbz x8, 0x60d0 ; <+188> at config_read.cpp 26 liblog.so[0x6070] <+92>: blr x8 28 liblog.so[0x6078] <+100>: ldr x8, [x21, #0x18] 29 liblog.so[0x607c] <+104>: cbz x8, 0x60c0 ; <+172> [inlined] __android_log_add_tra… [all …]
|
/external/llvm/test/CodeGen/AArch64/ |
D | arm64-fast-isel-intrinsic.ll | 8 ; ARM64: adrp x8, _message@PAGE 9 ; ARM64: add x0, x8, _message@PAGEOFF 22 ; ARM64: adrp x8, _temp@GOTPAGE 23 ; ARM64: ldr x0, [x8, _temp@GOTPAGEOFF] 24 ; ARM64: adrp x8, _message@PAGE 25 ; ARM64: add x1, x8, _message@PAGEOFF 36 ; ARM64: adrp x8, _temp@GOTPAGE 37 ; ARM64: ldr x0, [x8, _temp@GOTPAGEOFF] 38 ; ARM64: adrp x8, _message@PAGE 39 ; ARM64: add x1, x8, _message@PAGEOFF [all …]
|
/external/llvm/test/MC/AArch64/ |
D | arm64-leaf-compact-unwind.s | 87 mov x8, sp 90 str w9, [x8, x9, lsl #2] 99 ldr w10, [x8, x9] 137 adrp x8, _bar@GOTPAGE 139 ldr x8, [x8, _bar@GOTPAGEOFF] 140 ldr w9, [x8] 141 ldr w10, [x8] 142 ldr w11, [x8] 143 ldr w12, [x8] 144 ldr w13, [x8] [all …]
|
/external/llvm-project/llvm/test/MC/AArch64/ |
D | arm64-leaf-compact-unwind.s | 87 mov x8, sp 90 str w9, [x8, x9, lsl #2] 99 ldr w10, [x8, x9] 137 adrp x8, _bar@GOTPAGE 139 ldr x8, [x8, _bar@GOTPAGEOFF] 140 ldr w9, [x8] 141 ldr w10, [x8] 142 ldr w11, [x8] 143 ldr w12, [x8] 144 ldr w13, [x8] [all …]
|
/external/llvm-project/lld/test/COFF/ |
D | arm64-relocs-imports.test | 14 # BEFORE: 10: 00 01 40 39 ldrb w0, [x8] 15 # BEFORE: 14: 00 01 40 79 ldrh w0, [x8] 16 # BEFORE: 18: 00 01 40 b9 ldr w0, [x8] 17 # BEFORE: 1c: 00 01 40 f9 ldr x0, [x8] 18 # BEFORE: 20: 00 01 00 39 strb w0, [x8] 19 # BEFORE: 24: 00 01 00 79 strh w0, [x8] 20 # BEFORE: 28: 00 01 00 b9 str w0, [x8] 21 # BEFORE: 2c: 00 01 00 f9 str x0, [x8] 22 # BEFORE: 30: 00 01 40 3d ldr b0, [x8] 23 # BEFORE: 34: 00 01 40 7d ldr h0, [x8] [all …]
|
/external/boringssl/ios-aarch64/crypto/third_party/sike/asm/ |
D | fp-armv8.S | 43 ldp x7, x8, [x0,#32] 52 adcs x4, x4, x8 62 adc x8, xzr, xzr 65 and x9, x7, x8 68 sub x8, xzr, x8 77 and x20, x3, x8 78 and x21, x4, x8 79 and x22, x5, x8 80 and x23, x6, x8 104 and x8, x23, x19 [all …]
|
/external/libaom/libaom/av1/encoder/x86/ |
D | av1_fwd_txfm1d_sse4.c | 1080 __m128i x8[64]; in av1_fdct64_sse4_1() local 1081 x8[0] = x7[0]; in av1_fdct64_sse4_1() 1082 x8[1] = x7[1]; in av1_fdct64_sse4_1() 1083 x8[2] = x7[2]; in av1_fdct64_sse4_1() 1084 x8[3] = x7[3]; in av1_fdct64_sse4_1() 1085 x8[4] = x7[4]; in av1_fdct64_sse4_1() 1086 x8[5] = x7[5]; in av1_fdct64_sse4_1() 1087 x8[6] = x7[6]; in av1_fdct64_sse4_1() 1088 x8[7] = x7[7]; in av1_fdct64_sse4_1() 1089 btf_32_type1_sse4_1_new(cospi_p60, cospi_p04, x7[8], x7[15], x8[8], x8[15], in av1_fdct64_sse4_1() [all …]
|
/external/rust/crates/ring/pregenerated/ |
D | ecp_nistz256-armv8-ios64.S | 82 ldp x8,x9,[x2] 123 umulh x8,x4,x3 135 adds x15,x15,x8 // accumulate high parts of multiplication 136 lsl x8,x14,#32 142 subs x10,x14,x8 // "*0xffff0001" 144 adds x14,x15,x8 // +=acc[0]<<96 and omit acc[0] 145 mul x8,x4,x3 // lo(a[0]*b[i]) 154 adds x14,x14,x8 // accumulate low parts of multiplication 155 umulh x8,x4,x3 // hi(a[0]*b[i]) 164 adds x15,x15,x8 // accumulate high parts of multiplication [all …]
|