/external/llvm/test/MC/AMDGPU/ |
D | reg-syntax-extra.s | 10 s_mov_b64 [ttmp4,ttmp5], [ttmp2,ttmp3] label 14 s_mov_b64 ttmp[4:5], ttmp[2:3] label 18 s_mov_b64 [s6,s7], s[8:9] label 22 s_mov_b64 s[6:7], [s8,s9] label 26 s_mov_b64 [exec_lo,exec_hi], s[2:3] label 30 s_mov_b64 [flat_scratch_lo,flat_scratch_hi], s[2:3] label 34 s_mov_b64 [vcc_lo,vcc_hi], s[2:3] label 38 s_mov_b64 [tba_lo,tba_hi], s[2:3] label 42 s_mov_b64 [tma_lo,tma_hi], s[2:3] label
|
D | trap.s | 101 s_mov_b64 ttmp[4:5], exec label 105 s_mov_b64 [ttmp4,ttmp5], exec label 109 s_mov_b64 exec, [ttmp4,ttmp5] label 113 s_mov_b64 tba, ttmp[4:5] label 117 s_mov_b64 ttmp[4:5], tba label 121 s_mov_b64 tma, ttmp[4:5] label 125 s_mov_b64 ttmp[4:5], tma label
|
D | sop1-err.s | 23 s_mov_b64 s1, s[0:1] label 26 s_mov_b64 s[0:1], s1 label 34 s_mov_b64 s[0:1], 0xfffffffff label 37 s_mov_b64 s[0:1], 0x0000000200000000 label 53 s_mov_b64 s[102:103], -1 label
|
D | sop1.s | 28 s_mov_b64 s[2:3], s[4:5] label 32 s_mov_b64 s[2:3], 0xffffffffffffffff label 36 s_mov_b64 s[2:3], 0xffffffff label 40 s_mov_b64 s[0:1], 0x80000000 label 44 s_mov_b64 s[102:103], -1 label
|
D | flat-scratch.s | 7 s_mov_b64 flat_scratch, -1 label 23 s_mov_b64 flat_scratch_lo, -1 label 28 s_mov_b64 flat_scratch_hi, -1 label
|
D | out-of-range-registers.s | 16 s_mov_b64 s[0:17], -1 label 19 s_mov_b64 s[103:104], -1 label 22 s_mov_b64 s[104:105], -1 label
|
/external/llvm/test/CodeGen/AMDGPU/ |
D | write_register.ll | 15 ; CHECK: s_mov_b64 exec, 0 16 ; CHECK: s_mov_b64 exec, -1 17 ; CHECK: s_mov_b64 exec, s{{\[[0-9]+:[0-9]+\]}} 26 ; CHECK: s_mov_b64 flat_scratch, 0 27 ; CHECK: s_mov_b64 flat_scratch, -1 28 ; CHECK: s_mov_b64 flat_scratch, s{{\[[0-9]+:[0-9]+\]}}
|
D | wqm.ll | 37 ;CHECK-NEXT: s_mov_b64 [[ORIG:s\[[0-9]+:[0-9]+\]]], exec 59 ;CHECK-NEXT: s_mov_b64 [[ORIG:s\[[0-9]+:[0-9]+\]]], exec 84 ;CHECK-NEXT: s_mov_b64 [[ORIG:s\[[0-9]+:[0-9]+\]]], exec 89 ;CHECK: s_mov_b64 exec, [[SAVED]] 116 ;CHECK-NEXT: s_mov_b64 [[ORIG:s\[[0-9]+:[0-9]+\]]], exec 156 ;CHECK-NEXT: s_mov_b64 [[ORIG:s\[[0-9]+:[0-9]+\]]], exec 205 ;CHECK-NEXT: s_mov_b64 [[ORIG:s\[[0-9]+:[0-9]+\]]], exec 252 ;CHECK-NEXT: s_mov_b64 [[ORIG:s\[[0-9]+:[0-9]+\]]], exec 258 ;CHECK: s_mov_b64 exec, [[SAVE]] 281 ;CHECK-NEXT: s_mov_b64 [[ORIG:s\[[0-9]+:[0-9]+\]]], exec [all …]
|
D | smrd-vccz-bug.ll | 10 ; VCCZ-BUG: s_mov_b64 vcc, vcc 11 ; NOVCCZ-BUG-NOT: s_mov_b64 vcc, vcc
|
D | valu-i1.ll | 8 ; SI-NOT: s_mov_b64 s[{{[0-9]:[0-9]}}], -1 77 ; SI: s_mov_b64 {{s\[[0-9]+:[0-9]+\]}}, 0{{$}} 122 ; SI: s_mov_b64 [[ZERO:s\[[0-9]+:[0-9]+\]]], 0{{$}} 123 ; SI: s_mov_b64 [[COND_STATE:s\[[0-9]+:[0-9]+\]]], [[ZERO]]
|
D | llvm.amdgcn.ps.live.ll | 18 ; CHECK: s_mov_b64 [[LIVE:s\[[0-9]+:[0-9]+\]]], exec 33 ; CHECK: s_mov_b64 [[LIVE:s\[[0-9]+:[0-9]+\]]], exec
|
D | llvm.amdgpu.kilp.ll | 5 ; SI: s_mov_b64 exec, 0
|
D | skip-if-dead.ll | 13 ; CHECK-NEXT: s_mov_b64 exec, 0 24 ; CHECK-NEXT: s_mov_b64 exec, 0 26 ; CHECK-NEXT: s_mov_b64 exec, 0
|
D | indirect-addressing-si.ll | 210 ; CHECK: s_mov_b64 [[MASK:s\[[0-9]+:[0-9]+\]]], exec 222 ; CHECK: s_mov_b64 exec, [[MASK]] 223 ; CHECK: s_mov_b64 [[MASK2:s\[[0-9]+:[0-9]+\]]], exec 266 ; CHECK: s_mov_b64 [[MASK:s\[[0-9]+:[0-9]+\]]], exec 278 ; CHECK: s_mov_b64 exec, [[MASK]] 280 ; CHECK: s_mov_b64 [[MASK]], exec
|
D | llvm.AMDGPU.kill.ll | 6 ; SI: s_mov_b64 exec, 0
|
D | llvm.amdgcn.div.fmas.ll | 90 ; SI: s_mov_b64 vcc, 0 99 ; SI: s_mov_b64 vcc, -1
|
D | operand-folding.ll | 40 ; CHECK-NOT: s_mov_b64
|
D | si-annotate-cf.ll | 31 ; SI: s_mov_b64 [[ZERO:s\[[0-9]+:[0-9]+\]]], 0
|
D | salu-to-valu.ll | 17 ; Make sure we aren't using VGPRs for the source operand of s_mov_b64 18 ; GCN-NOT: s_mov_b64 s[{{[0-9]+:[0-9]+}}], v
|
/external/llvm/test/MC/Disassembler/AMDGPU/ |
D | trap_vi.txt | 77 # VI: s_mov_b64 ttmp[4:5], exec ; encoding: [0x7e,0x01,0xf4,0xbe] 80 # VI: s_mov_b64 ttmp[4:5], exec ; encoding: [0x7e,0x01,0xf4,0xbe] 83 # VI: s_mov_b64 exec, ttmp[4:5] ; encoding: [0x74,0x01,0xfe,0xbe] 86 # VI: s_mov_b64 tba, ttmp[4:5] ; encoding: [0x74,0x01,0xec,0xbe] 89 # VI: s_mov_b64 ttmp[4:5], tba ; encoding: [0x6c,0x01,0xf4,0xbe] 92 # VI: s_mov_b64 tma, ttmp[4:5] ; encoding: [0x74,0x01,0xee,0xbe] 95 # VI: s_mov_b64 ttmp[4:5], tma ; encoding: [0x6e,0x01,0xf4,0xbe]
|
D | sop1_vi.txt | 18 # VI: s_mov_b64 s[2:3], s[4:5] ; encoding: [0x04,0x01,0x82,0xbe] 21 # FIXME: s_mov_b64 s[2:3], -1 ; encoding: [0xc1,0x01,0x82,0xbe] 24 # VI: s_mov_b64 s[2:3], 0xffffffff ; encoding: [0xff,0x01,0x82,0xbe,0xff,0xff,0xff,0xff] 27 # VI: s_mov_b64 s[0:1], 0x80000000 ; encoding: [0xff,0x01,0x80,0xbe,0x00,0x00,0x00,0x80]
|
/external/clang/test/CodeGenOpenCL/ |
D | amdgcn-flat-scratch-name.cl | 7 // CHECK: tail call void asm sideeffect "s_mov_b64 flat_scratch, 0", "~{flat_scratch}"() 8 __asm__ volatile("s_mov_b64 flat_scratch, 0" : : : "flat_scratch");
|
/external/llvm/test/Object/AMDGPU/ |
D | objdump.s | 20 s_mov_b64 s[2:3], exec 21 s_mov_b64 s[10:11], exec
|
/external/llvm/lib/Target/AMDGPU/ |
D | SIRegisterInfo.td | 283 // Requires 2 s_mov_b64 to copy 301 // Requires 4 s_mov_b64 to copy 307 // Requires 8 s_mov_b64 to copy
|
D | SIInstructions.td | 93 defm S_MOV_B64 : SOP1_64 <sop1<0x04, 0x01>, "s_mov_b64", []>;
|