/external/llvm/test/Transforms/IndVarSimplify/ |
D | loop-invariant-conditions.ll | 13 ; CHECK: %cmp1 = icmp slt i64 %start, -1 14 %cmp1 = icmp slt i64 %indvars.iv, -1 15 br i1 %cmp1, label %for.end, label %loop 29 ; CHECK: %cmp1 = icmp sle i64 %start, -1 30 %cmp1 = icmp sle i64 %indvars.iv, -1 31 br i1 %cmp1, label %for.end, label %loop 52 ; CHECK: %cmp1 = icmp slt i64 %start, -1 53 %cmp1 = icmp slt i64 %indvars.iv, -1 54 br i1 %cmp1, label %for.end, label %loop 74 ; CHECK: %cmp1 = icmp sgt i64 %start, -1 [all …]
|
/external/llvm/test/CodeGen/PowerPC/ |
D | select-i1-vs-i1.ll | 11 %cmp1 = icmp eq i32 %c3, %c4 13 %cmp3 = icmp slt i1 %cmp3tmp, %cmp1 27 %cmp1 = icmp eq i32 %c3, %c4 29 %cmp3 = icmp ult i1 %cmp3tmp, %cmp1 43 %cmp1 = icmp eq i32 %c3, %c4 45 %cmp3 = icmp sle i1 %cmp3tmp, %cmp1 59 %cmp1 = icmp eq i32 %c3, %c4 61 %cmp3 = icmp ule i1 %cmp3tmp, %cmp1 75 %cmp1 = icmp eq i32 %c3, %c4 77 %cmp3 = icmp eq i1 %cmp3tmp, %cmp1 [all …]
|
/external/llvm/test/CodeGen/X86/ |
D | setcc-combine.ll | 11 %cmp1 = icmp eq <4 x i32> %sext, zeroinitializer 12 %0 = extractelement <4 x i1> %cmp1, i32 1 25 %cmp1 = icmp ne <4 x i32> %sext, zeroinitializer 26 %0 = extractelement <4 x i1> %cmp1, i32 1 38 %cmp1 = icmp sle <4 x i32> %sext, zeroinitializer 39 %0 = extractelement <4 x i1> %cmp1, i32 1 52 %cmp1 = icmp sge <4 x i32> %sext, zeroinitializer 53 %0 = extractelement <4 x i1> %cmp1, i32 1 66 %cmp1 = icmp slt <4 x i32> %sext, zeroinitializer 79 %cmp1 = icmp sgt <4 x i32> %sext, zeroinitializer [all …]
|
D | setcc-sentinals.ll | 9 %cmp1 = icmp ne i64 %x, -1 11 %.cmp1 = and i1 %cmp1, %not.cmp 12 ret i1 %.cmp1
|
D | cmov-double.ll | 15 %cmp1 = fcmp olt float %a2, %a3 16 %and = and i1 %cmp0, %cmp1 31 %cmp1 = fcmp olt float %a2, %a3 32 %and = or i1 %cmp0, %cmp1 46 %cmp1 = icmp ult i32 %a1, %a2 47 %or = or i1 %cmp0, %cmp1
|
D | vsplit-and.ll | 7 %cmp1 = icmp ne <2 x i64> %src1, zeroinitializer 9 %t1 = and <2 x i1> %cmp1, %cmp2 19 %cmp1 = icmp ne <3 x i64> %src1, zeroinitializer 21 %t1 = and <3 x i1> %cmp1, %cmp2
|
/external/llvm/test/Transforms/InstCombine/ |
D | abs_abs.ll | 7 %cmp1 = icmp sgt i32 %cond, -1 9 %cond18 = select i1 %cmp1, i32 %cond, i32 %sub16 22 %cmp1 = icmp sgt i32 %cond, -1 24 %cond18 = select i1 %cmp1, i32 %cond, i32 %sub16 37 %cmp1 = icmp sgt i32 %cond, -1 39 %cond18 = select i1 %cmp1, i32 %cond, i32 %sub16 52 %cmp1 = icmp sgt i32 %cond, -1 54 %cond18 = select i1 %cmp1, i32 %cond, i32 %sub16 67 %cmp1 = icmp sgt i32 %cond, 0 69 %cond18 = select i1 %cmp1, i32 %cond, i32 %sub16 [all …]
|
D | merge-icmp.ll | 6 %cmp1 = icmp eq i8 %trunc, 127 9 %or = and i1 %cmp1, %cmp2 20 %cmp1 = icmp eq i16 %and, 32512 23 %or = and i1 %cmp1, %cmp2
|
D | and2.ll | 51 %cmp1 = icmp ne i64 %x, -1 53 %.cmp1 = and i1 %cmp1, %not.cmp 54 %land.ext = zext i1 %.cmp1 to i32 63 %cmp1 = icmp slt i32 %i, 1 65 %and1 = and i1 %cmp1, %b 75 %cmp1 = icmp ne i32 %i, 0 77 %cond = and i1 %cmp1, %cmp2
|
/external/llvm/test/CodeGen/ARM/ |
D | vminmaxnm.ll | 155 %cmp1 = fcmp fast olt float %a, 12. 156 %cond1 = select i1 %cmp1, float %a, float 12. 166 %cmp1 = fcmp fast ole double %a, 34. 167 %cond1 = select i1 %cmp1, double %a, double 34. 177 %cmp1 = fcmp fast ogt float %a, 56. 178 %cond1 = select i1 %cmp1, float 56., float %a 188 %cmp1 = fcmp fast oge double %a, 78. 189 %cond1 = select i1 %cmp1, double 78., double %a 199 %cmp1 = fcmp fast ult float 12., %b 200 %cond1 = select i1 %cmp1, float 12., float %b [all …]
|
D | vminmaxnm-safe.ll | 177 %cmp1 = fcmp olt float %a, 12. 178 %cond1 = select i1 %cmp1, float %a, float 12. 188 %cmp1 = fcmp ole double %a, 34. 189 %cond1 = select i1 %cmp1, double %a, double 34. 199 %cmp1 = fcmp ogt float %a, 56. 200 %cond1 = select i1 %cmp1, float 56., float %a 210 %cmp1 = fcmp oge double %a, 78. 211 %cond1 = select i1 %cmp1, double 78., double %a 221 %cmp1 = fcmp ult float 12., %b 222 %cond1 = select i1 %cmp1, float 12., float %b [all …]
|
D | setcc-sentinals.ll | 10 %cmp1 = icmp ne i32 %x, -1 12 %.cmp1 = and i1 %cmp1, %not.cmp 13 ret i1 %.cmp1
|
D | movcc-double.ll | 14 %cmp1 = icmp ult i32 %a2, %a3 15 %and = and i1 %cmp0, %cmp1 29 %cmp1 = icmp ult i32 %a2, %a3 30 %and = or i1 %cmp0, %cmp1 44 %cmp1 = icmp ult i32 %a1, %a2 45 %or = or i1 %cmp0, %cmp1
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | highbd_quantize_intrin_sse2.c | 54 __m128i coeffs, cmp1, cmp2; in vpx_highbd_quantize_b_sse2() local 57 cmp1 = _mm_cmplt_epi32(coeffs, zbins[i != 0]); in vpx_highbd_quantize_b_sse2() 59 cmp1 = _mm_and_si128(cmp1, cmp2); in vpx_highbd_quantize_b_sse2() 60 test = _mm_movemask_epi8(cmp1); in vpx_highbd_quantize_b_sse2() 142 __m128i coeffs, cmp1, cmp2; in vpx_highbd_quantize_b_32x32_sse2() local 145 cmp1 = _mm_cmplt_epi32(coeffs, zbins[i != 0]); in vpx_highbd_quantize_b_32x32_sse2() 147 cmp1 = _mm_and_si128(cmp1, cmp2); in vpx_highbd_quantize_b_32x32_sse2() 148 test = _mm_movemask_epi8(cmp1); in vpx_highbd_quantize_b_32x32_sse2()
|
/external/llvm/test/Transforms/JumpThreading/ |
D | phi-known.ll | 15 %cmp1 = icmp eq i8* %p1, null 16 br i1 %cmp1, label %exit, label %loop 29 %cmp1 = icmp eq i8* %p1, null 30 br i1 %cmp1, label %exit, label %backedge 58 ; CHECK-NEXT: %cmp1 = icmp 59 ; CHECK-NEXT: br i1 %cmp1 61 %cmp1 = icmp ne i8* %p1, null 62 br i1 %cmp1, label %exit, label %loop
|
/external/llvm/test/Analysis/ValueTracking/ |
D | monotonic-phi.ll | 10 %cmp1 = icmp eq i8 %A, %n 11 br i1 %cmp1, label %exit, label %loop 26 %cmp1 = icmp eq i8 %A, %n 27 br i1 %cmp1, label %exit, label %loop 42 %cmp1 = icmp eq i8 %A, %n 43 br i1 %cmp1, label %exit, label %loop
|
/external/llvm/test/CodeGen/Mips/ |
D | selpat.ll | 46 %cmp1 = icmp eq i32 %0, 10 47 %cond5 = select i1 %cmp1, i32 %2, i32 %1 73 %cmp1 = icmp eq i32 %3, 0 74 %cond5 = select i1 %cmp1, i32 %2, i32 %1 96 %cmp1 = icmp sge i32 %1, %0 97 %cond5 = select i1 %cmp1, i32 %3, i32 %2 121 %cmp1 = icmp sgt i32 %1, %0 122 %cond5 = select i1 %cmp1, i32 %3, i32 %2 146 %cmp1 = icmp sle i32 %1, %0 147 %cond5 = select i1 %cmp1, i32 %3, i32 %2 [all …]
|
D | mips16fpe.ll | 266 %cmp1 = fcmp oge float %2, %0 267 %and3 = and i1 %cmp, %cmp1 281 %cmp1 = fcmp oge double %2, %0 282 %and3 = and i1 %cmp, %cmp1 296 %cmp1 = fcmp olt float %0, %2 297 %and2 = and i1 %lnot, %cmp1 312 %cmp1 = fcmp olt double %0, %2 313 %and2 = and i1 %lnot, %cmp1 328 %cmp1 = fcmp ole float %0, %2 329 %and3 = and i1 %cmp, %cmp1 [all …]
|
/external/libavc/common/x86/ |
D | ih264_resi_trans_quant_sse42.c | 114 __m128i sum0, sum1, sum2, cmp0, cmp1; in ih264_resi_trans_quant_4x4_sse42() local 290 cmp1 = _mm_cmpeq_epi16(temp2, zero_8x16b); in ih264_resi_trans_quant_4x4_sse42() 293 mask1 = _mm_movemask_epi8(cmp1); in ih264_resi_trans_quant_4x4_sse42() 314 cmp1 = _mm_and_si128(temp_1, cmp1); in ih264_resi_trans_quant_4x4_sse42() 315 sum0 = _mm_hadd_epi16(cmp1, zero_8x16b); in ih264_resi_trans_quant_4x4_sse42() 388 __m128i cmp0, cmp1, sum0, sum1, sum2; in ih264_resi_trans_quant_chroma_4x4_sse42() local 575 cmp1 = _mm_cmpeq_epi16(temp2, zero_8x16b); in ih264_resi_trans_quant_chroma_4x4_sse42() 578 mask1 = _mm_movemask_epi8(cmp1); in ih264_resi_trans_quant_chroma_4x4_sse42() 599 cmp1 = _mm_and_si128(temp_1, cmp1); in ih264_resi_trans_quant_chroma_4x4_sse42() 600 sum0 = _mm_hadd_epi16(cmp1, zero_8x16b); in ih264_resi_trans_quant_chroma_4x4_sse42() [all …]
|
/external/llvm/test/Transforms/GVN/ |
D | commute.ll | 18 %cmp1 = icmp ult i32 %x, %y 20 call void @vse(i1 %cmp1, i1 %cmp2) 21 ; CHECK: @vse(i1 %cmp1, i1 %cmp1)
|
/external/llvm/test/Transforms/LoopUnswitch/ |
D | 2015-06-17-Metadata.ll | 16 %cmp1 = icmp eq i32 %a, 12345 17 br i1 %cmp1, label %if.then, label %if.else, !prof !0 18 ; CHECK: %cmp1 = icmp eq i32 %a, 12345 19 ; CHECK-NEXT: br i1 %cmp1, label %if.then.us, label %if.else, !prof !0 55 ;CHECK-NEXT: %cmp1 = icmp eq i32 1, 2 56 ;CHECK-NEXT: br i1 %cmp1, label %for.body, label %for.cond.cleanup.split, !prof !1 62 %cmp1 = icmp eq i32 1, 2 63 br i1 %cmp1, label %if.then, label %for.cond.cleanup, !prof !0
|
/external/llvm/test/Transforms/SimplifyCFG/ |
D | clamp.ll | 6 ; CHECK: %cmp1 = fcmp olt float %a, %b 7 ; CHECK: %cond = select i1 %cmp1, float %b, float %a 15 %cmp1 = fcmp olt float %a, %b 16 %cond = select i1 %cmp1, float %b, float %a
|
/external/llvm/test/CodeGen/AArch64/ |
D | arm64-ccmp.ll | 14 %cmp1 = icmp eq i32 %b, 17 15 %or.cond = or i1 %cmp, %cmp1 37 %cmp1 = icmp ne i32 %b, 17 38 %or.cond = or i1 %cmp, %cmp1 61 %cmp1 = icmp slt i32 %b, 7 64 %cond = select i1 %cmp1, i32 %mul, i32 %add 88 %cmp1 = icmp slt i32 %b, 7 91 %cond = select i1 %cmp1, i32 %mul, i32 %add 122 %cmp1 = icmp slt i32 %div, 17 123 br i1 %cmp1, label %if.then, label %if.end [all …]
|
/external/llvm/test/CodeGen/AMDGPU/ |
D | setcc-equivalent.ll | 7 %cmp1 = icmp eq i32 %a, -1 9 %and = and i1 %cmp1, %cmp2 24 %cmp1 = icmp eq <4 x i32> %a, <i32 -1, i32 -1, i32 -1, i32 -1> 26 %and = and <4 x i1> %cmp1, %cmp2
|
/external/llvm/test/Transforms/LoopVectorize/ |
D | runtime-check-address-space.ll | 34 %cmp1 = icmp slt i32 0, %n 35 br i1 %cmp1, label %for.body, label %for.end 61 %cmp1 = icmp slt i32 0, %n 62 br i1 %cmp1, label %for.body, label %for.end 88 %cmp1 = icmp slt i32 0, %n 89 br i1 %cmp1, label %for.body, label %for.end 116 %cmp1 = icmp slt i32 0, %n 117 br i1 %cmp1, label %for.body, label %for.end 144 %cmp1 = icmp slt i32 0, %n 145 br i1 %cmp1, label %for.body, label %for.end [all …]
|