/external/llvm-project/llvm/test/Analysis/CostModel/SystemZ/ |
D | scalar-cmp-cmp-log-sel.ll | 9 %cmp0 = icmp eq i8 %val1, %val2 11 %and = and i1 %cmp0, %cmp1 16 ; CHECK: cost of 3 for instruction: %cmp0 = icmp eq i8 %val1, %val2 18 ; CHECK: cost of 1 for instruction: %and = and i1 %cmp0, %cmp1 24 %cmp0 = icmp eq i8 %val1, %val2 26 %and = and i1 %cmp0, %cmp1 31 ; CHECK: cost of 3 for instruction: %cmp0 = icmp eq i8 %val1, %val2 33 ; CHECK: cost of 1 for instruction: %and = and i1 %cmp0, %cmp1 39 %cmp0 = icmp eq i8 %val1, %val2 41 %and = and i1 %cmp0, %cmp1 [all …]
|
/external/llvm-project/llvm/test/CodeGen/AMDGPU/ |
D | sopk-compares.ll | 14 %cmp0 = icmp eq i32 %cond, 4 15 br i1 %cmp0, label %endif, label %if 30 %cmp0 = icmp eq i32 %cond, 32767 31 br i1 %cmp0, label %endif, label %if 46 %cmp0 = icmp eq i32 %cond, 32768 47 br i1 %cmp0, label %endif, label %if 62 %cmp0 = icmp ne i32 %cond, 32768 63 br i1 %cmp0, label %endif, label %if 78 %cmp0 = icmp eq i32 %cond, -32768 79 br i1 %cmp0, label %endif, label %if [all …]
|
D | uniform-cfg.ll | 17 %cmp0 = icmp eq i32 %cond, 0 18 br i1 %cmp0, label %if, label %else 45 %cmp0 = fcmp oeq float %cond, 0.0 46 br i1 %cmp0, label %if, label %else 73 %cmp0 = icmp eq i32 %cond, 0 74 br i1 %cmp0, label %else, label %if 101 %cmp0 = fcmp oeq float %cond, 0.0 102 br i1 %cmp0, label %else, label %if 264 %cmp0 = icmp sgt i32 %cond0, 0 266 br i1 %cmp0, label %bb2, label %bb9 [all …]
|
D | xfail.r600.bitcast.ll | 19 %cmp0 = icmp eq i32 %cond, 0 20 br i1 %cmp0, label %if, label %end 35 %cmp0 = icmp eq i32 %cond, 0 36 br i1 %cmp0, label %if, label %end
|
D | branch-relaxation-gfx10-branch-offset-bug.ll | 27 %cmp0 = icmp eq i32 %cnd0, 0 28 br i1 %cmp0, label %bb2, label %bb3 75 %cmp0 = icmp eq i32 %cnd0, 0 76 br i1 %cmp0, label %bb2, label %bb3
|
/external/llvm-project/llvm/test/CodeGen/SystemZ/ |
D | vec-cmp-cmp-logic-select.ll | 16 %cmp0 = icmp eq <2 x i8> %val1, %val2 18 %and = and <2 x i1> %cmp0, %cmp1 32 %cmp0 = icmp eq <2 x i8> %val1, %val2 34 %and = and <2 x i1> %cmp0, %cmp1 49 %cmp0 = icmp eq <16 x i8> %val1, %val2 51 %and = or <16 x i1> %cmp0, %cmp1 72 %cmp0 = icmp eq <16 x i8> %val1, %val2 74 %and = or <16 x i1> %cmp0, %cmp1 95 %cmp0 = icmp eq <32 x i8> %val1, %val2 97 %and = xor <32 x i1> %cmp0, %cmp1 [all …]
|
D | foldmemop-imm-01.ll | 25 %cmp0 = icmp ne i32 %tmp0, 2 26 %zxt0 = zext i1 %cmp0 to i32 50 %cmp0 = icmp ne i64 %tmp0, 2 51 %zxt0 = zext i1 %cmp0 to i64
|
/external/llvm/test/CodeGen/ARM/ |
D | movcc-double.ll | 13 %cmp0 = icmp ult i32 %a0, %a1 15 %and = and i1 %cmp0, %cmp1 28 %cmp0 = icmp ult i32 %a0, %a1 30 %and = or i1 %cmp0, %cmp1 43 %cmp0 = icmp ult i32 %a0, %a1 45 %or = or i1 %cmp0, %cmp1
|
/external/llvm-project/llvm/test/CodeGen/ARM/ |
D | movcc-double.ll | 13 %cmp0 = icmp ult i32 %a0, %a1 15 %and = and i1 %cmp0, %cmp1 28 %cmp0 = icmp ult i32 %a0, %a1 30 %and = or i1 %cmp0, %cmp1 43 %cmp0 = icmp ult i32 %a0, %a1 45 %or = or i1 %cmp0, %cmp1
|
/external/llvm/test/CodeGen/X86/ |
D | cmov-double.ll | 14 %cmp0 = icmp ult i32 %a0, %a1 16 %and = and i1 %cmp0, %cmp1 30 %cmp0 = icmp ult i32 %a0, %a1 32 %and = or i1 %cmp0, %cmp1 45 %cmp0 = icmp ult i32 %a0, %a1 47 %or = or i1 %cmp0, %cmp1
|
/external/llvm/test/Transforms/JumpThreading/ |
D | phi-known.ll | 8 %cmp0 = icmp eq i8* %p, null 9 br i1 %cmp0, label %exit, label %loop 25 %cmp0 = icmp eq i8* %p, null 26 br i1 %cmp0, label %exit, label %loop 53 %cmp0 = icmp eq i8* %p, null 54 br i1 %cmp0, label %exit, label %loop
|
/external/libavc/common/x86/ |
D | ih264_resi_trans_quant_sse42.c | 114 __m128i sum0, sum1, sum2, cmp0, cmp1; in ih264_resi_trans_quant_4x4_sse42() local 289 cmp0 = _mm_cmpeq_epi16(temp0, zero_8x16b); in ih264_resi_trans_quant_4x4_sse42() 292 mask0 = _mm_movemask_epi8(cmp0); in ih264_resi_trans_quant_4x4_sse42() 301 cmp0 = _mm_and_si128(temp_1, cmp0); in ih264_resi_trans_quant_4x4_sse42() 302 sum0 = _mm_hadd_epi16(cmp0, zero_8x16b); in ih264_resi_trans_quant_4x4_sse42() 388 __m128i cmp0, cmp1, sum0, sum1, sum2; in ih264_resi_trans_quant_chroma_4x4_sse42() local 574 cmp0 = _mm_cmpeq_epi16(temp0, zero_8x16b); in ih264_resi_trans_quant_chroma_4x4_sse42() 577 mask0 = _mm_movemask_epi8(cmp0); in ih264_resi_trans_quant_chroma_4x4_sse42() 586 cmp0 = _mm_and_si128(temp_1, cmp0); in ih264_resi_trans_quant_chroma_4x4_sse42() 587 sum0 = _mm_hadd_epi16(cmp0, zero_8x16b); in ih264_resi_trans_quant_chroma_4x4_sse42() [all …]
|
/external/libaom/libaom/aom_dsp/x86/ |
D | fwd_txfm_sse2.h | 38 __m128i cmp0 = _mm_or_si128(_mm_cmpeq_epi16(*preg0, max_overflow), in check_epi16_overflow_x2() local 42 cmp0 = _mm_or_si128(cmp0, cmp1); in check_epi16_overflow_x2() 43 return _mm_movemask_epi8(cmp0); in check_epi16_overflow_x2() 52 __m128i cmp0 = _mm_or_si128(_mm_cmpeq_epi16(*preg0, max_overflow), in check_epi16_overflow_x4() local 60 cmp0 = _mm_or_si128(_mm_or_si128(cmp0, cmp1), _mm_or_si128(cmp2, cmp3)); in check_epi16_overflow_x4() 61 return _mm_movemask_epi8(cmp0); in check_epi16_overflow_x4()
|
/external/llvm-project/llvm/test/CodeGen/X86/ |
D | cmov-double.ll | 17 %cmp0 = icmp ult i32 %a0, %a1 19 %and = and i1 %cmp0, %cmp1 36 %cmp0 = icmp ult i32 %a0, %a1 38 %and = or i1 %cmp0, %cmp1 61 %cmp0 = icmp ult i32 %a0, %a1 63 %or = or i1 %cmp0, %cmp1
|
/external/llvm-project/llvm/test/Transforms/JumpThreading/ |
D | phi-known.ll | 8 %cmp0 = icmp eq i8* %p, null 9 br i1 %cmp0, label %exit, label %loop 25 %cmp0 = icmp eq i8* %p, null 26 br i1 %cmp0, label %exit, label %loop 53 %cmp0 = icmp eq i8* %p, null 54 br i1 %cmp0, label %exit, label %loop
|
/external/llvm/test/Transforms/InstSimplify/ |
D | AndOrXor.ll | 260 %cmp0 = icmp eq i32 %i, 0 261 %conv0 = zext i1 %cmp0 to i32 274 %cmp0 = icmp eq <4 x i32> %i, zeroinitializer 275 %conv0 = zext <4 x i1> %cmp0 to <4 x i32> 288 %cmp0 = icmp eq i3 %i, 0 289 %conv0 = sext i1 %cmp0 to i5 302 %cmp0 = icmp sgt <3 x i65> %i, zeroinitializer 303 %conv0 = bitcast <3 x i1> %cmp0 to i3 321 %cmp0 = icmp eq i8 %i, 0 322 %conv0 = zext i1 %cmp0 to i16 [all …]
|
/external/oboe/samples/RhythmGame/third_party/glm/simd/ |
D | common.h | 102 glm_vec4 const cmp0 = _mm_cmplt_ps(x, zro0); in glm_vec4_sign() local 104 glm_vec4 const and0 = _mm_and_ps(cmp0, _mm_set1_ps(-1.0f)); in glm_vec4_sign() 130 glm_vec4 const cmp0 = _mm_cmplt_ps(x, rnd0); in glm_vec4_floor() 131 glm_vec4 const and0 = _mm_and_ps(cmp0, _mm_set1_ps(1.0f)); in glm_vec4_floor() 161 glm_vec4 const cmp0 = _mm_cmpgt_ps(x, rnd0); in glm_vec4_ceil() 162 glm_vec4 const and0 = _mm_and_ps(cmp0, _mm_set1_ps(1.0f)); in glm_vec4_ceil()
|
/external/libvpx/libvpx/vp9/encoder/x86/ |
D | vp9_highbd_block_error_intrin_sse2.c | 22 __m128i max, min, cmp0, cmp1, cmp2, cmp3; in vp9_highbd_block_error_sse2() local 36 cmp0 = _mm_xor_si128(_mm_cmpgt_epi32(mm_coeff, max), in vp9_highbd_block_error_sse2() 45 _mm_or_si128(_mm_or_si128(cmp0, cmp1), _mm_or_si128(cmp2, cmp3))); in vp9_highbd_block_error_sse2()
|
/external/libaom/libaom/av1/encoder/x86/ |
D | highbd_block_error_intrin_sse2.c | 23 __m128i max, min, cmp0, cmp1, cmp2, cmp3; in av1_highbd_block_error_sse2() local 37 cmp0 = _mm_xor_si128(_mm_cmpgt_epi32(mm_coeff, max), in av1_highbd_block_error_sse2() 46 _mm_or_si128(_mm_or_si128(cmp0, cmp1), _mm_or_si128(cmp2, cmp3))); in av1_highbd_block_error_sse2()
|
/external/llvm/test/Transforms/Util/ |
D | flattencfg.ll | 10 %cmp0 = icmp eq i32 %in_a, -1 12 %cond0 = and i1 %cmp0, %cmp1
|
/external/llvm-project/llvm/test/CodeGen/AArch64/ |
D | subs-to-sub-opt.ll | 18 %cmp0 = icmp eq i32 %s, -1 19 %cmp1 = sext i1 %cmp0 to i8
|
/external/llvm/test/CodeGen/AArch64/ |
D | subs-to-sub-opt.ll | 18 %cmp0 = icmp eq i32 %s, -1 19 %cmp1 = sext i1 %cmp0 to i8
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | fwd_txfm_sse2.h | 40 __m128i cmp0 = _mm_or_si128(_mm_cmpeq_epi16(*preg0, max_overflow), in check_epi16_overflow_x2() local 44 cmp0 = _mm_or_si128(cmp0, cmp1); in check_epi16_overflow_x2() 45 return _mm_movemask_epi8(cmp0); in check_epi16_overflow_x2() 54 __m128i cmp0 = _mm_or_si128(_mm_cmpeq_epi16(*preg0, max_overflow), in check_epi16_overflow_x4() local 62 cmp0 = _mm_or_si128(_mm_or_si128(cmp0, cmp1), _mm_or_si128(cmp2, cmp3)); in check_epi16_overflow_x4() 63 return _mm_movemask_epi8(cmp0); in check_epi16_overflow_x4()
|
/external/llvm-project/llvm/test/Transforms/SLPVectorizer/X86/ |
D | compare-reduce.ll | 97 %cmp0 = fcmp ogt float %x0, 1.0 101 %or03 = or i1 %cmp0, %cmp3wrong 124 %cmp0 = fcmp ogt float %x0, 1.0 128 %or03 = or i1 %cmp0, %cmp3 151 %cmp0 = icmp sgt i32 %x0, 1 155 %or03 = or i1 %cmp0, %cmp3 186 %cmp0 = icmp sgt i32 %x0, %y0 190 %or03 = or i1 %cmp0, %cmp3
|
/external/llvm-project/mlir/test/Dialect/Vector/ |
D | vector-transfer-full-partial-split.mlir | 36 // CHECK: %[[cmp0:.*]] = cmpi "sle", %[[idx0]], %[[d0]] : index 41 // CHECK: %[[cond:.*]] = and %[[cmp0]], %[[cmp1]] : i1 70 // LINALG: %[[cmp0:.*]] = cmpi "sle", %[[idx0]], %[[d0]] : index 75 // LINALG: %[[cond:.*]] = and %[[cmp0]], %[[cmp1]] : i1 124 // CHECK: %[[cmp0:.*]] = cmpi "sle", %[[idx0]], %[[c7]] : index 129 // CHECK: %[[cond:.*]] = and %[[cmp0]], %[[cmp1]] : i1 162 // LINALG: %[[cmp0:.*]] = cmpi "sle", %[[idx0]], %[[c7]] : index 167 // LINALG: %[[cond:.*]] = and %[[cmp0]], %[[cmp1]] : i1
|