/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/Hexagon/ |
D | Atomics.ll | 17 %t41 = atomicrmw and i64* @sll, i64 1 monotonic 46 %t41 = atomicrmw and i64* @sll, i64 11 monotonic 47 store i64 %t41, i64* @sll, align 8
|
D | rdf-filter-defs.ll | 116 %t41 = phi i8* [ %t38, %b16 ], [ %t39, %b17 ] 118 %t43 = getelementptr inbounds i8, i8* %t41, i32 %p1
|
/external/llvm/test/CodeGen/Hexagon/ |
D | Atomics.ll | 17 %t41 = atomicrmw and i64* @sll, i64 1 monotonic 46 %t41 = atomicrmw and i64* @sll, i64 11 monotonic 47 store i64 %t41, i64* @sll, align 8
|
/external/llvm/test/Transforms/IndVarSimplify/ |
D | eliminate-rem.ll | 74 %t22 = phi i64 [ %t41, %bb21 ], [ 0, %bb20 ] ; <i64> [#uses=4] 93 %t41 = add nsw i64 %t22, 1 ; <i64> [#uses=2] 94 %t42 = icmp slt i64 %t41, %arg1 ; <i1> [#uses=1]
|
D | ashr-tripcount.ll | 85 %t41 = fmul float %t25, %invQuantizer.0 ; <float> [#uses=1] 88 store float %t41, float* %t43, align 4
|
/external/swiftshader/third_party/LLVM/test/Transforms/IndVarSimplify/ |
D | eliminate-rem.ll | 74 %t22 = phi i64 [ %t41, %bb21 ], [ 0, %bb20 ] ; <i64> [#uses=4] 93 %t41 = add nsw i64 %t22, 1 ; <i64> [#uses=2] 94 %t42 = icmp slt i64 %t41, %arg1 ; <i1> [#uses=1]
|
D | ashr-tripcount.ll | 85 %t41 = fmul float %t25, %invQuantizer.0 ; <float> [#uses=1] 88 store float %t41, float* %t43, align 4
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/IndVarSimplify/ |
D | eliminate-rem.ll | 74 %t22 = phi i64 [ %t41, %bb21 ], [ 0, %bb20 ] ; <i64> [#uses=4] 93 %t41 = add nsw i64 %t22, 1 ; <i64> [#uses=2] 94 %t42 = icmp slt i64 %t41, %arg1 ; <i1> [#uses=1]
|
D | ashr-tripcount.ll | 87 %t41 = fmul float %t25, %invQuantizer.0 ; <float> [#uses=1] 90 store float %t41, float* %t43, align 4
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/Reassociate/ |
D | 2012-05-08-UndefLeak.ll | 53 %t41 = add i64 %t40, %t6 54 %t42 = add i64 %t41, %t7
|
/external/llvm/test/Transforms/Reassociate/ |
D | 2012-05-08-UndefLeak.ll | 53 %t41 = add i64 %t40, %t6 54 %t42 = add i64 %t41, %t7
|
/external/deqp-deps/glslang/Test/ |
D | 150.frag | 42 vec4 t41 = texelFetch( smsa, p3, samp);
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/X86/ |
D | pseudo_cmov_lower.ll | 191 %t41 = select i1 %cmp, double %v42, double %t43 202 store double %t41, double* %a41, align 8 259 %t41 = select i1 %cmp, <64 x i1> %v42, <64 x i1> %t43 264 store <64 x i1> %t41, <64 x i1>* %a41, align 16
|
D | lsr-reuse.ll | 542 %t41 = fadd float %t23, %t40 ; <float> [#uses=1] 543 %t42 = insertelement <4 x float> undef, float %t41, i32 0 ; <<4 x float>> [#uses=1]
|
/external/llvm/test/CodeGen/X86/ |
D | pseudo_cmov_lower.ll | 191 %t41 = select i1 %cmp, double %v42, double %t43 202 store double %t41, double* %a41, align 8 259 %t41 = select i1 %cmp, <64 x i1> %v42, <64 x i1> %t43 264 store <64 x i1> %t41, <64 x i1>* %a41, align 16
|
D | lsr-reuse.ll | 542 %t41 = fadd float %t23, %t40 ; <float> [#uses=1] 543 %t42 = insertelement <4 x float> undef, float %t41, i32 0 ; <<4 x float>> [#uses=1]
|
/external/clang/test/CodeGen/ |
D | ms-inline-asm.c | 523 void t41(unsigned short a) { 581 __asm call t41
|
/external/deqp-deps/glslang/Test/baseResults/ |
D | 150.frag.out | 82 0:42 't41' ( temp 4-component vector of float)
|
/external/swiftshader/third_party/LLVM/test/CodeGen/ARM/ |
D | lsr-on-unrolled-loops.ll | 82 %t41 = bitcast i8* %t40 to float* 185 %t122 = load float* %t41, align 4
|
/external/swiftshader/third_party/LLVM/test/CodeGen/X86/ |
D | lsr-reuse.ll | 541 %t41 = fadd float %t23, %t40 ; <float> [#uses=1] 542 %t42 = insertelement <4 x float> undef, float %t41, i32 0 ; <<4 x float>> [#uses=1]
|
/external/libcxxabi/test/ |
D | dynamic_cast3.pass.cpp | 2358 namespace t41 namespace 2459 t41::test(); in main()
|
/external/clang/test/SemaCXX/ |
D | type-traits.cpp | 945 int t41[F(__is_const(unsigned int))]; in is_const() local 994 int t41[F(__is_volatile(unsigned int))]; in is_volatile() local
|
/external/honggfuzz/examples/apache-httpd/corpus_http2/ |
D | 179aaccfe25dca3c899610668cd59155.00021d24.honggfuzz.cov | 169 ��^-�4�b�S �t41ػ|�s�9�,����%������d���m�0
|