/external/llvm/test/Transforms/GVN/ |
D | invariant.group.ll | 13 store i8 42, i8* %ptr, !invariant.group !0 16 %a = load i8, i8* %ptr, !invariant.group !0 17 %b = load i8, i8* %ptr, !invariant.group !0 18 %c = load i8, i8* %ptr, !invariant.group !0 27 store i8 42, i8* %ptr, !invariant.group !0 28 %ptr2 = call i8* @llvm.invariant.group.barrier(i8* %ptr) 29 %a = load i8, i8* %ptr, !invariant.group !0 40 store i8 42, i8* %ptr, !invariant.group !0 43 store i8 13, i8* %ptr ; can't use this store with invariant.group 48 %b = load i8, i8* %ptr, !invariant.group !0 [all …]
|
D | invariant-load.ll | 1 ; Test if the !invariant.load metadata is maintained by GVN. 6 ; CHECK: %x = load i32, i32* %p, align 4, !invariant.load !0 9 %x = load i32, i32* %p, align 4, !invariant.load !0 12 %y = load i32, i32* %p, align 4, !invariant.load !0 19 ; CHECK-NOT: !invariant.load 25 %y = load i32, i32* %p, align 4, !invariant.load !0 30 ; With the invariant.load metadata, what would otherwise 44 %v2 = load i32, i32* %p, !invariant.load !0 55 %v1 = load i32, i32* %p, !invariant.load !0 77 %v1 = load i32, i32* %p, !invariant.load !0 [all …]
|
/external/llvm/test/Transforms/GlobalOpt/ |
D | invariant.group.barrier.ll | 36 %barr = call i8* @llvm.invariant.group.barrier(i8* %0) 44 ; We can't step through invariant.group.barrier here, because that would change 46 ; val = load i32, i32* %ptrVal, !invariant.group !0 48 ; %val = load i32, i32* @tmp3, !invariant.group !0 54 store i32 13, i32* @tmp3, !invariant.group !0 57 %barr = call i8* @llvm.invariant.group.barrier(i8* %0) 61 store i32 42, i32* %1, !invariant.group !0 68 %val = load i32, i32* %ptrVal, !invariant.group !0 71 %val2 = load i32, i32* @tmp3, !invariant.group !0 77 declare i8* @llvm.invariant.group.barrier(i8*)
|
D | invariant.ll | 6 declare {}* @llvm.invariant.start(i64 %size, i8* nocapture %ptr) 9 call {}* @llvm.invariant.start(i64 4, i8* %ptr) 28 %B = call {}* @llvm.invariant.start(i64 4, i8* %A) 39 call {}* @llvm.invariant.start(i64 3, i8* %A) 49 call {}* @llvm.invariant.start(i64 -1, i8* %A)
|
D | invariant-nodatalayout.ll | 4 declare {}* @llvm.invariant.start(i64 %size, i8* nocapture %ptr) 11 call {}* @llvm.invariant.start(i64 4, i8* %ptr)
|
/external/llvm/test/Transforms/CodeGenPrepare/ |
D | invariant.group.ll | 8 ; CHECK-NOT: !invariant.group 9 ; CHECK-NOT: @llvm.invariant.group.barrier( 11 %val = load i8, i8* @tmp, !invariant.group !0 12 %ptr = call i8* @llvm.invariant.group.barrier(i8* @tmp) 15 store i8 42, i8* %ptr, !invariant.group !0 21 declare i8* @llvm.invariant.group.barrier(i8*)
|
/external/llvm/test/Transforms/InstCombine/ |
D | invariant.ll | 1 ; Test to make sure unused llvm.invariant.start calls are not trivially eliminated 6 declare {}* @llvm.invariant.start(i64, i8* nocapture) nounwind readonly 11 %i = call {}* @llvm.invariant.start(i64 1, i8* %a) ; <{}*> [#uses=0] 12 ; CHECK: call {}* @llvm.invariant.start
|
D | fold-phi-load-metadata.ll | 12 …-z0-9.]+}}, align 4, !tbaa ![[TBAA:[0-9]+]], !range ![[RANGE:[0-9]+]], !invariant.load ![[EMPTYNOD… 22 …%val = load i32, i32* %i, align 4, !tbaa !0, !alias.scope !13, !noalias !14, !invariant.load !17, … 28 …%val2 = load i32, i32* %i2, align 4, !tbaa !2, !alias.scope !15, !noalias !16, !invariant.load !17…
|
/external/deqp/data/gles2/shaders/ |
D | qualification_order.test | 16 invariant varying lowp float x0; 32 invariant varying lowp float x0; 86 invariant varying float x0; 102 invariant varying float x0; 125 invariant lowp varying float x0; 141 invariant lowp varying float x0; 160 varying invariant lowp float x0; 176 varying invariant lowp float x0; 195 varying lowp invariant float x0; 211 varying lowp invariant float x0; [all …]
|
/external/llvm/test/Analysis/BasicAA/ |
D | invariant_load.ll | 3 ; The input *.ll is obtained by manually annotating "invariant.load" to the 4 ; two loads. With "invariant.load" metadata, the second load is redundant. 13 %0 = load i32, i32* %p, align 4, !invariant.load !3 16 %1 = load i32, i32* %p, align 4, !invariant.load !3
|
/external/deqp/data/gles3/shaders/ |
D | qualification_order.test | 17 invariant smooth centroid out lowp float x0; 99 invariant smooth centroid out float x0; 140 invariant centroid out lowp float x0; 263 invariant centroid out float x0; 308 invariant smooth lowp centroid out float x0; 349 invariant centroid out smooth lowp float x0; 390 invariant centroid out lowp smooth float x0; 431 invariant lowp smooth centroid out float x0; 472 invariant lowp centroid out smooth float x0; 513 smooth invariant centroid out lowp float x0; [all …]
|
/external/llvm/test/Bitcode/ |
D | memInstructions.3.2.ll | 54 ; CHECK-NEXT: %res9 = load i8, i8* %ptr1, !invariant.load !1 55 %res9 = load i8, i8* %ptr1, !invariant.load !1 57 ; CHECK-NEXT: %res10 = load volatile i8, i8* %ptr1, !invariant.load !1 58 %res10 = load volatile i8, i8* %ptr1, !invariant.load !1 60 ; CHECK-NEXT: %res11 = load i8, i8* %ptr1, align 1, !invariant.load !1 61 %res11 = load i8, i8* %ptr1, align 1, !invariant.load !1 63 ; CHECK-NEXT: %res12 = load volatile i8, i8* %ptr1, align 1, !invariant.load !1 64 %res12 = load volatile i8, i8* %ptr1, align 1, !invariant.load !1 66 ; CHECK-NEXT: %res13 = load i8, i8* %ptr1, {{[(!nontemporal !0, !invariant.load !1) | (!invariant.l… 67 %res13 = load i8, i8* %ptr1, !nontemporal !0, !invariant.load !1 [all …]
|
/external/clang/test/CodeGenObjC/ |
D | ivar-base-as-invariant-load.m | 26 // CHECK: [[T1:%.*]] = load i64, i64* @"OBJC_IVAR_$_A._flags", align 8, !invariant.load ![[MD_NUM:[… 27 // CHECK: [[T2:%.*]] = load i64, i64* @"OBJC_IVAR_$_A._flags", align 8, !invariant.load ![[MD_NUM]] 28 // CHECK: [[T3:%.*]] = load i64, i64* @"OBJC_IVAR_$_A._flags", align 8, !invariant.load ![[MD_NUM]]
|
D | optimize-ivar-offset-load.m | 47 // CHECK: [[ZERO:%.*]] = load i8*, i8** @OBJC_SELECTOR_REFERENCES_, align 8, !invariant.load 48 // CHECK: [[IVAR:%.*]] = load i64, i64* @"OBJC_IVAR_$_SampleClass._value", align 8, !invariant.load 62 // CHECK: [[ZERO:%.*]] = load i8*, i8** @OBJC_SELECTOR_REFERENCES_, align 8, !invariant.load 63 // CHECK: [[IVAR:%.*]] = load i64, i64* @"OBJC_IVAR_$_SampleClass._value", align 8, !invariant.load
|
D | ivar-invariant.m | 32 // CHECK: [[IVAR:%.*]] = load i64, i64* @"OBJC_IVAR_$_Derived.member", align 8, !invariant.load 55 // CHECK: [[IVAR:%.*]] = load i64, i64* @"OBJC_IVAR_$_Derived.member", align 8, !invariant.load
|
/external/llvm/test/Feature/ |
D | memorymarkers.ll | 9 declare {}* @llvm.invariant.start(i64, i8* nocapture) readonly nounwind 10 declare void @llvm.invariant.end({}*, i64, i8* nocapture) nounwind 27 %inv = call {}* @llvm.invariant.start(i64 8, i8* %y) 32 call void @llvm.invariant.end({}* %inv, i64 8, i8* %y)
|
/external/llvm/test/Analysis/LoopAccessAnalysis/ |
D | store-to-invariant-check2.ll | 3 ; Test to confirm LAA will not find store to invariant address. 4 ; Inner loop has no store to invariant address. 12 ; CHECK: Store to invariant address was not found in loop. 13 ; CHECK-NOT: Store to invariant address was found in loop.
|
D | store-to-invariant-check1.ll | 3 ; Test to confirm LAA will find store to invariant address. 4 ; Inner loop has a store to invariant address. 12 ; CHECK: Store to invariant address was found in loop. 13 ; CHECK-NOT: Store to invariant address was not found in loop.
|
D | store-to-invariant-check3.ll | 3 ; Test to confirm LAA will find store to invariant address. 4 ; Inner loop has a store to invariant address. 12 ; CHECK: Store to invariant address was found in loop.
|
/external/llvm/test/CodeGen/AMDGPU/ |
D | invariant-load-no-alias-store.ll | 7 ; from constant/invariant memory. 14 %ptr = load i16 addrspace(1)*, i16 addrspace(1)* addrspace(1)* %in, !invariant.load !0 26 %ptr = load i16 addrspace(1)*, i16 addrspace(1)* addrspace(2)* %in, !invariant.load !0
|
/external/antlr/antlr-3.4/runtime/ActionScript/project/test/org/antlr/runtime/test/ |
D | TestANTLRStringStream.as | 24 assertEquals(stream.size, 3); // invariant 26 assertEquals(stream.line, 1); // invariant 34 assertEquals(stream.size, 3); // invariant 36 assertEquals(stream.line, 1); // invariant
|
/external/clang/test/CodeGenObjCXX/ |
D | property-lvalue-capture.mm | 27 // CHECK: [[TWO:%.*]] = load i8*, i8** @OBJC_SELECTOR_REFERENCES_, align 8, !invariant.load ![[MD… 30 // CHECK: [[FOUR:%.*]] = load i8*, i8** @OBJC_SELECTOR_REFERENCES_.2, align 8, !invariant.load ![… 50 // CHECK: [[TWO1:%.*]] = load i8*, i8** @OBJC_SELECTOR_REFERENCES_.5, align 8, !invariant.load ![…
|
/external/chromium-trace/catapult/third_party/gsutil/third_party/httplib2/python3/httplib2/ |
D | iri2uri.py | 86 invariant = [ 95 for uri in invariant:
|
/external/chromium-trace/catapult/third_party/gsutil/third_party/httplib2/python2/httplib2/ |
D | iri2uri.py | 86 invariant = [ 95 for uri in invariant:
|
/external/llvm/test/CodeGen/WebAssembly/ |
D | reg-stackify.ll | 23 %t = load volatile i32, i32* %q, !invariant.load !0 28 ; Yes because of invariant load and no side effects. 33 %t = load i32, i32* %q, !invariant.load !0
|