/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/ARM/ |
D | relax-per-target-feature.ll | 17 ; Function has thumb2 target-feature, tail call is allowed and must be widened. 27 ; generated as it cannot be widened.
|
D | widen-vmovs.ll | 8 ; The vmovs is first widened to a vmovd, and then converted to a vorr because of the v2f32 vadd.f32.
|
/external/llvm/lib/Target/SystemZ/ |
D | SystemZCallingConv.td | 24 // Match if this specific argument was widened from a short vector type. 54 // Sub-128 vectors are returned in the same way, but they're widened 93 // are passed in the same way, but they're widened to one of these types
|
/external/swiftshader/third_party/llvm-7.0/llvm/lib/Target/SystemZ/ |
D | SystemZCallingConv.td | 24 // Match if this specific argument was widened from a short vector type. 54 // Sub-128 vectors are returned in the same way, but they're widened 93 // are passed in the same way, but they're widened to one of these types
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/LoopVectorize/ |
D | vect-phiscev-sext-trunc.ll | 9 ; that these casts, do not get vectorized/scalarized/widened. 40 ; In the case of widened induction, this means that the induction phi 87 ; the induction is not vectorized but rather only the step is widened. 108 ; In the case of widened induction, this means that the induction phi
|
D | first-order-recurrence.ll | 495 ; widened, the second is a cast that will be widened and needs to sink after the 497 ; instead of widened. Although the cast and the first instruction will both be 498 ; widened, and are originally adjacent to each other, make sure the replicated 528 store i32 7, i32* %arraycidx ; 1st instruction, to be widened. 530 %1 = load i16, i16* %cur.index ; 3rd, first-order-recurring load not widened.
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/IndVarSimplify/ |
D | lftr-wide-trip-count.ll | 77 ; Trip count should be widened and LFTR should canonicalize the condition 114 ; Trip count should be widened and LFTR should canonicalize the condition 152 ; Trip count should be widened and LFTR should canonicalize the condition
|
D | elim-extend.ll | 71 ; %innercount is currently blocked by lcssa, so is not widened. 72 ; %inneriv can be widened only after proving it has no signed-overflow
|
D | iv-widen.ll | 120 ; check that the induction variable of the inner loop has been widened after indvars. 136 ; check that the induction variable of the inner loop has been widened after indvars.
|
/external/swiftshader/third_party/subzero/docs/ |
D | ASAN.rst | 42 contain widened loads that would cause false positives. To avoid reporting such 44 widened load and only check the first byte of the loaded word against shadow
|
/external/swiftshader/third_party/LLVM/test/Transforms/IndVarSimplify/ |
D | elim-extend.ll | 69 ; %innercount is currently blocked by lcssa, so is not widened. 70 ; %inneriv can be widened only after proving it has no signed-overflow
|
/external/llvm/test/Transforms/IndVarSimplify/ |
D | elim-extend.ll | 69 ; %innercount is currently blocked by lcssa, so is not widened. 70 ; %inneriv can be widened only after proving it has no signed-overflow
|
/external/swiftshader/third_party/LLVM/test/CodeGen/ARM/ |
D | widen-vmovs.ll | 8 ; The vmovs is first widened to a vmovd, and then converted to a vorr because of the v2f32 vadd.f32.
|
/external/llvm/test/CodeGen/ARM/ |
D | widen-vmovs.ll | 8 ; The vmovs is first widened to a vmovd, and then converted to a vorr because of the v2f32 vadd.f32.
|
/external/swiftshader/third_party/subzero/tests_lit/asan_tests/ |
D | wideloads.ll | 1 ; Test that potentially widened loads to not trigger an error report
|
/external/llvm/test/Instrumentation/AddressSanitizer/ |
D | asan-vs-gvn.ll | 2 ; ASAN conflicts with load widening iff the widened load accesses data out of bounds
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Instrumentation/AddressSanitizer/ |
D | asan-vs-gvn.ll | 2 ; ASAN conflicts with load widening iff the widened load accesses data out of bounds
|
/external/clang/test/OpenMP/ |
D | for_simd_codegen.cpp | 598 void widened(float *a, float *b, float *c, float *d) { in widened() function
|
D | parallel_for_simd_codegen.cpp | 551 void widened(float *a, float *b, float *c, float *d) { in widened() function
|
D | simd_codegen.cpp | 433 void widened(float *a, float *b, float *c, float *d) { in widened() function
|
/external/swiftshader/third_party/llvm-7.0/llvm/docs/Proposals/ |
D | VectorizationPlan.rst | 45 vectorizer. The output IR contains code that has been vectorized or "widened" 145 instructions; e.g., cloned once, replicated multiple times or widened 189 cast may sink past a later instruction and be widened to handle first-order
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/DebugInfo/Generic/ |
D | indvar-discriminator.ll | 3 ; When the induction variable is widened by indvars, check that the debug loc
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/IndVarSimplify/AMDGPU/ |
D | no-widen-to-i64.ll | 5 ; Induction variables should not be widened for 64-bit integers,
|
/external/llvm/test/Transforms/IndVarSimplify/AMDGPU/ |
D | no-widen-to-i64.ll | 5 ; Induction variables should not be widened for 64-bit integers,
|
/external/llvm/test/Transforms/GVN/ |
D | load-pre-nonlocal.ll | 53 ; %1 is partially redundant if %0 can be widened to a 64-bit load.
|