/external/v8/test/mjsunit/asm/ |
D | pointer-masking.js | 14 function load1(i) { function 25 return {load1: load1, store1: store1}; 28 assertEquals(0, pm1.load1(0)); 29 assertEquals(0, pm1.load1(1025)); 32 assertEquals(1, pm1.load1(0)); 33 assertEquals(1, pm1.load1(1024)); 34 assertEquals(127, pm1.load1(1)); 35 assertEquals(127, pm1.load1(1025));
|
/external/llvm/test/Transforms/SLPVectorizer/X86/ |
D | propagate_ir_flags.ll | 18 %load1 = load i32, i32* %idx1, align 4 23 %op1 = lshr exact i32 %load1, 1 44 %load1 = load i32, i32* %idx1, align 4 49 %op1 = lshr exact i32 %load1, 1 70 %load1 = load i32, i32* %idx1, align 4 75 %op1 = add nsw i32 %load1, 1 96 %load1 = load i32, i32* %idx1, align 4 101 %op1 = add nsw i32 %load1, 1 122 %load1 = load i32, i32* %idx1, align 4 127 %op1 = add nuw i32 %load1, 1 [all …]
|
/external/llvm/test/Transforms/LoadCombine/ |
D | load-combine-aa.ll | 11 %load1 = load i32, i32* %a, align 4 12 %conv = zext i32 %load1 to i64 14 store i32 %load1, i32* %b, align 4 29 %load1 = load i32, i32* %a, align 4 30 %conv = zext i32 %load1 to i64 32 store i32 %load1, i32* %b, align 4
|
D | load-combine-assume.ll | 15 %load1 = load i32, i32* %a, align 4 16 %conv = zext i32 %load1 to i64 34 %load1 = load i32, i32* %a, align 4 35 %conv = zext i32 %load1 to i64
|
/external/llvm/test/Transforms/CodeGenPrepare/AArch64/ |
D | free-zext.ll | 11 %load1 = load i32, i32* %ptr, align 4 23 %phi = phi i32 [ %load1, %bb1 ], [ %load2, %bb2 ] 36 %load1 = load i32, i32* %ptr, align 4 41 %trunc = trunc i32 %load1 to i16 46 %shl = shl i32 %load1, 16 53 %and = and i32 %load1, 65535 65 %load1 = load i32, i32* %ptr, align 4 70 %phi = phi i32 [ %load1, %bb1 ], [ %load2, %loop ]
|
/external/libvpx/libvpx/vpx_dsp/mips/ |
D | itrans32_cols_dspr2.c | 33 int load1, load2, load3, load4; in vpx_idct32_cols_add_blk_dspr2() local 104 : [load1] "=&r" (load1), [load2] "=&r" (load2), [load3] "=&r" (load3), in vpx_idct32_cols_add_blk_dspr2() 165 : [load1] "=&r" (load1), [load2] "=&r" (load2), [load3] "=&r" (load3), in vpx_idct32_cols_add_blk_dspr2() 226 : [load1] "=&r" (load1), [load2] "=&r" (load2), [load3] "=&r" (load3), in vpx_idct32_cols_add_blk_dspr2() 283 : [load1] "=&r" (load1), [load2] "=&r" (load2), [load3] "=&r" (load3), in vpx_idct32_cols_add_blk_dspr2() 340 : [load1] "=&r" (load1), [load2] "=&r" (load2), [load3] "=&r" (load3), in vpx_idct32_cols_add_blk_dspr2() 397 : [load1] "=&r" (load1), [load2] "=&r" (load2), [load3] "=&r" (load3), in vpx_idct32_cols_add_blk_dspr2() 591 : [load1] "=&r" (load1), [load2] "=&r" (load2), in vpx_idct32_cols_add_blk_dspr2() 652 : [load1] "=&r" (load1), [load2] "=&r" (load2), in vpx_idct32_cols_add_blk_dspr2()
|
D | itrans32_dspr2.c | 36 int load1, load2, load3, load4; in idct32_rows_dspr2() local 149 : [load1] "=&r" (load1), [load2] "=&r" (load2), in idct32_rows_dspr2() 211 : [load1] "=&r" (load1), [load2] "=&r" (load2), in idct32_rows_dspr2() 273 : [load1] "=&r" (load1), [load2] "=&r" (load2), in idct32_rows_dspr2() 335 : [load1] "=&r" (load1), [load2] "=&r" (load2), in idct32_rows_dspr2() 397 : [load1] "=&r" (load1), [load2] "=&r" (load2), in idct32_rows_dspr2() 459 : [load1] "=&r" (load1), [load2] "=&r" (load2), in idct32_rows_dspr2() 661 : [load1] "=&r" (load1), [load2] "=&r" (load2), in idct32_rows_dspr2() 727 : [load1] "=&r" (load1), [load2] "=&r" (load2), in idct32_rows_dspr2()
|
D | convolve2_vert_dspr2.c | 33 uint32_t load1, load2; in convolve_bi_vert_4_dspr2() local 101 : [load1] "=&r" (load1), [load2] "=&r" (load2), in convolve_bi_vert_4_dspr2() 130 uint32_t load1, load2; in convolve_bi_vert_64_dspr2() local 198 : [load1] "=&r" (load1), [load2] "=&r" (load2), in convolve_bi_vert_64_dspr2()
|
D | convolve2_avg_dspr2.c | 33 uint32_t load1, load2; in convolve_bi_avg_vert_4_dspr2() local 108 : [load1] "=&r" (load1), [load2] "=&r" (load2), in convolve_bi_avg_vert_4_dspr2() 137 uint32_t load1, load2; in convolve_bi_avg_vert_64_dspr2() local 213 : [load1] "=&r" (load1), [load2] "=&r" (load2), in convolve_bi_avg_vert_64_dspr2()
|
D | convolve8_vert_dspr2.c | 33 uint32_t load1, load2, load3, load4; in convolve_vert_4_dspr2() local 155 : [load1] "=&r" (load1), [load2] "=&r" (load2), in convolve_vert_4_dspr2() 187 uint32_t load1, load2, load3, load4; in convolve_vert_64_dspr2() local 310 : [load1] "=&r" (load1), [load2] "=&r" (load2), in convolve_vert_64_dspr2()
|
D | convolve8_avg_dspr2.c | 33 uint32_t load1, load2, load3, load4; in convolve_avg_vert_4_dspr2() local 163 : [load1] "=&r" (load1), [load2] "=&r" (load2), in convolve_avg_vert_4_dspr2() 194 uint32_t load1, load2, load3, load4; in convolve_avg_vert_64_dspr2() local 325 : [load1] "=&r" (load1), [load2] "=&r" (load2), in convolve_avg_vert_64_dspr2()
|
D | itrans16_dspr2.c | 25 int load1, load2, load3, load4, load5, load6, load7, load8; in idct16_rows_dspr2() local 67 : [load1] "=&r" (load1), [load2] "=&r" (load2), in idct16_rows_dspr2() 191 : [load1] "=&r" (load1), [load2] "=&r" (load2), in idct16_rows_dspr2() 411 int load1, load2, load3, load4, load5, load6, load7, load8; in idct16_cols_add_blk_dspr2() local 463 : [load1] "=&r" (load1), [load2] "=&r" (load2), in idct16_cols_add_blk_dspr2() 587 : [load1] "=&r" (load1), [load2] "=&r" (load2), in idct16_cols_add_blk_dspr2()
|
/external/llvm/test/CodeGen/X86/ |
D | addr-mode-matcher.ll | 24 ; %load1 = (load (and (shl %xor, 2), 1020)) 29 %load1 = load i32, i32* %tmp1704, align 4 40 ; While matching xor we address-match %load1. The and-of-shift reassocication 42 ; node becomes identical to %load2. CSE replaces %load1 which leaves its 44 %tmp1711 = xor i32 %load1, %tmp1710
|
D | add-nsw-sext.ll | 155 %load1 = load i32, i32* %gep1, align 4 162 %add3 = add i32 %load1, %load2
|
/external/llvm/test/CodeGen/AArch64/ |
D | ldst-opt.ll | 291 %load1 = load %pre.struct.i32*, %pre.struct.i32** %this 292 %gep1 = getelementptr inbounds %pre.struct.i32, %pre.struct.i32* %load1, i64 0, i32 1 309 %load1 = load %pre.struct.i64*, %pre.struct.i64** %this 310 %gep1 = getelementptr inbounds %pre.struct.i64, %pre.struct.i64* %load1, i64 0, i32 1 327 %load1 = load %pre.struct.i128*, %pre.struct.i128** %this 328 %gep1 = getelementptr inbounds %pre.struct.i128, %pre.struct.i128* %load1, i64 0, i32 1 345 %load1 = load %pre.struct.float*, %pre.struct.float** %this 346 %gep1 = getelementptr inbounds %pre.struct.float, %pre.struct.float* %load1, i64 0, i32 1 363 %load1 = load %pre.struct.double*, %pre.struct.double** %this 364 %gep1 = getelementptr inbounds %pre.struct.double, %pre.struct.double* %load1, i64 0, i32 1 [all …]
|
D | arm64-vabs.ll | 37 %load1 = load <16 x i8>, <16 x i8>* %A 39 …%tmp1 = shufflevector <16 x i8> %load1, <16 x i8> undef, <8 x i32> <i32 8, i32 9, i32 10, i32 11, … 49 %load1 = load <8 x i16>, <8 x i16>* %A 51 … %tmp1 = shufflevector <8 x i16> %load1, <8 x i16> undef, <4 x i32> <i32 4, i32 5, i32 6, i32 7> 61 %load1 = load <4 x i32>, <4 x i32>* %A 63 %tmp1 = shufflevector <4 x i32> %load1, <4 x i32> undef, <2 x i32> <i32 2, i32 3> 103 %load1 = load <16 x i8>, <16 x i8>* %A 105 …%tmp1 = shufflevector <16 x i8> %load1, <16 x i8> undef, <8 x i32> <i32 8, i32 9, i32 10, i32 11, … 116 %load1 = load <8 x i16>, <8 x i16>* %A 118 %tmp1 = shufflevector <8 x i16> %load1, <8 x i16> undef, <4 x i32> <i32 4, i32 5, i32 6, i32 7> [all …]
|
D | free-zext.ll | 60 %load1 = load i32, i32* %ptr, align 4 63 %phi = phi i32 [ %load1, %bb1 ], [ %load2, %loop ]
|
D | arm64-vmul.ll | 87 %load1 = load <8 x i16>, <8 x i16>* %A 89 %tmp1 = shufflevector <8 x i16> %load1, <8 x i16> undef, <4 x i32> <i32 4, i32 5, i32 6, i32 7> 98 %load1 = load <4 x i32>, <4 x i32>* %A 100 %tmp1 = shufflevector <4 x i32> %load1, <4 x i32> undef, <2 x i32> <i32 2, i32 3> 328 %load1 = load <8 x i16>, <8 x i16>* %A 331 %tmp1 = shufflevector <8 x i16> %load1, <8 x i16> undef, <4 x i32> <i32 4, i32 5, i32 6, i32 7> 341 %load1 = load <4 x i32>, <4 x i32>* %A 344 %tmp1 = shufflevector <4 x i32> %load1, <4 x i32> undef, <2 x i32> <i32 2, i32 3> 376 %load1 = load <8 x i16>, <8 x i16>* %A 379 %tmp1 = shufflevector <8 x i16> %load1, <8 x i16> undef, <4 x i32> <i32 4, i32 5, i32 6, i32 7> [all …]
|
/external/llvm/test/CodeGen/AMDGPU/ |
D | schedule-global-loads.ll | 19 %load1 = load i32, i32 addrspace(1)* %gep, align 4 21 store i32 %load1, i32 addrspace(1)* %out1, align 4
|
/external/llvm/test/CodeGen/PowerPC/ |
D | load-shift-combine.ll | 22 …%bf.load1 = load i96, i96* bitcast (%struct.S1847* getelementptr inbounds ([5 x %struct.S1847], [5… 23 %bf.clear2 = and i96 %bf.load1, 302231454903657293676543
|
/external/llvm/test/Instrumentation/AddressSanitizer/ |
D | experiment-call.ll | 7 define void @load1(i8* %p) sanitize_address { 11 ; CHECK-LABEL: define void @load1
|
D | experiment.ll | 7 define void @load1(i8* %p) sanitize_address { 11 ; CHECK-LABEL: define void @load1
|
/external/v8/test/unittests/compiler/ |
D | escape-analysis-unittest.cc | 412 Node* load1 = Load(FieldAccessAtIndex(0), allocation); in TEST_F() local 418 graph()->end()->AppendInput(zone(), load1); in TEST_F() 423 ExpectReplacement(load1, object1); in TEST_F()
|
/external/llvm/test/Transforms/EarlyCSE/ |
D | atomics.ll | 7 %load1 = load i32, i32* %P1 8 %sel = select i1 %B, i32 %load0, i32 %load1
|
D | basic.ll | 200 %load1 = load i32, i32* %P1 201 %sel = select i1 %B, i32 %load0, i32 %load1
|