Home
last modified time | relevance | path

Searched refs:l13 (Results 1 – 25 of 37) sorted by relevance

12

/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/X86/
Dpr30562.ll12 %l13 = extractelement <2 x i64> %vec.ind, i32 %n
13 %l14 = getelementptr inbounds i64, i64* %perm, i64 %l13
16 %niter.ncmp.3 = icmp eq i64 %l13, 0
/external/llvm/test/CodeGen/SystemZ/
Dframe-18.ll31 %l13 = load volatile i32 , i32 *%ptr
36 store volatile i32 %l13, i32 *%ptr
73 %l13 = load volatile i64 , i64 *%ptr
78 store volatile i64 %l13, i64 *%ptr
Dframe-17.ll47 %l13 = load volatile float , float *%ptr
54 store volatile float %l13, float *%ptr
108 %l13 = load volatile double , double *%ptr
115 store volatile double %l13, double *%ptr
165 %l13 = load volatile fp128 , fp128 *%ptr
168 store volatile fp128 %l13, fp128 *%ptr
Dframe-06.ll44 %l13 = load volatile i64 , i64 *%ptr
58 %add13 = add i64 %l13, %add12
111 %l13 = load volatile i64 , i64 *%ptr
124 %add13 = add i64 %l13, %add12
Dframe-05.ll47 %l13 = load volatile i32 , i32 *%ptr
61 %add13 = add i32 %l13, %add12
114 %l13 = load volatile i32 , i32 *%ptr
127 %add13 = add i32 %l13, %add12
Dframe-03.ll55 %l13 = load volatile double , double *%ptr
71 %add13 = fadd double %l13, %add12
137 %l13 = load volatile double , double *%ptr
152 %add13 = fadd double %l13, %add12
Dframe-02.ll53 %l13 = load volatile float , float *%ptr
69 %add13 = fadd float %l13, %add12
135 %l13 = load volatile float , float *%ptr
150 %add13 = fadd float %l13, %add12
Dframe-08.ll39 %l13 = load volatile i32 , i32 *%ptr
52 %add13 = add i32 %l13, %add12
142 %l13 = load volatile i32 , i32 *%ptr
155 %add13 = add i32 %l13, %add12
Dframe-07.ll84 %l13 = load volatile double , double *%ptr
100 %add13 = fadd double %l13, %add12
213 %l13 = load volatile double , double *%ptr
229 %add13 = fadd double %l13, %add12
Dframe-09.ll78 %l13 = load volatile i32 , i32 *%ptr
91 %add13 = add i32 %l13, %add12
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/SystemZ/
Dframe-18.ll31 %l13 = load volatile i32, i32 *%ptr
36 store volatile i32 %l13, i32 *%ptr
73 %l13 = load volatile i64, i64 *%ptr
78 store volatile i64 %l13, i64 *%ptr
Dframe-17.ll47 %l13 = load volatile float, float *%ptr
54 store volatile float %l13, float *%ptr
108 %l13 = load volatile double, double *%ptr
115 store volatile double %l13, double *%ptr
165 %l13 = load volatile fp128, fp128 *%ptr
168 store volatile fp128 %l13, fp128 *%ptr
Dframe-06.ll44 %l13 = load volatile i64, i64 *%ptr
58 %add13 = add i64 %l13, %add12
111 %l13 = load volatile i64, i64 *%ptr
124 %add13 = add i64 %l13, %add12
Dframe-05.ll47 %l13 = load volatile i32, i32 *%ptr
61 %add13 = add i32 %l13, %add12
114 %l13 = load volatile i32, i32 *%ptr
127 %add13 = add i32 %l13, %add12
Dframe-03.ll55 %l13 = load volatile double, double *%ptr
71 %add13 = fadd double %l13, %add12
137 %l13 = load volatile double, double *%ptr
152 %add13 = fadd double %l13, %add12
Dframe-02.ll53 %l13 = load volatile float, float *%ptr
69 %add13 = fadd float %l13, %add12
135 %l13 = load volatile float, float *%ptr
150 %add13 = fadd float %l13, %add12
Dframe-07.ll84 %l13 = load volatile double, double *%ptr
100 %add13 = fadd double %l13, %add12
213 %l13 = load volatile double, double *%ptr
229 %add13 = fadd double %l13, %add12
Dframe-08.ll39 %l13 = load volatile i32, i32 *%ptr
52 %add13 = add i32 %l13, %add12
142 %l13 = load volatile i32, i32 *%ptr
155 %add13 = add i32 %l13, %add12
Dframe-09.ll78 %l13 = load volatile i32, i32 *%ptr
91 %add13 = add i32 %l13, %add12
/external/libvpx/libvpx/vp9/encoder/mips/msa/
Dvp9_fdct16x16_msa.c158 v8i16 l0, l1, l2, l3, l4, l5, l6, l7, l8, l9, l10, l11, l12, l13, l14, l15; in fadst16_transpose_postproc_msa() local
171 LD_SH8(input + 8, 16, l8, l9, l10, l11, l12, l13, l14, l15); in fadst16_transpose_postproc_msa()
172 TRANSPOSE8x8_SH_SH(l8, l9, l10, l11, l12, l13, l14, l15, r8, r9, r10, r11, in fadst16_transpose_postproc_msa()
193 LD_SH8(input + 8, 16, l8, l9, l10, l11, l12, l13, l14, l15); in fadst16_transpose_postproc_msa()
194 TRANSPOSE8x8_SH_SH(l8, l9, l10, l11, l12, l13, l14, l15, r8, r9, r10, r11, in fadst16_transpose_postproc_msa()
339 v8i16 l0, l1, l2, l3, l4, l5, l6, l7, l8, l9, l10, l11, l12, l13, l14, l15; in fadst16_transpose_msa() local
342 LD_SH16(input, 8, l0, l8, l1, l9, l2, l10, l3, l11, l4, l12, l5, l13, l6, l14, in fadst16_transpose_msa()
346 TRANSPOSE8x8_SH_SH(l8, l9, l10, l11, l12, l13, l14, l15, r8, r9, r10, r11, in fadst16_transpose_msa()
354 LD_SH16(input, 8, l0, l8, l1, l9, l2, l10, l3, l11, l4, l12, l5, l13, l6, l14, in fadst16_transpose_msa()
358 TRANSPOSE8x8_SH_SH(l8, l9, l10, l11, l12, l13, l14, l15, r8, r9, r10, r11, in fadst16_transpose_msa()
/external/libvpx/libvpx/vpx_dsp/mips/
Didct16x16_msa.c294 v8i16 l0, l1, l2, l3, l4, l5, l6, l7, l8, l9, l10, l11, l12, l13, l14, l15; in vpx_iadst16_1d_rows_msa() local
297 LD_SH16(input, 8, l0, l8, l1, l9, l2, l10, l3, l11, l4, l12, l5, l13, l6, l14, in vpx_iadst16_1d_rows_msa()
301 TRANSPOSE8x8_SH_SH(l8, l9, l10, l11, l12, l13, l14, l15, l8, l9, l10, l11, in vpx_iadst16_1d_rows_msa()
302 l12, l13, l14, l15); in vpx_iadst16_1d_rows_msa()
305 VP9_IADST8x16_1D(l0, l1, l2, l3, l4, l5, l6, l7, l8, l9, l10, l11, l12, l13, in vpx_iadst16_1d_rows_msa()
311 l13 = -r13; in vpx_iadst16_1d_rows_msa()
317 TRANSPOSE8x8_SH_SH(r3, r11, r15, r7, r5, l13, r9, l15, l8, l9, l10, l11, l12, in vpx_iadst16_1d_rows_msa()
318 l13, l14, l15); in vpx_iadst16_1d_rows_msa()
319 ST_SH8(l8, l9, l10, l11, l12, l13, l14, l15, (output + 8), 16); in vpx_iadst16_1d_rows_msa()
/external/llvm/test/CodeGen/PowerPC/
Dppc64-stackmap.ll261 …l5, i64 %l6, i64 %l7, i64 %l8, i64 %l9, i64 %l10, i64 %l11, i64 %l12, i64 %l13, i64 %l14, i64 %l15…
263 …l5, i64 %l6, i64 %l7, i64 %l8, i64 %l9, i64 %l10, i64 %l11, i64 %l12, i64 %l13, i64 %l14, i64 %l15…
280 …l5, i64 %l6, i64 %l7, i64 %l8, i64 %l9, i64 %l10, i64 %l11, i64 %l12, i64 %l13, i64 %l14, i64 %l15…
282 …l5, i64 %l6, i64 %l7, i64 %l8, i64 %l9, i64 %l10, i64 %l11, i64 %l12, i64 %l13, i64 %l14, i64 %l15…
/external/llvm/test/CodeGen/AArch64/
Darm64-stackmap.ll227 …l5, i64 %l6, i64 %l7, i64 %l8, i64 %l9, i64 %l10, i64 %l11, i64 %l12, i64 %l13, i64 %l14, i64 %l15…
229 …l5, i64 %l6, i64 %l7, i64 %l8, i64 %l9, i64 %l10, i64 %l11, i64 %l12, i64 %l13, i64 %l14, i64 %l15…
246 …l5, i64 %l6, i64 %l7, i64 %l8, i64 %l9, i64 %l10, i64 %l11, i64 %l12, i64 %l13, i64 %l14, i64 %l15…
248 …l5, i64 %l6, i64 %l7, i64 %l8, i64 %l9, i64 %l10, i64 %l11, i64 %l12, i64 %l13, i64 %l14, i64 %l15…
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/AArch64/
Darm64-stackmap.ll269 …l5, i64 %l6, i64 %l7, i64 %l8, i64 %l9, i64 %l10, i64 %l11, i64 %l12, i64 %l13, i64 %l14, i64 %l15…
271 …l5, i64 %l6, i64 %l7, i64 %l8, i64 %l9, i64 %l10, i64 %l11, i64 %l12, i64 %l13, i64 %l14, i64 %l15…
291 …l5, i64 %l6, i64 %l7, i64 %l8, i64 %l9, i64 %l10, i64 %l11, i64 %l12, i64 %l13, i64 %l14, i64 %l15…
293 …l5, i64 %l6, i64 %l7, i64 %l8, i64 %l9, i64 %l10, i64 %l11, i64 %l12, i64 %l13, i64 %l14, i64 %l15…
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/PowerPC/
Dppc64-stackmap.ll303 …l5, i64 %l6, i64 %l7, i64 %l8, i64 %l9, i64 %l10, i64 %l11, i64 %l12, i64 %l13, i64 %l14, i64 %l15…
305 …l5, i64 %l6, i64 %l7, i64 %l8, i64 %l9, i64 %l10, i64 %l11, i64 %l12, i64 %l13, i64 %l14, i64 %l15…
325 …l5, i64 %l6, i64 %l7, i64 %l8, i64 %l9, i64 %l10, i64 %l11, i64 %l12, i64 %l13, i64 %l14, i64 %l15…
327 …l5, i64 %l6, i64 %l7, i64 %l8, i64 %l9, i64 %l10, i64 %l11, i64 %l12, i64 %l13, i64 %l14, i64 %l15…

12