Home
last modified time | relevance | path

Searched refs:step1 (Results 1 – 25 of 44) sorted by relevance

12

/external/libvpx/libvpx/vpx_dsp/x86/
Dhighbd_idct32x32_add_sse4.c22 __m128i *const step1 /*step1[16]*/, __m128i *const out /*out[16]*/) { in highbd_idct32_4x32_quarter_2_stage_4_to_6() argument
26 step2[8] = step1[8]; in highbd_idct32_4x32_quarter_2_stage_4_to_6()
27 step2[15] = step1[15]; in highbd_idct32_4x32_quarter_2_stage_4_to_6()
28 highbd_butterfly_sse4_1(step1[14], step1[9], cospi_24_64, cospi_8_64, in highbd_idct32_4x32_quarter_2_stage_4_to_6()
30 highbd_butterfly_sse4_1(step1[13], step1[10], -cospi_8_64, cospi_24_64, in highbd_idct32_4x32_quarter_2_stage_4_to_6()
32 step2[11] = step1[11]; in highbd_idct32_4x32_quarter_2_stage_4_to_6()
33 step2[12] = step1[12]; in highbd_idct32_4x32_quarter_2_stage_4_to_6()
36 step1[8] = _mm_add_epi32(step2[8], step2[11]); in highbd_idct32_4x32_quarter_2_stage_4_to_6()
37 step1[9] = _mm_add_epi32(step2[9], step2[10]); in highbd_idct32_4x32_quarter_2_stage_4_to_6()
38 step1[10] = _mm_sub_epi32(step2[9], step2[10]); in highbd_idct32_4x32_quarter_2_stage_4_to_6()
[all …]
Dhighbd_idct32x32_add_sse2.c18 __m128i *const step1 /*step1[16]*/, __m128i *const out /*out[16]*/) { in highbd_idct32_4x32_quarter_2_stage_4_to_6() argument
22 step2[8] = step1[8]; in highbd_idct32_4x32_quarter_2_stage_4_to_6()
23 step2[15] = step1[15]; in highbd_idct32_4x32_quarter_2_stage_4_to_6()
24 highbd_butterfly_sse2(step1[14], step1[9], cospi_24_64, cospi_8_64, &step2[9], in highbd_idct32_4x32_quarter_2_stage_4_to_6()
26 highbd_butterfly_sse2(step1[10], step1[13], cospi_8_64, cospi_24_64, in highbd_idct32_4x32_quarter_2_stage_4_to_6()
28 step2[11] = step1[11]; in highbd_idct32_4x32_quarter_2_stage_4_to_6()
29 step2[12] = step1[12]; in highbd_idct32_4x32_quarter_2_stage_4_to_6()
32 step1[8] = _mm_add_epi32(step2[8], step2[11]); in highbd_idct32_4x32_quarter_2_stage_4_to_6()
33 step1[9] = _mm_add_epi32(step2[9], step2[10]); in highbd_idct32_4x32_quarter_2_stage_4_to_6()
34 step1[10] = _mm_sub_epi32(step2[9], step2[10]); in highbd_idct32_4x32_quarter_2_stage_4_to_6()
[all …]
Dinv_txfm_ssse3.c57 __m128i step1[8], step2[8]; in idct32_34_8x32_quarter_1() local
60 partial_butterfly_ssse3(in[4], cospi_28_64, cospi_4_64, &step1[4], &step1[7]); in idct32_34_8x32_quarter_1()
64 step2[4] = step1[4]; in idct32_34_8x32_quarter_1()
65 step2[5] = step1[4]; in idct32_34_8x32_quarter_1()
66 step2[6] = step1[7]; in idct32_34_8x32_quarter_1()
67 step2[7] = step1[7]; in idct32_34_8x32_quarter_1()
70 step1[0] = step2[0]; in idct32_34_8x32_quarter_1()
71 step1[1] = step2[0]; in idct32_34_8x32_quarter_1()
72 step1[2] = step2[0]; in idct32_34_8x32_quarter_1()
73 step1[3] = step2[0]; in idct32_34_8x32_quarter_1()
[all …]
Dinv_txfm_sse2.h252 __m128i step1[8], step2[8]; in idct8() local
255 butterfly(in[1], in[7], cospi_28_64, cospi_4_64, &step1[4], &step1[7]); in idct8()
256 butterfly(in[5], in[3], cospi_12_64, cospi_20_64, &step1[5], &step1[6]); in idct8()
262 step2[4] = _mm_add_epi16(step1[4], step1[5]); in idct8()
263 step2[5] = _mm_sub_epi16(step1[4], step1[5]); in idct8()
264 step2[6] = _mm_sub_epi16(step1[7], step1[6]); in idct8()
265 step2[7] = _mm_add_epi16(step1[7], step1[6]); in idct8()
268 step1[0] = _mm_add_epi16(step2[0], step2[3]); in idct8()
269 step1[1] = _mm_add_epi16(step2[1], step2[2]); in idct8()
270 step1[2] = _mm_sub_epi16(step2[1], step2[2]); in idct8()
[all …]
Dhighbd_idct16x16_add_sse4.c57 __m128i step1[16], step2[16]; in vpx_highbd_idct16_4col_sse4_1() local
70 highbd_butterfly_sse4_1(io[2], io[14], cospi_28_64, cospi_4_64, &step1[4], in vpx_highbd_idct16_4col_sse4_1()
71 &step1[7]); in vpx_highbd_idct16_4col_sse4_1()
72 highbd_butterfly_sse4_1(io[10], io[6], cospi_12_64, cospi_20_64, &step1[5], in vpx_highbd_idct16_4col_sse4_1()
73 &step1[6]); in vpx_highbd_idct16_4col_sse4_1()
74 step1[8] = _mm_add_epi32(step2[8], step2[9]); in vpx_highbd_idct16_4col_sse4_1()
75 step1[9] = _mm_sub_epi32(step2[8], step2[9]); in vpx_highbd_idct16_4col_sse4_1()
76 step1[10] = _mm_sub_epi32(step2[11], step2[10]); in vpx_highbd_idct16_4col_sse4_1()
77 step1[11] = _mm_add_epi32(step2[11], step2[10]); in vpx_highbd_idct16_4col_sse4_1()
78 step1[12] = _mm_add_epi32(step2[12], step2[13]); in vpx_highbd_idct16_4col_sse4_1()
[all …]
Dhighbd_idct16x16_add_sse2.c56 __m128i step1[16], step2[16]; in highbd_idct16_4col() local
69 highbd_butterfly_sse2(io[2], io[14], cospi_28_64, cospi_4_64, &step1[4], in highbd_idct16_4col()
70 &step1[7]); in highbd_idct16_4col()
71 highbd_butterfly_sse2(io[10], io[6], cospi_12_64, cospi_20_64, &step1[5], in highbd_idct16_4col()
72 &step1[6]); in highbd_idct16_4col()
73 step1[8] = _mm_add_epi32(step2[8], step2[9]); in highbd_idct16_4col()
74 step1[9] = _mm_sub_epi32(step2[8], step2[9]); in highbd_idct16_4col()
75 step1[10] = _mm_sub_epi32(step2[10], step2[11]); // step1[10] = -step1[10] in highbd_idct16_4col()
76 step1[11] = _mm_add_epi32(step2[10], step2[11]); in highbd_idct16_4col()
77 step1[12] = _mm_add_epi32(step2[13], step2[12]); in highbd_idct16_4col()
[all …]
Dinv_txfm_ssse3.h34 __m128i step1[8], step2[8], tmp[4]; in idct8x8_12_add_kernel_ssse3() local
47 step1[4] = _mm_mulhrs_epi16(tmp[1], cp_28d_4d); // step1 4&7 in idct8x8_12_add_kernel_ssse3()
48 step1[5] = _mm_mulhrs_epi16(tmp[3], cp_n20d_12d); // step1 5&6 in idct8x8_12_add_kernel_ssse3()
53 step2[4] = _mm_add_epi16(step1[4], step1[5]); // step2 4&7 in idct8x8_12_add_kernel_ssse3()
54 step2[5] = _mm_sub_epi16(step1[4], step1[5]); // step2 5&6 in idct8x8_12_add_kernel_ssse3()
59 step1[5] = idct_calc_wraplow_sse2(cp_16_n16, cp_16_16, tmp[0]); // step1 5&6 in idct8x8_12_add_kernel_ssse3()
62 step1[2] = _mm_unpackhi_epi64(tmp[1], tmp[0]); // step1 2&1 in idct8x8_12_add_kernel_ssse3()
63 step1[3] = _mm_unpacklo_epi64(tmp[1], tmp[0]); // step1 3&0 in idct8x8_12_add_kernel_ssse3()
66 tmp[0] = _mm_add_epi16(step1[3], step2[4]); // output 3&0 in idct8x8_12_add_kernel_ssse3()
67 tmp[1] = _mm_add_epi16(step1[2], step1[5]); // output 2&1 in idct8x8_12_add_kernel_ssse3()
[all …]
Dhighbd_idct8x8_add_sse4.c21 __m128i step1[8], step2[8]; in vpx_highbd_idct8x8_half1d_sse4_1() local
26 step1[0] = io[0]; in vpx_highbd_idct8x8_half1d_sse4_1()
27 step1[2] = io[4]; in vpx_highbd_idct8x8_half1d_sse4_1()
28 step1[1] = io[2]; in vpx_highbd_idct8x8_half1d_sse4_1()
29 step1[3] = io[6]; in vpx_highbd_idct8x8_half1d_sse4_1()
30 highbd_butterfly_sse4_1(io[1], io[7], cospi_28_64, cospi_4_64, &step1[4], in vpx_highbd_idct8x8_half1d_sse4_1()
31 &step1[7]); in vpx_highbd_idct8x8_half1d_sse4_1()
32 highbd_butterfly_sse4_1(io[5], io[3], cospi_12_64, cospi_20_64, &step1[5], in vpx_highbd_idct8x8_half1d_sse4_1()
33 &step1[6]); in vpx_highbd_idct8x8_half1d_sse4_1()
36 highbd_butterfly_cospi16_sse4_1(step1[0], step1[2], &step2[0], &step2[1]); in vpx_highbd_idct8x8_half1d_sse4_1()
[all …]
Dhighbd_idct8x8_add_sse2.c19 __m128i step1[8], step2[8]; in highbd_idct8x8_half1d() local
24 step1[0] = io[0]; in highbd_idct8x8_half1d()
25 step1[2] = io[4]; in highbd_idct8x8_half1d()
26 step1[1] = io[2]; in highbd_idct8x8_half1d()
27 step1[3] = io[6]; in highbd_idct8x8_half1d()
28 highbd_butterfly_sse2(io[1], io[7], cospi_28_64, cospi_4_64, &step1[4], in highbd_idct8x8_half1d()
29 &step1[7]); in highbd_idct8x8_half1d()
30 highbd_butterfly_sse2(io[5], io[3], cospi_12_64, cospi_20_64, &step1[5], in highbd_idct8x8_half1d()
31 &step1[6]); in highbd_idct8x8_half1d()
34 highbd_butterfly_cospi16_sse2(step1[0], step1[2], &step2[0], &step2[1]); in highbd_idct8x8_half1d()
[all …]
Dinv_txfm_sse2.c893 __m128i step1[8], step2[8]; in idct32_34_8x32_quarter_1() local
896 butterfly(in[4], zero, cospi_28_64, cospi_4_64, &step1[4], &step1[7]); in idct32_34_8x32_quarter_1()
900 step2[4] = step1[4]; in idct32_34_8x32_quarter_1()
901 step2[5] = step1[4]; in idct32_34_8x32_quarter_1()
902 step2[6] = step1[7]; in idct32_34_8x32_quarter_1()
903 step2[7] = step1[7]; in idct32_34_8x32_quarter_1()
906 step1[0] = step2[0]; in idct32_34_8x32_quarter_1()
907 step1[1] = step2[0]; in idct32_34_8x32_quarter_1()
908 step1[2] = step2[0]; in idct32_34_8x32_quarter_1()
909 step1[3] = step2[0]; in idct32_34_8x32_quarter_1()
[all …]
Dfwd_dct32x32_impl_sse2.h115 __m128i step1[32]; in FDCT32x32_2D() local
130 __m128i *step1a = &step1[0]; in FDCT32x32_2D()
131 __m128i *step1b = &step1[31]; in FDCT32x32_2D()
160 __m128i *step1a = &step1[4]; in FDCT32x32_2D()
161 __m128i *step1b = &step1[27]; in FDCT32x32_2D()
190 __m128i *step1a = &step1[8]; in FDCT32x32_2D()
191 __m128i *step1b = &step1[23]; in FDCT32x32_2D()
220 __m128i *step1a = &step1[12]; in FDCT32x32_2D()
221 __m128i *step1b = &step1[19]; in FDCT32x32_2D()
264 step1[0] = ADD_EPI16(in00, in31); in FDCT32x32_2D()
[all …]
Dfwd_dct32x32_impl_avx2.h100 __m256i step1[32]; in FDCT32x32_2D_AVX2() local
115 __m256i *step1a = &step1[0]; in FDCT32x32_2D_AVX2()
116 __m256i *step1b = &step1[31]; in FDCT32x32_2D_AVX2()
151 __m256i *step1a = &step1[4]; in FDCT32x32_2D_AVX2()
152 __m256i *step1b = &step1[27]; in FDCT32x32_2D_AVX2()
187 __m256i *step1a = &step1[8]; in FDCT32x32_2D_AVX2()
188 __m256i *step1b = &step1[23]; in FDCT32x32_2D_AVX2()
223 __m256i *step1a = &step1[12]; in FDCT32x32_2D_AVX2()
224 __m256i *step1b = &step1[19]; in FDCT32x32_2D_AVX2()
273 step1[0] = _mm256_add_epi16(in00, in31); in FDCT32x32_2D_AVX2()
[all …]
/external/libvpx/libvpx/vpx_dsp/arm/
Didct16x16_add_neon.c75 int16x8_t in[16], step1[16], step2[16], out[16]; in vpx_idct16x16_256_add_half1d() local
153 step1[0] = in[0 / 2]; in vpx_idct16x16_256_add_half1d()
154 step1[1] = in[16 / 2]; in vpx_idct16x16_256_add_half1d()
155 step1[2] = in[8 / 2]; in vpx_idct16x16_256_add_half1d()
156 step1[3] = in[24 / 2]; in vpx_idct16x16_256_add_half1d()
157 step1[4] = in[4 / 2]; in vpx_idct16x16_256_add_half1d()
158 step1[5] = in[20 / 2]; in vpx_idct16x16_256_add_half1d()
159 step1[6] = in[12 / 2]; in vpx_idct16x16_256_add_half1d()
160 step1[7] = in[28 / 2]; in vpx_idct16x16_256_add_half1d()
161 step1[8] = in[2 / 2]; in vpx_idct16x16_256_add_half1d()
[all …]
Dhighbd_idct16x16_add_neon.c526 int32x4x2_t in[16], step1[16], step2[16], out[16]; in vpx_highbd_idct16x16_256_add_half1d() local
584 step1[0] = in[0 / 2]; in vpx_highbd_idct16x16_256_add_half1d()
585 step1[1] = in[16 / 2]; in vpx_highbd_idct16x16_256_add_half1d()
586 step1[2] = in[8 / 2]; in vpx_highbd_idct16x16_256_add_half1d()
587 step1[3] = in[24 / 2]; in vpx_highbd_idct16x16_256_add_half1d()
588 step1[4] = in[4 / 2]; in vpx_highbd_idct16x16_256_add_half1d()
589 step1[5] = in[20 / 2]; in vpx_highbd_idct16x16_256_add_half1d()
590 step1[6] = in[12 / 2]; in vpx_highbd_idct16x16_256_add_half1d()
591 step1[7] = in[28 / 2]; in vpx_highbd_idct16x16_256_add_half1d()
592 step1[8] = in[2 / 2]; in vpx_highbd_idct16x16_256_add_half1d()
[all …]
Dhighbd_idct8x8_add_neon.c76 int32x4_t step1[8], step2[8]; in idct8x8_12_half1d_bd10() local
81 step1[4] = vmulq_lane_s32(*io1, vget_high_s32(cospis1), 1); in idct8x8_12_half1d_bd10()
82 step1[5] = vmulq_lane_s32(*io3, vget_high_s32(cospis1), 0); in idct8x8_12_half1d_bd10()
83 step1[6] = vmulq_lane_s32(*io3, vget_low_s32(cospis1), 1); in idct8x8_12_half1d_bd10()
84 step1[7] = vmulq_lane_s32(*io1, vget_low_s32(cospis1), 0); in idct8x8_12_half1d_bd10()
85 step1[4] = vrshrq_n_s32(step1[4], DCT_CONST_BITS); in idct8x8_12_half1d_bd10()
86 step1[5] = vrshrq_n_s32(step1[5], DCT_CONST_BITS); in idct8x8_12_half1d_bd10()
87 step1[6] = vrshrq_n_s32(step1[6], DCT_CONST_BITS); in idct8x8_12_half1d_bd10()
88 step1[7] = vrshrq_n_s32(step1[7], DCT_CONST_BITS); in idct8x8_12_half1d_bd10()
98 step2[4] = vaddq_s32(step1[4], step1[5]); in idct8x8_12_half1d_bd10()
[all …]
Dhighbd_idct_neon.h171 int32x4_t step1[8], step2[8]; in idct8x8_64_half1d_bd10() local
176 step1[4] = vmulq_lane_s32(*io1, vget_high_s32(cospis1), 1); in idct8x8_64_half1d_bd10()
177 step1[5] = vmulq_lane_s32(*io3, vget_high_s32(cospis1), 0); in idct8x8_64_half1d_bd10()
178 step1[6] = vmulq_lane_s32(*io3, vget_low_s32(cospis1), 1); in idct8x8_64_half1d_bd10()
179 step1[7] = vmulq_lane_s32(*io1, vget_low_s32(cospis1), 0); in idct8x8_64_half1d_bd10()
181 step1[4] = vmlsq_lane_s32(step1[4], *io7, vget_low_s32(cospis1), 0); in idct8x8_64_half1d_bd10()
182 step1[5] = vmlaq_lane_s32(step1[5], *io5, vget_low_s32(cospis1), 1); in idct8x8_64_half1d_bd10()
183 step1[6] = vmlsq_lane_s32(step1[6], *io5, vget_high_s32(cospis1), 0); in idct8x8_64_half1d_bd10()
184 step1[7] = vmlaq_lane_s32(step1[7], *io7, vget_high_s32(cospis1), 1); in idct8x8_64_half1d_bd10()
186 step1[4] = vrshrq_n_s32(step1[4], DCT_CONST_BITS); in idct8x8_64_half1d_bd10()
[all …]
Didct_neon.h323 int16x4_t step1[8], step2[8]; in idct8x8_12_pass1_bd8() local
329 step1[4] = vqrdmulh_lane_s16(io[1], cospisd1, 3); in idct8x8_12_pass1_bd8()
330 step1[5] = vqrdmulh_lane_s16(io[3], cospisd1, 2); in idct8x8_12_pass1_bd8()
331 step1[6] = vqrdmulh_lane_s16(io[3], cospisd1, 1); in idct8x8_12_pass1_bd8()
332 step1[7] = vqrdmulh_lane_s16(io[1], cospisd1, 0); in idct8x8_12_pass1_bd8()
339 step2[4] = vadd_s16(step1[4], step1[5]); in idct8x8_12_pass1_bd8()
340 step2[5] = vsub_s16(step1[4], step1[5]); in idct8x8_12_pass1_bd8()
341 step2[6] = vsub_s16(step1[7], step1[6]); in idct8x8_12_pass1_bd8()
342 step2[7] = vadd_s16(step1[7], step1[6]); in idct8x8_12_pass1_bd8()
345 step1[0] = vadd_s16(step2[1], step2[3]); in idct8x8_12_pass1_bd8()
[all …]
/external/libaom/libaom/av1/common/arm/
Dav1_inv_txfm_neon.c397 int16x8_t step1[8], step2[8]; in idct8_new_neon() local
406 btf_16_lane_0_1_neon(in[1], in[7], c0, &step1[7], &step1[4]); in idct8_new_neon()
407 btf_16_lane_2_3_neon(in[5], in[3], c0, &step1[6], &step1[5]); in idct8_new_neon()
412 step2[4] = vqaddq_s16(step1[4], step1[5]); in idct8_new_neon()
413 step2[5] = vqsubq_s16(step1[4], step1[5]); in idct8_new_neon()
414 step2[6] = vqsubq_s16(step1[7], step1[6]); in idct8_new_neon()
415 step2[7] = vqaddq_s16(step1[7], step1[6]); in idct8_new_neon()
418 step1[0] = vqaddq_s16(step2[0], step2[3]); in idct8_new_neon()
419 step1[1] = vqaddq_s16(step2[1], step2[2]); in idct8_new_neon()
420 step1[2] = vqsubq_s16(step2[1], step2[2]); in idct8_new_neon()
[all …]
/external/libvpx/libvpx/vpx_dsp/
Dinv_txfm.c272 int16_t step1[8], step2[8]; in idct8_c() local
276 step1[0] = (int16_t)input[0]; in idct8_c()
277 step1[2] = (int16_t)input[4]; in idct8_c()
278 step1[1] = (int16_t)input[2]; in idct8_c()
279 step1[3] = (int16_t)input[6]; in idct8_c()
282 step1[4] = WRAPLOW(dct_const_round_shift(temp1)); in idct8_c()
283 step1[7] = WRAPLOW(dct_const_round_shift(temp2)); in idct8_c()
286 step1[5] = WRAPLOW(dct_const_round_shift(temp1)); in idct8_c()
287 step1[6] = WRAPLOW(dct_const_round_shift(temp2)); in idct8_c()
290 temp1 = (step1[0] + step1[2]) * cospi_16_64; in idct8_c()
[all …]
Dfwd_txfm.c197 tran_high_t step1[8]; // canbe16 in vpx_fdct16x16_c() local
215 step1[0] = (input[7 * stride] - input[8 * stride]) * 4; in vpx_fdct16x16_c()
216 step1[1] = (input[6 * stride] - input[9 * stride]) * 4; in vpx_fdct16x16_c()
217 step1[2] = (input[5 * stride] - input[10 * stride]) * 4; in vpx_fdct16x16_c()
218 step1[3] = (input[4 * stride] - input[11 * stride]) * 4; in vpx_fdct16x16_c()
219 step1[4] = (input[3 * stride] - input[12 * stride]) * 4; in vpx_fdct16x16_c()
220 step1[5] = (input[2 * stride] - input[13 * stride]) * 4; in vpx_fdct16x16_c()
221 step1[6] = (input[1 * stride] - input[14 * stride]) * 4; in vpx_fdct16x16_c()
222 step1[7] = (input[0 * stride] - input[15 * stride]) * 4; in vpx_fdct16x16_c()
235 step1[0] = ((in_low[7 * 16] + 1) >> 2) - ((in_low[8 * 16] + 1) >> 2); in vpx_fdct16x16_c()
[all …]
/external/libvpx/libvpx/vp9/encoder/
Dvp9_dct.c95 tran_high_t step1[8]; // canbe16 in fdct16() local
111 step1[0] = in[7] - in[8]; in fdct16()
112 step1[1] = in[6] - in[9]; in fdct16()
113 step1[2] = in[5] - in[10]; in fdct16()
114 step1[3] = in[4] - in[11]; in fdct16()
115 step1[4] = in[3] - in[12]; in fdct16()
116 step1[5] = in[2] - in[13]; in fdct16()
117 step1[6] = in[1] - in[14]; in fdct16()
118 step1[7] = in[0] - in[15]; in fdct16()
174 temp1 = (step1[5] - step1[2]) * cospi_16_64; in fdct16()
[all …]
/external/autotest/client/tests/error_skip_step/
Dcontrol7 DOC = """Raise TestNAError during step1. step0, 2 and 3 should run."""
16 job.next_step('step1')
24 def step1():
34 print 'at least I can run. unlike step1.'
/external/autotest/frontend/afe/doctests/
D003_misc_rpc_features.txt19 job.next_step('step1')
27 def step1():
49 job.next_step('step1')
56 def step1():
75 job.next_step('step1')
82 def step1():
/external/llvm/test/Transforms/IndVarSimplify/
Dno-iv-rewrite.ll282 ; Two increments should remain, one by %step and one by %step1.
303 %step1 = add i32 %step, 1
304 %init1 = add i32 %init, %step1
305 %l.0 = sub i32 %init1, %step1
314 %ii.next = add i32 %ii, %step1
315 %j.next = add i32 %j, %step1
316 %k.next = add i32 %k, %step1
/external/swiftshader/third_party/LLVM/test/Transforms/IndVarSimplify/
Dno-iv-rewrite.ll277 ; Two increments should remain, one by %step and one by %step1.
298 %step1 = add i32 %step, 1
299 %init1 = add i32 %init, %step1
300 %l.0 = sub i32 %init1, %step1
309 %ii.next = add i32 %ii, %step1
310 %j.next = add i32 %j, %step1
311 %k.next = add i32 %k, %step1

12