Home
last modified time | relevance | path

Searched full:vs2 (Results 1 – 25 of 296) sorted by relevance

12345678910>>...12

/external/rust/crates/libz-sys/src/zlib-ng/arch/x86/
Dadler32_avx512_vnni.c52 __m512i vs1, vs2; in adler32_avx512_vnni() local
56 vs2 = _mm512_zextsi128_si512(_mm_cvtsi32_si128(adler1)); in adler32_avx512_vnni()
76 vs2 = _mm512_dpbusd_epi32(vs2, vbuf1, dot2v); in adler32_avx512_vnni()
84 vs2 = sum2 + 64 vs1 + sum( (64-i+1) c[i] ) in adler32_avx512_vnni()
96 vs2 = _mm512_dpbusd_epi32(vs2, vbuf0, dot2v); in adler32_avx512_vnni()
106 vs2 = _mm512_add_epi32(vs2, vs3); in adler32_avx512_vnni()
107 vs2 = _mm512_add_epi32(vs2, vs2_1); in adler32_avx512_vnni()
110 adler1 = _mm512_reduce_add_epu32(vs2) % BASE; in adler32_avx512_vnni()
149 __m256i vs1, vs2; in adler32_fold_copy_avx512_vnni() local
153 vs2 = _mm256_zextsi128_si256(_mm_cvtsi32_si128(adler1)); in adler32_fold_copy_avx512_vnni()
[all …]
Dadler32_sse42.c28 __m128i vs1_0, vs3, vs1, vs2, vs2_0, v_sad_sum1, v_short_sum2, v_short_sum2_0, in adler32_fold_copy_sse42() local
43 vs2 = _mm_cvtsi32_si128(adler1); in adler32_fold_copy_sse42()
52 vs2 = sum2 + 16 vs1 + sum( (16-i+1) c[i] ) in adler32_fold_copy_sse42()
74 vs2 = _mm_add_epi32(vsum2, vs2); in adler32_fold_copy_sse42()
79 vs2 = _mm_add_epi32(vs2_0, vs2); in adler32_fold_copy_sse42()
81 vs2 = _mm_add_epi32(vs3, vs2); in adler32_fold_copy_sse42()
87 vs2 = sum2 + 16 vs1 + sum( (16-i+1) c[i] ) in adler32_fold_copy_sse42()
99 vs2 = _mm_add_epi32(vsum2, vs2); in adler32_fold_copy_sse42()
107 vs2 = _mm_add_epi32(vs2, vs3); in adler32_fold_copy_sse42()
110 adler1 = hsum(vs2) % BASE; in adler32_fold_copy_sse42()
Dadler32_ssse3.c41 __m128i vbuf, vs1_0, vs3, vs1, vs2, vs2_0, v_sad_sum1, v_short_sum2, v_short_sum2_0, in adler32_ssse3() local
63 vs2 = _mm_cvtsi32_si128(sum2); in adler32_ssse3()
83 vs2 = _mm_cvtsi32_si128(sum2); in adler32_ssse3()
95 vs2 = sum2 + 16 vs1 + sum( (16-i+1) c[i] ) in adler32_ssse3()
111 vs2 = _mm_add_epi32(vsum2, vs2); in adler32_ssse3()
117 vs2 = _mm_add_epi32(vs2_0, vs2); in adler32_ssse3()
119 vs2 = _mm_add_epi32(vs3, vs2); in adler32_ssse3()
125 vs2 = sum2 + 16 vs1 + sum( (16-i+1) c[i] ) in adler32_ssse3()
137 vs2 = _mm_add_epi32(vsum2, vs2); in adler32_ssse3()
142 vs2 = _mm_add_epi32(vs2, vs3); in adler32_ssse3()
[all …]
Dadler32_avx512_tpl.h59 __m512i vs2 = _mm512_zextsi128_si512(_mm_cvtsi32_si128(adler1)); local
70 vs2 = sum2 + 64 vs1 + sum( (64-i+1) c[i] )
85 vs2 = _mm512_add_epi32(vsum2, vs2);
90 vs2 = _mm512_add_epi32(vs2, vs3);
93 adler1 = _mm512_reduce_add_epu32(vs2) % BASE;
Dadler32_avx2_tpl.h50 __m256i vs1, vs2; local
59 vs2 = _mm256_zextsi128_si256(_mm_cvtsi32_si128(adler1));
70 vs2 = sum2 + 32 vs1 + sum( (32-i+1) c[i] )
86 vs2 = _mm256_add_epi32(vsum2, vs2);
92 vs2 = _mm256_add_epi32(vs2, vs3);
130 adler1 = hsum256(vs2) % BASE;
/external/rust/crates/libz-sys/src/zlib-ng/arch/power/
Dadler32_power8.c78 vector unsigned int vs2 = { 0 }; in adler32_power8() local
85 vs2[0] = s2; in adler32_power8()
100 vs2 = vec_add(vsum2, vs2); in adler32_power8()
107 vs2 = vec_add(vs1_save, vs2); in adler32_power8()
108 vs2 = vec_sumsu(vs2, vsum2); in adler32_power8()
112 /* vs2[0] = s2_i + 16*s1_save + in adler32_power8()
114 vs2[0] = vs2[0] % BASE; in adler32_power8()
117 vs2 = vec_and(vs2, vmask); in adler32_power8()
135 vs2 = vec_add(vsum2, vs2); in adler32_power8()
142 vs2 = vec_add(vs1_save, vs2); in adler32_power8()
[all …]
/external/swiftshader/third_party/llvm-16.0/llvm/lib/Target/RISCV/
DRISCVInstrInfoV.td227 // indexed load vd, (rs1), vs2, vm
231 (ins GPRMem:$rs1, VR:$vs2, VMaskOp:$vm), opcodestr,
232 "$vd, (${rs1}), $vs2$vm">;
253 // indexed segment load vd, (rs1), vs2, vm
258 (ins GPRMem:$rs1, VR:$vs2, VMaskOp:$vm), opcodestr,
259 "$vd, (${rs1}), $vs2$vm">;
291 // indexed store vd, vs3, (rs1), vs2, vm
294 (ins VR:$vs3, GPRMem:$rs1, VR:$vs2, VMaskOp:$vm),
295 opcodestr, "$vs3, (${rs1}), $vs2$vm">;
309 // segment store vd, vs3, (rs1), vs2, vm
[all …]
DRISCVInstrFormatsV.td110 bits<5> vs2;
117 let Inst{24-20} = vs2;
130 bits<5> vs2;
137 let Inst{24-20} = vs2;
147 class RVInstV2<bits<6> funct6, bits<5> vs2, RISCVVFormat opv, dag outs, dag ins,
156 let Inst{24-20} = vs2;
169 bits<5> vs2;
176 let Inst{24-20} = vs2;
189 bits<5> vs2;
195 let Inst{24-20} = vs2;
[all …]
DRISCVInstrInfoXTHead.td33 // op vd, vs1, vs2, vm (reverse the order of vs1 and vs2)
36 (ins VR:$vs1, VR:$vs2, VMaskOp:$vm),
37 opcodestr, "$vd, $vs1, $vs2$vm">;
39 // op vd, rs1, vs2, vm (reverse the order of rs1 and vs2)
42 (ins GPR:$rs1, VR:$vs2, VMaskOp:$vm),
43 opcodestr, "$vd, $rs1, $vs2$vm">;
/external/go-cmp/cmp/internal/value/
Dsort.go25 vs2 := vs[:1]
27 if isLess(vs2[len(vs2)-1], v) {
28 vs2 = append(vs2, v)
31 return vs2
/external/XNNPACK/src/f32-velu/gen/
Dvelu-scalar-rr2-lut16-p3-x3.c70 float vs2 = uint32_as_float(xnn_table_exp2minus_k_over_16[vidx2] + ven2); in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x3() local
84 vs2 = 0.0f; in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x3()
100 vt2 *= vs2; in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x3()
101 vs2 -= vone; in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x3()
111 const float ve2 = (vp2 + vs2) * valpha; in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x3()
Dvelu-scalar-rr2-p6-x3.c59 float vs2 = uint32_as_float(float_as_uint32(vn2) << 23); in xnn_f32_velu_ukernel__scalar_rr2_p6_x3() local
79 vs2 = 0.0f; in xnn_f32_velu_ukernel__scalar_rr2_p6_x3()
107 vt2 *= vs2; in xnn_f32_velu_ukernel__scalar_rr2_p6_x3()
108 vs2 -= vone; in xnn_f32_velu_ukernel__scalar_rr2_p6_x3()
118 const float ve2 = (vp2 + vs2) * valpha; in xnn_f32_velu_ukernel__scalar_rr2_p6_x3()
Dvelu-scalar-rr2-lut16-p3-x4.c76 float vs2 = uint32_as_float(xnn_table_exp2minus_k_over_16[vidx2] + ven2); in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x4() local
92 vs2 = 0.0f; in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x4()
115 vt2 *= vs2; in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x4()
116 vs2 -= vone; in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x4()
129 const float ve2 = (vp2 + vs2) * valpha; in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x4()
Dvelu-scalar-rr2-p6-x4.c62 float vs2 = uint32_as_float(float_as_uint32(vn2) << 23); in xnn_f32_velu_ukernel__scalar_rr2_p6_x4() local
86 vs2 = 0.0f; in xnn_f32_velu_ukernel__scalar_rr2_p6_x4()
123 vt2 *= vs2; in xnn_f32_velu_ukernel__scalar_rr2_p6_x4()
124 vs2 -= vone; in xnn_f32_velu_ukernel__scalar_rr2_p6_x4()
137 const float ve2 = (vp2 + vs2) * valpha; in xnn_f32_velu_ukernel__scalar_rr2_p6_x4()
Dvelu-scalar-rr2-lut16-p3-x5.c82 float vs2 = uint32_as_float(xnn_table_exp2minus_k_over_16[vidx2] + ven2); in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x5() local
100 vs2 = 0.0f; in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x5()
130 vt2 *= vs2; in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x5()
131 vs2 -= vone; in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x5()
147 const float ve2 = (vp2 + vs2) * valpha; in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x5()
Dvelu-avx512f-rr1-p6-x48.c59 __m512 vs2 = _mm512_castsi512_ps(_mm512_slli_epi32(_mm512_castps_si512(vn2), 23)); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x48() local
87 vt2 = _mm512_mul_ps(vt2, vs2); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x48()
91 vs2 = _mm512_fmsub_ps(vs2, valpha, valpha); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x48()
102 __m512 vy2 = _mm512_fmadd_ps(vp2, valpha, vs2); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x48()
Dvelu-avx512f-rr1-lut16-p3-perm-x48.c64 __m512 vs2 = _mm512_castsi512_ps(_mm512_add_epi32(vl2, ven2)); in xnn_f32_velu_ukernel__avx512f_rr1_lut16_p3_perm_x48() local
80 vt2 = _mm512_mul_ps(vt2, vs2); in xnn_f32_velu_ukernel__avx512f_rr1_lut16_p3_perm_x48()
84 vs2 = _mm512_fmsub_ps(vs2, valpha, valpha); in xnn_f32_velu_ukernel__avx512f_rr1_lut16_p3_perm_x48()
95 __m512 vy2 = _mm512_fmadd_ps(vp2, valpha, vs2); in xnn_f32_velu_ukernel__avx512f_rr1_lut16_p3_perm_x48()
Dvelu-scalar-rr2-p6-x5.c65 float vs2 = uint32_as_float(float_as_uint32(vn2) << 23); in xnn_f32_velu_ukernel__scalar_rr2_p6_x5() local
93 vs2 = 0.0f; in xnn_f32_velu_ukernel__scalar_rr2_p6_x5()
139 vt2 *= vs2; in xnn_f32_velu_ukernel__scalar_rr2_p6_x5()
140 vs2 -= vone; in xnn_f32_velu_ukernel__scalar_rr2_p6_x5()
156 const float ve2 = (vp2 + vs2) * valpha; in xnn_f32_velu_ukernel__scalar_rr2_p6_x5()
Dvelu-wasm-rr2-lut16-p3-x3.c70 float vs2 = uint32_as_float(xnn_table_exp2minus_k_over_16[vidx2] + ven2); in xnn_f32_velu_ukernel__wasm_rr2_lut16_p3_x3() local
88 vt2 *= vs2; in xnn_f32_velu_ukernel__wasm_rr2_lut16_p3_x3()
89 vs2 -= vone; in xnn_f32_velu_ukernel__wasm_rr2_lut16_p3_x3()
99 const float ve2 = (vp2 + vs2) * valpha; in xnn_f32_velu_ukernel__wasm_rr2_lut16_p3_x3()
Dvelu-scalar-rr2-lut16-p3-x6.c88 float vs2 = uint32_as_float(xnn_table_exp2minus_k_over_16[vidx2] + ven2); in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x6() local
108 vs2 = 0.0f; in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x6()
145 vt2 *= vs2; in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x6()
146 vs2 -= vone; in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x6()
165 const float ve2 = (vp2 + vs2) * valpha; in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x6()
Dvelu-avx2-rr1-lut4-p4-perm-x24.c66 __m256 vs2 = _mm256_castsi256_ps(_mm256_add_epi32(vl2, ven2)); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x24() local
82 vt2 = _mm256_mul_ps(vt2, vs2); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x24()
88 vs2 = _mm256_fmsub_ps(vs2, valpha, valpha); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x24()
95 const __m256 ve2 = _mm256_fmadd_ps(vp2, valpha, vs2); in xnn_f32_velu_ukernel__avx2_rr1_lut4_p4_perm_x24()
Dvelu-scalar-rr2-p6-x6.c68 float vs2 = uint32_as_float(float_as_uint32(vn2) << 23); in xnn_f32_velu_ukernel__scalar_rr2_p6_x6() local
100 vs2 = 0.0f; in xnn_f32_velu_ukernel__scalar_rr2_p6_x6()
155 vt2 *= vs2; in xnn_f32_velu_ukernel__scalar_rr2_p6_x6()
156 vs2 -= vone; in xnn_f32_velu_ukernel__scalar_rr2_p6_x6()
175 const float ve2 = (vp2 + vs2) * valpha; in xnn_f32_velu_ukernel__scalar_rr2_p6_x6()
Dvelu-avx2-rr1-lut8-p4-perm-x24.c66 __m256 vs2 = _mm256_castsi256_ps(_mm256_add_epi32(vl2, ven2)); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x24() local
82 vt2 = _mm256_mul_ps(vt2, vs2); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x24()
88 vs2 = _mm256_fmsub_ps(vs2, valpha, valpha); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x24()
95 const __m256 ve2 = _mm256_fmadd_ps(vp2, valpha, vs2); in xnn_f32_velu_ukernel__avx2_rr1_lut8_p4_perm_x24()
Dvelu-avx2-rr1-lut16-p3-gather-x24.c71 __m256 vs2 = _mm256_castsi256_ps(_mm256_add_epi32(vl2, ven2)); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x24() local
83 vt2 = _mm256_mul_ps(vt2, vs2); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x24()
89 vs2 = _mm256_fmsub_ps(vs2, valpha, valpha); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x24()
96 const __m256 ve2 = _mm256_fmadd_ps(vp2, valpha, vs2); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x24()
/external/tensorflow/tensorflow/python/kernel_tests/variables/
Dpartitioned_variables_test.py350 vs2 = partitioned_variables.create_partitioned_variables([2, 4], [1, 2],
354 var2_name = vs2[0]._save_slice_info.full_name
359 self.assertEqual(var2_name + "/part_0:0", vs2[0].name)
360 self.assertEqual(var2_name + "/part_1:0", vs2[1].name)
370 vs2 = partitioned_variables.create_partitioned_variables(
374 var2_name = vs2[0]._save_slice_info.full_name
379 self.assertEqual(var2_name + "/part_0:0", vs2[0].name)
380 self.assertEqual(var2_name + "/part_1:0", vs2[1].name)
387 vs2 = partitioned_variables.create_partitioned_variables([2, 4], [1, 2],
391 var2_name = vs2[0]._save_slice_info.full_name
[all …]

12345678910>>...12