/external/XNNPACK/src/f32-velu/gen/ |
D | velu-wasm-rr2-p6-x2.c | 67 float vp1 = vc6 * vt1 + vc5; in xnn_f32_velu_ukernel__wasm_rr2_p6_x2() local 70 vp1 = vp1 * vt1 + vc4; in xnn_f32_velu_ukernel__wasm_rr2_p6_x2() 73 vp1 = vp1 * vt1 + vc3; in xnn_f32_velu_ukernel__wasm_rr2_p6_x2() 76 vp1 = vp1 * vt1 + vc2; in xnn_f32_velu_ukernel__wasm_rr2_p6_x2() 79 vp1 *= vt1; in xnn_f32_velu_ukernel__wasm_rr2_p6_x2() 87 vp1 = vp1 * vt1 + vt1; in xnn_f32_velu_ukernel__wasm_rr2_p6_x2() 91 const float ve1 = (vp1 + vs1) * valpha; in xnn_f32_velu_ukernel__wasm_rr2_p6_x2()
|
D | velu-scalar-rr2-p6-x2.c | 75 float vp1 = vc6 * vt1 + vc5; in xnn_f32_velu_ukernel__scalar_rr2_p6_x2() local 78 vp1 = vp1 * vt1 + vc4; in xnn_f32_velu_ukernel__scalar_rr2_p6_x2() 81 vp1 = vp1 * vt1 + vc3; in xnn_f32_velu_ukernel__scalar_rr2_p6_x2() 84 vp1 = vp1 * vt1 + vc2; in xnn_f32_velu_ukernel__scalar_rr2_p6_x2() 87 vp1 *= vt1; in xnn_f32_velu_ukernel__scalar_rr2_p6_x2() 95 vp1 = vp1 * vt1 + vt1; in xnn_f32_velu_ukernel__scalar_rr2_p6_x2() 99 const float ve1 = (vp1 + vs1) * valpha; in xnn_f32_velu_ukernel__scalar_rr2_p6_x2()
|
D | velu-scalar-rr2-p6-x3.c | 86 float vp1 = vc6 * vt1 + vc5; in xnn_f32_velu_ukernel__scalar_rr2_p6_x3() local 90 vp1 = vp1 * vt1 + vc4; in xnn_f32_velu_ukernel__scalar_rr2_p6_x3() 94 vp1 = vp1 * vt1 + vc3; in xnn_f32_velu_ukernel__scalar_rr2_p6_x3() 98 vp1 = vp1 * vt1 + vc2; in xnn_f32_velu_ukernel__scalar_rr2_p6_x3() 102 vp1 *= vt1; in xnn_f32_velu_ukernel__scalar_rr2_p6_x3() 113 vp1 = vp1 * vt1 + vt1; in xnn_f32_velu_ukernel__scalar_rr2_p6_x3() 118 const float ve1 = (vp1 + vs1) * valpha; in xnn_f32_velu_ukernel__scalar_rr2_p6_x3()
|
D | velu-wasm-rr2-p6-x3.c | 74 float vp1 = vc6 * vt1 + vc5; in xnn_f32_velu_ukernel__wasm_rr2_p6_x3() local 78 vp1 = vp1 * vt1 + vc4; in xnn_f32_velu_ukernel__wasm_rr2_p6_x3() 82 vp1 = vp1 * vt1 + vc3; in xnn_f32_velu_ukernel__wasm_rr2_p6_x3() 86 vp1 = vp1 * vt1 + vc2; in xnn_f32_velu_ukernel__wasm_rr2_p6_x3() 90 vp1 *= vt1; in xnn_f32_velu_ukernel__wasm_rr2_p6_x3() 101 vp1 = vp1 * vt1 + vt1; in xnn_f32_velu_ukernel__wasm_rr2_p6_x3() 106 const float ve1 = (vp1 + vs1) * valpha; in xnn_f32_velu_ukernel__wasm_rr2_p6_x3()
|
D | velu-wasm-rr2-p6-x4.c | 81 float vp1 = vc6 * vt1 + vc5; in xnn_f32_velu_ukernel__wasm_rr2_p6_x4() local 86 vp1 = vp1 * vt1 + vc4; in xnn_f32_velu_ukernel__wasm_rr2_p6_x4() 91 vp1 = vp1 * vt1 + vc3; in xnn_f32_velu_ukernel__wasm_rr2_p6_x4() 96 vp1 = vp1 * vt1 + vc2; in xnn_f32_velu_ukernel__wasm_rr2_p6_x4() 101 vp1 *= vt1; in xnn_f32_velu_ukernel__wasm_rr2_p6_x4() 115 vp1 = vp1 * vt1 + vt1; in xnn_f32_velu_ukernel__wasm_rr2_p6_x4() 121 const float ve1 = (vp1 + vs1) * valpha; in xnn_f32_velu_ukernel__wasm_rr2_p6_x4()
|
D | velu-scalar-rr2-p6-x4.c | 97 float vp1 = vc6 * vt1 + vc5; in xnn_f32_velu_ukernel__scalar_rr2_p6_x4() local 102 vp1 = vp1 * vt1 + vc4; in xnn_f32_velu_ukernel__scalar_rr2_p6_x4() 107 vp1 = vp1 * vt1 + vc3; in xnn_f32_velu_ukernel__scalar_rr2_p6_x4() 112 vp1 = vp1 * vt1 + vc2; in xnn_f32_velu_ukernel__scalar_rr2_p6_x4() 117 vp1 *= vt1; in xnn_f32_velu_ukernel__scalar_rr2_p6_x4() 131 vp1 = vp1 * vt1 + vt1; in xnn_f32_velu_ukernel__scalar_rr2_p6_x4() 137 const float ve1 = (vp1 + vs1) * valpha; in xnn_f32_velu_ukernel__scalar_rr2_p6_x4()
|
D | velu-avx2-rr1-p6-x16.c | 62 __m256 vp1 = _mm256_fmadd_ps(vc6, vt1, vc5); in xnn_f32_velu_ukernel__avx2_rr1_p6_x16() local 65 vp1 = _mm256_fmadd_ps(vp1, vt1, vc4); in xnn_f32_velu_ukernel__avx2_rr1_p6_x16() 68 vp1 = _mm256_fmadd_ps(vp1, vt1, vc3); in xnn_f32_velu_ukernel__avx2_rr1_p6_x16() 71 vp1 = _mm256_fmadd_ps(vp1, vt1, vc2); in xnn_f32_velu_ukernel__avx2_rr1_p6_x16() 75 vp1 = _mm256_mul_ps(vp1, vt1); in xnn_f32_velu_ukernel__avx2_rr1_p6_x16() 81 vp1 = _mm256_fmadd_ps(vp1, vt1, vt1); in xnn_f32_velu_ukernel__avx2_rr1_p6_x16() 85 const __m256 ve1 = _mm256_fmadd_ps(vp1, valpha, vs1); in xnn_f32_velu_ukernel__avx2_rr1_p6_x16()
|
D | velu-avx512f-rr1-p6-x32.c | 62 __m512 vp1 = _mm512_fmadd_ps(vc6, vt1, vc5); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x32() local 65 vp1 = _mm512_fmadd_ps(vp1, vt1, vc4); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x32() 68 vp1 = _mm512_fmadd_ps(vp1, vt1, vc3); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x32() 71 vp1 = _mm512_fmadd_ps(vp1, vt1, vc2); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x32() 75 vp1 = _mm512_mul_ps(vp1, vt1); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x32() 82 vp1 = _mm512_fmadd_ps(vp1, vt1, vt1); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x32() 87 __m512 vy1 = _mm512_fmadd_ps(vp1, valpha, vs1); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x32()
|
D | velu-wasm-rr2-p6-x5.c | 88 float vp1 = vc6 * vt1 + vc5; in xnn_f32_velu_ukernel__wasm_rr2_p6_x5() local 94 vp1 = vp1 * vt1 + vc4; in xnn_f32_velu_ukernel__wasm_rr2_p6_x5() 100 vp1 = vp1 * vt1 + vc3; in xnn_f32_velu_ukernel__wasm_rr2_p6_x5() 106 vp1 = vp1 * vt1 + vc2; in xnn_f32_velu_ukernel__wasm_rr2_p6_x5() 112 vp1 *= vt1; in xnn_f32_velu_ukernel__wasm_rr2_p6_x5() 129 vp1 = vp1 * vt1 + vt1; in xnn_f32_velu_ukernel__wasm_rr2_p6_x5() 136 const float ve1 = (vp1 + vs1) * valpha; in xnn_f32_velu_ukernel__wasm_rr2_p6_x5()
|
D | velu-scalar-rr2-p6-x5.c | 108 float vp1 = vc6 * vt1 + vc5; in xnn_f32_velu_ukernel__scalar_rr2_p6_x5() local 114 vp1 = vp1 * vt1 + vc4; in xnn_f32_velu_ukernel__scalar_rr2_p6_x5() 120 vp1 = vp1 * vt1 + vc3; in xnn_f32_velu_ukernel__scalar_rr2_p6_x5() 126 vp1 = vp1 * vt1 + vc2; in xnn_f32_velu_ukernel__scalar_rr2_p6_x5() 132 vp1 *= vt1; in xnn_f32_velu_ukernel__scalar_rr2_p6_x5() 149 vp1 = vp1 * vt1 + vt1; in xnn_f32_velu_ukernel__scalar_rr2_p6_x5() 156 const float ve1 = (vp1 + vs1) * valpha; in xnn_f32_velu_ukernel__scalar_rr2_p6_x5()
|
/external/XNNPACK/src/f32-raddstoreexpminusmax/gen/ |
D | scalar-p5-x2.c | 79 float vp1 = vc5 * vt1 + vc4; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_p5_x2() local 82 vp1 = vp1 * vt1 + vc3; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_p5_x2() 85 vp1 = vp1 * vt1 + vc2; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_p5_x2() 88 vp1 = vp1 * vt1 + vc1; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_p5_x2() 98 float vf1 = vt1 * vp1 + vs1; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_p5_x2()
|
D | scalar-p5-x2-acc2.c | 80 float vp1 = vc5 * vt1 + vc4; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_p5_x2_acc2() local 83 vp1 = vp1 * vt1 + vc3; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_p5_x2_acc2() 86 vp1 = vp1 * vt1 + vc2; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_p5_x2_acc2() 89 vp1 = vp1 * vt1 + vc1; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_p5_x2_acc2() 99 float vf1 = vt1 * vp1 + vs1; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_p5_x2_acc2()
|
D | scalar-p5-x4-acc2.c | 94 float vp1 = vc5 * vt1 + vc4; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_p5_x4_acc2() local 99 vp1 = vp1 * vt1 + vc3; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_p5_x4_acc2() 104 vp1 = vp1 * vt1 + vc2; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_p5_x4_acc2() 109 vp1 = vp1 * vt1 + vc1; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_p5_x4_acc2() 123 float vf1 = vt1 * vp1 + vs1; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_p5_x4_acc2()
|
/external/XNNPACK/src/f32-sigmoid/gen/ |
D | scalar-p5-div-x2.c | 63 float vp1 = vt1 * vc5 + vc4; in xnn_f32_sigmoid_ukernel__scalar_p5_div_x2() local 66 vp1 = vt1 * vp1 + vc3; in xnn_f32_sigmoid_ukernel__scalar_p5_div_x2() 69 vp1 = vt1 * vp1 + vc2; in xnn_f32_sigmoid_ukernel__scalar_p5_div_x2() 72 vp1 = vt1 * vp1 + vc1; in xnn_f32_sigmoid_ukernel__scalar_p5_div_x2() 78 const float ve1 = vt1 * vp1 + vs1; in xnn_f32_sigmoid_ukernel__scalar_p5_div_x2()
|
D | avx512f-rr1-p5-scalef-div-x32.c | 55 __m512 vp1 = _mm512_fmadd_ps(vc5, vt1, vc4); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x32() local 58 vp1 = _mm512_fmadd_ps(vp1, vt1, vc3); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x32() 61 vp1 = _mm512_fmadd_ps(vp1, vt1, vc2); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x32() 64 vp1 = _mm512_fmadd_ps(vp1, vt1, vc1); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x32() 67 vp1 = _mm512_fmadd_ps(vp1, vt1, vone); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x32() 70 const __m512 ve1 = _mm512_scalef_ps(vp1, vn1); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x32()
|
/external/XNNPACK/src/f32-vscaleexpminusmax/gen/ |
D | avx512f-p5-scalef-x32.c | 65 __m512 vp1 = _mm512_fmadd_ps(vc5, vt1, vc4); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x32() local 68 vp1 = _mm512_fmadd_ps(vp1, vt1, vc3); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x32() 71 vp1 = _mm512_fmadd_ps(vp1, vt1, vc2); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x32() 74 vp1 = _mm512_fmadd_ps(vp1, vt1, vc1); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x32() 77 vp1 = _mm512_fmadd_ps(vp1, vt1, vc0); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x32() 83 __m512 vf1 = _mm512_scalef_ps(vp1, vn1); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x32()
|
D | avx512f-p5-scalef-x48.c | 70 __m512 vp1 = _mm512_fmadd_ps(vc5, vt1, vc4); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x48() local 74 vp1 = _mm512_fmadd_ps(vp1, vt1, vc3); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x48() 78 vp1 = _mm512_fmadd_ps(vp1, vt1, vc2); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x48() 82 vp1 = _mm512_fmadd_ps(vp1, vt1, vc1); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x48() 86 vp1 = _mm512_fmadd_ps(vp1, vt1, vc0); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x48() 93 __m512 vf1 = _mm512_scalef_ps(vp1, vn1); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x48()
|
/external/XNNPACK/src/f32-vscaleextexp/gen/ |
D | avx512f-p5-scalef-x32.c | 62 __m512 vp1 = _mm512_fmadd_ps(vc5, vt1, vc4); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x32() local 65 vp1 = _mm512_fmadd_ps(vp1, vt1, vc3); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x32() 68 vp1 = _mm512_fmadd_ps(vp1, vt1, vc2); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x32() 71 vp1 = _mm512_fmadd_ps(vp1, vt1, vc1); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x32() 74 vp1 = _mm512_fmadd_ps(vp1, vt1, vc0); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x32() 83 __m512 vf1 = _mm512_mul_ps(vp1, vscalev); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x32()
|
D | avx512f-p5-scalef-x48.c | 66 __m512 vp1 = _mm512_fmadd_ps(vc5, vt1, vc4); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x48() local 70 vp1 = _mm512_fmadd_ps(vp1, vt1, vc3); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x48() 74 vp1 = _mm512_fmadd_ps(vp1, vt1, vc2); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x48() 78 vp1 = _mm512_fmadd_ps(vp1, vt1, vc1); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x48() 82 vp1 = _mm512_fmadd_ps(vp1, vt1, vc0); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x48() 92 __m512 vf1 = _mm512_mul_ps(vp1, vscalev); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x48()
|
D | avx2-p5-x16.c | 68 __m256 vp1 = _mm256_fmadd_ps(vc5, vt1, vc4); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x16() local 71 vp1 = _mm256_fmadd_ps(vp1, vt1, vc3); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x16() 74 vp1 = _mm256_fmadd_ps(vp1, vt1, vc2); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x16() 77 vp1 = _mm256_fmadd_ps(vp1, vt1, vc1); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x16() 80 vp1 = _mm256_fmadd_ps(vp1, vt1, vc0); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x16() 89 __m256 vf1 = _mm256_mul_ps(vp1, vscalev); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x16()
|
/external/clang/test/SemaCXX/ |
D | address-space-conversion.cpp | 50 void test_static_cast(void_ptr vp, void_ptr_1 vp1, void_ptr_2 vp2, in test_static_cast() argument 68 (void)static_cast<A_ptr_1>(vp1); in test_static_cast() 94 (void)static_cast<A_ptr>(vp1); // expected-error{{casts away qualifiers}} in test_static_cast() 99 (void)static_cast<A_ptr_2>(vp1); // expected-error{{casts away qualifiers}} in test_static_cast() 131 void test_reinterpret_cast(void_ptr vp, void_ptr_1 vp1, void_ptr_2 vp2, in test_reinterpret_cast() argument 142 (void)reinterpret_cast<A_ptr>(vp1); in test_reinterpret_cast() 150 (void)reinterpret_cast<A_ptr_1>(vp1); in test_reinterpret_cast() 157 void test_cstyle_cast(void_ptr vp, void_ptr_1 vp1, void_ptr_2 vp2, in test_cstyle_cast() argument 168 (void)(A_ptr)(vp1); in test_cstyle_cast() 176 (void)(A_ptr_1)(vp1); in test_cstyle_cast() [all …]
|
/external/llvm-project/clang/test/SemaCXX/ |
D | address-space-conversion.cpp | 50 void test_static_cast(void_ptr vp, void_ptr_1 vp1, void_ptr_2 vp2, in test_static_cast() argument 68 (void)static_cast<A_ptr_1>(vp1); in test_static_cast() 94 (void)static_cast<A_ptr>(vp1); // expected-error{{casts away qualifiers}} in test_static_cast() 99 (void)static_cast<A_ptr_2>(vp1); // expected-error{{casts away qualifiers}} in test_static_cast() 131 void test_reinterpret_cast(void_ptr vp, void_ptr_1 vp1, void_ptr_2 vp2, in test_reinterpret_cast() argument 142 …(void)reinterpret_cast<A_ptr>(vp1); // expected-error{{reinterpret_cast from 'void_ptr_1' (aka '… in test_reinterpret_cast() 150 (void)reinterpret_cast<A_ptr_1>(vp1); in test_reinterpret_cast() 157 void test_cstyle_cast(void_ptr vp, void_ptr_1 vp1, void_ptr_2 vp2, in test_cstyle_cast() argument 168 (void)(A_ptr)(vp1); in test_cstyle_cast() 176 (void)(A_ptr_1)(vp1); in test_cstyle_cast() [all …]
|
/external/clang/test/CodeGen/ |
D | enable_if.c | 21 void *vp1 = (void*)&foo; in test1() local 25 vp1 = (void*)&foo; in test1() 27 vp1 = (void*)foo; in test1() 44 void *vp1 = (void*)&bar; in test2() local 48 vp1 = (void*)&bar; in test2() 50 vp1 = (void*)bar; in test2()
|
/external/llvm-project/clang/test/CodeGen/ |
D | enable_if.c | 21 void *vp1 = (void*)&foo; in test1() local 25 vp1 = (void*)&foo; in test1() 27 vp1 = (void*)foo; in test1() 44 void *vp1 = (void*)&bar; in test2() local 48 vp1 = (void*)&bar; in test2() 50 vp1 = (void*)bar; in test2()
|
/external/llvm-project/clang/test/Sema/ |
D | ppc-mma-types.c | 247 __vector_pair vp1 = *vpp; in testVPLocal() local 278 __vector_pair vp1 = *(vpp + 0); in testVPOperators1() local 281 …if (vp1) // expected-error {{statement requires expression of scalar type ('__vector_pair' invalid… in testVPOperators1() 282 *(vpp + 10) = vp1; in testVPOperators1() 285 …int c1 = vp1 && vp2; // expected-error {{invalid operands to binary expression ('__vector_pair' an… in testVPOperators1() 287 …int c3 = vp2 < vp1; // expected-error {{invalid operands to binary expression ('__vector_pair' an… in testVPOperators1() 293 __vector_pair vp1 = *(vpp + 0); in testVPOperators2() local 296 vp1 = -vp1; // expected-error {{invalid argument type '__vector_pair' to unary expression}} in testVPOperators2() 297 …vp2 = vp1 + vp3; // expected-error {{invalid operands to binary expression ('__vector_pair' and '_… in testVPOperators2() 301 *(vpp + 10) = vp1; in testVPOperators2() [all …]
|