| /external/clang/test/CodeGen/ |
| D | enable_if.c | 21 void *vp1 = (void*)&foo; in test1() local 44 void *vp1 = (void*)&bar; in test2() local
|
| D | overloadable.c | 51 void *vp1 = (void (*)(char *)) & addrof_many; in foo() local
|
| /external/clang/test/SemaCXX/ |
| D | address-space-conversion.cpp | 50 void test_static_cast(void_ptr vp, void_ptr_1 vp1, void_ptr_2 vp2, in test_static_cast() 131 void test_reinterpret_cast(void_ptr vp, void_ptr_1 vp1, void_ptr_2 vp2, in test_reinterpret_cast() 157 void test_cstyle_cast(void_ptr vp, void_ptr_1 vp1, void_ptr_2 vp2, in test_cstyle_cast() 181 void test_implicit_conversion(void_ptr vp, void_ptr_1 vp1, void_ptr_2 vp2, in test_implicit_conversion()
|
| /external/clang/test/PCH/ |
| D | exprs.c | 72 void_ptr vp1 = &integer; variable
|
| /external/XNNPACK/src/f32-raddstoreexpminusmax/gen/ |
| D | scalar-rr2-p5-x2.c | 77 float vp1 = vc5 * vt1 + vc4; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_rr2_p5_x2() local
|
| D | scalar-rr2-p5-x2-acc2.c | 78 float vp1 = vc5 * vt1 + vc4; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_rr2_p5_x2_acc2() local
|
| D | scalar-rr2-lut64-p2-x2-acc2.c | 95 float vp1 = vt1 * vc2; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_rr2_lut64_p2_x2_acc2() local
|
| D | scalar-rr2-lut64-p2-x2.c | 94 float vp1 = vt1 * vc2; in xnn_f32_raddstoreexpminusmax_ukernel__scalar_rr2_lut64_p2_x2() local
|
| /external/XNNPACK/src/f32-vscaleextexp/gen/ |
| D | avx512f-p5-scalef-x32.c | 62 __m512 vp1 = _mm512_fmadd_ps(vc5, vt1, vc4); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x32() local
|
| D | avx512f-p5-scalef-x48.c | 66 __m512 vp1 = _mm512_fmadd_ps(vc5, vt1, vc4); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x48() local
|
| D | avx2-p5-x16.c | 68 __m256 vp1 = _mm256_fmadd_ps(vc5, vt1, vc4); in xnn_f32_vscaleextexp_ukernel__avx2_p5_x16() local
|
| /external/XNNPACK/src/f32-vscaleexpminusmax/gen/ |
| D | avx512f-p5-scalef-x32.c | 65 __m512 vp1 = _mm512_fmadd_ps(vc5, vt1, vc4); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x32() local
|
| D | avx2-p5-x16.c | 78 __m256 vp1 = _mm256_fmadd_ps(vc5, vt1, vc4); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x16() local
|
| D | avx512f-p5-scalef-x48.c | 70 __m512 vp1 = _mm512_fmadd_ps(vc5, vt1, vc4); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x48() local
|
| /external/XNNPACK/src/f32-velu/gen/ |
| D | velu-scalar-rr2-p6-x2.c | 73 float vp1 = vc6 * vt1 + vc5; in xnn_f32_velu_ukernel__scalar_rr2_p6_x2() local
|
| D | velu-wasm-rr2-p6-x2.c | 65 float vp1 = vc6 * vt1 + vc5; in xnn_f32_velu_ukernel__wasm_rr2_p6_x2() local
|
| D | velu-scalar-rr2-lut16-p3-x2.c | 76 float vp1 = vc3 * vt1 + vc2; in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x2() local
|
| D | velu-wasm-rr2-lut16-p3-x2.c | 68 float vp1 = vc3 * vt1 + vc2; in xnn_f32_velu_ukernel__wasm_rr2_lut16_p3_x2() local
|
| D | velu-avx2-rr1-p6-x16.c | 59 __m256 vp1 = _mm256_fmadd_ps(vc6, vt1, vc5); in xnn_f32_velu_ukernel__avx2_rr1_p6_x16() local
|
| D | velu-scalar-rr2-p6-x3.c | 84 float vp1 = vc6 * vt1 + vc5; in xnn_f32_velu_ukernel__scalar_rr2_p6_x3() local
|
| /external/XNNPACK/src/f32-vsigmoid/gen/ |
| D | vsigmoid-scalar-rr2-p5-div-x2.c | 62 float vp1 = vt1 * vc5 + vc4; in xnn_f32_vsigmoid_ukernel__scalar_rr2_p5_div_x2() local
|
| D | vsigmoid-scalar-rr2-lut64-p2-div-x2.c | 67 float vp1 = vt1 * vc2; in xnn_f32_vsigmoid_ukernel__scalar_rr2_lut64_p2_div_x2() local
|
| D | vsigmoid-scalar-rr2-lut2048-p1-div-x2.c | 67 const float vp1 = vt1 * vc1; in xnn_f32_vsigmoid_ukernel__scalar_rr2_lut2048_p1_div_x2() local
|
| D | vsigmoid-avx512f-rr1-p5-scalef-div-x32.c | 55 __m512 vp1 = _mm512_fmadd_ps(vc5, vt1, vc4); in xnn_f32_vsigmoid_ukernel__avx512f_rr1_p5_scalef_div_x32() local
|
| D | vsigmoid-avx512f-rr1-p5-scalef-nr1fma-x32.c | 55 __m512 vp1 = _mm512_fmadd_ps(vc5, vt1, vc4); in xnn_f32_vsigmoid_ukernel__avx512f_rr1_p5_scalef_nr1fma_x32() local
|