/external/XNNPACK/src/f32-argmaxpool/ |
D | 9p8x-sse2-c4.c | 183 const __m128i vidx4 = _mm_add_epi32(vidx3, v1); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() local 298 const __m128i vidx4 = _mm_add_epi32(vidx3, v1); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() local 355 const __m128i vidx4 = _mm_add_epi32(vidx3, v1); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() local
|
D | 9p8x-neon-c4.c | 164 const uint32x4_t vidx4 = vaddq_u32(vidx3, v1); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4() local 267 const uint32x4_t vidx4 = vaddq_u32(vidx3, v1); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4() local 322 const uint32x4_t vidx4 = vaddq_u32(vidx3, v1); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4() local
|
D | 9p8x-wasmsimd-c4.c | 183 const v128_t vidx4 = wasm_i32x4_add(vidx3, v1); in xnn_f32_argmaxpool_ukernel_9p8x__wasmsimd_c4() local 298 const v128_t vidx4 = wasm_i32x4_add(vidx3, v1); in xnn_f32_argmaxpool_ukernel_9p8x__wasmsimd_c4() local 355 const v128_t vidx4 = wasm_i32x4_add(vidx3, v1); in xnn_f32_argmaxpool_ukernel_9p8x__wasmsimd_c4() local
|
/external/XNNPACK/src/f32-velu/gen/ |
D | velu-wasm-rr2-lut16-p3-x5.c | 75 const uint32_t vidx4 = fp32_to_bits(vn4) & vindex_mask; in xnn_f32_velu_ukernel__wasm_rr2_lut16_p3_x5() local
|
D | velu-scalar-rr2-lut16-p3-x5.c | 75 const uint32_t vidx4 = fp32_to_bits(vn4) & vindex_mask; in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x5() local
|
D | velu-wasm-rr2-lut16-p3-x6.c | 78 const uint32_t vidx4 = fp32_to_bits(vn4) & vindex_mask; in xnn_f32_velu_ukernel__wasm_rr2_lut16_p3_x6() local
|
D | velu-scalar-rr2-lut16-p3-x6.c | 78 const uint32_t vidx4 = fp32_to_bits(vn4) & vindex_mask; in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x6() local
|
D | velu-avx2-rr1-lut16-p3-gather-x40.c | 67 const __m256i vidx4 = _mm256_and_si256(_mm256_castps_si256(vn4), vindex_mask); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x40() local
|
D | velu-avx2-rr1-lut16-p3-gather-x48.c | 70 const __m256i vidx4 = _mm256_and_si256(_mm256_castps_si256(vn4), vindex_mask); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x48() local
|
D | velu-sse41-rr2-lut16-p3-x8.c | 85 const uint32_t vidx4 = (uint32_t) _mm_cvtsi128_si32(vidx4567); in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x8() local
|
D | velu-avx2-rr1-lut16-p3-gather-x56.c | 73 const __m256i vidx4 = _mm256_and_si256(_mm256_castps_si256(vn4), vindex_mask); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x56() local
|
D | velu-avx2-rr1-lut16-p3-gather-x64.c | 76 const __m256i vidx4 = _mm256_and_si256(_mm256_castps_si256(vn4), vindex_mask); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x64() local
|
D | velu-sse2-rr2-lut16-p3-x8.c | 91 const uint32_t vidx4 = (uint32_t) _mm_cvtsi128_si32(vidx4567); in xnn_f32_velu_ukernel__sse2_rr2_lut16_p3_x8() local
|
D | velu-sse41-rr2-lut16-p3-x12.c | 97 const uint32_t vidx4 = (uint32_t) _mm_cvtsi128_si32(vidx4567); in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x12() local
|
D | velu-avx2-rr1-lut16-p3-gather-x72.c | 79 const __m256i vidx4 = _mm256_and_si256(_mm256_castps_si256(vn4), vindex_mask); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x72() local
|
D | velu-avx2-rr1-lut16-p3-gather-x80.c | 82 const __m256i vidx4 = _mm256_and_si256(_mm256_castps_si256(vn4), vindex_mask); in xnn_f32_velu_ukernel__avx2_rr1_lut16_p3_gather_x80() local
|
D | velu-sse2-rr2-lut16-p3-x12.c | 105 const uint32_t vidx4 = (uint32_t) _mm_cvtsi128_si32(vidx4567); in xnn_f32_velu_ukernel__sse2_rr2_lut16_p3_x12() local
|
D | velu-sse41-rr2-lut16-p3-x16.c | 109 const uint32_t vidx4 = (uint32_t) _mm_cvtsi128_si32(vidx4567); in xnn_f32_velu_ukernel__sse41_rr2_lut16_p3_x16() local
|
/external/XNNPACK/src/f32-vsigmoid/gen/ |
D | vsigmoid-sse41-rr2-lut64-p2-div-x8.c | 80 const uint32_t vidx4 = (uint32_t) _mm_cvtsi128_si32(vidx4567); in xnn_f32_vsigmoid_ukernel__sse41_rr2_lut64_p2_div_x8() local
|
D | vsigmoid-sse2-rr2-lut64-p2-div-x8.c | 86 const uint32_t vidx4 = (uint32_t) _mm_cvtsi128_si32(vidx4567); in xnn_f32_vsigmoid_ukernel__sse2_rr2_lut64_p2_div_x8() local
|
D | vsigmoid-sse41-rr2-lut64-p2-div-x12.c | 92 const uint32_t vidx4 = (uint32_t) _mm_cvtsi128_si32(vidx4567); in xnn_f32_vsigmoid_ukernel__sse41_rr2_lut64_p2_div_x12() local
|
D | vsigmoid-sse2-rr2-lut64-p2-div-x12.c | 100 const uint32_t vidx4 = (uint32_t) _mm_cvtsi128_si32(vidx4567); in xnn_f32_vsigmoid_ukernel__sse2_rr2_lut64_p2_div_x12() local
|
D | vsigmoid-sse41-rr2-lut64-p2-div-x16.c | 104 const uint32_t vidx4 = (uint32_t) _mm_cvtsi128_si32(vidx4567); in xnn_f32_vsigmoid_ukernel__sse41_rr2_lut64_p2_div_x16() local
|
D | vsigmoid-sse41-rr2-lut64-p2-div-x20.c | 116 const uint32_t vidx4 = (uint32_t) _mm_cvtsi128_si32(vidx4567); in xnn_f32_vsigmoid_ukernel__sse41_rr2_lut64_p2_div_x20() local
|
D | vsigmoid-sse2-rr2-lut64-p2-div-x16.c | 114 const uint32_t vidx4 = (uint32_t) _mm_cvtsi128_si32(vidx4567); in xnn_f32_vsigmoid_ukernel__sse2_rr2_lut64_p2_div_x16() local
|