| /external/XNNPACK/src/f32-gemm/gen/ | 
| D | 4x2-minmax-neon-lane-ld64.c | 69       const float32x2_t va1 = vld1_f32(a1); a1 += 2;  in xnn_f32_gemm_minmax_ukernel_4x2__neon_lane_ld64()  local 88       const float32x2_t va1 = vld1_dup_f32(a1); a1 += 1;  in xnn_f32_gemm_minmax_ukernel_4x2__neon_lane_ld64()  local
  | 
| D | 4x2-minmax-neonfma-lane-ld64.c | 69       const float32x2_t va1 = vld1_f32(a1); a1 += 2;  in xnn_f32_gemm_minmax_ukernel_4x2__neonfma_lane_ld64()  local 110       const float32x2_t va1 = vld1_dup_f32(a1); a1 += 1;  in xnn_f32_gemm_minmax_ukernel_4x2__neonfma_lane_ld64()  local
  | 
| D | 6x2-minmax-neon-lane-ld64.c | 83       const float32x2_t va1 = vld1_f32(a1); a1 += 2;  in xnn_f32_gemm_minmax_ukernel_6x2__neon_lane_ld64()  local 108       const float32x2_t va1 = vld1_dup_f32(a1); a1 += 1;  in xnn_f32_gemm_minmax_ukernel_6x2__neon_lane_ld64()  local
  | 
| D | 3x8s4-wasmrelaxedsimd-fma.c | 66       v128_t va1 = wasm_v128_load(a1);  in xnn_f32_gemm_ukernel_3x8s4__wasmrelaxedsimd_fma()  local 131       v128_t va1 = wasm_v128_load(a1);  in xnn_f32_gemm_ukernel_3x8s4__wasmrelaxedsimd_fma()  local
  | 
| /external/XNNPACK/src/qs8-gemm/gen/ | 
| D | 2x4c2-xw-minmax-fp32-sse2.c | 61       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qs8_gemm_xw_minmax_fp32_ukernel_2x4c2__sse2()  local 97       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qs8_gemm_xw_minmax_fp32_ukernel_2x4c2__sse2()  local
  | 
| D | 2x4c2-xw-minmax-fp32-sse41.c | 61       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qs8_gemm_xw_minmax_fp32_ukernel_2x4c2__sse41()  local 97       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qs8_gemm_xw_minmax_fp32_ukernel_2x4c2__sse41()  local
  | 
| D | 2x4c2-xw-minmax-fp32-xop.c | 66       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qs8_gemm_xw_minmax_fp32_ukernel_2x4c2__xop()  local 102       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qs8_gemm_xw_minmax_fp32_ukernel_2x4c2__xop()  local
  | 
| D | 2x4c2-xw-minmax-fp32-avx.c | 61       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qs8_gemm_xw_minmax_fp32_ukernel_2x4c2__avx()  local 97       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qs8_gemm_xw_minmax_fp32_ukernel_2x4c2__avx()  local
  | 
| D | 2x4c2-minmax-fp32-sse41-ld128.c | 61       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qs8_gemm_minmax_fp32_ukernel_2x4c2__sse41_ld128()  local 99       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qs8_gemm_minmax_fp32_ukernel_2x4c2__sse41_ld128()  local
  | 
| D | 2x4c2-minmax-fp32-xop-ld128.c | 66       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qs8_gemm_minmax_fp32_ukernel_2x4c2__xop_ld128()  local 104       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qs8_gemm_minmax_fp32_ukernel_2x4c2__xop_ld128()  local
  | 
| D | 2x4c2-minmax-fp32-avx-ld128.c | 61       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qs8_gemm_minmax_fp32_ukernel_2x4c2__avx_ld128()  local 99       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qs8_gemm_minmax_fp32_ukernel_2x4c2__avx_ld128()  local
  | 
| D | 2x4c2-minmax-fp32-xop-ld64.c | 66       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qs8_gemm_minmax_fp32_ukernel_2x4c2__xop_ld64()  local 106       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qs8_gemm_minmax_fp32_ukernel_2x4c2__xop_ld64()  local
  | 
| D | 2x4c2-minmax-fp32-sse41-ld64.c | 61       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qs8_gemm_minmax_fp32_ukernel_2x4c2__sse41_ld64()  local 101       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qs8_gemm_minmax_fp32_ukernel_2x4c2__sse41_ld64()  local
  | 
| /external/XNNPACK/src/f32-igemm/gen/ | 
| D | 4x2-minmax-neon-lane-ld64.c | 91         const float32x2_t va1 = vld1_f32(a1); a1 += 2;  in xnn_f32_igemm_minmax_ukernel_4x2__neon_lane_ld64()  local 110         const float32x2_t va1 = vld1_dup_f32(a1);  in xnn_f32_igemm_minmax_ukernel_4x2__neon_lane_ld64()  local
  | 
| D | 4x4-minmax-neon-lane-ld64.c | 91         const float32x2_t va1 = vld1_f32(a1); a1 += 2;  in xnn_f32_igemm_minmax_ukernel_4x4__neon_lane_ld64()  local 111         const float32x4_t va1 = vld1q_dup_f32(a1);  in xnn_f32_igemm_minmax_ukernel_4x4__neon_lane_ld64()  local
  | 
| D | 4x4-minmax-neonfma-lane-ld64.c | 91         const float32x2_t va1 = vld1_f32(a1); a1 += 2;  in xnn_f32_igemm_minmax_ukernel_4x4__neonfma_lane_ld64()  local 111         const float32x4_t va1 = vld1q_dup_f32(a1);  in xnn_f32_igemm_minmax_ukernel_4x4__neonfma_lane_ld64()  local
  | 
| /external/XNNPACK/src/qc8-gemm/gen/ | 
| D | 2x4c2-minmax-fp32-sse41-ld128.c | 61       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qc8_gemm_minmax_fp32_ukernel_2x4c2__sse41_ld128()  local 99       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qc8_gemm_minmax_fp32_ukernel_2x4c2__sse41_ld128()  local
  | 
| D | 2x4c2-minmax-fp32-avx-ld128.c | 61       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qc8_gemm_minmax_fp32_ukernel_2x4c2__avx_ld128()  local 99       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qc8_gemm_minmax_fp32_ukernel_2x4c2__avx_ld128()  local
  | 
| D | 2x4c2-minmax-fp32-xop-ld128.c | 66       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qc8_gemm_minmax_fp32_ukernel_2x4c2__xop_ld128()  local 104       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qc8_gemm_minmax_fp32_ukernel_2x4c2__xop_ld128()  local
  | 
| D | 2x4c2-minmax-fp32-sse41-ld64.c | 61       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qc8_gemm_minmax_fp32_ukernel_2x4c2__sse41_ld64()  local 101       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qc8_gemm_minmax_fp32_ukernel_2x4c2__sse41_ld64()  local
  | 
| D | 2x4c2-minmax-fp32-avx-ld64.c | 61       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qc8_gemm_minmax_fp32_ukernel_2x4c2__avx_ld64()  local 101       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qc8_gemm_minmax_fp32_ukernel_2x4c2__avx_ld64()  local
  | 
| D | 2x4c2-minmax-fp32-xop-ld64.c | 66       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qc8_gemm_minmax_fp32_ukernel_2x4c2__xop_ld64()  local 106       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qc8_gemm_minmax_fp32_ukernel_2x4c2__xop_ld64()  local
  | 
| /external/XNNPACK/src/qu8-gemm/gen/ | 
| D | 2x4c2-minmax-fp32-xop-ld128.c | 68       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qu8_gemm_minmax_fp32_ukernel_2x4c2__xop_ld128()  local 106       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qu8_gemm_minmax_fp32_ukernel_2x4c2__xop_ld128()  local
  | 
| D | 2x4c2-minmax-fp32-sse2-ld128.c | 63       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qu8_gemm_minmax_fp32_ukernel_2x4c2__sse2_ld128()  local 101       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qu8_gemm_minmax_fp32_ukernel_2x4c2__sse2_ld128()  local
  | 
| D | 2x4c2-minmax-fp32-avx-ld128.c | 63       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qu8_gemm_minmax_fp32_ukernel_2x4c2__avx_ld128()  local 101       const __m128i va1 = _mm_loadl_epi64((const __m128i*) a1);  in xnn_qu8_gemm_minmax_fp32_ukernel_2x4c2__avx_ld128()  local
  |