/external/XNNPACK/src/f32-vbinary/gen/ |
D | vmax-wasmsimd-x86-x16.c | 42 const v128_t vbCDEF = wasm_v128_load(b + 12); in xnn_f32_vmax_ukernel__wasmsimd_x86_x16() local 48 const v128_t vmCDEF = wasm_f32x4_le(vaCDEF, vbCDEF); in xnn_f32_vmax_ukernel__wasmsimd_x86_x16() 53 v128_t vyCDEF = wasm_v128_bitselect(vbCDEF, vaCDEF, vmCDEF); in xnn_f32_vmax_ukernel__wasmsimd_x86_x16()
|
D | vmin-wasmsimd-x86-x16.c | 42 const v128_t vbCDEF = wasm_v128_load(b + 12); in xnn_f32_vmin_ukernel__wasmsimd_x86_x16() local 48 const v128_t vmCDEF = wasm_f32x4_lt(vaCDEF, vbCDEF); in xnn_f32_vmin_ukernel__wasmsimd_x86_x16() 53 v128_t vyCDEF = wasm_v128_bitselect(vaCDEF, vbCDEF, vmCDEF); in xnn_f32_vmin_ukernel__wasmsimd_x86_x16()
|
D | vsub-wasmsimd-x16.c | 42 const v128_t vbCDEF = wasm_v128_load(b + 12); in xnn_f32_vsub_ukernel__wasmsimd_x16() local 48 v128_t vyCDEF = wasm_f32x4_sub(vaCDEF, vbCDEF); in xnn_f32_vsub_ukernel__wasmsimd_x16()
|
D | vdiv-wasmsimd-x16.c | 42 const v128_t vbCDEF = wasm_v128_load(b + 12); in xnn_f32_vdiv_ukernel__wasmsimd_x16() local 48 v128_t vyCDEF = wasm_f32x4_div(vaCDEF, vbCDEF); in xnn_f32_vdiv_ukernel__wasmsimd_x16()
|
D | vmin-wasmsimd-arm-x16.c | 42 const v128_t vbCDEF = wasm_v128_load(b + 12); in xnn_f32_vmin_ukernel__wasmsimd_arm_x16() local 48 v128_t vyCDEF = wasm_f32x4_min(vaCDEF, vbCDEF); in xnn_f32_vmin_ukernel__wasmsimd_arm_x16()
|
D | vmax-wasmsimd-arm-x16.c | 42 const v128_t vbCDEF = wasm_v128_load(b + 12); in xnn_f32_vmax_ukernel__wasmsimd_arm_x16() local 48 v128_t vyCDEF = wasm_f32x4_max(vaCDEF, vbCDEF); in xnn_f32_vmax_ukernel__wasmsimd_arm_x16()
|
D | vadd-wasmsimd-x16.c | 42 const v128_t vbCDEF = wasm_v128_load(b + 12); in xnn_f32_vadd_ukernel__wasmsimd_x16() local 48 v128_t vyCDEF = wasm_f32x4_add(vaCDEF, vbCDEF); in xnn_f32_vadd_ukernel__wasmsimd_x16()
|
D | vmul-wasmsimd-x16.c | 42 const v128_t vbCDEF = wasm_v128_load(b + 12); in xnn_f32_vmul_ukernel__wasmsimd_x16() local 48 v128_t vyCDEF = wasm_f32x4_mul(vaCDEF, vbCDEF); in xnn_f32_vmul_ukernel__wasmsimd_x16()
|
D | vsqrdiff-wasmsimd-x16.c | 42 const v128_t vbCDEF = wasm_v128_load(b + 12); in xnn_f32_vsqrdiff_ukernel__wasmsimd_x16() local 48 v128_t vyCDEF = wasm_f32x4_sub(vaCDEF, vbCDEF); in xnn_f32_vsqrdiff_ukernel__wasmsimd_x16()
|
D | vadd-relu-wasmsimd-x16.c | 43 const v128_t vbCDEF = wasm_v128_load(b + 12); in xnn_f32_vadd_relu_ukernel__wasmsimd_x16() local 49 v128_t vyCDEF = wasm_f32x4_add(vaCDEF, vbCDEF); in xnn_f32_vadd_relu_ukernel__wasmsimd_x16()
|
D | vsub-relu-wasmsimd-x16.c | 43 const v128_t vbCDEF = wasm_v128_load(b + 12); in xnn_f32_vsub_relu_ukernel__wasmsimd_x16() local 49 v128_t vyCDEF = wasm_f32x4_sub(vaCDEF, vbCDEF); in xnn_f32_vsub_relu_ukernel__wasmsimd_x16()
|
D | vdiv-relu-wasmsimd-x16.c | 43 const v128_t vbCDEF = wasm_v128_load(b + 12); in xnn_f32_vdiv_relu_ukernel__wasmsimd_x16() local 49 v128_t vyCDEF = wasm_f32x4_div(vaCDEF, vbCDEF); in xnn_f32_vdiv_relu_ukernel__wasmsimd_x16()
|
D | vmul-relu-wasmsimd-x16.c | 43 const v128_t vbCDEF = wasm_v128_load(b + 12); in xnn_f32_vmul_relu_ukernel__wasmsimd_x16() local 49 v128_t vyCDEF = wasm_f32x4_mul(vaCDEF, vbCDEF); in xnn_f32_vmul_relu_ukernel__wasmsimd_x16()
|
D | vdiv-minmax-wasmsimd-arm-x16.c | 44 const v128_t vbCDEF = wasm_v128_load(b + 12); in xnn_f32_vdiv_minmax_ukernel__wasmsimd_arm_x16() local 50 v128_t vyCDEF = wasm_f32x4_div(vaCDEF, vbCDEF); in xnn_f32_vdiv_minmax_ukernel__wasmsimd_arm_x16()
|
D | vsub-minmax-wasmsimd-arm-x16.c | 44 const v128_t vbCDEF = wasm_v128_load(b + 12); in xnn_f32_vsub_minmax_ukernel__wasmsimd_arm_x16() local 50 v128_t vyCDEF = wasm_f32x4_sub(vaCDEF, vbCDEF); in xnn_f32_vsub_minmax_ukernel__wasmsimd_arm_x16()
|
D | vmul-minmax-wasmsimd-arm-x16.c | 44 const v128_t vbCDEF = wasm_v128_load(b + 12); in xnn_f32_vmul_minmax_ukernel__wasmsimd_arm_x16() local 50 v128_t vyCDEF = wasm_f32x4_mul(vaCDEF, vbCDEF); in xnn_f32_vmul_minmax_ukernel__wasmsimd_arm_x16()
|
D | vadd-minmax-wasmsimd-arm-x16.c | 44 const v128_t vbCDEF = wasm_v128_load(b + 12); in xnn_f32_vadd_minmax_ukernel__wasmsimd_arm_x16() local 50 v128_t vyCDEF = wasm_f32x4_add(vaCDEF, vbCDEF); in xnn_f32_vadd_minmax_ukernel__wasmsimd_arm_x16()
|
D | vsub-minmax-wasmsimd-x86-x16.c | 44 const v128_t vbCDEF = wasm_v128_load(b + 12); in xnn_f32_vsub_minmax_ukernel__wasmsimd_x86_x16() local 50 v128_t vyCDEF = wasm_f32x4_sub(vaCDEF, vbCDEF); in xnn_f32_vsub_minmax_ukernel__wasmsimd_x86_x16()
|
D | vmul-minmax-wasmsimd-x86-x16.c | 44 const v128_t vbCDEF = wasm_v128_load(b + 12); in xnn_f32_vmul_minmax_ukernel__wasmsimd_x86_x16() local 50 v128_t vyCDEF = wasm_f32x4_mul(vaCDEF, vbCDEF); in xnn_f32_vmul_minmax_ukernel__wasmsimd_x86_x16()
|
D | vadd-minmax-wasmsimd-x86-x16.c | 44 const v128_t vbCDEF = wasm_v128_load(b + 12); in xnn_f32_vadd_minmax_ukernel__wasmsimd_x86_x16() local 50 v128_t vyCDEF = wasm_f32x4_add(vaCDEF, vbCDEF); in xnn_f32_vadd_minmax_ukernel__wasmsimd_x86_x16()
|
/external/XNNPACK/src/qs8-gemm/gen/ |
D | 4x16c8-minmax-avx512skx.c | 120 …const __m512i vbCDEF = _mm512_cvtepi8_epi16(_mm256_load_si256((const __m256i*) ((uintptr_t) w + 96… in xnn_qs8_gemm_minmax_ukernel_4x16c8__avx512skx() local 122 vacc0xCDEF = _mm512_add_epi32(vacc0xCDEF, _mm512_madd_epi16(va0, vbCDEF)); in xnn_qs8_gemm_minmax_ukernel_4x16c8__avx512skx() 123 vacc1xCDEF = _mm512_add_epi32(vacc1xCDEF, _mm512_madd_epi16(va1, vbCDEF)); in xnn_qs8_gemm_minmax_ukernel_4x16c8__avx512skx() 124 vacc2xCDEF = _mm512_add_epi32(vacc2xCDEF, _mm512_madd_epi16(va2, vbCDEF)); in xnn_qs8_gemm_minmax_ukernel_4x16c8__avx512skx() 125 vacc3xCDEF = _mm512_add_epi32(vacc3xCDEF, _mm512_madd_epi16(va3, vbCDEF)); in xnn_qs8_gemm_minmax_ukernel_4x16c8__avx512skx()
|
D | 3x16c8-minmax-avx512skx.c | 105 …const __m512i vbCDEF = _mm512_cvtepi8_epi16(_mm256_load_si256((const __m256i*) ((uintptr_t) w + 96… in xnn_qs8_gemm_minmax_ukernel_3x16c8__avx512skx() local 107 vacc0xCDEF = _mm512_add_epi32(vacc0xCDEF, _mm512_madd_epi16(va0, vbCDEF)); in xnn_qs8_gemm_minmax_ukernel_3x16c8__avx512skx() 108 vacc1xCDEF = _mm512_add_epi32(vacc1xCDEF, _mm512_madd_epi16(va1, vbCDEF)); in xnn_qs8_gemm_minmax_ukernel_3x16c8__avx512skx() 109 vacc2xCDEF = _mm512_add_epi32(vacc2xCDEF, _mm512_madd_epi16(va2, vbCDEF)); in xnn_qs8_gemm_minmax_ukernel_3x16c8__avx512skx()
|
D | 2x16c8-minmax-avx512skx.c | 90 …const __m512i vbCDEF = _mm512_cvtepi8_epi16(_mm256_load_si256((const __m256i*) ((uintptr_t) w + 96… in xnn_qs8_gemm_minmax_ukernel_2x16c8__avx512skx() local 92 vacc0xCDEF = _mm512_add_epi32(vacc0xCDEF, _mm512_madd_epi16(va0, vbCDEF)); in xnn_qs8_gemm_minmax_ukernel_2x16c8__avx512skx() 93 vacc1xCDEF = _mm512_add_epi32(vacc1xCDEF, _mm512_madd_epi16(va1, vbCDEF)); in xnn_qs8_gemm_minmax_ukernel_2x16c8__avx512skx()
|
/external/XNNPACK/src/qs8-igemm/gen/ |
D | 4x16c8-minmax-avx512skx.c | 135 …const __m512i vbCDEF = _mm512_cvtepi8_epi16(_mm256_load_si256((const __m256i*) ((uintptr_t) w + 96… in xnn_qs8_igemm_minmax_ukernel_4x16c8__avx512skx() local 137 vacc0xCDEF = _mm512_add_epi32(vacc0xCDEF, _mm512_madd_epi16(va0, vbCDEF)); in xnn_qs8_igemm_minmax_ukernel_4x16c8__avx512skx() 138 vacc1xCDEF = _mm512_add_epi32(vacc1xCDEF, _mm512_madd_epi16(va1, vbCDEF)); in xnn_qs8_igemm_minmax_ukernel_4x16c8__avx512skx() 139 vacc2xCDEF = _mm512_add_epi32(vacc2xCDEF, _mm512_madd_epi16(va2, vbCDEF)); in xnn_qs8_igemm_minmax_ukernel_4x16c8__avx512skx() 140 vacc3xCDEF = _mm512_add_epi32(vacc3xCDEF, _mm512_madd_epi16(va3, vbCDEF)); in xnn_qs8_igemm_minmax_ukernel_4x16c8__avx512skx()
|
D | 3x16c8-minmax-avx512skx.c | 118 …const __m512i vbCDEF = _mm512_cvtepi8_epi16(_mm256_load_si256((const __m256i*) ((uintptr_t) w + 96… in xnn_qs8_igemm_minmax_ukernel_3x16c8__avx512skx() local 120 vacc0xCDEF = _mm512_add_epi32(vacc0xCDEF, _mm512_madd_epi16(va0, vbCDEF)); in xnn_qs8_igemm_minmax_ukernel_3x16c8__avx512skx() 121 vacc1xCDEF = _mm512_add_epi32(vacc1xCDEF, _mm512_madd_epi16(va1, vbCDEF)); in xnn_qs8_igemm_minmax_ukernel_3x16c8__avx512skx() 122 vacc2xCDEF = _mm512_add_epi32(vacc2xCDEF, _mm512_madd_epi16(va2, vbCDEF)); in xnn_qs8_igemm_minmax_ukernel_3x16c8__avx512skx()
|