/external/XNNPACK/src/f32-gemm/gen/ |
D | 3x8-wasmsimd-splat.c | 112 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_gemm_ukernel_3x8__wasmsimd_splat() local 119 vacc2x0123 = wasm_f32x4_add(vacc2x0123, wasm_f32x4_mul(va2c3, vb0123c3)); in xnn_f32_gemm_ukernel_3x8__wasmsimd_splat() 122 vacc2x4567 = wasm_f32x4_add(vacc2x4567, wasm_f32x4_mul(va2c3, vb4567c3)); in xnn_f32_gemm_ukernel_3x8__wasmsimd_splat()
|
D | 3x8-relu-wasmrelaxedsimd-fma-splat.c | 112 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_gemm_relu_ukernel_3x8__wasmrelaxedsimd_fma_splat() local 119 vacc2x0123 = wasm_f32x4_add(vacc2x0123, wasm_f32x4_mul(va2c3, vb0123c3)); in xnn_f32_gemm_relu_ukernel_3x8__wasmrelaxedsimd_fma_splat() 122 vacc2x4567 = wasm_f32x4_add(vacc2x4567, wasm_f32x4_mul(va2c3, vb4567c3)); in xnn_f32_gemm_relu_ukernel_3x8__wasmrelaxedsimd_fma_splat()
|
D | 3x8-wasmrelaxedsimd-fma-splat.c | 112 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_gemm_ukernel_3x8__wasmrelaxedsimd_fma_splat() local 119 vacc2x0123 = wasm_f32x4_add(vacc2x0123, wasm_f32x4_mul(va2c3, vb0123c3)); in xnn_f32_gemm_ukernel_3x8__wasmrelaxedsimd_fma_splat() 122 vacc2x4567 = wasm_f32x4_add(vacc2x4567, wasm_f32x4_mul(va2c3, vb4567c3)); in xnn_f32_gemm_ukernel_3x8__wasmrelaxedsimd_fma_splat()
|
D | 3x8-minmax-wasmrelaxedsimd-fma-splat.c | 114 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_gemm_minmax_ukernel_3x8__wasmrelaxedsimd_fma_splat() local 121 vacc2x0123 = wasm_f32x4_add(vacc2x0123, wasm_f32x4_mul(va2c3, vb0123c3)); in xnn_f32_gemm_minmax_ukernel_3x8__wasmrelaxedsimd_fma_splat() 124 vacc2x4567 = wasm_f32x4_add(vacc2x4567, wasm_f32x4_mul(va2c3, vb4567c3)); in xnn_f32_gemm_minmax_ukernel_3x8__wasmrelaxedsimd_fma_splat()
|
D | 3x8-minmax-wasmrelaxedsimd-splat.c | 114 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_gemm_minmax_ukernel_3x8__wasmrelaxedsimd_splat() local 121 vacc2x0123 = wasm_f32x4_add(vacc2x0123, wasm_f32x4_mul(va2c3, vb0123c3)); in xnn_f32_gemm_minmax_ukernel_3x8__wasmrelaxedsimd_splat() 124 vacc2x4567 = wasm_f32x4_add(vacc2x4567, wasm_f32x4_mul(va2c3, vb4567c3)); in xnn_f32_gemm_minmax_ukernel_3x8__wasmrelaxedsimd_splat()
|
D | 3x8-minmax-wasmsimd-x86-splat.c | 114 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_gemm_minmax_ukernel_3x8__wasmsimd_x86_splat() local 121 vacc2x0123 = wasm_f32x4_add(vacc2x0123, wasm_f32x4_mul(va2c3, vb0123c3)); in xnn_f32_gemm_minmax_ukernel_3x8__wasmsimd_x86_splat() 124 vacc2x4567 = wasm_f32x4_add(vacc2x4567, wasm_f32x4_mul(va2c3, vb4567c3)); in xnn_f32_gemm_minmax_ukernel_3x8__wasmsimd_x86_splat()
|
D | 3x8-minmax-wasmsimd-arm-splat.c | 114 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_gemm_minmax_ukernel_3x8__wasmsimd_arm_splat() local 121 vacc2x0123 = wasm_f32x4_add(vacc2x0123, wasm_f32x4_mul(va2c3, vb0123c3)); in xnn_f32_gemm_minmax_ukernel_3x8__wasmsimd_arm_splat() 124 vacc2x4567 = wasm_f32x4_add(vacc2x4567, wasm_f32x4_mul(va2c3, vb4567c3)); in xnn_f32_gemm_minmax_ukernel_3x8__wasmsimd_arm_splat()
|
D | 3x8-relu-wasmsimd-splat.c | 112 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_gemm_relu_ukernel_3x8__wasmsimd_splat() local 119 vacc2x0123 = wasm_f32x4_add(vacc2x0123, wasm_f32x4_mul(va2c3, vb0123c3)); in xnn_f32_gemm_relu_ukernel_3x8__wasmsimd_splat() 122 vacc2x4567 = wasm_f32x4_add(vacc2x4567, wasm_f32x4_mul(va2c3, vb4567c3)); in xnn_f32_gemm_relu_ukernel_3x8__wasmsimd_splat()
|
D | 4x8-wasmsimd-splat.c | 131 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_gemm_ukernel_4x8__wasmsimd_splat() local 139 vacc2x0123 = wasm_f32x4_add(vacc2x0123, wasm_f32x4_mul(va2c3, vb0123c3)); in xnn_f32_gemm_ukernel_4x8__wasmsimd_splat() 143 vacc2x4567 = wasm_f32x4_add(vacc2x4567, wasm_f32x4_mul(va2c3, vb4567c3)); in xnn_f32_gemm_ukernel_4x8__wasmsimd_splat()
|
D | 4x8-wasmrelaxedsimd-fma-splat.c | 131 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_gemm_ukernel_4x8__wasmrelaxedsimd_fma_splat() local 139 vacc2x0123 = wasm_f32x4_add(vacc2x0123, wasm_f32x4_mul(va2c3, vb0123c3)); in xnn_f32_gemm_ukernel_4x8__wasmrelaxedsimd_fma_splat() 143 vacc2x4567 = wasm_f32x4_add(vacc2x4567, wasm_f32x4_mul(va2c3, vb4567c3)); in xnn_f32_gemm_ukernel_4x8__wasmrelaxedsimd_fma_splat()
|
/external/XNNPACK/src/f32-igemm/gen/ |
D | 3x8-wasmrelaxedsimd-fma-splat.c | 131 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_igemm_ukernel_3x8__wasmrelaxedsimd_fma_splat() local 138 vacc2x0123 = __builtin_wasm_fma_f32x4(vacc2x0123, va2c3, vb0123c3); in xnn_f32_igemm_ukernel_3x8__wasmrelaxedsimd_fma_splat() 141 vacc2x4567 = __builtin_wasm_fma_f32x4(vacc2x4567, va2c3, vb4567c3); in xnn_f32_igemm_ukernel_3x8__wasmrelaxedsimd_fma_splat()
|
D | 3x8-minmax-wasmrelaxedsimd-fma-splat.c | 133 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_igemm_minmax_ukernel_3x8__wasmrelaxedsimd_fma_splat() local 140 vacc2x0123 = __builtin_wasm_fma_f32x4(vacc2x0123, va2c3, vb0123c3); in xnn_f32_igemm_minmax_ukernel_3x8__wasmrelaxedsimd_fma_splat() 143 vacc2x4567 = __builtin_wasm_fma_f32x4(vacc2x4567, va2c3, vb4567c3); in xnn_f32_igemm_minmax_ukernel_3x8__wasmrelaxedsimd_fma_splat()
|
D | 3x8-wasmsimd-splat.c | 131 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_igemm_ukernel_3x8__wasmsimd_splat() local 138 vacc2x0123 = wasm_f32x4_add(vacc2x0123, wasm_f32x4_mul(va2c3, vb0123c3)); in xnn_f32_igemm_ukernel_3x8__wasmsimd_splat() 141 vacc2x4567 = wasm_f32x4_add(vacc2x4567, wasm_f32x4_mul(va2c3, vb4567c3)); in xnn_f32_igemm_ukernel_3x8__wasmsimd_splat()
|
D | 3x8-relu-wasmsimd-splat.c | 131 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_igemm_relu_ukernel_3x8__wasmsimd_splat() local 138 vacc2x0123 = wasm_f32x4_add(vacc2x0123, wasm_f32x4_mul(va2c3, vb0123c3)); in xnn_f32_igemm_relu_ukernel_3x8__wasmsimd_splat() 141 vacc2x4567 = wasm_f32x4_add(vacc2x4567, wasm_f32x4_mul(va2c3, vb4567c3)); in xnn_f32_igemm_relu_ukernel_3x8__wasmsimd_splat()
|
D | 3x8-relu-wasmrelaxedsimd-fma-splat.c | 131 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_igemm_relu_ukernel_3x8__wasmrelaxedsimd_fma_splat() local 138 vacc2x0123 = __builtin_wasm_fma_f32x4(vacc2x0123, va2c3, vb0123c3); in xnn_f32_igemm_relu_ukernel_3x8__wasmrelaxedsimd_fma_splat() 141 vacc2x4567 = __builtin_wasm_fma_f32x4(vacc2x4567, va2c3, vb4567c3); in xnn_f32_igemm_relu_ukernel_3x8__wasmrelaxedsimd_fma_splat()
|
D | 3x8-minmax-wasmsimd-arm-splat.c | 133 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_igemm_minmax_ukernel_3x8__wasmsimd_arm_splat() local 140 vacc2x0123 = wasm_f32x4_add(vacc2x0123, wasm_f32x4_mul(va2c3, vb0123c3)); in xnn_f32_igemm_minmax_ukernel_3x8__wasmsimd_arm_splat() 143 vacc2x4567 = wasm_f32x4_add(vacc2x4567, wasm_f32x4_mul(va2c3, vb4567c3)); in xnn_f32_igemm_minmax_ukernel_3x8__wasmsimd_arm_splat()
|
D | 4x8-wasmrelaxedsimd-fma-splat.c | 153 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_igemm_ukernel_4x8__wasmrelaxedsimd_fma_splat() local 161 vacc2x0123 = __builtin_wasm_fma_f32x4(vacc2x0123, va2c3, vb0123c3); in xnn_f32_igemm_ukernel_4x8__wasmrelaxedsimd_fma_splat() 165 vacc2x4567 = __builtin_wasm_fma_f32x4(vacc2x4567, va2c3, vb4567c3); in xnn_f32_igemm_ukernel_4x8__wasmrelaxedsimd_fma_splat()
|
D | 3x8-minmax-wasmsimd-x86-splat.c | 133 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_igemm_minmax_ukernel_3x8__wasmsimd_x86_splat() local 140 vacc2x0123 = wasm_f32x4_add(vacc2x0123, wasm_f32x4_mul(va2c3, vb0123c3)); in xnn_f32_igemm_minmax_ukernel_3x8__wasmsimd_x86_splat() 143 vacc2x4567 = wasm_f32x4_add(vacc2x4567, wasm_f32x4_mul(va2c3, vb4567c3)); in xnn_f32_igemm_minmax_ukernel_3x8__wasmsimd_x86_splat()
|
D | 3x8-minmax-wasmrelaxedsimd-splat.c | 133 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_igemm_minmax_ukernel_3x8__wasmrelaxedsimd_splat() local 140 vacc2x0123 = wasm_f32x4_add(vacc2x0123, wasm_f32x4_mul(va2c3, vb0123c3)); in xnn_f32_igemm_minmax_ukernel_3x8__wasmrelaxedsimd_splat() 143 vacc2x4567 = wasm_f32x4_add(vacc2x4567, wasm_f32x4_mul(va2c3, vb4567c3)); in xnn_f32_igemm_minmax_ukernel_3x8__wasmrelaxedsimd_splat()
|
D | 4x8-relu-wasmrelaxedsimd-fma-splat.c | 153 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_igemm_relu_ukernel_4x8__wasmrelaxedsimd_fma_splat() local 161 vacc2x0123 = __builtin_wasm_fma_f32x4(vacc2x0123, va2c3, vb0123c3); in xnn_f32_igemm_relu_ukernel_4x8__wasmrelaxedsimd_fma_splat() 165 vacc2x4567 = __builtin_wasm_fma_f32x4(vacc2x4567, va2c3, vb4567c3); in xnn_f32_igemm_relu_ukernel_4x8__wasmrelaxedsimd_fma_splat()
|
/external/XNNPACK/src/f32-gemm/gen-inc/ |
D | 3x8inc-minmax-wasmrelaxedsimd-fma-splat.c | 116 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_gemminc_minmax_ukernel_3x8__wasmrelaxedsimd_fma_splat() local 123 vacc2x0123 = wasm_f32x4_add(vacc2x0123, wasm_f32x4_mul(va2c3, vb0123c3)); in xnn_f32_gemminc_minmax_ukernel_3x8__wasmrelaxedsimd_fma_splat() 126 vacc2x4567 = wasm_f32x4_add(vacc2x4567, wasm_f32x4_mul(va2c3, vb4567c3)); in xnn_f32_gemminc_minmax_ukernel_3x8__wasmrelaxedsimd_fma_splat()
|
D | 3x8inc-minmax-wasmrelaxedsimd-splat.c | 116 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_gemminc_minmax_ukernel_3x8__wasmrelaxedsimd_splat() local 123 vacc2x0123 = wasm_f32x4_add(vacc2x0123, wasm_f32x4_mul(va2c3, vb0123c3)); in xnn_f32_gemminc_minmax_ukernel_3x8__wasmrelaxedsimd_splat() 126 vacc2x4567 = wasm_f32x4_add(vacc2x4567, wasm_f32x4_mul(va2c3, vb4567c3)); in xnn_f32_gemminc_minmax_ukernel_3x8__wasmrelaxedsimd_splat()
|
D | 3x8inc-minmax-wasmsimd-arm-splat.c | 116 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_gemminc_minmax_ukernel_3x8__wasmsimd_arm_splat() local 123 vacc2x0123 = wasm_f32x4_add(vacc2x0123, wasm_f32x4_mul(va2c3, vb0123c3)); in xnn_f32_gemminc_minmax_ukernel_3x8__wasmsimd_arm_splat() 126 vacc2x4567 = wasm_f32x4_add(vacc2x4567, wasm_f32x4_mul(va2c3, vb4567c3)); in xnn_f32_gemminc_minmax_ukernel_3x8__wasmsimd_arm_splat()
|
D | 3x8inc-minmax-relaxedwasmsimd-splat.c | 116 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_gemminc_minmax_ukernel_3x8__relaxedwasmsimd_splat() local 123 vacc2x0123 = wasm_f32x4_add(vacc2x0123, wasm_f32x4_mul(va2c3, vb0123c3)); in xnn_f32_gemminc_minmax_ukernel_3x8__relaxedwasmsimd_splat() 126 vacc2x4567 = wasm_f32x4_add(vacc2x4567, wasm_f32x4_mul(va2c3, vb4567c3)); in xnn_f32_gemminc_minmax_ukernel_3x8__relaxedwasmsimd_splat()
|
D | 3x8inc-minmax-wasmsimd-x86-splat.c | 116 const v128_t va2c3 = wasm_v32x4_shuffle(va2, va2, 3, 3, 3, 3); in xnn_f32_gemminc_minmax_ukernel_3x8__wasmsimd_x86_splat() local 123 vacc2x0123 = wasm_f32x4_add(vacc2x0123, wasm_f32x4_mul(va2c3, vb0123c3)); in xnn_f32_gemminc_minmax_ukernel_3x8__wasmsimd_x86_splat() 126 vacc2x4567 = wasm_f32x4_add(vacc2x4567, wasm_f32x4_mul(va2c3, vb4567c3)); in xnn_f32_gemminc_minmax_ukernel_3x8__wasmsimd_x86_splat()
|