/external/XNNPACK/src/qs8-gemm/gen/ |
D | 1x4c8-minmax-wasmsimd-ld64.c | 45 v128_t vacc0x0 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[0]); in xnn_qs8_gemm_minmax_ukernel_1x4c8__wasmsimd_ld64() 46 v128_t vacc0x1 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[1]); in xnn_qs8_gemm_minmax_ukernel_1x4c8__wasmsimd_ld64() 47 v128_t vacc0x2 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[2]); in xnn_qs8_gemm_minmax_ukernel_1x4c8__wasmsimd_ld64() 48 v128_t vacc0x3 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[3]); in xnn_qs8_gemm_minmax_ukernel_1x4c8__wasmsimd_ld64()
|
D | 1x4c8-minmax-wasmsimd-ld128.c | 45 v128_t vacc0x0 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[0]); in xnn_qs8_gemm_minmax_ukernel_1x4c8__wasmsimd_ld128() 46 v128_t vacc0x1 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[1]); in xnn_qs8_gemm_minmax_ukernel_1x4c8__wasmsimd_ld128() 47 v128_t vacc0x2 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[2]); in xnn_qs8_gemm_minmax_ukernel_1x4c8__wasmsimd_ld128() 48 v128_t vacc0x3 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[3]); in xnn_qs8_gemm_minmax_ukernel_1x4c8__wasmsimd_ld128()
|
D | 1x4c8-xw-minmax-wasmsimd.c | 45 v128_t vacc0x0 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[0]); in xnn_qs8_gemm_xw_minmax_ukernel_1x4c8__wasmsimd() 46 v128_t vacc0x1 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[1]); in xnn_qs8_gemm_xw_minmax_ukernel_1x4c8__wasmsimd() 47 v128_t vacc0x2 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[2]); in xnn_qs8_gemm_xw_minmax_ukernel_1x4c8__wasmsimd() 48 v128_t vacc0x3 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[3]); in xnn_qs8_gemm_xw_minmax_ukernel_1x4c8__wasmsimd()
|
D | 2x4c8-xw-minmax-wasmsimd.c | 51 v128_t vacc0x0 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[0]); in xnn_qs8_gemm_xw_minmax_ukernel_2x4c8__wasmsimd() 52 v128_t vacc0x1 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[1]); in xnn_qs8_gemm_xw_minmax_ukernel_2x4c8__wasmsimd() 53 v128_t vacc0x2 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[2]); in xnn_qs8_gemm_xw_minmax_ukernel_2x4c8__wasmsimd() 54 v128_t vacc0x3 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[3]); in xnn_qs8_gemm_xw_minmax_ukernel_2x4c8__wasmsimd()
|
D | 2x4c8-minmax-wasmsimd-ld128.c | 51 v128_t vacc0x0 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[0]); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld128() 52 v128_t vacc0x1 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[1]); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld128() 53 v128_t vacc0x2 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[2]); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld128() 54 v128_t vacc0x3 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[3]); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld128()
|
D | 2x4c8-minmax-wasmsimd-ld64.c | 51 v128_t vacc0x0 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[0]); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld64() 52 v128_t vacc0x1 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[1]); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld64() 53 v128_t vacc0x2 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[2]); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld64() 54 v128_t vacc0x3 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[3]); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld64()
|
D | 3x4c8-xw-minmax-wasmsimd.c | 57 v128_t vacc0x0 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[0]); in xnn_qs8_gemm_xw_minmax_ukernel_3x4c8__wasmsimd() 58 v128_t vacc0x1 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[1]); in xnn_qs8_gemm_xw_minmax_ukernel_3x4c8__wasmsimd() 59 v128_t vacc0x2 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[2]); in xnn_qs8_gemm_xw_minmax_ukernel_3x4c8__wasmsimd() 60 v128_t vacc0x3 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[3]); in xnn_qs8_gemm_xw_minmax_ukernel_3x4c8__wasmsimd()
|
D | 3x4c8-minmax-wasmsimd-ld128.c | 57 v128_t vacc0x0 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[0]); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld128() 58 v128_t vacc0x1 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[1]); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld128() 59 v128_t vacc0x2 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[2]); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld128() 60 v128_t vacc0x3 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[3]); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld128()
|
D | 3x4c8-minmax-wasmsimd-ld64.c | 57 v128_t vacc0x0 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[0]); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld64() 58 v128_t vacc0x1 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[1]); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld64() 59 v128_t vacc0x2 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[2]); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld64() 60 v128_t vacc0x3 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[3]); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld64()
|
/external/XNNPACK/src/qs8-igemm/gen/ |
D | 1x4c8-minmax-wasmsimd-ld64.c | 48 v128_t vacc0x0 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[0]); in xnn_qs8_igemm_minmax_ukernel_1x4c8__wasmsimd_ld64() 49 v128_t vacc0x1 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[1]); in xnn_qs8_igemm_minmax_ukernel_1x4c8__wasmsimd_ld64() 50 v128_t vacc0x2 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[2]); in xnn_qs8_igemm_minmax_ukernel_1x4c8__wasmsimd_ld64() 51 v128_t vacc0x3 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[3]); in xnn_qs8_igemm_minmax_ukernel_1x4c8__wasmsimd_ld64()
|
D | 1x4c8-minmax-wasmsimd-ld128.c | 48 v128_t vacc0x0 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[0]); in xnn_qs8_igemm_minmax_ukernel_1x4c8__wasmsimd_ld128() 49 v128_t vacc0x1 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[1]); in xnn_qs8_igemm_minmax_ukernel_1x4c8__wasmsimd_ld128() 50 v128_t vacc0x2 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[2]); in xnn_qs8_igemm_minmax_ukernel_1x4c8__wasmsimd_ld128() 51 v128_t vacc0x3 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[3]); in xnn_qs8_igemm_minmax_ukernel_1x4c8__wasmsimd_ld128()
|
D | 2x4c8-minmax-wasmsimd-ld128.c | 52 v128_t vacc0x0 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[0]); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld128() 53 v128_t vacc0x1 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[1]); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld128() 54 v128_t vacc0x2 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[2]); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld128() 55 v128_t vacc0x3 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[3]); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld128()
|
D | 2x4c8-minmax-wasmsimd-ld64.c | 52 v128_t vacc0x0 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[0]); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld64() 53 v128_t vacc0x1 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[1]); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld64() 54 v128_t vacc0x2 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[2]); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld64() 55 v128_t vacc0x3 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[3]); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld64()
|
D | 3x4c8-minmax-wasmsimd-ld128.c | 56 v128_t vacc0x0 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[0]); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld128() 57 v128_t vacc0x1 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[1]); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld128() 58 v128_t vacc0x2 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[2]); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld128() 59 v128_t vacc0x3 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[3]); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld128()
|
D | 3x4c8-minmax-wasmsimd-ld64.c | 56 v128_t vacc0x0 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[0]); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld64() 57 v128_t vacc0x1 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[1]); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld64() 58 v128_t vacc0x2 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[2]); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld64() 59 v128_t vacc0x3 = wasm_f32x4_replace_lane(vzero, 0, ((const float*) w)[3]); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld64()
|
/external/llvm-project/clang/test/Headers/ |
D | wasm.c | 35 v = wasm_f32x4_replace_lane(v, 0, 42.0); in main()
|
/external/XNNPACK/src/f32-gemm/gen/ |
D | 4x2c4-wasmsimd.c | 60 v128_t vacc0x0c4 = wasm_f32x4_replace_lane(wasm_f32x4_splat(0.0f), 0, w[0]); in xnn_f32_gemm_ukernel_4x2c4__wasmsimd() 61 v128_t vacc0x1c4 = wasm_f32x4_replace_lane(vacc0x0c4, 0, w[1]); in xnn_f32_gemm_ukernel_4x2c4__wasmsimd()
|
D | 4x2c4-relu-wasmsimd.c | 60 v128_t vacc0x0c4 = wasm_f32x4_replace_lane(wasm_f32x4_splat(0.0f), 0, w[0]); in xnn_f32_gemm_relu_ukernel_4x2c4__wasmsimd() 61 v128_t vacc0x1c4 = wasm_f32x4_replace_lane(vacc0x0c4, 0, w[1]); in xnn_f32_gemm_relu_ukernel_4x2c4__wasmsimd()
|
D | 4x2c4-minmax-wasmsimd-arm.c | 62 v128_t vacc0x0c4 = wasm_f32x4_replace_lane(wasm_f32x4_splat(0.0f), 0, w[0]); in xnn_f32_gemm_minmax_ukernel_4x2c4__wasmsimd_arm() 63 v128_t vacc0x1c4 = wasm_f32x4_replace_lane(vacc0x0c4, 0, w[1]); in xnn_f32_gemm_minmax_ukernel_4x2c4__wasmsimd_arm()
|
D | 4x2c4-minmax-wasmsimd-x86.c | 60 v128_t vacc0x0c4 = wasm_f32x4_replace_lane(wasm_f32x4_splat(0.0f), 0, w[0]); in xnn_f32_gemm_minmax_ukernel_4x2c4__wasmsimd_x86() 61 v128_t vacc0x1c4 = wasm_f32x4_replace_lane(vacc0x0c4, 0, w[1]); in xnn_f32_gemm_minmax_ukernel_4x2c4__wasmsimd_x86()
|
/external/XNNPACK/src/f32-igemm/gen/ |
D | 4x2c4-relu-wasmsimd.c | 58 v128_t vacc0x0c4 = wasm_f32x4_replace_lane(wasm_f32x4_splat(0.0f), 0, w[0]); in xnn_f32_igemm_relu_ukernel_4x2c4__wasmsimd() 59 v128_t vacc0x1c4 = wasm_f32x4_replace_lane(vacc0x0c4, 0, w[1]); in xnn_f32_igemm_relu_ukernel_4x2c4__wasmsimd()
|
D | 4x2c4-wasmsimd.c | 58 v128_t vacc0x0c4 = wasm_f32x4_replace_lane(wasm_f32x4_splat(0.0f), 0, w[0]); in xnn_f32_igemm_ukernel_4x2c4__wasmsimd() 59 v128_t vacc0x1c4 = wasm_f32x4_replace_lane(vacc0x0c4, 0, w[1]); in xnn_f32_igemm_ukernel_4x2c4__wasmsimd()
|
D | 4x2c4-minmax-wasmsimd-x86.c | 58 v128_t vacc0x0c4 = wasm_f32x4_replace_lane(wasm_f32x4_splat(0.0f), 0, w[0]); in xnn_f32_igemm_minmax_ukernel_4x2c4__wasmsimd_x86() 59 v128_t vacc0x1c4 = wasm_f32x4_replace_lane(vacc0x0c4, 0, w[1]); in xnn_f32_igemm_minmax_ukernel_4x2c4__wasmsimd_x86()
|
D | 4x2c4-minmax-wasmsimd-arm.c | 60 v128_t vacc0x0c4 = wasm_f32x4_replace_lane(wasm_f32x4_splat(0.0f), 0, w[0]); in xnn_f32_igemm_minmax_ukernel_4x2c4__wasmsimd_arm() 61 v128_t vacc0x1c4 = wasm_f32x4_replace_lane(vacc0x0c4, 0, w[1]); in xnn_f32_igemm_minmax_ukernel_4x2c4__wasmsimd_arm()
|
/external/XNNPACK/src/f32-igemm/ |
D | MRx2c4-wasmsimd.c.in | 65 v128_t vacc0x0c4 = wasm_f32x4_replace_lane(wasm_f32x4_splat(0.0f), 0, w[0]); 67 v128_t vacc0x${N}c4 = wasm_f32x4_replace_lane(vacc0x0c4, 0, w[${N}]);
|