Home
last modified time | relevance | path

Searched refs:wasm_i16x8_mul (Results 1 – 23 of 23) sorted by relevance

/external/XNNPACK/src/qs8-dwconv/gen/
Dup24x9-minmax-wasmsimd-mul16.c99 const v128_t vprod0x01234567 = wasm_i16x8_mul(vi0x01234567, vk0x01234567); in xnn_qs8_dwconv_minmax_ukernel_up24x9__wasmsimd_mul16()
100 const v128_t vprod0x89ABCDEF = wasm_i16x8_mul(vi0x89ABCDEF, vk0x89ABCDEF); in xnn_qs8_dwconv_minmax_ukernel_up24x9__wasmsimd_mul16()
101 const v128_t vprod0xGHIJKLMN = wasm_i16x8_mul(vi0xGHIJKLMN, vk0xGHIJKLMN); in xnn_qs8_dwconv_minmax_ukernel_up24x9__wasmsimd_mul16()
118 const v128_t vprod1x01234567 = wasm_i16x8_mul(vi1x01234567, vk1x01234567); in xnn_qs8_dwconv_minmax_ukernel_up24x9__wasmsimd_mul16()
119 const v128_t vprod1x89ABCDEF = wasm_i16x8_mul(vi1x89ABCDEF, vk1x89ABCDEF); in xnn_qs8_dwconv_minmax_ukernel_up24x9__wasmsimd_mul16()
120 const v128_t vprod1xGHIJKLMN = wasm_i16x8_mul(vi1xGHIJKLMN, vk1xGHIJKLMN); in xnn_qs8_dwconv_minmax_ukernel_up24x9__wasmsimd_mul16()
137 const v128_t vprod2x01234567 = wasm_i16x8_mul(vi2x01234567, vk2x01234567); in xnn_qs8_dwconv_minmax_ukernel_up24x9__wasmsimd_mul16()
138 const v128_t vprod2x89ABCDEF = wasm_i16x8_mul(vi2x89ABCDEF, vk2x89ABCDEF); in xnn_qs8_dwconv_minmax_ukernel_up24x9__wasmsimd_mul16()
139 const v128_t vprod2xGHIJKLMN = wasm_i16x8_mul(vi2xGHIJKLMN, vk2xGHIJKLMN); in xnn_qs8_dwconv_minmax_ukernel_up24x9__wasmsimd_mul16()
156 const v128_t vprod3x01234567 = wasm_i16x8_mul(vi3x01234567, vk3x01234567); in xnn_qs8_dwconv_minmax_ukernel_up24x9__wasmsimd_mul16()
[all …]
Dup16x9-minmax-wasmsimd-mul16.c95 const v128_t vprod0x01234567 = wasm_i16x8_mul(vi0x01234567, vk0x01234567); in xnn_qs8_dwconv_minmax_ukernel_up16x9__wasmsimd_mul16()
96 const v128_t vprod0x89ABCDEF = wasm_i16x8_mul(vi0x89ABCDEF, vk0x89ABCDEF); in xnn_qs8_dwconv_minmax_ukernel_up16x9__wasmsimd_mul16()
109 const v128_t vprod1x01234567 = wasm_i16x8_mul(vi1x01234567, vk1x01234567); in xnn_qs8_dwconv_minmax_ukernel_up16x9__wasmsimd_mul16()
110 const v128_t vprod1x89ABCDEF = wasm_i16x8_mul(vi1x89ABCDEF, vk1x89ABCDEF); in xnn_qs8_dwconv_minmax_ukernel_up16x9__wasmsimd_mul16()
123 const v128_t vprod2x01234567 = wasm_i16x8_mul(vi2x01234567, vk2x01234567); in xnn_qs8_dwconv_minmax_ukernel_up16x9__wasmsimd_mul16()
124 const v128_t vprod2x89ABCDEF = wasm_i16x8_mul(vi2x89ABCDEF, vk2x89ABCDEF); in xnn_qs8_dwconv_minmax_ukernel_up16x9__wasmsimd_mul16()
137 const v128_t vprod3x01234567 = wasm_i16x8_mul(vi3x01234567, vk3x01234567); in xnn_qs8_dwconv_minmax_ukernel_up16x9__wasmsimd_mul16()
138 const v128_t vprod3x89ABCDEF = wasm_i16x8_mul(vi3x89ABCDEF, vk3x89ABCDEF); in xnn_qs8_dwconv_minmax_ukernel_up16x9__wasmsimd_mul16()
151 const v128_t vprod4x01234567 = wasm_i16x8_mul(vi4x01234567, vk4x01234567); in xnn_qs8_dwconv_minmax_ukernel_up16x9__wasmsimd_mul16()
152 const v128_t vprod4x89ABCDEF = wasm_i16x8_mul(vi4x89ABCDEF, vk4x89ABCDEF); in xnn_qs8_dwconv_minmax_ukernel_up16x9__wasmsimd_mul16()
[all …]
Dup8x9-minmax-wasmsimd-mul16.c91 const v128_t vprod0x01234567 = wasm_i16x8_mul(vi0x01234567, vk0x01234567); in xnn_qs8_dwconv_minmax_ukernel_up8x9__wasmsimd_mul16()
100 const v128_t vprod1x01234567 = wasm_i16x8_mul(vi1x01234567, vk1x01234567); in xnn_qs8_dwconv_minmax_ukernel_up8x9__wasmsimd_mul16()
109 const v128_t vprod2x01234567 = wasm_i16x8_mul(vi2x01234567, vk2x01234567); in xnn_qs8_dwconv_minmax_ukernel_up8x9__wasmsimd_mul16()
118 const v128_t vprod3x01234567 = wasm_i16x8_mul(vi3x01234567, vk3x01234567); in xnn_qs8_dwconv_minmax_ukernel_up8x9__wasmsimd_mul16()
127 const v128_t vprod4x01234567 = wasm_i16x8_mul(vi4x01234567, vk4x01234567); in xnn_qs8_dwconv_minmax_ukernel_up8x9__wasmsimd_mul16()
136 const v128_t vprod5x01234567 = wasm_i16x8_mul(vi5x01234567, vk5x01234567); in xnn_qs8_dwconv_minmax_ukernel_up8x9__wasmsimd_mul16()
145 const v128_t vprod6x01234567 = wasm_i16x8_mul(vi6x01234567, vk6x01234567); in xnn_qs8_dwconv_minmax_ukernel_up8x9__wasmsimd_mul16()
154 const v128_t vprod7x01234567 = wasm_i16x8_mul(vi7x01234567, vk7x01234567); in xnn_qs8_dwconv_minmax_ukernel_up8x9__wasmsimd_mul16()
163 const v128_t vprod8x01234567 = wasm_i16x8_mul(vi8x01234567, vk8x01234567); in xnn_qs8_dwconv_minmax_ukernel_up8x9__wasmsimd_mul16()
216 const v128_t vprod0x01234567 = wasm_i16x8_mul(vi0x01234567, vk0x01234567); in xnn_qs8_dwconv_minmax_ukernel_up8x9__wasmsimd_mul16()
[all …]
/external/XNNPACK/src/qs8-gemm/gen/
D3x4c8-xw-minmax-wasmsimd.c82 const v128_t vprod0x0 = wasm_i16x8_mul(vxa0, vxb0); in xnn_qs8_gemm_xw_minmax_ukernel_3x4c8__wasmsimd()
85 const v128_t vprod1x0 = wasm_i16x8_mul(vxa1, vxb0); in xnn_qs8_gemm_xw_minmax_ukernel_3x4c8__wasmsimd()
88 const v128_t vprod2x0 = wasm_i16x8_mul(vxa2, vxb0); in xnn_qs8_gemm_xw_minmax_ukernel_3x4c8__wasmsimd()
93 const v128_t vprod0x1 = wasm_i16x8_mul(vxa0, vxb1); in xnn_qs8_gemm_xw_minmax_ukernel_3x4c8__wasmsimd()
96 const v128_t vprod1x1 = wasm_i16x8_mul(vxa1, vxb1); in xnn_qs8_gemm_xw_minmax_ukernel_3x4c8__wasmsimd()
99 const v128_t vprod2x1 = wasm_i16x8_mul(vxa2, vxb1); in xnn_qs8_gemm_xw_minmax_ukernel_3x4c8__wasmsimd()
104 const v128_t vprod0x2 = wasm_i16x8_mul(vxa0, vxb2); in xnn_qs8_gemm_xw_minmax_ukernel_3x4c8__wasmsimd()
107 const v128_t vprod1x2 = wasm_i16x8_mul(vxa1, vxb2); in xnn_qs8_gemm_xw_minmax_ukernel_3x4c8__wasmsimd()
110 const v128_t vprod2x2 = wasm_i16x8_mul(vxa2, vxb2); in xnn_qs8_gemm_xw_minmax_ukernel_3x4c8__wasmsimd()
115 const v128_t vprod0x3 = wasm_i16x8_mul(vxa0, vxb3); in xnn_qs8_gemm_xw_minmax_ukernel_3x4c8__wasmsimd()
[all …]
D3x4c8-minmax-wasmsimd-ld128.c84 const v128_t vprod0x0 = wasm_i16x8_mul(vxb0, vxa0); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld128()
86 const v128_t vprod1x0 = wasm_i16x8_mul(vxb0, vxa1); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld128()
88 const v128_t vprod2x0 = wasm_i16x8_mul(vxb0, vxa2); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld128()
91 const v128_t vprod0x1 = wasm_i16x8_mul(vxb1, vxa0); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld128()
94 const v128_t vprod1x1 = wasm_i16x8_mul(vxb1, vxa1); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld128()
97 const v128_t vprod2x1 = wasm_i16x8_mul(vxb1, vxa2); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld128()
108 const v128_t vprod0x2 = wasm_i16x8_mul(vxb2, vxa0); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld128()
110 const v128_t vprod1x2 = wasm_i16x8_mul(vxb2, vxa1); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld128()
112 const v128_t vprod2x2 = wasm_i16x8_mul(vxb2, vxa2); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld128()
115 const v128_t vprod0x3 = wasm_i16x8_mul(vxb3, vxa0); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld128()
[all …]
D3x4c8-minmax-wasmsimd-ld64.c82 const v128_t vprod0x0 = wasm_i16x8_mul(vxa0, vxb0); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld64()
85 const v128_t vprod1x0 = wasm_i16x8_mul(vxa1, vxb0); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld64()
88 const v128_t vprod2x0 = wasm_i16x8_mul(vxa2, vxb0); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld64()
93 const v128_t vprod0x1 = wasm_i16x8_mul(vxa0, vxb1); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld64()
96 const v128_t vprod1x1 = wasm_i16x8_mul(vxa1, vxb1); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld64()
99 const v128_t vprod2x1 = wasm_i16x8_mul(vxa2, vxb1); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld64()
104 const v128_t vprod0x2 = wasm_i16x8_mul(vxa0, vxb2); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld64()
107 const v128_t vprod1x2 = wasm_i16x8_mul(vxa1, vxb2); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld64()
110 const v128_t vprod2x2 = wasm_i16x8_mul(vxa2, vxb2); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld64()
115 const v128_t vprod0x3 = wasm_i16x8_mul(vxa0, vxb3); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld64()
[all …]
D2x4c8-xw-minmax-wasmsimd.c70 const v128_t vprod0x0 = wasm_i16x8_mul(vxa0, vxb0); in xnn_qs8_gemm_xw_minmax_ukernel_2x4c8__wasmsimd()
73 const v128_t vprod1x0 = wasm_i16x8_mul(vxa1, vxb0); in xnn_qs8_gemm_xw_minmax_ukernel_2x4c8__wasmsimd()
78 const v128_t vprod0x1 = wasm_i16x8_mul(vxa0, vxb1); in xnn_qs8_gemm_xw_minmax_ukernel_2x4c8__wasmsimd()
81 const v128_t vprod1x1 = wasm_i16x8_mul(vxa1, vxb1); in xnn_qs8_gemm_xw_minmax_ukernel_2x4c8__wasmsimd()
86 const v128_t vprod0x2 = wasm_i16x8_mul(vxa0, vxb2); in xnn_qs8_gemm_xw_minmax_ukernel_2x4c8__wasmsimd()
89 const v128_t vprod1x2 = wasm_i16x8_mul(vxa1, vxb2); in xnn_qs8_gemm_xw_minmax_ukernel_2x4c8__wasmsimd()
94 const v128_t vprod0x3 = wasm_i16x8_mul(vxa0, vxb3); in xnn_qs8_gemm_xw_minmax_ukernel_2x4c8__wasmsimd()
97 const v128_t vprod1x3 = wasm_i16x8_mul(vxa1, vxb3); in xnn_qs8_gemm_xw_minmax_ukernel_2x4c8__wasmsimd()
D2x4c8-minmax-wasmsimd-ld128.c72 const v128_t vprod0x0 = wasm_i16x8_mul(vxb0, vxa0); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld128()
74 const v128_t vprod1x0 = wasm_i16x8_mul(vxb0, vxa1); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld128()
77 const v128_t vprod0x1 = wasm_i16x8_mul(vxb1, vxa0); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld128()
80 const v128_t vprod1x1 = wasm_i16x8_mul(vxb1, vxa1); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld128()
90 const v128_t vprod0x2 = wasm_i16x8_mul(vxb2, vxa0); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld128()
92 const v128_t vprod1x2 = wasm_i16x8_mul(vxb2, vxa1); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld128()
95 const v128_t vprod0x3 = wasm_i16x8_mul(vxb3, vxa0); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld128()
98 const v128_t vprod1x3 = wasm_i16x8_mul(vxb3, vxa1); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld128()
D2x4c8-minmax-wasmsimd-ld64.c70 const v128_t vprod0x0 = wasm_i16x8_mul(vxa0, vxb0); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld64()
73 const v128_t vprod1x0 = wasm_i16x8_mul(vxa1, vxb0); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld64()
78 const v128_t vprod0x1 = wasm_i16x8_mul(vxa0, vxb1); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld64()
81 const v128_t vprod1x1 = wasm_i16x8_mul(vxa1, vxb1); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld64()
86 const v128_t vprod0x2 = wasm_i16x8_mul(vxa0, vxb2); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld64()
89 const v128_t vprod1x2 = wasm_i16x8_mul(vxa1, vxb2); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld64()
94 const v128_t vprod0x3 = wasm_i16x8_mul(vxa0, vxb3); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld64()
97 const v128_t vprod1x3 = wasm_i16x8_mul(vxa1, vxb3); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld64()
D1x4c8-minmax-wasmsimd-ld64.c58 const v128_t vprod0x0 = wasm_i16x8_mul(vxa0, vxb0); in xnn_qs8_gemm_minmax_ukernel_1x4c8__wasmsimd_ld64()
63 const v128_t vprod0x1 = wasm_i16x8_mul(vxa0, vxb1); in xnn_qs8_gemm_minmax_ukernel_1x4c8__wasmsimd_ld64()
68 const v128_t vprod0x2 = wasm_i16x8_mul(vxa0, vxb2); in xnn_qs8_gemm_minmax_ukernel_1x4c8__wasmsimd_ld64()
73 const v128_t vprod0x3 = wasm_i16x8_mul(vxa0, vxb3); in xnn_qs8_gemm_minmax_ukernel_1x4c8__wasmsimd_ld64()
D1x4c8-minmax-wasmsimd-ld128.c60 const v128_t vprod0x0 = wasm_i16x8_mul(vxb0, vxa0); in xnn_qs8_gemm_minmax_ukernel_1x4c8__wasmsimd_ld128()
63 const v128_t vprod0x1 = wasm_i16x8_mul(vxb1, vxa0); in xnn_qs8_gemm_minmax_ukernel_1x4c8__wasmsimd_ld128()
72 const v128_t vprod0x2 = wasm_i16x8_mul(vxb2, vxa0); in xnn_qs8_gemm_minmax_ukernel_1x4c8__wasmsimd_ld128()
75 const v128_t vprod0x3 = wasm_i16x8_mul(vxb3, vxa0); in xnn_qs8_gemm_minmax_ukernel_1x4c8__wasmsimd_ld128()
D1x4c8-xw-minmax-wasmsimd.c58 const v128_t vprod0x0 = wasm_i16x8_mul(vxa0, vxb0); in xnn_qs8_gemm_xw_minmax_ukernel_1x4c8__wasmsimd()
63 const v128_t vprod0x1 = wasm_i16x8_mul(vxa0, vxb1); in xnn_qs8_gemm_xw_minmax_ukernel_1x4c8__wasmsimd()
68 const v128_t vprod0x2 = wasm_i16x8_mul(vxa0, vxb2); in xnn_qs8_gemm_xw_minmax_ukernel_1x4c8__wasmsimd()
73 const v128_t vprod0x3 = wasm_i16x8_mul(vxa0, vxb3); in xnn_qs8_gemm_xw_minmax_ukernel_1x4c8__wasmsimd()
/external/XNNPACK/src/qs8-igemm/gen/
D3x4c8-minmax-wasmsimd-ld128.c99 const v128_t vprod0x0 = wasm_i16x8_mul(vxb0, vxa0); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld128()
101 const v128_t vprod1x0 = wasm_i16x8_mul(vxb0, vxa1); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld128()
103 const v128_t vprod2x0 = wasm_i16x8_mul(vxb0, vxa2); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld128()
106 const v128_t vprod0x1 = wasm_i16x8_mul(vxb1, vxa0); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld128()
109 const v128_t vprod1x1 = wasm_i16x8_mul(vxb1, vxa1); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld128()
112 const v128_t vprod2x1 = wasm_i16x8_mul(vxb1, vxa2); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld128()
123 const v128_t vprod0x2 = wasm_i16x8_mul(vxb2, vxa0); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld128()
125 const v128_t vprod1x2 = wasm_i16x8_mul(vxb2, vxa1); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld128()
127 const v128_t vprod2x2 = wasm_i16x8_mul(vxb2, vxa2); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld128()
130 const v128_t vprod0x3 = wasm_i16x8_mul(vxb3, vxa0); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld128()
[all …]
D3x4c8-minmax-wasmsimd-ld64.c97 const v128_t vprod0x0 = wasm_i16x8_mul(vxa0, vxb0); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld64()
100 const v128_t vprod1x0 = wasm_i16x8_mul(vxa1, vxb0); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld64()
103 const v128_t vprod2x0 = wasm_i16x8_mul(vxa2, vxb0); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld64()
108 const v128_t vprod0x1 = wasm_i16x8_mul(vxa0, vxb1); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld64()
111 const v128_t vprod1x1 = wasm_i16x8_mul(vxa1, vxb1); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld64()
114 const v128_t vprod2x1 = wasm_i16x8_mul(vxa2, vxb1); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld64()
119 const v128_t vprod0x2 = wasm_i16x8_mul(vxa0, vxb2); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld64()
122 const v128_t vprod1x2 = wasm_i16x8_mul(vxa1, vxb2); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld64()
125 const v128_t vprod2x2 = wasm_i16x8_mul(vxa2, vxb2); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld64()
130 const v128_t vprod0x3 = wasm_i16x8_mul(vxa0, vxb3); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld64()
[all …]
D2x4c8-minmax-wasmsimd-ld128.c85 const v128_t vprod0x0 = wasm_i16x8_mul(vxb0, vxa0); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld128()
87 const v128_t vprod1x0 = wasm_i16x8_mul(vxb0, vxa1); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld128()
90 const v128_t vprod0x1 = wasm_i16x8_mul(vxb1, vxa0); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld128()
93 const v128_t vprod1x1 = wasm_i16x8_mul(vxb1, vxa1); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld128()
103 const v128_t vprod0x2 = wasm_i16x8_mul(vxb2, vxa0); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld128()
105 const v128_t vprod1x2 = wasm_i16x8_mul(vxb2, vxa1); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld128()
108 const v128_t vprod0x3 = wasm_i16x8_mul(vxb3, vxa0); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld128()
111 const v128_t vprod1x3 = wasm_i16x8_mul(vxb3, vxa1); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld128()
D2x4c8-minmax-wasmsimd-ld64.c83 const v128_t vprod0x0 = wasm_i16x8_mul(vxa0, vxb0); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld64()
86 const v128_t vprod1x0 = wasm_i16x8_mul(vxa1, vxb0); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld64()
91 const v128_t vprod0x1 = wasm_i16x8_mul(vxa0, vxb1); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld64()
94 const v128_t vprod1x1 = wasm_i16x8_mul(vxa1, vxb1); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld64()
99 const v128_t vprod0x2 = wasm_i16x8_mul(vxa0, vxb2); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld64()
102 const v128_t vprod1x2 = wasm_i16x8_mul(vxa1, vxb2); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld64()
107 const v128_t vprod0x3 = wasm_i16x8_mul(vxa0, vxb3); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld64()
110 const v128_t vprod1x3 = wasm_i16x8_mul(vxa1, vxb3); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld64()
D1x4c8-minmax-wasmsimd-ld64.c69 const v128_t vprod0x0 = wasm_i16x8_mul(vxa0, vxb0); in xnn_qs8_igemm_minmax_ukernel_1x4c8__wasmsimd_ld64()
74 const v128_t vprod0x1 = wasm_i16x8_mul(vxa0, vxb1); in xnn_qs8_igemm_minmax_ukernel_1x4c8__wasmsimd_ld64()
79 const v128_t vprod0x2 = wasm_i16x8_mul(vxa0, vxb2); in xnn_qs8_igemm_minmax_ukernel_1x4c8__wasmsimd_ld64()
84 const v128_t vprod0x3 = wasm_i16x8_mul(vxa0, vxb3); in xnn_qs8_igemm_minmax_ukernel_1x4c8__wasmsimd_ld64()
D1x4c8-minmax-wasmsimd-ld128.c71 const v128_t vprod0x0 = wasm_i16x8_mul(vxb0, vxa0); in xnn_qs8_igemm_minmax_ukernel_1x4c8__wasmsimd_ld128()
74 const v128_t vprod0x1 = wasm_i16x8_mul(vxb1, vxa0); in xnn_qs8_igemm_minmax_ukernel_1x4c8__wasmsimd_ld128()
83 const v128_t vprod0x2 = wasm_i16x8_mul(vxb2, vxa0); in xnn_qs8_igemm_minmax_ukernel_1x4c8__wasmsimd_ld128()
86 const v128_t vprod0x3 = wasm_i16x8_mul(vxb3, vxa0); in xnn_qs8_igemm_minmax_ukernel_1x4c8__wasmsimd_ld128()
/external/XNNPACK/src/qs8-gemm/
DMRx4c8-wasmsimd.c.in86 const v128_t vprod${M}x${N} = wasm_i16x8_mul(vxb${N}, vxa${M});
90 const v128_t vprod${M}x${N+1} = wasm_i16x8_mul(vxb${N+1}, vxa${M});
110 const v128_t vprod${M}x${N} = wasm_i16x8_mul(vxa${M}, vxb${N});
/external/XNNPACK/src/qs8-igemm/
DMRx4c8-wasmsimd.c.in94 const v128_t vprod${M}x${N} = wasm_i16x8_mul(vxb${N}, vxa${M});
98 const v128_t vprod${M}x${N+1} = wasm_i16x8_mul(vxb${N+1}, vxa${M});
118 const v128_t vprod${M}x${N} = wasm_i16x8_mul(vxa${M}, vxb${N});
/external/XNNPACK/src/qs8-dwconv/
Dunipass-wasmsimd-mul16.c.in59 … const v128_t vprod${K}x${ABC[C:C+8]} = wasm_i16x8_mul(vi${K}x${ABC[C:C+8]}, vk${K}x${ABC[C:C+8]});
134 … const v128_t vprod${K}x${ABC[0:8]} = wasm_i16x8_mul(vi${K}x${ABC[0:8]}, vk${K}x${ABC[0:8]});
/external/gemmlowp/fixedpoint/
Dfixedpoint_wasmsimd.h114 return to_int16x8_v128_t(wasm_i16x8_mul(a.v, b.v));
/external/llvm-project/clang/lib/Headers/
Dwasm_simd128.h736 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_mul(v128_t __a, in wasm_i16x8_mul() function