/external/XNNPACK/src/f32-gemm/gen/ |
D | 4x4-wasm.c | 87 const float vb3 = w[3]; in xnn_f32_gemm_ukernel_4x4__wasm() local 93 vacc03 += va0 * vb3; in xnn_f32_gemm_ukernel_4x4__wasm() 97 vacc13 += va1 * vb3; in xnn_f32_gemm_ukernel_4x4__wasm() 101 vacc23 += va2 * vb3; in xnn_f32_gemm_ukernel_4x4__wasm() 105 vacc33 += va3 * vb3; in xnn_f32_gemm_ukernel_4x4__wasm()
|
D | 4x4-scalar.c | 87 const float vb3 = w[3]; in xnn_f32_gemm_ukernel_4x4__scalar() local 93 vacc03 += va0 * vb3; in xnn_f32_gemm_ukernel_4x4__scalar() 97 vacc13 += va1 * vb3; in xnn_f32_gemm_ukernel_4x4__scalar() 101 vacc23 += va2 * vb3; in xnn_f32_gemm_ukernel_4x4__scalar() 105 vacc33 += va3 * vb3; in xnn_f32_gemm_ukernel_4x4__scalar()
|
D | 4x4-relu-wasm.c | 87 const float vb3 = w[3]; in xnn_f32_gemm_relu_ukernel_4x4__wasm() local 93 vacc03 += va0 * vb3; in xnn_f32_gemm_relu_ukernel_4x4__wasm() 97 vacc13 += va1 * vb3; in xnn_f32_gemm_relu_ukernel_4x4__wasm() 101 vacc23 += va2 * vb3; in xnn_f32_gemm_relu_ukernel_4x4__wasm() 105 vacc33 += va3 * vb3; in xnn_f32_gemm_relu_ukernel_4x4__wasm()
|
D | 4x4-relu-scalar.c | 87 const float vb3 = w[3]; in xnn_f32_gemm_relu_ukernel_4x4__scalar() local 93 vacc03 += va0 * vb3; in xnn_f32_gemm_relu_ukernel_4x4__scalar() 97 vacc13 += va1 * vb3; in xnn_f32_gemm_relu_ukernel_4x4__scalar() 101 vacc23 += va2 * vb3; in xnn_f32_gemm_relu_ukernel_4x4__scalar() 105 vacc33 += va3 * vb3; in xnn_f32_gemm_relu_ukernel_4x4__scalar()
|
D | 4x4-minmax-scalar.c | 89 const float vb3 = w[3]; in xnn_f32_gemm_minmax_ukernel_4x4__scalar() local 95 vacc03 += va0 * vb3; in xnn_f32_gemm_minmax_ukernel_4x4__scalar() 99 vacc13 += va1 * vb3; in xnn_f32_gemm_minmax_ukernel_4x4__scalar() 103 vacc23 += va2 * vb3; in xnn_f32_gemm_minmax_ukernel_4x4__scalar() 107 vacc33 += va3 * vb3; in xnn_f32_gemm_minmax_ukernel_4x4__scalar()
|
D | 4x4-minmax-wasm.c | 89 const float vb3 = w[3]; in xnn_f32_gemm_minmax_ukernel_4x4__wasm() local 95 vacc03 += va0 * vb3; in xnn_f32_gemm_minmax_ukernel_4x4__wasm() 99 vacc13 += va1 * vb3; in xnn_f32_gemm_minmax_ukernel_4x4__wasm() 103 vacc23 += va2 * vb3; in xnn_f32_gemm_minmax_ukernel_4x4__wasm() 107 vacc33 += va3 * vb3; in xnn_f32_gemm_minmax_ukernel_4x4__wasm()
|
D | 2x4-wasm.c | 65 const float vb3 = w[3]; in xnn_f32_gemm_ukernel_2x4__wasm() local 71 vacc03 += va0 * vb3; in xnn_f32_gemm_ukernel_2x4__wasm() 75 vacc13 += va1 * vb3; in xnn_f32_gemm_ukernel_2x4__wasm()
|
D | 2x4-scalar.c | 65 const float vb3 = w[3]; in xnn_f32_gemm_ukernel_2x4__scalar() local 71 vacc03 += va0 * vb3; in xnn_f32_gemm_ukernel_2x4__scalar() 75 vacc13 += va1 * vb3; in xnn_f32_gemm_ukernel_2x4__scalar()
|
D | 2x4-relu-scalar.c | 65 const float vb3 = w[3]; in xnn_f32_gemm_relu_ukernel_2x4__scalar() local 71 vacc03 += va0 * vb3; in xnn_f32_gemm_relu_ukernel_2x4__scalar() 75 vacc13 += va1 * vb3; in xnn_f32_gemm_relu_ukernel_2x4__scalar()
|
D | 2x4-relu-wasm.c | 65 const float vb3 = w[3]; in xnn_f32_gemm_relu_ukernel_2x4__wasm() local 71 vacc03 += va0 * vb3; in xnn_f32_gemm_relu_ukernel_2x4__wasm() 75 vacc13 += va1 * vb3; in xnn_f32_gemm_relu_ukernel_2x4__wasm()
|
/external/XNNPACK/src/f32-igemm/gen/ |
D | 4x4-scalar.c | 109 const float vb3 = w[3]; in xnn_f32_igemm_ukernel_4x4__scalar() local 115 vacc03 += va0 * vb3; in xnn_f32_igemm_ukernel_4x4__scalar() 119 vacc13 += va1 * vb3; in xnn_f32_igemm_ukernel_4x4__scalar() 123 vacc23 += va2 * vb3; in xnn_f32_igemm_ukernel_4x4__scalar() 127 vacc33 += va3 * vb3; in xnn_f32_igemm_ukernel_4x4__scalar()
|
D | 4x4-wasm.c | 109 const float vb3 = w[3]; in xnn_f32_igemm_ukernel_4x4__wasm() local 115 vacc03 += va0 * vb3; in xnn_f32_igemm_ukernel_4x4__wasm() 119 vacc13 += va1 * vb3; in xnn_f32_igemm_ukernel_4x4__wasm() 123 vacc23 += va2 * vb3; in xnn_f32_igemm_ukernel_4x4__wasm() 127 vacc33 += va3 * vb3; in xnn_f32_igemm_ukernel_4x4__wasm()
|
D | 4x4-relu-scalar.c | 109 const float vb3 = w[3]; in xnn_f32_igemm_relu_ukernel_4x4__scalar() local 115 vacc03 += va0 * vb3; in xnn_f32_igemm_relu_ukernel_4x4__scalar() 119 vacc13 += va1 * vb3; in xnn_f32_igemm_relu_ukernel_4x4__scalar() 123 vacc23 += va2 * vb3; in xnn_f32_igemm_relu_ukernel_4x4__scalar() 127 vacc33 += va3 * vb3; in xnn_f32_igemm_relu_ukernel_4x4__scalar()
|
D | 4x4-relu-wasm.c | 109 const float vb3 = w[3]; in xnn_f32_igemm_relu_ukernel_4x4__wasm() local 115 vacc03 += va0 * vb3; in xnn_f32_igemm_relu_ukernel_4x4__wasm() 119 vacc13 += va1 * vb3; in xnn_f32_igemm_relu_ukernel_4x4__wasm() 123 vacc23 += va2 * vb3; in xnn_f32_igemm_relu_ukernel_4x4__wasm() 127 vacc33 += va3 * vb3; in xnn_f32_igemm_relu_ukernel_4x4__wasm()
|
D | 4x4-minmax-wasm.c | 111 const float vb3 = w[3]; in xnn_f32_igemm_minmax_ukernel_4x4__wasm() local 117 vacc03 += va0 * vb3; in xnn_f32_igemm_minmax_ukernel_4x4__wasm() 121 vacc13 += va1 * vb3; in xnn_f32_igemm_minmax_ukernel_4x4__wasm() 125 vacc23 += va2 * vb3; in xnn_f32_igemm_minmax_ukernel_4x4__wasm() 129 vacc33 += va3 * vb3; in xnn_f32_igemm_minmax_ukernel_4x4__wasm()
|
D | 4x4-minmax-scalar.c | 111 const float vb3 = w[3]; in xnn_f32_igemm_minmax_ukernel_4x4__scalar() local 117 vacc03 += va0 * vb3; in xnn_f32_igemm_minmax_ukernel_4x4__scalar() 121 vacc13 += va1 * vb3; in xnn_f32_igemm_minmax_ukernel_4x4__scalar() 125 vacc23 += va2 * vb3; in xnn_f32_igemm_minmax_ukernel_4x4__scalar() 129 vacc33 += va3 * vb3; in xnn_f32_igemm_minmax_ukernel_4x4__scalar()
|
D | 2x4-scalar.c | 81 const float vb3 = w[3]; in xnn_f32_igemm_ukernel_2x4__scalar() local 87 vacc03 += va0 * vb3; in xnn_f32_igemm_ukernel_2x4__scalar() 91 vacc13 += va1 * vb3; in xnn_f32_igemm_ukernel_2x4__scalar()
|
D | 2x4-wasm.c | 81 const float vb3 = w[3]; in xnn_f32_igemm_ukernel_2x4__wasm() local 87 vacc03 += va0 * vb3; in xnn_f32_igemm_ukernel_2x4__wasm() 91 vacc13 += va1 * vb3; in xnn_f32_igemm_ukernel_2x4__wasm()
|
D | 2x4-relu-scalar.c | 81 const float vb3 = w[3]; in xnn_f32_igemm_relu_ukernel_2x4__scalar() local 87 vacc03 += va0 * vb3; in xnn_f32_igemm_relu_ukernel_2x4__scalar() 91 vacc13 += va1 * vb3; in xnn_f32_igemm_relu_ukernel_2x4__scalar()
|
D | 2x4-relu-wasm.c | 81 const float vb3 = w[3]; in xnn_f32_igemm_relu_ukernel_2x4__wasm() local 87 vacc03 += va0 * vb3; in xnn_f32_igemm_relu_ukernel_2x4__wasm() 91 vacc13 += va1 * vb3; in xnn_f32_igemm_relu_ukernel_2x4__wasm()
|
/external/XNNPACK/src/f32-ppmm/gen/ |
D | 4x4-minmax-scalar.c | 77 const float vb3 = w[3]; in xnn_f32_ppmm_minmax_ukernel_4x4__scalar() local 92 vacc0x3 += va0 * vb3; in xnn_f32_ppmm_minmax_ukernel_4x4__scalar() 93 vacc1x3 += va1 * vb3; in xnn_f32_ppmm_minmax_ukernel_4x4__scalar() 94 vacc2x3 += va2 * vb3; in xnn_f32_ppmm_minmax_ukernel_4x4__scalar() 95 vacc3x3 += va3 * vb3; in xnn_f32_ppmm_minmax_ukernel_4x4__scalar()
|
D | 2x4-minmax-scalar.c | 59 const float vb3 = w[3]; in xnn_f32_ppmm_minmax_ukernel_2x4__scalar() local 68 vacc0x3 += va0 * vb3; in xnn_f32_ppmm_minmax_ukernel_2x4__scalar() 69 vacc1x3 += va1 * vb3; in xnn_f32_ppmm_minmax_ukernel_2x4__scalar()
|
/external/XNNPACK/src/f32-gemm/gen-inc/ |
D | 4x4inc-minmax-wasm.c | 91 const float vb3 = w[3]; in xnn_f32_gemminc_minmax_ukernel_4x4__wasm() local 97 vacc03 += va0 * vb3; in xnn_f32_gemminc_minmax_ukernel_4x4__wasm() 101 vacc13 += va1 * vb3; in xnn_f32_gemminc_minmax_ukernel_4x4__wasm() 105 vacc23 += va2 * vb3; in xnn_f32_gemminc_minmax_ukernel_4x4__wasm() 109 vacc33 += va3 * vb3; in xnn_f32_gemminc_minmax_ukernel_4x4__wasm()
|
D | 4x4inc-minmax-scalar.c | 91 const float vb3 = w[3]; in xnn_f32_gemminc_minmax_ukernel_4x4__scalar() local 97 vacc03 += va0 * vb3; in xnn_f32_gemminc_minmax_ukernel_4x4__scalar() 101 vacc13 += va1 * vb3; in xnn_f32_gemminc_minmax_ukernel_4x4__scalar() 105 vacc23 += va2 * vb3; in xnn_f32_gemminc_minmax_ukernel_4x4__scalar() 109 vacc33 += va3 * vb3; in xnn_f32_gemminc_minmax_ukernel_4x4__scalar()
|
/external/XNNPACK/src/f32-vbinary/gen/ |
D | vsub-scalar-x4.c | 41 const float vb3 = b[3]; in xnn_f32_vsub_ukernel__scalar_x4() local 47 float vy3 = va3 - vb3; in xnn_f32_vsub_ukernel__scalar_x4()
|