/external/XNNPACK/src/f32-spmm/gen/ |
D | 32x4-minmax-neonfma.c | 124 float32x4_t vout0123n0 = vminq_f32(vacc0123n0, vmax); in xnn_f32_spmm_minmax_ukernel_32x4__neonfma() 125 float32x4_t vout4567n0 = vminq_f32(vacc4567n0, vmax); in xnn_f32_spmm_minmax_ukernel_32x4__neonfma() 126 float32x4_t vout89ABn0 = vminq_f32(vacc89ABn0, vmax); in xnn_f32_spmm_minmax_ukernel_32x4__neonfma() 127 float32x4_t voutCDEFn0 = vminq_f32(vaccCDEFn0, vmax); in xnn_f32_spmm_minmax_ukernel_32x4__neonfma() 128 float32x4_t voutGHIJn0 = vminq_f32(vaccGHIJn0, vmax); in xnn_f32_spmm_minmax_ukernel_32x4__neonfma() 129 float32x4_t voutKLMNn0 = vminq_f32(vaccKLMNn0, vmax); in xnn_f32_spmm_minmax_ukernel_32x4__neonfma() 130 float32x4_t voutOPQRn0 = vminq_f32(vaccOPQRn0, vmax); in xnn_f32_spmm_minmax_ukernel_32x4__neonfma() 131 float32x4_t voutSTUVn0 = vminq_f32(vaccSTUVn0, vmax); in xnn_f32_spmm_minmax_ukernel_32x4__neonfma() 132 float32x4_t vout0123n1 = vminq_f32(vacc0123n1, vmax); in xnn_f32_spmm_minmax_ukernel_32x4__neonfma() 133 float32x4_t vout4567n1 = vminq_f32(vacc4567n1, vmax); in xnn_f32_spmm_minmax_ukernel_32x4__neonfma() [all …]
|
D | 32x2-minmax-neonfma.c | 92 float32x4_t vout0123n0 = vminq_f32(vacc0123n0, vmax); in xnn_f32_spmm_minmax_ukernel_32x2__neonfma() 93 float32x4_t vout4567n0 = vminq_f32(vacc4567n0, vmax); in xnn_f32_spmm_minmax_ukernel_32x2__neonfma() 94 float32x4_t vout89ABn0 = vminq_f32(vacc89ABn0, vmax); in xnn_f32_spmm_minmax_ukernel_32x2__neonfma() 95 float32x4_t voutCDEFn0 = vminq_f32(vaccCDEFn0, vmax); in xnn_f32_spmm_minmax_ukernel_32x2__neonfma() 96 float32x4_t voutGHIJn0 = vminq_f32(vaccGHIJn0, vmax); in xnn_f32_spmm_minmax_ukernel_32x2__neonfma() 97 float32x4_t voutKLMNn0 = vminq_f32(vaccKLMNn0, vmax); in xnn_f32_spmm_minmax_ukernel_32x2__neonfma() 98 float32x4_t voutOPQRn0 = vminq_f32(vaccOPQRn0, vmax); in xnn_f32_spmm_minmax_ukernel_32x2__neonfma() 99 float32x4_t voutSTUVn0 = vminq_f32(vaccSTUVn0, vmax); in xnn_f32_spmm_minmax_ukernel_32x2__neonfma() 100 float32x4_t vout0123n1 = vminq_f32(vacc0123n1, vmax); in xnn_f32_spmm_minmax_ukernel_32x2__neonfma() 101 float32x4_t vout4567n1 = vminq_f32(vacc4567n1, vmax); in xnn_f32_spmm_minmax_ukernel_32x2__neonfma() [all …]
|
D | 16x4-minmax-neonfma.c | 87 float32x4_t vout0123n0 = vminq_f32(vacc0123n0, vmax); in xnn_f32_spmm_minmax_ukernel_16x4__neonfma() 88 float32x4_t vout4567n0 = vminq_f32(vacc4567n0, vmax); in xnn_f32_spmm_minmax_ukernel_16x4__neonfma() 89 float32x4_t vout89ABn0 = vminq_f32(vacc89ABn0, vmax); in xnn_f32_spmm_minmax_ukernel_16x4__neonfma() 90 float32x4_t voutCDEFn0 = vminq_f32(vaccCDEFn0, vmax); in xnn_f32_spmm_minmax_ukernel_16x4__neonfma() 91 float32x4_t vout0123n1 = vminq_f32(vacc0123n1, vmax); in xnn_f32_spmm_minmax_ukernel_16x4__neonfma() 92 float32x4_t vout4567n1 = vminq_f32(vacc4567n1, vmax); in xnn_f32_spmm_minmax_ukernel_16x4__neonfma() 93 float32x4_t vout89ABn1 = vminq_f32(vacc89ABn1, vmax); in xnn_f32_spmm_minmax_ukernel_16x4__neonfma() 94 float32x4_t voutCDEFn1 = vminq_f32(vaccCDEFn1, vmax); in xnn_f32_spmm_minmax_ukernel_16x4__neonfma() 95 float32x4_t vout0123n2 = vminq_f32(vacc0123n2, vmax); in xnn_f32_spmm_minmax_ukernel_16x4__neonfma() 96 float32x4_t vout4567n2 = vminq_f32(vacc4567n2, vmax); in xnn_f32_spmm_minmax_ukernel_16x4__neonfma() [all …]
|
D | 12x4-minmax-neonfma.c | 78 float32x4_t vout0123n0 = vminq_f32(vacc0123n0, vmax); in xnn_f32_spmm_minmax_ukernel_12x4__neonfma() 79 float32x4_t vout4567n0 = vminq_f32(vacc4567n0, vmax); in xnn_f32_spmm_minmax_ukernel_12x4__neonfma() 80 float32x4_t vout89ABn0 = vminq_f32(vacc89ABn0, vmax); in xnn_f32_spmm_minmax_ukernel_12x4__neonfma() 81 float32x4_t vout0123n1 = vminq_f32(vacc0123n1, vmax); in xnn_f32_spmm_minmax_ukernel_12x4__neonfma() 82 float32x4_t vout4567n1 = vminq_f32(vacc4567n1, vmax); in xnn_f32_spmm_minmax_ukernel_12x4__neonfma() 83 float32x4_t vout89ABn1 = vminq_f32(vacc89ABn1, vmax); in xnn_f32_spmm_minmax_ukernel_12x4__neonfma() 84 float32x4_t vout0123n2 = vminq_f32(vacc0123n2, vmax); in xnn_f32_spmm_minmax_ukernel_12x4__neonfma() 85 float32x4_t vout4567n2 = vminq_f32(vacc4567n2, vmax); in xnn_f32_spmm_minmax_ukernel_12x4__neonfma() 86 float32x4_t vout89ABn2 = vminq_f32(vacc89ABn2, vmax); in xnn_f32_spmm_minmax_ukernel_12x4__neonfma() 87 float32x4_t vout0123n3 = vminq_f32(vacc0123n3, vmax); in xnn_f32_spmm_minmax_ukernel_12x4__neonfma() [all …]
|
D | 32x1-minmax-neon.c | 76 float32x4_t vout0123 = vminq_f32(vacc0123, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon() 77 float32x4_t vout4567 = vminq_f32(vacc4567, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon() 78 float32x4_t vout89AB = vminq_f32(vacc89AB, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon() 79 float32x4_t voutCDEF = vminq_f32(vaccCDEF, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon() 80 float32x4_t voutGHIJ = vminq_f32(vaccGHIJ, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon() 81 float32x4_t voutKLMN = vminq_f32(vaccKLMN, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon() 82 float32x4_t voutOPQR = vminq_f32(vaccOPQR, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon() 83 float32x4_t voutSTUV = vminq_f32(vaccSTUV, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon() 134 float32x4_t vout0123 = vminq_f32(vacc0123, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon() 135 float32x4_t vout4567 = vminq_f32(vacc4567, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon() [all …]
|
D | 32x1-minmax-neonfma.c | 76 float32x4_t vout0123 = vminq_f32(vacc0123, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma() 77 float32x4_t vout4567 = vminq_f32(vacc4567, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma() 78 float32x4_t vout89AB = vminq_f32(vacc89AB, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma() 79 float32x4_t voutCDEF = vminq_f32(vaccCDEF, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma() 80 float32x4_t voutGHIJ = vminq_f32(vaccGHIJ, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma() 81 float32x4_t voutKLMN = vminq_f32(vaccKLMN, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma() 82 float32x4_t voutOPQR = vminq_f32(vaccOPQR, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma() 83 float32x4_t voutSTUV = vminq_f32(vaccSTUV, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma() 134 float32x4_t vout0123 = vminq_f32(vacc0123, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma() 135 float32x4_t vout4567 = vminq_f32(vacc4567, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma() [all …]
|
D | 16x2-minmax-neonfma.c | 71 float32x4_t vout0123n0 = vminq_f32(vacc0123n0, vmax); in xnn_f32_spmm_minmax_ukernel_16x2__neonfma() 72 float32x4_t vout4567n0 = vminq_f32(vacc4567n0, vmax); in xnn_f32_spmm_minmax_ukernel_16x2__neonfma() 73 float32x4_t vout89ABn0 = vminq_f32(vacc89ABn0, vmax); in xnn_f32_spmm_minmax_ukernel_16x2__neonfma() 74 float32x4_t voutCDEFn0 = vminq_f32(vaccCDEFn0, vmax); in xnn_f32_spmm_minmax_ukernel_16x2__neonfma() 75 float32x4_t vout0123n1 = vminq_f32(vacc0123n1, vmax); in xnn_f32_spmm_minmax_ukernel_16x2__neonfma() 76 float32x4_t vout4567n1 = vminq_f32(vacc4567n1, vmax); in xnn_f32_spmm_minmax_ukernel_16x2__neonfma() 77 float32x4_t vout89ABn1 = vminq_f32(vacc89ABn1, vmax); in xnn_f32_spmm_minmax_ukernel_16x2__neonfma() 78 float32x4_t voutCDEFn1 = vminq_f32(vaccCDEFn1, vmax); in xnn_f32_spmm_minmax_ukernel_16x2__neonfma() 127 float32x4_t vout0123 = vminq_f32(vacc0123, vmax); in xnn_f32_spmm_minmax_ukernel_16x2__neonfma() 128 float32x4_t vout4567 = vminq_f32(vacc4567, vmax); in xnn_f32_spmm_minmax_ukernel_16x2__neonfma() [all …]
|
D | 32x1-minmax-neon-pipelined.c | 87 float32x4_t vout0123 = vminq_f32(vacc0123, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon_pipelined() 88 float32x4_t vout4567 = vminq_f32(vacc4567, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon_pipelined() 89 float32x4_t vout89AB = vminq_f32(vacc89AB, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon_pipelined() 90 float32x4_t voutCDEF = vminq_f32(vaccCDEF, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon_pipelined() 91 float32x4_t voutGHIJ = vminq_f32(vaccGHIJ, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon_pipelined() 92 float32x4_t voutKLMN = vminq_f32(vaccKLMN, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon_pipelined() 93 float32x4_t voutOPQR = vminq_f32(vaccOPQR, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon_pipelined() 94 float32x4_t voutSTUV = vminq_f32(vaccSTUV, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon_pipelined() 148 float32x4_t vout0123 = vminq_f32(vacc0123, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon_pipelined() 149 float32x4_t vout4567 = vminq_f32(vacc4567, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon_pipelined() [all …]
|
D | 32x1-minmax-neonfma-pipelined.c | 87 float32x4_t vout0123 = vminq_f32(vacc0123, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma_pipelined() 88 float32x4_t vout4567 = vminq_f32(vacc4567, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma_pipelined() 89 float32x4_t vout89AB = vminq_f32(vacc89AB, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma_pipelined() 90 float32x4_t voutCDEF = vminq_f32(vaccCDEF, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma_pipelined() 91 float32x4_t voutGHIJ = vminq_f32(vaccGHIJ, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma_pipelined() 92 float32x4_t voutKLMN = vminq_f32(vaccKLMN, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma_pipelined() 93 float32x4_t voutOPQR = vminq_f32(vaccOPQR, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma_pipelined() 94 float32x4_t voutSTUV = vminq_f32(vaccSTUV, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma_pipelined() 148 float32x4_t vout0123 = vminq_f32(vacc0123, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma_pipelined() 149 float32x4_t vout4567 = vminq_f32(vacc4567, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma_pipelined() [all …]
|
D | 12x2-minmax-neonfma.c | 66 float32x4_t vout0123n0 = vminq_f32(vacc0123n0, vmax); in xnn_f32_spmm_minmax_ukernel_12x2__neonfma() 67 float32x4_t vout4567n0 = vminq_f32(vacc4567n0, vmax); in xnn_f32_spmm_minmax_ukernel_12x2__neonfma() 68 float32x4_t vout89ABn0 = vminq_f32(vacc89ABn0, vmax); in xnn_f32_spmm_minmax_ukernel_12x2__neonfma() 69 float32x4_t vout0123n1 = vminq_f32(vacc0123n1, vmax); in xnn_f32_spmm_minmax_ukernel_12x2__neonfma() 70 float32x4_t vout4567n1 = vminq_f32(vacc4567n1, vmax); in xnn_f32_spmm_minmax_ukernel_12x2__neonfma() 71 float32x4_t vout89ABn1 = vminq_f32(vacc89ABn1, vmax); in xnn_f32_spmm_minmax_ukernel_12x2__neonfma() 113 float32x4_t vout0123 = vminq_f32(vacc0123, vmax); in xnn_f32_spmm_minmax_ukernel_12x2__neonfma() 114 float32x4_t vout4567 = vminq_f32(vacc4567, vmax); in xnn_f32_spmm_minmax_ukernel_12x2__neonfma() 115 float32x4_t vout89AB = vminq_f32(vacc89AB, vmax); in xnn_f32_spmm_minmax_ukernel_12x2__neonfma() 159 float32x4_t vout0123n0 = vminq_f32(vacc0123n0, vmax); in xnn_f32_spmm_minmax_ukernel_12x2__neonfma() [all …]
|
D | 8x4-minmax-neonfma.c | 69 float32x4_t vout0123n0 = vminq_f32(vacc0123n0, vmax); in xnn_f32_spmm_minmax_ukernel_8x4__neonfma() 70 float32x4_t vout4567n0 = vminq_f32(vacc4567n0, vmax); in xnn_f32_spmm_minmax_ukernel_8x4__neonfma() 71 float32x4_t vout0123n1 = vminq_f32(vacc0123n1, vmax); in xnn_f32_spmm_minmax_ukernel_8x4__neonfma() 72 float32x4_t vout4567n1 = vminq_f32(vacc4567n1, vmax); in xnn_f32_spmm_minmax_ukernel_8x4__neonfma() 73 float32x4_t vout0123n2 = vminq_f32(vacc0123n2, vmax); in xnn_f32_spmm_minmax_ukernel_8x4__neonfma() 74 float32x4_t vout4567n2 = vminq_f32(vacc4567n2, vmax); in xnn_f32_spmm_minmax_ukernel_8x4__neonfma() 75 float32x4_t vout0123n3 = vminq_f32(vacc0123n3, vmax); in xnn_f32_spmm_minmax_ukernel_8x4__neonfma() 76 float32x4_t vout4567n3 = vminq_f32(vacc4567n3, vmax); in xnn_f32_spmm_minmax_ukernel_8x4__neonfma() 121 float32x4_t vout0123 = vminq_f32(vacc0123, vmax); in xnn_f32_spmm_minmax_ukernel_8x4__neonfma() 122 float32x4_t vout4567 = vminq_f32(vacc4567, vmax); in xnn_f32_spmm_minmax_ukernel_8x4__neonfma() [all …]
|
D | 32x1-minmax-neon-x2.c | 147 float32x4_t vout0123 = vminq_f32(vacc0123, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon_x2() 148 float32x4_t vout4567 = vminq_f32(vacc4567, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon_x2() 149 float32x4_t vout89AB = vminq_f32(vacc89AB, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon_x2() 150 float32x4_t voutCDEF = vminq_f32(vaccCDEF, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon_x2() 151 float32x4_t voutGHIJ = vminq_f32(vaccGHIJ, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon_x2() 152 float32x4_t voutKLMN = vminq_f32(vaccKLMN, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon_x2() 153 float32x4_t voutOPQR = vminq_f32(vaccOPQR, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon_x2() 154 float32x4_t voutSTUV = vminq_f32(vaccSTUV, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon_x2() 205 float32x4_t vout0123 = vminq_f32(vacc0123, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon_x2() 206 float32x4_t vout4567 = vminq_f32(vacc4567, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neon_x2() [all …]
|
D | 32x1-minmax-neonfma-x2.c | 147 float32x4_t vout0123 = vminq_f32(vacc0123, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma_x2() 148 float32x4_t vout4567 = vminq_f32(vacc4567, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma_x2() 149 float32x4_t vout89AB = vminq_f32(vacc89AB, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma_x2() 150 float32x4_t voutCDEF = vminq_f32(vaccCDEF, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma_x2() 151 float32x4_t voutGHIJ = vminq_f32(vaccGHIJ, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma_x2() 152 float32x4_t voutKLMN = vminq_f32(vaccKLMN, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma_x2() 153 float32x4_t voutOPQR = vminq_f32(vaccOPQR, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma_x2() 154 float32x4_t voutSTUV = vminq_f32(vaccSTUV, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma_x2() 205 float32x4_t vout0123 = vminq_f32(vacc0123, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma_x2() 206 float32x4_t vout4567 = vminq_f32(vacc4567, vmax); in xnn_f32_spmm_minmax_ukernel_32x1__neonfma_x2() [all …]
|
/external/XNNPACK/src/f32-ppmm/gen/ |
D | 8x8-minmax-neon.c | 112 vacc0x0123 = vminq_f32(vacc0x0123, vmax); in xnn_f32_ppmm_minmax_ukernel_8x8__neon() 113 vacc1x0123 = vminq_f32(vacc1x0123, vmax); in xnn_f32_ppmm_minmax_ukernel_8x8__neon() 114 vacc2x0123 = vminq_f32(vacc2x0123, vmax); in xnn_f32_ppmm_minmax_ukernel_8x8__neon() 115 vacc3x0123 = vminq_f32(vacc3x0123, vmax); in xnn_f32_ppmm_minmax_ukernel_8x8__neon() 116 vacc4x0123 = vminq_f32(vacc4x0123, vmax); in xnn_f32_ppmm_minmax_ukernel_8x8__neon() 117 vacc5x0123 = vminq_f32(vacc5x0123, vmax); in xnn_f32_ppmm_minmax_ukernel_8x8__neon() 118 vacc6x0123 = vminq_f32(vacc6x0123, vmax); in xnn_f32_ppmm_minmax_ukernel_8x8__neon() 119 vacc7x0123 = vminq_f32(vacc7x0123, vmax); in xnn_f32_ppmm_minmax_ukernel_8x8__neon() 120 vacc0x4567 = vminq_f32(vacc0x4567, vmax); in xnn_f32_ppmm_minmax_ukernel_8x8__neon() 121 vacc1x4567 = vminq_f32(vacc1x4567, vmax); in xnn_f32_ppmm_minmax_ukernel_8x8__neon() [all …]
|
D | 8x8-minmax-neonfma.c | 140 vacc0x0123 = vminq_f32(vacc0x0123, vmax); in xnn_f32_ppmm_minmax_ukernel_8x8__neonfma() 141 vacc1x0123 = vminq_f32(vacc1x0123, vmax); in xnn_f32_ppmm_minmax_ukernel_8x8__neonfma() 142 vacc2x0123 = vminq_f32(vacc2x0123, vmax); in xnn_f32_ppmm_minmax_ukernel_8x8__neonfma() 143 vacc3x0123 = vminq_f32(vacc3x0123, vmax); in xnn_f32_ppmm_minmax_ukernel_8x8__neonfma() 144 vacc4x0123 = vminq_f32(vacc4x0123, vmax); in xnn_f32_ppmm_minmax_ukernel_8x8__neonfma() 145 vacc5x0123 = vminq_f32(vacc5x0123, vmax); in xnn_f32_ppmm_minmax_ukernel_8x8__neonfma() 146 vacc6x0123 = vminq_f32(vacc6x0123, vmax); in xnn_f32_ppmm_minmax_ukernel_8x8__neonfma() 147 vacc7x0123 = vminq_f32(vacc7x0123, vmax); in xnn_f32_ppmm_minmax_ukernel_8x8__neonfma() 148 vacc0x4567 = vminq_f32(vacc0x4567, vmax); in xnn_f32_ppmm_minmax_ukernel_8x8__neonfma() 149 vacc1x4567 = vminq_f32(vacc1x4567, vmax); in xnn_f32_ppmm_minmax_ukernel_8x8__neonfma() [all …]
|
D | 4x8-minmax-neon.c | 79 vacc0x0123 = vminq_f32(vacc0x0123, vmax); in xnn_f32_ppmm_minmax_ukernel_4x8__neon() 80 vacc1x0123 = vminq_f32(vacc1x0123, vmax); in xnn_f32_ppmm_minmax_ukernel_4x8__neon() 81 vacc2x0123 = vminq_f32(vacc2x0123, vmax); in xnn_f32_ppmm_minmax_ukernel_4x8__neon() 82 vacc3x0123 = vminq_f32(vacc3x0123, vmax); in xnn_f32_ppmm_minmax_ukernel_4x8__neon() 83 vacc0x4567 = vminq_f32(vacc0x4567, vmax); in xnn_f32_ppmm_minmax_ukernel_4x8__neon() 84 vacc1x4567 = vminq_f32(vacc1x4567, vmax); in xnn_f32_ppmm_minmax_ukernel_4x8__neon() 85 vacc2x4567 = vminq_f32(vacc2x4567, vmax); in xnn_f32_ppmm_minmax_ukernel_4x8__neon() 86 vacc3x4567 = vminq_f32(vacc3x4567, vmax); in xnn_f32_ppmm_minmax_ukernel_4x8__neon()
|
D | 4x8-minmax-neonfma.c | 95 vacc0x0123 = vminq_f32(vacc0x0123, vmax); in xnn_f32_ppmm_minmax_ukernel_4x8__neonfma() 96 vacc1x0123 = vminq_f32(vacc1x0123, vmax); in xnn_f32_ppmm_minmax_ukernel_4x8__neonfma() 97 vacc2x0123 = vminq_f32(vacc2x0123, vmax); in xnn_f32_ppmm_minmax_ukernel_4x8__neonfma() 98 vacc3x0123 = vminq_f32(vacc3x0123, vmax); in xnn_f32_ppmm_minmax_ukernel_4x8__neonfma() 99 vacc0x4567 = vminq_f32(vacc0x4567, vmax); in xnn_f32_ppmm_minmax_ukernel_4x8__neonfma() 100 vacc1x4567 = vminq_f32(vacc1x4567, vmax); in xnn_f32_ppmm_minmax_ukernel_4x8__neonfma() 101 vacc2x4567 = vminq_f32(vacc2x4567, vmax); in xnn_f32_ppmm_minmax_ukernel_4x8__neonfma() 102 vacc3x4567 = vminq_f32(vacc3x4567, vmax); in xnn_f32_ppmm_minmax_ukernel_4x8__neonfma()
|
/external/XNNPACK/src/f32-vmulcaddc/gen/ |
D | c8-minmax-neonfma-2x.c | 72 vacc0x0123 = vminq_f32(vacc0x0123, vmax); in xnn_f32_vmulcaddc_minmax_ukernel_c8__neonfma_2x() 73 vacc0x4567 = vminq_f32(vacc0x4567, vmax); in xnn_f32_vmulcaddc_minmax_ukernel_c8__neonfma_2x() 74 vacc1x0123 = vminq_f32(vacc1x0123, vmax); in xnn_f32_vmulcaddc_minmax_ukernel_c8__neonfma_2x() 75 vacc1x4567 = vminq_f32(vacc1x4567, vmax); in xnn_f32_vmulcaddc_minmax_ukernel_c8__neonfma_2x() 97 vacc0x0123 = vminq_f32(vacc0x0123, vmax); in xnn_f32_vmulcaddc_minmax_ukernel_c8__neonfma_2x() 98 vacc1x0123 = vminq_f32(vacc1x0123, vmax); in xnn_f32_vmulcaddc_minmax_ukernel_c8__neonfma_2x() 118 vacc0x0123 = vminq_f32(vacc0x0123, vmax); in xnn_f32_vmulcaddc_minmax_ukernel_c8__neonfma_2x() 119 vacc1x0123 = vminq_f32(vacc1x0123, vmax); in xnn_f32_vmulcaddc_minmax_ukernel_c8__neonfma_2x()
|
D | c8-minmax-neon-2x.c | 76 vacc0x0123 = vminq_f32(vacc0x0123, vmax); in xnn_f32_vmulcaddc_minmax_ukernel_c8__neon_2x() 77 vacc0x4567 = vminq_f32(vacc0x4567, vmax); in xnn_f32_vmulcaddc_minmax_ukernel_c8__neon_2x() 78 vacc1x0123 = vminq_f32(vacc1x0123, vmax); in xnn_f32_vmulcaddc_minmax_ukernel_c8__neon_2x() 79 vacc1x4567 = vminq_f32(vacc1x4567, vmax); in xnn_f32_vmulcaddc_minmax_ukernel_c8__neon_2x() 103 vacc0x0123 = vminq_f32(vacc0x0123, vmax); in xnn_f32_vmulcaddc_minmax_ukernel_c8__neon_2x() 104 vacc1x0123 = vminq_f32(vacc1x0123, vmax); in xnn_f32_vmulcaddc_minmax_ukernel_c8__neon_2x() 126 vacc0x0123 = vminq_f32(vacc0x0123, vmax); in xnn_f32_vmulcaddc_minmax_ukernel_c8__neon_2x() 127 vacc1x0123 = vminq_f32(vacc1x0123, vmax); in xnn_f32_vmulcaddc_minmax_ukernel_c8__neon_2x()
|
/external/XNNPACK/src/f32-igemm/gen/ |
D | 6x8-minmax-neonfma-lane-ld64.c | 183 vacc0x0123 = vminq_f32(vacc0x0123, vmax); in xnn_f32_igemm_minmax_ukernel_6x8__neonfma_lane_ld64() 184 vacc1x0123 = vminq_f32(vacc1x0123, vmax); in xnn_f32_igemm_minmax_ukernel_6x8__neonfma_lane_ld64() 185 vacc2x0123 = vminq_f32(vacc2x0123, vmax); in xnn_f32_igemm_minmax_ukernel_6x8__neonfma_lane_ld64() 186 vacc3x0123 = vminq_f32(vacc3x0123, vmax); in xnn_f32_igemm_minmax_ukernel_6x8__neonfma_lane_ld64() 187 vacc4x0123 = vminq_f32(vacc4x0123, vmax); in xnn_f32_igemm_minmax_ukernel_6x8__neonfma_lane_ld64() 188 vacc5x0123 = vminq_f32(vacc5x0123, vmax); in xnn_f32_igemm_minmax_ukernel_6x8__neonfma_lane_ld64() 189 vacc0x4567 = vminq_f32(vacc0x4567, vmax); in xnn_f32_igemm_minmax_ukernel_6x8__neonfma_lane_ld64() 190 vacc1x4567 = vminq_f32(vacc1x4567, vmax); in xnn_f32_igemm_minmax_ukernel_6x8__neonfma_lane_ld64() 191 vacc2x4567 = vminq_f32(vacc2x4567, vmax); in xnn_f32_igemm_minmax_ukernel_6x8__neonfma_lane_ld64() 192 vacc3x4567 = vminq_f32(vacc3x4567, vmax); in xnn_f32_igemm_minmax_ukernel_6x8__neonfma_lane_ld64() [all …]
|
D | 6x8-minmax-neon-lane-ld64.c | 183 vacc0x0123 = vminq_f32(vacc0x0123, vmax); in xnn_f32_igemm_minmax_ukernel_6x8__neon_lane_ld64() 184 vacc1x0123 = vminq_f32(vacc1x0123, vmax); in xnn_f32_igemm_minmax_ukernel_6x8__neon_lane_ld64() 185 vacc2x0123 = vminq_f32(vacc2x0123, vmax); in xnn_f32_igemm_minmax_ukernel_6x8__neon_lane_ld64() 186 vacc3x0123 = vminq_f32(vacc3x0123, vmax); in xnn_f32_igemm_minmax_ukernel_6x8__neon_lane_ld64() 187 vacc4x0123 = vminq_f32(vacc4x0123, vmax); in xnn_f32_igemm_minmax_ukernel_6x8__neon_lane_ld64() 188 vacc5x0123 = vminq_f32(vacc5x0123, vmax); in xnn_f32_igemm_minmax_ukernel_6x8__neon_lane_ld64() 189 vacc0x4567 = vminq_f32(vacc0x4567, vmax); in xnn_f32_igemm_minmax_ukernel_6x8__neon_lane_ld64() 190 vacc1x4567 = vminq_f32(vacc1x4567, vmax); in xnn_f32_igemm_minmax_ukernel_6x8__neon_lane_ld64() 191 vacc2x4567 = vminq_f32(vacc2x4567, vmax); in xnn_f32_igemm_minmax_ukernel_6x8__neon_lane_ld64() 192 vacc3x4567 = vminq_f32(vacc3x4567, vmax); in xnn_f32_igemm_minmax_ukernel_6x8__neon_lane_ld64() [all …]
|
/external/XNNPACK/src/f32-gemm/gen/ |
D | 6x8-minmax-neonfma-lane-ld64.c | 151 vacc0x0123 = vminq_f32(vacc0x0123, vmax); in xnn_f32_gemm_minmax_ukernel_6x8__neonfma_lane_ld64() 152 vacc1x0123 = vminq_f32(vacc1x0123, vmax); in xnn_f32_gemm_minmax_ukernel_6x8__neonfma_lane_ld64() 153 vacc2x0123 = vminq_f32(vacc2x0123, vmax); in xnn_f32_gemm_minmax_ukernel_6x8__neonfma_lane_ld64() 154 vacc3x0123 = vminq_f32(vacc3x0123, vmax); in xnn_f32_gemm_minmax_ukernel_6x8__neonfma_lane_ld64() 155 vacc4x0123 = vminq_f32(vacc4x0123, vmax); in xnn_f32_gemm_minmax_ukernel_6x8__neonfma_lane_ld64() 156 vacc5x0123 = vminq_f32(vacc5x0123, vmax); in xnn_f32_gemm_minmax_ukernel_6x8__neonfma_lane_ld64() 157 vacc0x4567 = vminq_f32(vacc0x4567, vmax); in xnn_f32_gemm_minmax_ukernel_6x8__neonfma_lane_ld64() 158 vacc1x4567 = vminq_f32(vacc1x4567, vmax); in xnn_f32_gemm_minmax_ukernel_6x8__neonfma_lane_ld64() 159 vacc2x4567 = vminq_f32(vacc2x4567, vmax); in xnn_f32_gemm_minmax_ukernel_6x8__neonfma_lane_ld64() 160 vacc3x4567 = vminq_f32(vacc3x4567, vmax); in xnn_f32_gemm_minmax_ukernel_6x8__neonfma_lane_ld64() [all …]
|
D | 6x8-minmax-neon-lane-ld64.c | 151 vacc0x0123 = vminq_f32(vacc0x0123, vmax); in xnn_f32_gemm_minmax_ukernel_6x8__neon_lane_ld64() 152 vacc1x0123 = vminq_f32(vacc1x0123, vmax); in xnn_f32_gemm_minmax_ukernel_6x8__neon_lane_ld64() 153 vacc2x0123 = vminq_f32(vacc2x0123, vmax); in xnn_f32_gemm_minmax_ukernel_6x8__neon_lane_ld64() 154 vacc3x0123 = vminq_f32(vacc3x0123, vmax); in xnn_f32_gemm_minmax_ukernel_6x8__neon_lane_ld64() 155 vacc4x0123 = vminq_f32(vacc4x0123, vmax); in xnn_f32_gemm_minmax_ukernel_6x8__neon_lane_ld64() 156 vacc5x0123 = vminq_f32(vacc5x0123, vmax); in xnn_f32_gemm_minmax_ukernel_6x8__neon_lane_ld64() 157 vacc0x4567 = vminq_f32(vacc0x4567, vmax); in xnn_f32_gemm_minmax_ukernel_6x8__neon_lane_ld64() 158 vacc1x4567 = vminq_f32(vacc1x4567, vmax); in xnn_f32_gemm_minmax_ukernel_6x8__neon_lane_ld64() 159 vacc2x4567 = vminq_f32(vacc2x4567, vmax); in xnn_f32_gemm_minmax_ukernel_6x8__neon_lane_ld64() 160 vacc3x4567 = vminq_f32(vacc3x4567, vmax); in xnn_f32_gemm_minmax_ukernel_6x8__neon_lane_ld64() [all …]
|
/external/XNNPACK/src/f32-gemm/gen-inc/ |
D | 6x8inc-minmax-neon-lane-ld64.c | 153 vacc0x0123 = vminq_f32(vacc0x0123, vmax); in xnn_f32_gemminc_minmax_ukernel_6x8__neon_lane_ld64() 154 vacc1x0123 = vminq_f32(vacc1x0123, vmax); in xnn_f32_gemminc_minmax_ukernel_6x8__neon_lane_ld64() 155 vacc2x0123 = vminq_f32(vacc2x0123, vmax); in xnn_f32_gemminc_minmax_ukernel_6x8__neon_lane_ld64() 156 vacc3x0123 = vminq_f32(vacc3x0123, vmax); in xnn_f32_gemminc_minmax_ukernel_6x8__neon_lane_ld64() 157 vacc4x0123 = vminq_f32(vacc4x0123, vmax); in xnn_f32_gemminc_minmax_ukernel_6x8__neon_lane_ld64() 158 vacc5x0123 = vminq_f32(vacc5x0123, vmax); in xnn_f32_gemminc_minmax_ukernel_6x8__neon_lane_ld64() 159 vacc0x4567 = vminq_f32(vacc0x4567, vmax); in xnn_f32_gemminc_minmax_ukernel_6x8__neon_lane_ld64() 160 vacc1x4567 = vminq_f32(vacc1x4567, vmax); in xnn_f32_gemminc_minmax_ukernel_6x8__neon_lane_ld64() 161 vacc2x4567 = vminq_f32(vacc2x4567, vmax); in xnn_f32_gemminc_minmax_ukernel_6x8__neon_lane_ld64() 162 vacc3x4567 = vminq_f32(vacc3x4567, vmax); in xnn_f32_gemminc_minmax_ukernel_6x8__neon_lane_ld64() [all …]
|
D | 6x8inc-minmax-neonfma-lane-ld64.c | 153 vacc0x0123 = vminq_f32(vacc0x0123, vmax); in xnn_f32_gemminc_minmax_ukernel_6x8__neonfma_lane_ld64() 154 vacc1x0123 = vminq_f32(vacc1x0123, vmax); in xnn_f32_gemminc_minmax_ukernel_6x8__neonfma_lane_ld64() 155 vacc2x0123 = vminq_f32(vacc2x0123, vmax); in xnn_f32_gemminc_minmax_ukernel_6x8__neonfma_lane_ld64() 156 vacc3x0123 = vminq_f32(vacc3x0123, vmax); in xnn_f32_gemminc_minmax_ukernel_6x8__neonfma_lane_ld64() 157 vacc4x0123 = vminq_f32(vacc4x0123, vmax); in xnn_f32_gemminc_minmax_ukernel_6x8__neonfma_lane_ld64() 158 vacc5x0123 = vminq_f32(vacc5x0123, vmax); in xnn_f32_gemminc_minmax_ukernel_6x8__neonfma_lane_ld64() 159 vacc0x4567 = vminq_f32(vacc0x4567, vmax); in xnn_f32_gemminc_minmax_ukernel_6x8__neonfma_lane_ld64() 160 vacc1x4567 = vminq_f32(vacc1x4567, vmax); in xnn_f32_gemminc_minmax_ukernel_6x8__neonfma_lane_ld64() 161 vacc2x4567 = vminq_f32(vacc2x4567, vmax); in xnn_f32_gemminc_minmax_ukernel_6x8__neonfma_lane_ld64() 162 vacc3x4567 = vminq_f32(vacc3x4567, vmax); in xnn_f32_gemminc_minmax_ukernel_6x8__neonfma_lane_ld64() [all …]
|