/external/XNNPACK/src/f32-raddstoreexpminusmax/gen/ |
D | sse2-p5-x20.c | 139 vf0123 = _mm_andnot_ps(_mm_cmplt_ps(vx0123, vdenorm_cutoff), vf0123); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x20() 140 vf4567 = _mm_andnot_ps(_mm_cmplt_ps(vx4567, vdenorm_cutoff), vf4567); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x20() 141 vf89AB = _mm_andnot_ps(_mm_cmplt_ps(vx89AB, vdenorm_cutoff), vf89AB); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x20() 142 vfCDEF = _mm_andnot_ps(_mm_cmplt_ps(vxCDEF, vdenorm_cutoff), vfCDEF); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x20() 143 vfGHIJ = _mm_andnot_ps(_mm_cmplt_ps(vxGHIJ, vdenorm_cutoff), vfGHIJ); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x20() 200 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x20() 248 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x20()
|
D | sse2-p5-x20-acc2.c | 140 vf0123 = _mm_andnot_ps(_mm_cmplt_ps(vx0123, vdenorm_cutoff), vf0123); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x20_acc2() 141 vf4567 = _mm_andnot_ps(_mm_cmplt_ps(vx4567, vdenorm_cutoff), vf4567); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x20_acc2() 142 vf89AB = _mm_andnot_ps(_mm_cmplt_ps(vx89AB, vdenorm_cutoff), vf89AB); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x20_acc2() 143 vfCDEF = _mm_andnot_ps(_mm_cmplt_ps(vxCDEF, vdenorm_cutoff), vfCDEF); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x20_acc2() 144 vfGHIJ = _mm_andnot_ps(_mm_cmplt_ps(vxGHIJ, vdenorm_cutoff), vfGHIJ); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x20_acc2() 203 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x20_acc2() 251 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x20_acc2()
|
D | sse2-p5-x20-acc5.c | 143 vf0123 = _mm_andnot_ps(_mm_cmplt_ps(vx0123, vdenorm_cutoff), vf0123); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x20_acc5() 144 vf4567 = _mm_andnot_ps(_mm_cmplt_ps(vx4567, vdenorm_cutoff), vf4567); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x20_acc5() 145 vf89AB = _mm_andnot_ps(_mm_cmplt_ps(vx89AB, vdenorm_cutoff), vf89AB); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x20_acc5() 146 vfCDEF = _mm_andnot_ps(_mm_cmplt_ps(vxCDEF, vdenorm_cutoff), vfCDEF); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x20_acc5() 147 vfGHIJ = _mm_andnot_ps(_mm_cmplt_ps(vxGHIJ, vdenorm_cutoff), vfGHIJ); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x20_acc5() 209 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x20_acc5() 257 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x20_acc5()
|
D | sse2-p5-x16.c | 126 vf0123 = _mm_andnot_ps(_mm_cmplt_ps(vx0123, vdenorm_cutoff), vf0123); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x16() 127 vf4567 = _mm_andnot_ps(_mm_cmplt_ps(vx4567, vdenorm_cutoff), vf4567); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x16() 128 vf89AB = _mm_andnot_ps(_mm_cmplt_ps(vx89AB, vdenorm_cutoff), vf89AB); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x16() 129 vfCDEF = _mm_andnot_ps(_mm_cmplt_ps(vxCDEF, vdenorm_cutoff), vfCDEF); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x16() 184 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x16() 232 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x16()
|
D | sse2-p5-x16-acc4.c | 129 vf0123 = _mm_andnot_ps(_mm_cmplt_ps(vx0123, vdenorm_cutoff), vf0123); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x16_acc4() 130 vf4567 = _mm_andnot_ps(_mm_cmplt_ps(vx4567, vdenorm_cutoff), vf4567); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x16_acc4() 131 vf89AB = _mm_andnot_ps(_mm_cmplt_ps(vx89AB, vdenorm_cutoff), vf89AB); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x16_acc4() 132 vfCDEF = _mm_andnot_ps(_mm_cmplt_ps(vxCDEF, vdenorm_cutoff), vfCDEF); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x16_acc4() 191 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x16_acc4() 239 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x16_acc4()
|
D | sse2-p5-x16-acc2.c | 127 vf0123 = _mm_andnot_ps(_mm_cmplt_ps(vx0123, vdenorm_cutoff), vf0123); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x16_acc2() 128 vf4567 = _mm_andnot_ps(_mm_cmplt_ps(vx4567, vdenorm_cutoff), vf4567); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x16_acc2() 129 vf89AB = _mm_andnot_ps(_mm_cmplt_ps(vx89AB, vdenorm_cutoff), vf89AB); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x16_acc2() 130 vfCDEF = _mm_andnot_ps(_mm_cmplt_ps(vxCDEF, vdenorm_cutoff), vfCDEF); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x16_acc2() 187 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x16_acc2() 235 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x16_acc2()
|
D | sse2-p5-x12.c | 113 vf0123 = _mm_andnot_ps(_mm_cmplt_ps(vx0123, vdenorm_cutoff), vf0123); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x12() 114 vf4567 = _mm_andnot_ps(_mm_cmplt_ps(vx4567, vdenorm_cutoff), vf4567); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x12() 115 vf89AB = _mm_andnot_ps(_mm_cmplt_ps(vx89AB, vdenorm_cutoff), vf89AB); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x12() 168 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x12() 216 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x12()
|
D | sse2-p5-x12-acc3.c | 115 vf0123 = _mm_andnot_ps(_mm_cmplt_ps(vx0123, vdenorm_cutoff), vf0123); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x12_acc3() 116 vf4567 = _mm_andnot_ps(_mm_cmplt_ps(vx4567, vdenorm_cutoff), vf4567); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x12_acc3() 117 vf89AB = _mm_andnot_ps(_mm_cmplt_ps(vx89AB, vdenorm_cutoff), vf89AB); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x12_acc3() 173 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x12_acc3() 221 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x12_acc3()
|
D | sse2-p5-x12-acc2.c | 114 vf0123 = _mm_andnot_ps(_mm_cmplt_ps(vx0123, vdenorm_cutoff), vf0123); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x12_acc2() 115 vf4567 = _mm_andnot_ps(_mm_cmplt_ps(vx4567, vdenorm_cutoff), vf4567); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x12_acc2() 116 vf89AB = _mm_andnot_ps(_mm_cmplt_ps(vx89AB, vdenorm_cutoff), vf89AB); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x12_acc2() 171 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x12_acc2() 219 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x12_acc2()
|
D | sse2-p5-x8.c | 100 vf0123 = _mm_andnot_ps(_mm_cmplt_ps(vx0123, vdenorm_cutoff), vf0123); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x8() 101 vf4567 = _mm_andnot_ps(_mm_cmplt_ps(vx4567, vdenorm_cutoff), vf4567); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x8() 152 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x8() 200 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x8()
|
D | sse2-p5-x8-acc2.c | 101 vf0123 = _mm_andnot_ps(_mm_cmplt_ps(vx0123, vdenorm_cutoff), vf0123); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x8_acc2() 102 vf4567 = _mm_andnot_ps(_mm_cmplt_ps(vx4567, vdenorm_cutoff), vf4567); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x8_acc2() 155 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x8_acc2() 203 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x8_acc2()
|
D | sse2-p5-x4.c | 87 vf0123 = _mm_andnot_ps(_mm_cmplt_ps(vx0123, vdenorm_cutoff), vf0123); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x4() 136 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x4() 184 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_f32_raddstoreexpminusmax_ukernel__sse2_p5_x4()
|
/external/XNNPACK/src/f32-sigmoid/gen/ |
D | sse41-p5-div-x24.c | 176 vf0123 = _mm_andnot_ps(_mm_cmplt_ps(vz0123, vdenorm_cutoff), vf0123); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x24() 177 vf4567 = _mm_andnot_ps(_mm_cmplt_ps(vz4567, vdenorm_cutoff), vf4567); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x24() 178 vf89AB = _mm_andnot_ps(_mm_cmplt_ps(vz89AB, vdenorm_cutoff), vf89AB); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x24() 179 vfCDEF = _mm_andnot_ps(_mm_cmplt_ps(vzCDEF, vdenorm_cutoff), vfCDEF); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x24() 180 vfGHIJ = _mm_andnot_ps(_mm_cmplt_ps(vzGHIJ, vdenorm_cutoff), vfGHIJ); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x24() 181 vfKLMN = _mm_andnot_ps(_mm_cmplt_ps(vzKLMN, vdenorm_cutoff), vfKLMN); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x24() 254 vf = _mm_andnot_ps(_mm_cmplt_ps(vz, vdenorm_cutoff), vf); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x24() 317 vf = _mm_andnot_ps(_mm_cmplt_ps(vz, vdenorm_cutoff), vf); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x24()
|
D | sse41-p5-div-x20.c | 161 vf0123 = _mm_andnot_ps(_mm_cmplt_ps(vz0123, vdenorm_cutoff), vf0123); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x20() 162 vf4567 = _mm_andnot_ps(_mm_cmplt_ps(vz4567, vdenorm_cutoff), vf4567); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x20() 163 vf89AB = _mm_andnot_ps(_mm_cmplt_ps(vz89AB, vdenorm_cutoff), vf89AB); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x20() 164 vfCDEF = _mm_andnot_ps(_mm_cmplt_ps(vzCDEF, vdenorm_cutoff), vfCDEF); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x20() 165 vfGHIJ = _mm_andnot_ps(_mm_cmplt_ps(vzGHIJ, vdenorm_cutoff), vfGHIJ); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x20() 236 vf = _mm_andnot_ps(_mm_cmplt_ps(vz, vdenorm_cutoff), vf); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x20() 299 vf = _mm_andnot_ps(_mm_cmplt_ps(vz, vdenorm_cutoff), vf); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x20()
|
D | sse2-p5-div-x24.c | 176 vf0123 = _mm_andnot_ps(_mm_cmplt_ps(vz0123, vdenorm_cutoff), vf0123); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24() 177 vf4567 = _mm_andnot_ps(_mm_cmplt_ps(vz4567, vdenorm_cutoff), vf4567); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24() 178 vf89AB = _mm_andnot_ps(_mm_cmplt_ps(vz89AB, vdenorm_cutoff), vf89AB); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24() 179 vfCDEF = _mm_andnot_ps(_mm_cmplt_ps(vzCDEF, vdenorm_cutoff), vfCDEF); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24() 180 vfGHIJ = _mm_andnot_ps(_mm_cmplt_ps(vzGHIJ, vdenorm_cutoff), vfGHIJ); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24() 181 vfKLMN = _mm_andnot_ps(_mm_cmplt_ps(vzKLMN, vdenorm_cutoff), vfKLMN); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24() 261 vf = _mm_andnot_ps(_mm_cmplt_ps(vz, vdenorm_cutoff), vf); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24() 325 vf = _mm_andnot_ps(_mm_cmplt_ps(vz, vdenorm_cutoff), vf); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24()
|
D | sse2-p5-div-x20.c | 161 vf0123 = _mm_andnot_ps(_mm_cmplt_ps(vz0123, vdenorm_cutoff), vf0123); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20() 162 vf4567 = _mm_andnot_ps(_mm_cmplt_ps(vz4567, vdenorm_cutoff), vf4567); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20() 163 vf89AB = _mm_andnot_ps(_mm_cmplt_ps(vz89AB, vdenorm_cutoff), vf89AB); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20() 164 vfCDEF = _mm_andnot_ps(_mm_cmplt_ps(vzCDEF, vdenorm_cutoff), vfCDEF); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20() 165 vfGHIJ = _mm_andnot_ps(_mm_cmplt_ps(vzGHIJ, vdenorm_cutoff), vfGHIJ); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20() 242 vf = _mm_andnot_ps(_mm_cmplt_ps(vz, vdenorm_cutoff), vf); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20() 306 vf = _mm_andnot_ps(_mm_cmplt_ps(vz, vdenorm_cutoff), vf); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20()
|
D | sse41-p5-div-x16.c | 146 vf0123 = _mm_andnot_ps(_mm_cmplt_ps(vz0123, vdenorm_cutoff), vf0123); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x16() 147 vf4567 = _mm_andnot_ps(_mm_cmplt_ps(vz4567, vdenorm_cutoff), vf4567); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x16() 148 vf89AB = _mm_andnot_ps(_mm_cmplt_ps(vz89AB, vdenorm_cutoff), vf89AB); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x16() 149 vfCDEF = _mm_andnot_ps(_mm_cmplt_ps(vzCDEF, vdenorm_cutoff), vfCDEF); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x16() 218 vf = _mm_andnot_ps(_mm_cmplt_ps(vz, vdenorm_cutoff), vf); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x16() 281 vf = _mm_andnot_ps(_mm_cmplt_ps(vz, vdenorm_cutoff), vf); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x16()
|
D | sse41-p5-div-x12.c | 131 vf0123 = _mm_andnot_ps(_mm_cmplt_ps(vz0123, vdenorm_cutoff), vf0123); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x12() 132 vf4567 = _mm_andnot_ps(_mm_cmplt_ps(vz4567, vdenorm_cutoff), vf4567); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x12() 133 vf89AB = _mm_andnot_ps(_mm_cmplt_ps(vz89AB, vdenorm_cutoff), vf89AB); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x12() 200 vf = _mm_andnot_ps(_mm_cmplt_ps(vz, vdenorm_cutoff), vf); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x12() 263 vf = _mm_andnot_ps(_mm_cmplt_ps(vz, vdenorm_cutoff), vf); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x12()
|
D | sse2-p5-div-x16.c | 146 vf0123 = _mm_andnot_ps(_mm_cmplt_ps(vz0123, vdenorm_cutoff), vf0123); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x16() 147 vf4567 = _mm_andnot_ps(_mm_cmplt_ps(vz4567, vdenorm_cutoff), vf4567); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x16() 148 vf89AB = _mm_andnot_ps(_mm_cmplt_ps(vz89AB, vdenorm_cutoff), vf89AB); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x16() 149 vfCDEF = _mm_andnot_ps(_mm_cmplt_ps(vzCDEF, vdenorm_cutoff), vfCDEF); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x16() 223 vf = _mm_andnot_ps(_mm_cmplt_ps(vz, vdenorm_cutoff), vf); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x16() 287 vf = _mm_andnot_ps(_mm_cmplt_ps(vz, vdenorm_cutoff), vf); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x16()
|
D | sse2-p5-div-x12.c | 131 vf0123 = _mm_andnot_ps(_mm_cmplt_ps(vz0123, vdenorm_cutoff), vf0123); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x12() 132 vf4567 = _mm_andnot_ps(_mm_cmplt_ps(vz4567, vdenorm_cutoff), vf4567); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x12() 133 vf89AB = _mm_andnot_ps(_mm_cmplt_ps(vz89AB, vdenorm_cutoff), vf89AB); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x12() 204 vf = _mm_andnot_ps(_mm_cmplt_ps(vz, vdenorm_cutoff), vf); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x12() 268 vf = _mm_andnot_ps(_mm_cmplt_ps(vz, vdenorm_cutoff), vf); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x12()
|
D | sse41-p5-div-x8.c | 116 vf0123 = _mm_andnot_ps(_mm_cmplt_ps(vz0123, vdenorm_cutoff), vf0123); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x8() 117 vf4567 = _mm_andnot_ps(_mm_cmplt_ps(vz4567, vdenorm_cutoff), vf4567); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x8() 182 vf = _mm_andnot_ps(_mm_cmplt_ps(vz, vdenorm_cutoff), vf); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x8() 245 vf = _mm_andnot_ps(_mm_cmplt_ps(vz, vdenorm_cutoff), vf); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x8()
|
D | sse2-p5-div-x8.c | 116 vf0123 = _mm_andnot_ps(_mm_cmplt_ps(vz0123, vdenorm_cutoff), vf0123); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x8() 117 vf4567 = _mm_andnot_ps(_mm_cmplt_ps(vz4567, vdenorm_cutoff), vf4567); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x8() 185 vf = _mm_andnot_ps(_mm_cmplt_ps(vz, vdenorm_cutoff), vf); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x8() 249 vf = _mm_andnot_ps(_mm_cmplt_ps(vz, vdenorm_cutoff), vf); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x8()
|
D | sse41-p5-div-x4.c | 96 vf = _mm_andnot_ps(_mm_cmplt_ps(vz, vdenorm_cutoff), vf); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x4() 159 vf = _mm_andnot_ps(_mm_cmplt_ps(vz, vdenorm_cutoff), vf); in xnn_f32_sigmoid_ukernel__sse41_p5_div_x4()
|
D | sse2-p5-div-x4.c | 96 vf = _mm_andnot_ps(_mm_cmplt_ps(vz, vdenorm_cutoff), vf); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x4() 160 vf = _mm_andnot_ps(_mm_cmplt_ps(vz, vdenorm_cutoff), vf); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x4()
|
/external/XNNPACK/src/math/ |
D | expminus-sse2-p5.c | 72 vf = _mm_andnot_ps(_mm_cmplt_ps(vx, vdenorm_cutoff), vf); in xnn_math_f32_expminus__sse2_p5()
|