/external/webrtc/webrtc/modules/audio_processing/aec/ |
D | aec_rdft_neon.c | 34 float32x4_t a01v = vcombine_f32(vget_low_f32(a00v), vget_low_f32(a08v)); in cft1st_128_neon() 35 float32x4_t a23v = vcombine_f32(vget_high_f32(a00v), vget_high_f32(a08v)); in cft1st_128_neon() 36 float32x4_t a45v = vcombine_f32(vget_low_f32(a04v), vget_low_f32(a12v)); in cft1st_128_neon() 37 float32x4_t a67v = vcombine_f32(vget_high_f32(a04v), vget_high_f32(a12v)); in cft1st_128_neon() 63 a00v = vcombine_f32(vget_low_f32(a01v), vget_low_f32(a23v)); in cft1st_128_neon() 64 a04v = vcombine_f32(vget_low_f32(a45v), vget_low_f32(a67v)); in cft1st_128_neon() 65 a08v = vcombine_f32(vget_high_f32(a01v), vget_high_f32(a23v)); in cft1st_128_neon() 66 a12v = vcombine_f32(vget_high_f32(a45v), vget_high_f32(a67v)); in cft1st_128_neon() 85 const float32x4_t a_00_32 = vcombine_f32(a_00, a_32); in cftmdl_128_neon() 86 const float32x4_t a_08_40 = vcombine_f32(a_08, a_40); in cftmdl_128_neon() [all …]
|
/external/XNNPACK/src/f32-sigmoid/gen/ |
D | neonfma-rr1-lut64-p2-div-x24.c | 128 const float32x4_t vl0123 = vcombine_f32(vl01, vl23); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x24() 131 const float32x4_t vl4567 = vcombine_f32(vl45, vl67); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x24() 134 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x24() 137 const float32x4_t vlCDEF = vcombine_f32(vlCD, vlEF); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x24() 140 const float32x4_t vlGHIJ = vcombine_f32(vlGH, vlIJ); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x24() 143 const float32x4_t vlKLMN = vcombine_f32(vlKL, vlMN); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x24() 285 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x24() 364 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x24()
|
D | neonfma-rr1-lut2048-p1-div-x24.c | 128 const float32x4_t vl0123 = vcombine_f32(vl01, vl23); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x24() 131 const float32x4_t vl4567 = vcombine_f32(vl45, vl67); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x24() 134 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x24() 137 const float32x4_t vlCDEF = vcombine_f32(vlCD, vlEF); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x24() 140 const float32x4_t vlGHIJ = vcombine_f32(vlGH, vlIJ); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x24() 143 const float32x4_t vlKLMN = vcombine_f32(vlKL, vlMN); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x24() 277 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x24() 354 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x24()
|
D | neonfma-rr1-lut2048-p1-div-x20.c | 119 const float32x4_t vl0123 = vcombine_f32(vl01, vl23); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x20() 122 const float32x4_t vl4567 = vcombine_f32(vl45, vl67); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x20() 125 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x20() 128 const float32x4_t vlCDEF = vcombine_f32(vlCD, vlEF); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x20() 131 const float32x4_t vlGHIJ = vcombine_f32(vlGH, vlIJ); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x20() 254 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x20() 331 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x20()
|
D | neonfma-rr1-lut2048-p1-div-x16.c | 110 const float32x4_t vl0123 = vcombine_f32(vl01, vl23); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x16() 113 const float32x4_t vl4567 = vcombine_f32(vl45, vl67); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x16() 116 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x16() 119 const float32x4_t vlCDEF = vcombine_f32(vlCD, vlEF); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x16() 231 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x16() 308 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_div_x16()
|
D | neonfma-rr1-lut64-p2-div-x16.c | 110 const float32x4_t vl0123 = vcombine_f32(vl01, vl23); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x16() 113 const float32x4_t vl4567 = vcombine_f32(vl45, vl67); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x16() 116 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x16() 119 const float32x4_t vlCDEF = vcombine_f32(vlCD, vlEF); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x16() 237 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x16() 316 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x16()
|
D | neonfma-rr1-lut64-p2-div-x20.c | 119 const float32x4_t vl0123 = vcombine_f32(vl01, vl23); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x20() 122 const float32x4_t vl4567 = vcombine_f32(vl45, vl67); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x20() 125 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x20() 128 const float32x4_t vlCDEF = vcombine_f32(vlCD, vlEF); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x20() 131 const float32x4_t vlGHIJ = vcombine_f32(vlGH, vlIJ); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x20() 261 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x20() 340 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_div_x20()
|
D | neonfma-rr1-lut64-p2-nr1recps1fma-x24.c | 128 const float32x4_t vl0123 = vcombine_f32(vl01, vl23); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr1recps1fma_x24() 131 const float32x4_t vl4567 = vcombine_f32(vl45, vl67); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr1recps1fma_x24() 134 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr1recps1fma_x24() 137 const float32x4_t vlCDEF = vcombine_f32(vlCD, vlEF); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr1recps1fma_x24() 140 const float32x4_t vlGHIJ = vcombine_f32(vlGH, vlIJ); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr1recps1fma_x24() 143 const float32x4_t vlKLMN = vcombine_f32(vlKL, vlMN); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr1recps1fma_x24() 309 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr1recps1fma_x24() 397 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr1recps1fma_x24()
|
D | neonfma-rr1-lut2048-p1-nr2fma-x24.c | 128 const float32x4_t vl0123 = vcombine_f32(vl01, vl23); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2fma_x24() 131 const float32x4_t vl4567 = vcombine_f32(vl45, vl67); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2fma_x24() 134 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2fma_x24() 137 const float32x4_t vlCDEF = vcombine_f32(vlCD, vlEF); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2fma_x24() 140 const float32x4_t vlGHIJ = vcombine_f32(vlGH, vlIJ); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2fma_x24() 143 const float32x4_t vlKLMN = vcombine_f32(vlKL, vlMN); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2fma_x24() 301 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2fma_x24() 387 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2fma_x24()
|
D | neonfma-rr1-lut64-p2-nr2fma-x24.c | 128 const float32x4_t vl0123 = vcombine_f32(vl01, vl23); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2fma_x24() 131 const float32x4_t vl4567 = vcombine_f32(vl45, vl67); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2fma_x24() 134 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2fma_x24() 137 const float32x4_t vlCDEF = vcombine_f32(vlCD, vlEF); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2fma_x24() 140 const float32x4_t vlGHIJ = vcombine_f32(vlGH, vlIJ); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2fma_x24() 143 const float32x4_t vlKLMN = vcombine_f32(vlKL, vlMN); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2fma_x24() 309 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2fma_x24() 397 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2fma_x24()
|
D | neonfma-rr1-lut2048-p1-nr2recps-x24.c | 128 const float32x4_t vl0123 = vcombine_f32(vl01, vl23); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2recps_x24() 131 const float32x4_t vl4567 = vcombine_f32(vl45, vl67); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2recps_x24() 134 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2recps_x24() 137 const float32x4_t vlCDEF = vcombine_f32(vlCD, vlEF); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2recps_x24() 140 const float32x4_t vlGHIJ = vcombine_f32(vlGH, vlIJ); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2recps_x24() 143 const float32x4_t vlKLMN = vcombine_f32(vlKL, vlMN); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2recps_x24() 301 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2recps_x24() 387 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut2048_p1_nr2recps_x24()
|
D | neonfma-rr1-lut64-p2-nr2recps-x24.c | 128 const float32x4_t vl0123 = vcombine_f32(vl01, vl23); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2recps_x24() 131 const float32x4_t vl4567 = vcombine_f32(vl45, vl67); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2recps_x24() 134 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2recps_x24() 137 const float32x4_t vlCDEF = vcombine_f32(vlCD, vlEF); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2recps_x24() 140 const float32x4_t vlGHIJ = vcombine_f32(vlGH, vlIJ); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2recps_x24() 143 const float32x4_t vlKLMN = vcombine_f32(vlKL, vlMN); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2recps_x24() 309 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2recps_x24() 397 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_sigmoid_ukernel__neonfma_rr1_lut64_p2_nr2recps_x24()
|
D | neon-rr2-lut64-p2-nr2recps-x24.c | 130 const float32x4_t vl0123 = vcombine_f32(vl01, vl23); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x24() 133 const float32x4_t vl4567 = vcombine_f32(vl45, vl67); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x24() 136 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x24() 139 const float32x4_t vlCDEF = vcombine_f32(vlCD, vlEF); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x24() 142 const float32x4_t vlGHIJ = vcombine_f32(vlGH, vlIJ); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x24() 145 const float32x4_t vlKLMN = vcombine_f32(vlKL, vlMN); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x24() 319 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x24() 409 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_sigmoid_ukernel__neon_rr2_lut64_p2_nr2recps_x24()
|
/external/XNNPACK/src/f32-raddstoreexpminusmax/gen/ |
D | neonfma-lut64-p2-x20.c | 117 const float32x4_t vl0123 = vcombine_f32(vl01, vl23); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x20() 120 const float32x4_t vl4567 = vcombine_f32(vl45, vl67); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x20() 123 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x20() 126 const float32x4_t vlCDEF = vcombine_f32(vlCD, vlEF); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x20() 129 const float32x4_t vlGHIJ = vcombine_f32(vlGH, vlIJ); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x20() 243 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x20() 319 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x20()
|
D | neonfma-lut64-p2-x20-acc5.c | 121 const float32x4_t vl0123 = vcombine_f32(vl01, vl23); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x20_acc5() 124 const float32x4_t vl4567 = vcombine_f32(vl45, vl67); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x20_acc5() 127 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x20_acc5() 130 const float32x4_t vlCDEF = vcombine_f32(vlCD, vlEF); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x20_acc5() 133 const float32x4_t vlGHIJ = vcombine_f32(vlGH, vlIJ); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x20_acc5() 252 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x20_acc5() 328 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x20_acc5()
|
D | neonfma-lut64-p2-x20-acc2.c | 118 const float32x4_t vl0123 = vcombine_f32(vl01, vl23); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x20_acc2() 121 const float32x4_t vl4567 = vcombine_f32(vl45, vl67); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x20_acc2() 124 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x20_acc2() 127 const float32x4_t vlCDEF = vcombine_f32(vlCD, vlEF); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x20_acc2() 130 const float32x4_t vlGHIJ = vcombine_f32(vlGH, vlIJ); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x20_acc2() 246 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x20_acc2() 322 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x20_acc2()
|
D | neon-lut64-p2-x20-acc2.c | 119 const float32x4_t vl0123 = vcombine_f32(vl01, vl23); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x20_acc2() 122 const float32x4_t vl4567 = vcombine_f32(vl45, vl67); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x20_acc2() 125 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x20_acc2() 128 const float32x4_t vlCDEF = vcombine_f32(vlCD, vlEF); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x20_acc2() 131 const float32x4_t vlGHIJ = vcombine_f32(vlGH, vlIJ); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x20_acc2() 247 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x20_acc2() 323 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x20_acc2()
|
D | neon-lut64-p2-x20.c | 118 const float32x4_t vl0123 = vcombine_f32(vl01, vl23); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x20() 121 const float32x4_t vl4567 = vcombine_f32(vl45, vl67); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x20() 124 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x20() 127 const float32x4_t vlCDEF = vcombine_f32(vlCD, vlEF); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x20() 130 const float32x4_t vlGHIJ = vcombine_f32(vlGH, vlIJ); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x20() 244 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x20() 320 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x20()
|
D | neon-lut64-p2-x20-acc5.c | 122 const float32x4_t vl0123 = vcombine_f32(vl01, vl23); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x20_acc5() 125 const float32x4_t vl4567 = vcombine_f32(vl45, vl67); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x20_acc5() 128 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x20_acc5() 131 const float32x4_t vlCDEF = vcombine_f32(vlCD, vlEF); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x20_acc5() 134 const float32x4_t vlGHIJ = vcombine_f32(vlGH, vlIJ); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x20_acc5() 253 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x20_acc5() 329 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x20_acc5()
|
D | neonfma-lut64-p2-x16-acc4.c | 111 const float32x4_t vl0123 = vcombine_f32(vl01, vl23); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x16_acc4() 114 const float32x4_t vl4567 = vcombine_f32(vl45, vl67); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x16_acc4() 117 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x16_acc4() 120 const float32x4_t vlCDEF = vcombine_f32(vlCD, vlEF); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x16_acc4() 228 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x16_acc4() 304 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x16_acc4()
|
D | neon-lut64-p2-x16-acc2.c | 110 const float32x4_t vl0123 = vcombine_f32(vl01, vl23); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x16_acc2() 113 const float32x4_t vl4567 = vcombine_f32(vl45, vl67); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x16_acc2() 116 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x16_acc2() 119 const float32x4_t vlCDEF = vcombine_f32(vlCD, vlEF); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x16_acc2() 225 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x16_acc2() 301 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x16_acc2()
|
D | neon-lut64-p2-x16.c | 109 const float32x4_t vl0123 = vcombine_f32(vl01, vl23); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x16() 112 const float32x4_t vl4567 = vcombine_f32(vl45, vl67); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x16() 115 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x16() 118 const float32x4_t vlCDEF = vcombine_f32(vlCD, vlEF); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x16() 222 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x16() 298 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x16()
|
D | neonfma-lut64-p2-x16-acc2.c | 109 const float32x4_t vl0123 = vcombine_f32(vl01, vl23); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x16_acc2() 112 const float32x4_t vl4567 = vcombine_f32(vl45, vl67); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x16_acc2() 115 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x16_acc2() 118 const float32x4_t vlCDEF = vcombine_f32(vlCD, vlEF); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x16_acc2() 224 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x16_acc2() 300 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x16_acc2()
|
D | neon-lut64-p2-x16-acc4.c | 112 const float32x4_t vl0123 = vcombine_f32(vl01, vl23); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x16_acc4() 115 const float32x4_t vl4567 = vcombine_f32(vl45, vl67); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x16_acc4() 118 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x16_acc4() 121 const float32x4_t vlCDEF = vcombine_f32(vlCD, vlEF); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x16_acc4() 229 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x16_acc4() 305 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_raddstoreexpminusmax_ukernel__neon_lut64_p2_x16_acc4()
|
D | neonfma-lut64-p2-x16.c | 108 const float32x4_t vl0123 = vcombine_f32(vl01, vl23); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x16() 111 const float32x4_t vl4567 = vcombine_f32(vl45, vl67); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x16() 114 const float32x4_t vl89AB = vcombine_f32(vl89, vlAB); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x16() 117 const float32x4_t vlCDEF = vcombine_f32(vlCD, vlEF); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x16() 221 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x16() 297 const float32x4_t vl = vcombine_f32(vl_lo, vl_hi); in xnn_f32_raddstoreexpminusmax_ukernel__neonfma_lut64_p2_x16()
|