/external/XNNPACK/src/f32-argmaxpool/ |
D | 9p8x-neon-c4.c | 70 vmax = vbslq_f32(vm1, vi1, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4() 74 vmax = vbslq_f32(vm2, vi2, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4() 78 vmax = vbslq_f32(vm3, vi3, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4() 82 vmax = vbslq_f32(vm4, vi4, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4() 86 vmax = vbslq_f32(vm5, vi5, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4() 90 vmax = vbslq_f32(vm6, vi6, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4() 94 vmax = vbslq_f32(vm7, vi7, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4() 98 vmax = vbslq_f32(vm8, vi8, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4() 145 vmax = vbslq_f32(vm0, vi0, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4() 150 vmax = vbslq_f32(vm1, vi1, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4() [all …]
|
D | 9x-neon-c4.c | 89 vmax = vbslq_f32(vm1, vi1, vmax); in xnn_f32_argmaxpool_ukernel_9x__neon_c4() 93 vmax = vbslq_f32(vm2, vi2, vmax); in xnn_f32_argmaxpool_ukernel_9x__neon_c4() 97 vmax = vbslq_f32(vm3, vi3, vmax); in xnn_f32_argmaxpool_ukernel_9x__neon_c4() 101 vmax = vbslq_f32(vm4, vi4, vmax); in xnn_f32_argmaxpool_ukernel_9x__neon_c4() 105 vmax = vbslq_f32(vm5, vi5, vmax); in xnn_f32_argmaxpool_ukernel_9x__neon_c4() 109 vmax = vbslq_f32(vm6, vi6, vmax); in xnn_f32_argmaxpool_ukernel_9x__neon_c4() 113 vmax = vbslq_f32(vm7, vi7, vmax); in xnn_f32_argmaxpool_ukernel_9x__neon_c4() 117 vmax = vbslq_f32(vm8, vi8, vmax); in xnn_f32_argmaxpool_ukernel_9x__neon_c4() 138 vmax = vbslq_f32(vm1, vi1, vmax); in xnn_f32_argmaxpool_ukernel_9x__neon_c4() 142 vmax = vbslq_f32(vm2, vi2, vmax); in xnn_f32_argmaxpool_ukernel_9x__neon_c4() [all …]
|
D | 4x-neon-c4.c | 59 vmax = vbslq_f32(vm1, vi1, vmax); in xnn_f32_argmaxpool_ukernel_4x__neon_c4() 63 vmax = vbslq_f32(vm2, vi2, vmax); in xnn_f32_argmaxpool_ukernel_4x__neon_c4() 67 vmax = vbslq_f32(vm3, vi3, vmax); in xnn_f32_argmaxpool_ukernel_4x__neon_c4() 83 vmax = vbslq_f32(vm1, vi1, vmax); in xnn_f32_argmaxpool_ukernel_4x__neon_c4() 87 vmax = vbslq_f32(vm2, vi2, vmax); in xnn_f32_argmaxpool_ukernel_4x__neon_c4() 91 vmax = vbslq_f32(vm3, vi3, vmax); in xnn_f32_argmaxpool_ukernel_4x__neon_c4()
|
/external/XNNPACK/src/f32-prelu/gen/ |
D | neon-4x16.c | 114 vacc0x0123 = vbslq_f32(vm0x0123, vacc0x0123, vi0x0123); in xnn_f32_prelu_ukernel__neon_4x16() 115 vacc0x4567 = vbslq_f32(vm0x4567, vacc0x4567, vi0x4567); in xnn_f32_prelu_ukernel__neon_4x16() 116 vacc0x89AB = vbslq_f32(vm0x89AB, vacc0x89AB, vi0x89AB); in xnn_f32_prelu_ukernel__neon_4x16() 117 vacc0xCDEF = vbslq_f32(vm0xCDEF, vacc0xCDEF, vi0xCDEF); in xnn_f32_prelu_ukernel__neon_4x16() 118 vacc1x0123 = vbslq_f32(vm1x0123, vacc1x0123, vi1x0123); in xnn_f32_prelu_ukernel__neon_4x16() 119 vacc1x4567 = vbslq_f32(vm1x4567, vacc1x4567, vi1x4567); in xnn_f32_prelu_ukernel__neon_4x16() 120 vacc1x89AB = vbslq_f32(vm1x89AB, vacc1x89AB, vi1x89AB); in xnn_f32_prelu_ukernel__neon_4x16() 121 vacc1xCDEF = vbslq_f32(vm1xCDEF, vacc1xCDEF, vi1xCDEF); in xnn_f32_prelu_ukernel__neon_4x16() 122 vacc2x0123 = vbslq_f32(vm2x0123, vacc2x0123, vi2x0123); in xnn_f32_prelu_ukernel__neon_4x16() 123 vacc2x4567 = vbslq_f32(vm2x4567, vacc2x4567, vi2x4567); in xnn_f32_prelu_ukernel__neon_4x16() [all …]
|
D | neon-4x8.c | 88 vacc0x0123 = vbslq_f32(vm0x0123, vacc0x0123, vi0x0123); in xnn_f32_prelu_ukernel__neon_4x8() 89 vacc0x4567 = vbslq_f32(vm0x4567, vacc0x4567, vi0x4567); in xnn_f32_prelu_ukernel__neon_4x8() 90 vacc1x0123 = vbslq_f32(vm1x0123, vacc1x0123, vi1x0123); in xnn_f32_prelu_ukernel__neon_4x8() 91 vacc1x4567 = vbslq_f32(vm1x4567, vacc1x4567, vi1x4567); in xnn_f32_prelu_ukernel__neon_4x8() 92 vacc2x0123 = vbslq_f32(vm2x0123, vacc2x0123, vi2x0123); in xnn_f32_prelu_ukernel__neon_4x8() 93 vacc2x4567 = vbslq_f32(vm2x4567, vacc2x4567, vi2x4567); in xnn_f32_prelu_ukernel__neon_4x8() 94 vacc3x0123 = vbslq_f32(vm3x0123, vacc3x0123, vi3x0123); in xnn_f32_prelu_ukernel__neon_4x8() 95 vacc3x4567 = vbslq_f32(vm3x4567, vacc3x4567, vi3x4567); in xnn_f32_prelu_ukernel__neon_4x8() 127 vacc0x0123 = vbslq_f32(vm0x0123, vacc0x0123, vi0x0123); in xnn_f32_prelu_ukernel__neon_4x8() 128 vacc1x0123 = vbslq_f32(vm1x0123, vacc1x0123, vi1x0123); in xnn_f32_prelu_ukernel__neon_4x8() [all …]
|
D | neon-2x16.c | 78 vacc0x0123 = vbslq_f32(vm0x0123, vacc0x0123, vi0x0123); in xnn_f32_prelu_ukernel__neon_2x16() 79 vacc0x4567 = vbslq_f32(vm0x4567, vacc0x4567, vi0x4567); in xnn_f32_prelu_ukernel__neon_2x16() 80 vacc0x89AB = vbslq_f32(vm0x89AB, vacc0x89AB, vi0x89AB); in xnn_f32_prelu_ukernel__neon_2x16() 81 vacc0xCDEF = vbslq_f32(vm0xCDEF, vacc0xCDEF, vi0xCDEF); in xnn_f32_prelu_ukernel__neon_2x16() 82 vacc1x0123 = vbslq_f32(vm1x0123, vacc1x0123, vi1x0123); in xnn_f32_prelu_ukernel__neon_2x16() 83 vacc1x4567 = vbslq_f32(vm1x4567, vacc1x4567, vi1x4567); in xnn_f32_prelu_ukernel__neon_2x16() 84 vacc1x89AB = vbslq_f32(vm1x89AB, vacc1x89AB, vi1x89AB); in xnn_f32_prelu_ukernel__neon_2x16() 85 vacc1xCDEF = vbslq_f32(vm1xCDEF, vacc1xCDEF, vi1xCDEF); in xnn_f32_prelu_ukernel__neon_2x16() 109 vacc0x0123 = vbslq_f32(vm0x0123, vacc0x0123, vi0x0123); in xnn_f32_prelu_ukernel__neon_2x16() 110 vacc1x0123 = vbslq_f32(vm1x0123, vacc1x0123, vi1x0123); in xnn_f32_prelu_ukernel__neon_2x16() [all …]
|
D | neon-2x8.c | 64 vacc0x0123 = vbslq_f32(vm0x0123, vacc0x0123, vi0x0123); in xnn_f32_prelu_ukernel__neon_2x8() 65 vacc0x4567 = vbslq_f32(vm0x4567, vacc0x4567, vi0x4567); in xnn_f32_prelu_ukernel__neon_2x8() 66 vacc1x0123 = vbslq_f32(vm1x0123, vacc1x0123, vi1x0123); in xnn_f32_prelu_ukernel__neon_2x8() 67 vacc1x4567 = vbslq_f32(vm1x4567, vacc1x4567, vi1x4567); in xnn_f32_prelu_ukernel__neon_2x8() 87 vacc0x0123 = vbslq_f32(vm0x0123, vacc0x0123, vi0x0123); in xnn_f32_prelu_ukernel__neon_2x8() 88 vacc1x0123 = vbslq_f32(vm1x0123, vacc1x0123, vi1x0123); in xnn_f32_prelu_ukernel__neon_2x8() 106 vacc0x0123 = vbslq_f32(vm0x0123, vacc0x0123, vi0x0123); in xnn_f32_prelu_ukernel__neon_2x8() 107 vacc1x0123 = vbslq_f32(vm1x0123, vacc1x0123, vi1x0123); in xnn_f32_prelu_ukernel__neon_2x8()
|
D | neon-4x4.c | 75 vacc0x0123 = vbslq_f32(vm0x0123, vacc0x0123, vi0x0123); in xnn_f32_prelu_ukernel__neon_4x4() 76 vacc1x0123 = vbslq_f32(vm1x0123, vacc1x0123, vi1x0123); in xnn_f32_prelu_ukernel__neon_4x4() 77 vacc2x0123 = vbslq_f32(vm2x0123, vacc2x0123, vi2x0123); in xnn_f32_prelu_ukernel__neon_4x4() 78 vacc3x0123 = vbslq_f32(vm3x0123, vacc3x0123, vi3x0123); in xnn_f32_prelu_ukernel__neon_4x4() 106 vacc0x0123 = vbslq_f32(vm0x0123, vacc0x0123, vi0x0123); in xnn_f32_prelu_ukernel__neon_4x4() 107 vacc1x0123 = vbslq_f32(vm1x0123, vacc1x0123, vi1x0123); in xnn_f32_prelu_ukernel__neon_4x4() 108 vacc2x0123 = vbslq_f32(vm2x0123, vacc2x0123, vi2x0123); in xnn_f32_prelu_ukernel__neon_4x4() 109 vacc3x0123 = vbslq_f32(vm3x0123, vacc3x0123, vi3x0123); in xnn_f32_prelu_ukernel__neon_4x4()
|
D | neon-1x16.c | 60 vacc0x0123 = vbslq_f32(vm0x0123, vacc0x0123, vi0x0123); in xnn_f32_prelu_ukernel__neon_1x16() 61 vacc0x4567 = vbslq_f32(vm0x4567, vacc0x4567, vi0x4567); in xnn_f32_prelu_ukernel__neon_1x16() 62 vacc0x89AB = vbslq_f32(vm0x89AB, vacc0x89AB, vi0x89AB); in xnn_f32_prelu_ukernel__neon_1x16() 63 vacc0xCDEF = vbslq_f32(vm0xCDEF, vacc0xCDEF, vi0xCDEF); in xnn_f32_prelu_ukernel__neon_1x16() 79 vacc0x0123 = vbslq_f32(vm0x0123, vacc0x0123, vi0x0123); in xnn_f32_prelu_ukernel__neon_1x16() 92 vacc0x0123 = vbslq_f32(vm0x0123, vacc0x0123, vi0x0123); in xnn_f32_prelu_ukernel__neon_1x16()
|
D | neon-1x8.c | 52 vacc0x0123 = vbslq_f32(vm0x0123, vacc0x0123, vi0x0123); in xnn_f32_prelu_ukernel__neon_1x8() 53 vacc0x4567 = vbslq_f32(vm0x4567, vacc0x4567, vi0x4567); in xnn_f32_prelu_ukernel__neon_1x8() 67 vacc0x0123 = vbslq_f32(vm0x0123, vacc0x0123, vi0x0123); in xnn_f32_prelu_ukernel__neon_1x8() 80 vacc0x0123 = vbslq_f32(vm0x0123, vacc0x0123, vi0x0123); in xnn_f32_prelu_ukernel__neon_1x8()
|
D | neon-2x4.c | 57 vacc0x0123 = vbslq_f32(vm0x0123, vacc0x0123, vi0x0123); in xnn_f32_prelu_ukernel__neon_2x4() 58 vacc1x0123 = vbslq_f32(vm1x0123, vacc1x0123, vi1x0123); in xnn_f32_prelu_ukernel__neon_2x4() 76 vacc0x0123 = vbslq_f32(vm0x0123, vacc0x0123, vi0x0123); in xnn_f32_prelu_ukernel__neon_2x4() 77 vacc1x0123 = vbslq_f32(vm1x0123, vacc1x0123, vi1x0123); in xnn_f32_prelu_ukernel__neon_2x4()
|
/external/XNNPACK/src/f32-vrnd/gen/ |
D | vrndu-neon-x8.c | 46 const float32x4_t vrndx0123 = vbslq_f32(vrndmask0123, vprerndx0123, vx0123); in xnn_f32_vrndu_ukernel__neon_x8() 47 const float32x4_t vrndx4567 = vbslq_f32(vrndmask4567, vprerndx4567, vx4567); in xnn_f32_vrndu_ukernel__neon_x8() 58 const float32x4_t vy0123 = vbslq_f32(vadjmask0123, vrndx0123, vadjrndx0123); in xnn_f32_vrndu_ukernel__neon_x8() 59 const float32x4_t vy4567 = vbslq_f32(vadjmask4567, vrndx4567, vadjrndx4567); in xnn_f32_vrndu_ukernel__neon_x8() 70 const float32x4_t vrndx = vbslq_f32(vrndmask, vprerndx, vx); in xnn_f32_vrndu_ukernel__neon_x8() 74 const float32x4_t vy = vbslq_f32(vadjmask, vrndx, vadjrndx); in xnn_f32_vrndu_ukernel__neon_x8() 83 const float32x4_t vrndx = vbslq_f32(vrndmask, vprerndx, vx); in xnn_f32_vrndu_ukernel__neon_x8() 87 const float32x4_t vy = vbslq_f32(vadjmask, vrndx, vadjrndx); in xnn_f32_vrndu_ukernel__neon_x8()
|
D | vrndu-neon-x4.c | 41 const float32x4_t vrndx0123 = vbslq_f32(vrndmask0123, vprerndx0123, vx0123); in xnn_f32_vrndu_ukernel__neon_x4() 49 const float32x4_t vy0123 = vbslq_f32(vadjmask0123, vrndx0123, vadjrndx0123); in xnn_f32_vrndu_ukernel__neon_x4() 59 const float32x4_t vrndx = vbslq_f32(vrndmask, vprerndx, vx); in xnn_f32_vrndu_ukernel__neon_x4() 63 const float32x4_t vy = vbslq_f32(vadjmask, vrndx, vadjrndx); in xnn_f32_vrndu_ukernel__neon_x4()
|
D | vrndz-neon-x8.c | 45 const float32x4_t vy0123 = vbslq_f32(vrndmask0123, vrndx0123, vx0123); in xnn_f32_vrndz_ukernel__neon_x8() 46 const float32x4_t vy4567 = vbslq_f32(vrndmask4567, vrndx4567, vx4567); in xnn_f32_vrndz_ukernel__neon_x8() 57 const float32x4_t vy = vbslq_f32(vrndmask, vrndx, vx); in xnn_f32_vrndz_ukernel__neon_x8() 66 const float32x4_t vy = vbslq_f32(vrndmask, vrndx, vx); in xnn_f32_vrndz_ukernel__neon_x8()
|
D | vrndne-neon-x8.c | 47 const float32x4_t vy0123 = vbslq_f32(vrndmask0123, vx0123, vrndabsx0123); in xnn_f32_vrndne_ukernel__neon_x8() 48 const float32x4_t vy4567 = vbslq_f32(vrndmask4567, vx4567, vrndabsx4567); in xnn_f32_vrndne_ukernel__neon_x8() 60 const float32x4_t vy = vbslq_f32(vrndmask, vx, vrndabsx); in xnn_f32_vrndne_ukernel__neon_x8() 70 const float32x4_t vy = vbslq_f32(vrndmask, vx, vrndabsx); in xnn_f32_vrndne_ukernel__neon_x8()
|
D | vrndd-neon-x8.c | 46 const float32x4_t vrndx0123 = vbslq_f32(vrndmask0123, vprerndx0123, vx0123); in xnn_f32_vrndd_ukernel__neon_x8() 47 const float32x4_t vrndx4567 = vbslq_f32(vrndmask4567, vprerndx4567, vx4567); in xnn_f32_vrndd_ukernel__neon_x8() 67 const float32x4_t vrndx = vbslq_f32(vrndmask, vprerndx, vx); in xnn_f32_vrndd_ukernel__neon_x8() 79 const float32x4_t vrndx = vbslq_f32(vrndmask, vprerndx, vx); in xnn_f32_vrndd_ukernel__neon_x8()
|
/external/XNNPACK/src/f32-vlrelu/gen/ |
D | vlrelu-neon-x8.c | 38 vacc0123 = vbslq_f32(vmask0123, vacc0123, vx0123); in xnn_f32_vlrelu_ukernel__neon_x8() 39 vacc4567 = vbslq_f32(vmask4567, vacc4567, vx4567); in xnn_f32_vlrelu_ukernel__neon_x8() 48 vacc = vbslq_f32(vmask, vacc, vx); in xnn_f32_vlrelu_ukernel__neon_x8() 55 vacc = vbslq_f32(vmask, vacc, vx); in xnn_f32_vlrelu_ukernel__neon_x8()
|
/external/XNNPACK/src/f32-vrnd/ |
D | vrndu-neon.c.in | 46 …const float32x4_t vrndx${ABC[N:N+4]} = vbslq_f32(vrndmask${ABC[N:N+4]}, vprerndx${ABC[N:N+4]}, vx$… 58 …const float32x4_t vy${ABC[N:N+4]} = vbslq_f32(vadjmask${ABC[N:N+4]}, vrndx${ABC[N:N+4]}, vadjrndx$… 70 const float32x4_t vrndx = vbslq_f32(vrndmask, vprerndx, vx); 74 const float32x4_t vy = vbslq_f32(vadjmask, vrndx, vadjrndx); 83 const float32x4_t vrndx = vbslq_f32(vrndmask, vprerndx, vx); 87 const float32x4_t vy = vbslq_f32(vadjmask, vrndx, vadjrndx);
|
/external/XNNPACK/src/f32-sigmoid/gen/ |
D | neonfma-rr1-p5-div-x24.c | 150 vf0123 = vbslq_f32(vm0123, vf0123, vsubq_f32(vone, vf0123)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x24() 151 vf4567 = vbslq_f32(vm4567, vf4567, vsubq_f32(vone, vf4567)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x24() 152 vf89AB = vbslq_f32(vm89AB, vf89AB, vsubq_f32(vone, vf89AB)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x24() 153 vfCDEF = vbslq_f32(vmCDEF, vfCDEF, vsubq_f32(vone, vfCDEF)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x24() 154 vfGHIJ = vbslq_f32(vmGHIJ, vfGHIJ, vsubq_f32(vone, vfGHIJ)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x24() 155 vfKLMN = vbslq_f32(vmKLMN, vfKLMN, vsubq_f32(vone, vfKLMN)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x24() 186 vf = vbslq_f32(vm, vf, vsubq_f32(vone, vf)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x24() 212 vf = vbslq_f32(vm, vf, vsubq_f32(vone, vf)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x24()
|
D | neonfma-rr1-p5-div-x20.c | 134 vf0123 = vbslq_f32(vm0123, vf0123, vsubq_f32(vone, vf0123)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x20() 135 vf4567 = vbslq_f32(vm4567, vf4567, vsubq_f32(vone, vf4567)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x20() 136 vf89AB = vbslq_f32(vm89AB, vf89AB, vsubq_f32(vone, vf89AB)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x20() 137 vfCDEF = vbslq_f32(vmCDEF, vfCDEF, vsubq_f32(vone, vfCDEF)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x20() 138 vfGHIJ = vbslq_f32(vmGHIJ, vfGHIJ, vsubq_f32(vone, vfGHIJ)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x20() 168 vf = vbslq_f32(vm, vf, vsubq_f32(vone, vf)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x20() 194 vf = vbslq_f32(vm, vf, vsubq_f32(vone, vf)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x20()
|
D | neonfma-rr1-p5-div-x16.c | 118 vf0123 = vbslq_f32(vm0123, vf0123, vsubq_f32(vone, vf0123)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x16() 119 vf4567 = vbslq_f32(vm4567, vf4567, vsubq_f32(vone, vf4567)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x16() 120 vf89AB = vbslq_f32(vm89AB, vf89AB, vsubq_f32(vone, vf89AB)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x16() 121 vfCDEF = vbslq_f32(vmCDEF, vfCDEF, vsubq_f32(vone, vfCDEF)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x16() 150 vf = vbslq_f32(vm, vf, vsubq_f32(vone, vf)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x16() 176 vf = vbslq_f32(vm, vf, vsubq_f32(vone, vf)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_div_x16()
|
D | neonfma-rr1-p5-nr2recps-x24.c | 171 vf0123 = vbslq_f32(vm0123, vf0123, vsubq_f32(vone, vf0123)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr2recps_x24() 172 vf4567 = vbslq_f32(vm4567, vf4567, vsubq_f32(vone, vf4567)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr2recps_x24() 173 vf89AB = vbslq_f32(vm89AB, vf89AB, vsubq_f32(vone, vf89AB)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr2recps_x24() 174 vfCDEF = vbslq_f32(vmCDEF, vfCDEF, vsubq_f32(vone, vfCDEF)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr2recps_x24() 175 vfGHIJ = vbslq_f32(vmGHIJ, vfGHIJ, vsubq_f32(vone, vfGHIJ)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr2recps_x24() 176 vfKLMN = vbslq_f32(vmKLMN, vfKLMN, vsubq_f32(vone, vfKLMN)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr2recps_x24() 211 vf = vbslq_f32(vm, vf, vsubq_f32(vone, vf)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr2recps_x24() 241 vf = vbslq_f32(vm, vf, vsubq_f32(vone, vf)); in xnn_f32_sigmoid_ukernel__neonfma_rr1_p5_nr2recps_x24()
|
/external/XNNPACK/src/f32-velu/gen/ |
D | velu-neon-rr2-p6-x24.c | 169 const float32x4_t vy0123 = vbslq_f32(vm0123, ve0123, vx0123); in xnn_f32_velu_ukernel__neon_rr2_p6_x24() 170 const float32x4_t vy4567 = vbslq_f32(vm4567, ve4567, vx4567); in xnn_f32_velu_ukernel__neon_rr2_p6_x24() 171 const float32x4_t vy89AB = vbslq_f32(vm89AB, ve89AB, vx89AB); in xnn_f32_velu_ukernel__neon_rr2_p6_x24() 172 const float32x4_t vyCDEF = vbslq_f32(vmCDEF, veCDEF, vxCDEF); in xnn_f32_velu_ukernel__neon_rr2_p6_x24() 173 const float32x4_t vyGHIJ = vbslq_f32(vmGHIJ, veGHIJ, vxGHIJ); in xnn_f32_velu_ukernel__neon_rr2_p6_x24() 174 const float32x4_t vyKLMN = vbslq_f32(vmKLMN, veKLMN, vxKLMN); in xnn_f32_velu_ukernel__neon_rr2_p6_x24() 208 const float32x4_t vy = vbslq_f32(vm, ve, vx); in xnn_f32_velu_ukernel__neon_rr2_p6_x24() 237 const float32x4_t vy = vbslq_f32(vm, ve, vx); in xnn_f32_velu_ukernel__neon_rr2_p6_x24()
|
D | velu-neonfma-rr1-p6-x24.c | 161 const float32x4_t vy0123 = vbslq_f32(vm0123, ve0123, vx0123); in xnn_f32_velu_ukernel__neonfma_rr1_p6_x24() 162 const float32x4_t vy4567 = vbslq_f32(vm4567, ve4567, vx4567); in xnn_f32_velu_ukernel__neonfma_rr1_p6_x24() 163 const float32x4_t vy89AB = vbslq_f32(vm89AB, ve89AB, vx89AB); in xnn_f32_velu_ukernel__neonfma_rr1_p6_x24() 164 const float32x4_t vyCDEF = vbslq_f32(vmCDEF, veCDEF, vxCDEF); in xnn_f32_velu_ukernel__neonfma_rr1_p6_x24() 165 const float32x4_t vyGHIJ = vbslq_f32(vmGHIJ, veGHIJ, vxGHIJ); in xnn_f32_velu_ukernel__neonfma_rr1_p6_x24() 166 const float32x4_t vyKLMN = vbslq_f32(vmKLMN, veKLMN, vxKLMN); in xnn_f32_velu_ukernel__neonfma_rr1_p6_x24() 199 const float32x4_t vy = vbslq_f32(vm, ve, vx); in xnn_f32_velu_ukernel__neonfma_rr1_p6_x24() 227 const float32x4_t vy = vbslq_f32(vm, ve, vx); in xnn_f32_velu_ukernel__neonfma_rr1_p6_x24()
|
D | velu-neonfma-rr1-p6-x20.c | 144 const float32x4_t vy0123 = vbslq_f32(vm0123, ve0123, vx0123); in xnn_f32_velu_ukernel__neonfma_rr1_p6_x20() 145 const float32x4_t vy4567 = vbslq_f32(vm4567, ve4567, vx4567); in xnn_f32_velu_ukernel__neonfma_rr1_p6_x20() 146 const float32x4_t vy89AB = vbslq_f32(vm89AB, ve89AB, vx89AB); in xnn_f32_velu_ukernel__neonfma_rr1_p6_x20() 147 const float32x4_t vyCDEF = vbslq_f32(vmCDEF, veCDEF, vxCDEF); in xnn_f32_velu_ukernel__neonfma_rr1_p6_x20() 148 const float32x4_t vyGHIJ = vbslq_f32(vmGHIJ, veGHIJ, vxGHIJ); in xnn_f32_velu_ukernel__neonfma_rr1_p6_x20() 180 const float32x4_t vy = vbslq_f32(vm, ve, vx); in xnn_f32_velu_ukernel__neonfma_rr1_p6_x20() 208 const float32x4_t vy = vbslq_f32(vm, ve, vx); in xnn_f32_velu_ukernel__neonfma_rr1_p6_x20()
|