/external/XNNPACK/src/qs8-gavgpool/gen/ |
D | 7p7x-minmax-fp32-neon-c24.c | 56 int16x8_t vsumGHIJKLMN = vaddl_s8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c24() local 63 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c24() 69 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c24() 75 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c24() 81 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c24() 84 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c24() 90 const int32x4_t vaccGHIJ = vaddw_s16(vinit_bias, vget_low_s16(vsumGHIJKLMN)); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c24() 91 const int32x4_t vaccKLMN = vaddw_s16(vinit_bias, vget_high_s16(vsumGHIJKLMN)); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c24() 151 int16x8_t vsumGHIJKLMN = vaddl_s8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c24() local 158 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c24() [all …]
|
D | 7p7x-minmax-fp32-neonv8-c24.c | 57 int16x8_t vsumGHIJKLMN = vaddl_s8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c24() local 64 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c24() 70 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c24() 76 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c24() 82 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c24() 85 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c24() 91 const int32x4_t vaccGHIJ = vaddw_s16(vinit_bias, vget_low_s16(vsumGHIJKLMN)); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c24() 92 const int32x4_t vaccKLMN = vaddw_s16(vinit_bias, vget_high_s16(vsumGHIJKLMN)); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c24() 152 int16x8_t vsumGHIJKLMN = vaddl_s8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c24() local 159 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c24() [all …]
|
D | 7p7x-minmax-rndnu-neon-c24.c | 56 int16x8_t vsumGHIJKLMN = vaddl_s8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c24() local 63 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c24() 69 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c24() 75 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c24() 81 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c24() 84 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c24() 90 const int32x4_t vaccGHIJ = vaddw_s16(vinit_bias, vget_low_s16(vsumGHIJKLMN)); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c24() 91 const int32x4_t vaccKLMN = vaddw_s16(vinit_bias, vget_high_s16(vsumGHIJKLMN)); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c24() 151 int16x8_t vsumGHIJKLMN = vaddl_s8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c24() local 158 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c24() [all …]
|
D | 7p7x-minmax-rndnu-neon-c32.c | 58 int16x8_t vsumGHIJKLMN = vaddl_s8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c32() local 67 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c32() 75 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c32() 83 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c32() 91 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c32() 96 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c32() 103 const int32x4_t vaccGHIJ = vaddw_s16(vinit_bias, vget_low_s16(vsumGHIJKLMN)); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c32() 104 const int32x4_t vaccKLMN = vaddw_s16(vinit_bias, vget_high_s16(vsumGHIJKLMN)); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c32() 170 int16x8_t vsumGHIJKLMN = vaddl_s8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c32() local 179 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c32() [all …]
|
D | 7p7x-minmax-fp32-neonv8-c32.c | 59 int16x8_t vsumGHIJKLMN = vaddl_s8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c32() local 68 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c32() 76 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c32() 84 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c32() 92 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c32() 97 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c32() 104 const int32x4_t vaccGHIJ = vaddw_s16(vinit_bias, vget_low_s16(vsumGHIJKLMN)); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c32() 105 const int32x4_t vaccKLMN = vaddw_s16(vinit_bias, vget_high_s16(vsumGHIJKLMN)); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c32() 171 int16x8_t vsumGHIJKLMN = vaddl_s8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c32() local 180 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c32() [all …]
|
D | 7p7x-minmax-fp32-neon-c32.c | 58 int16x8_t vsumGHIJKLMN = vaddl_s8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c32() local 67 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c32() 75 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c32() 83 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c32() 91 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c32() 96 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c32() 103 const int32x4_t vaccGHIJ = vaddw_s16(vinit_bias, vget_low_s16(vsumGHIJKLMN)); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c32() 104 const int32x4_t vaccKLMN = vaddw_s16(vinit_bias, vget_high_s16(vsumGHIJKLMN)); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c32() 170 int16x8_t vsumGHIJKLMN = vaddl_s8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c32() local 179 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c32() [all …]
|
D | 7x-minmax-fp32-neonv8-c24.c | 75 int16x8_t vsumGHIJKLMN = vaddl_s8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c24() local 82 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c24() 88 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c24() 94 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c24() 100 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c24() 103 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c24() 109 int32x4_t vaccGHIJ = vaddw_s16(vinit_bias, vget_low_s16(vsumGHIJKLMN)); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c24() 110 int32x4_t vaccKLMN = vaddw_s16(vinit_bias, vget_high_s16(vsumGHIJKLMN)); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c24()
|
D | 7x-minmax-fp32-neon-c24.c | 75 int16x8_t vsumGHIJKLMN = vaddl_s8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neon_c24() local 82 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neon_c24() 88 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neon_c24() 94 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neon_c24() 100 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neon_c24() 103 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neon_c24() 109 int32x4_t vaccGHIJ = vaddw_s16(vinit_bias, vget_low_s16(vsumGHIJKLMN)); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neon_c24() 110 int32x4_t vaccKLMN = vaddw_s16(vinit_bias, vget_high_s16(vsumGHIJKLMN)); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neon_c24()
|
D | 7x-minmax-rndnu-neon-c24.c | 76 int16x8_t vsumGHIJKLMN = vaddl_s8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7x__neon_c24() local 83 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7x__neon_c24() 89 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7x__neon_c24() 95 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7x__neon_c24() 101 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7x__neon_c24() 104 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7x__neon_c24() 110 int32x4_t vaccGHIJ = vaddw_s16(vinit_bias, vget_low_s16(vsumGHIJKLMN)); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7x__neon_c24() 111 int32x4_t vaccKLMN = vaddw_s16(vinit_bias, vget_high_s16(vsumGHIJKLMN)); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7x__neon_c24()
|
D | 7x-minmax-fp32-neonv8-c32.c | 77 int16x8_t vsumGHIJKLMN = vaddl_s8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c32() local 86 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c32() 94 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c32() 102 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c32() 110 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c32() 115 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c32() 122 int32x4_t vaccGHIJ = vaddw_s16(vinit_bias, vget_low_s16(vsumGHIJKLMN)); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c32() 123 int32x4_t vaccKLMN = vaddw_s16(vinit_bias, vget_high_s16(vsumGHIJKLMN)); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c32()
|
D | 7x-minmax-fp32-neon-c32.c | 77 int16x8_t vsumGHIJKLMN = vaddl_s8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neon_c32() local 86 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neon_c32() 94 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neon_c32() 102 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neon_c32() 110 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neon_c32() 115 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neon_c32() 122 int32x4_t vaccGHIJ = vaddw_s16(vinit_bias, vget_low_s16(vsumGHIJKLMN)); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neon_c32() 123 int32x4_t vaccKLMN = vaddw_s16(vinit_bias, vget_high_s16(vsumGHIJKLMN)); in xnn_qs8_gavgpool_minmax_fp32_ukernel_7x__neon_c32()
|
D | 7x-minmax-rndnu-neon-c32.c | 78 int16x8_t vsumGHIJKLMN = vaddl_s8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7x__neon_c32() local 87 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7x__neon_c32() 95 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7x__neon_c32() 103 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7x__neon_c32() 111 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7x__neon_c32() 116 vsumGHIJKLMN = vaddw_s8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7x__neon_c32() 123 int32x4_t vaccGHIJ = vaddw_s16(vinit_bias, vget_low_s16(vsumGHIJKLMN)); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7x__neon_c32() 124 int32x4_t vaccKLMN = vaddw_s16(vinit_bias, vget_high_s16(vsumGHIJKLMN)); in xnn_qs8_gavgpool_minmax_rndnu_ukernel_7x__neon_c32()
|
/external/XNNPACK/src/qu8-gavgpool/gen/ |
D | 7p7x-minmax-fp32-neonv8-c24.c | 57 uint16x8_t vsumGHIJKLMN = vaddl_u8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c24() local 64 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c24() 70 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c24() 76 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c24() 82 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c24() 85 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c24() 91 … = vreinterpretq_s32_u32(vaddw_u16(vreinterpretq_u32_s32(vinit_bias), vget_low_u16(vsumGHIJKLMN))); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c24() 92 …= vreinterpretq_s32_u32(vaddw_u16(vreinterpretq_u32_s32(vinit_bias), vget_high_u16(vsumGHIJKLMN))); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c24() 152 uint16x8_t vsumGHIJKLMN = vaddl_u8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c24() local 159 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c24() [all …]
|
D | 7p7x-minmax-fp32-neon-c24.c | 56 uint16x8_t vsumGHIJKLMN = vaddl_u8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c24() local 63 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c24() 69 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c24() 75 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c24() 81 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c24() 84 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c24() 90 … = vreinterpretq_s32_u32(vaddw_u16(vreinterpretq_u32_s32(vinit_bias), vget_low_u16(vsumGHIJKLMN))); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c24() 91 …= vreinterpretq_s32_u32(vaddw_u16(vreinterpretq_u32_s32(vinit_bias), vget_high_u16(vsumGHIJKLMN))); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c24() 151 uint16x8_t vsumGHIJKLMN = vaddl_u8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c24() local 158 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c24() [all …]
|
D | 7p7x-minmax-rndnu-neon-c24.c | 56 uint16x8_t vsumGHIJKLMN = vaddl_u8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c24() local 63 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c24() 69 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c24() 75 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c24() 81 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c24() 84 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c24() 90 … = vreinterpretq_s32_u32(vaddw_u16(vreinterpretq_u32_s32(vinit_bias), vget_low_u16(vsumGHIJKLMN))); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c24() 91 …= vreinterpretq_s32_u32(vaddw_u16(vreinterpretq_u32_s32(vinit_bias), vget_high_u16(vsumGHIJKLMN))); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c24() 151 uint16x8_t vsumGHIJKLMN = vaddl_u8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c24() local 158 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c24() [all …]
|
D | 7p7x-minmax-fp32-neonv8-c32.c | 59 uint16x8_t vsumGHIJKLMN = vaddl_u8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c32() local 68 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c32() 76 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c32() 84 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c32() 92 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c32() 97 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c32() 104 … = vreinterpretq_s32_u32(vaddw_u16(vreinterpretq_u32_s32(vinit_bias), vget_low_u16(vsumGHIJKLMN))); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c32() 105 …= vreinterpretq_s32_u32(vaddw_u16(vreinterpretq_u32_s32(vinit_bias), vget_high_u16(vsumGHIJKLMN))); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c32() 171 uint16x8_t vsumGHIJKLMN = vaddl_u8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c32() local 180 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neonv8_c32() [all …]
|
D | 7p7x-minmax-fp32-neon-c32.c | 58 uint16x8_t vsumGHIJKLMN = vaddl_u8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c32() local 67 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c32() 75 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c32() 83 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c32() 91 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c32() 96 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c32() 103 … = vreinterpretq_s32_u32(vaddw_u16(vreinterpretq_u32_s32(vinit_bias), vget_low_u16(vsumGHIJKLMN))); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c32() 104 …= vreinterpretq_s32_u32(vaddw_u16(vreinterpretq_u32_s32(vinit_bias), vget_high_u16(vsumGHIJKLMN))); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c32() 170 uint16x8_t vsumGHIJKLMN = vaddl_u8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c32() local 179 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7p7x__neon_c32() [all …]
|
D | 7p7x-minmax-rndnu-neon-c32.c | 58 uint16x8_t vsumGHIJKLMN = vaddl_u8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c32() local 67 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c32() 75 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c32() 83 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c32() 91 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c32() 96 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c32() 103 … = vreinterpretq_s32_u32(vaddw_u16(vreinterpretq_u32_s32(vinit_bias), vget_low_u16(vsumGHIJKLMN))); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c32() 104 …= vreinterpretq_s32_u32(vaddw_u16(vreinterpretq_u32_s32(vinit_bias), vget_high_u16(vsumGHIJKLMN))); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c32() 170 uint16x8_t vsumGHIJKLMN = vaddl_u8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c32() local 179 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7p7x__neon_c32() [all …]
|
D | 7x-minmax-rndnu-neon-c24.c | 76 uint16x8_t vsumGHIJKLMN = vaddl_u8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7x__neon_c24() local 83 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7x__neon_c24() 89 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7x__neon_c24() 95 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7x__neon_c24() 101 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7x__neon_c24() 104 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7x__neon_c24() 110 … = vreinterpretq_s32_u32(vaddw_u16(vreinterpretq_u32_s32(vinit_bias), vget_low_u16(vsumGHIJKLMN))); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7x__neon_c24() 111 …= vreinterpretq_s32_u32(vaddw_u16(vreinterpretq_u32_s32(vinit_bias), vget_high_u16(vsumGHIJKLMN))); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7x__neon_c24()
|
D | 7x-minmax-fp32-neon-c24.c | 75 uint16x8_t vsumGHIJKLMN = vaddl_u8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neon_c24() local 82 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neon_c24() 88 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neon_c24() 94 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neon_c24() 100 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neon_c24() 103 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neon_c24() 109 … = vreinterpretq_s32_u32(vaddw_u16(vreinterpretq_u32_s32(vinit_bias), vget_low_u16(vsumGHIJKLMN))); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neon_c24() 110 …= vreinterpretq_s32_u32(vaddw_u16(vreinterpretq_u32_s32(vinit_bias), vget_high_u16(vsumGHIJKLMN))); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neon_c24()
|
D | 7x-minmax-fp32-neonv8-c24.c | 75 uint16x8_t vsumGHIJKLMN = vaddl_u8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c24() local 82 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c24() 88 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c24() 94 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c24() 100 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c24() 103 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c24() 109 … = vreinterpretq_s32_u32(vaddw_u16(vreinterpretq_u32_s32(vinit_bias), vget_low_u16(vsumGHIJKLMN))); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c24() 110 …= vreinterpretq_s32_u32(vaddw_u16(vreinterpretq_u32_s32(vinit_bias), vget_high_u16(vsumGHIJKLMN))); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c24()
|
D | 7x-minmax-fp32-neon-c32.c | 77 uint16x8_t vsumGHIJKLMN = vaddl_u8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neon_c32() local 86 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neon_c32() 94 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neon_c32() 102 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neon_c32() 110 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neon_c32() 115 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neon_c32() 122 … = vreinterpretq_s32_u32(vaddw_u16(vreinterpretq_u32_s32(vinit_bias), vget_low_u16(vsumGHIJKLMN))); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neon_c32() 123 …= vreinterpretq_s32_u32(vaddw_u16(vreinterpretq_u32_s32(vinit_bias), vget_high_u16(vsumGHIJKLMN))); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neon_c32()
|
D | 7x-minmax-rndnu-neon-c32.c | 78 uint16x8_t vsumGHIJKLMN = vaddl_u8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7x__neon_c32() local 87 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7x__neon_c32() 95 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7x__neon_c32() 103 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7x__neon_c32() 111 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7x__neon_c32() 116 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7x__neon_c32() 123 … = vreinterpretq_s32_u32(vaddw_u16(vreinterpretq_u32_s32(vinit_bias), vget_low_u16(vsumGHIJKLMN))); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7x__neon_c32() 124 …= vreinterpretq_s32_u32(vaddw_u16(vreinterpretq_u32_s32(vinit_bias), vget_high_u16(vsumGHIJKLMN))); in xnn_qu8_gavgpool_minmax_rndnu_ukernel_7x__neon_c32()
|
D | 7x-minmax-fp32-neonv8-c32.c | 77 uint16x8_t vsumGHIJKLMN = vaddl_u8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c32() local 86 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c32() 94 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c32() 102 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c32() 110 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c32() 115 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c32() 122 … = vreinterpretq_s32_u32(vaddw_u16(vreinterpretq_u32_s32(vinit_bias), vget_low_u16(vsumGHIJKLMN))); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c32() 123 …= vreinterpretq_s32_u32(vaddw_u16(vreinterpretq_u32_s32(vinit_bias), vget_high_u16(vsumGHIJKLMN))); in xnn_qu8_gavgpool_minmax_fp32_ukernel_7x__neonv8_c32()
|
/external/XNNPACK/src/qu8-dwconv/gen/ |
D | up24x25-minmax-rndnu-neon-mul8.c | 209 uint16x8_t vsumGHIJKLMN = vaddl_u8(vi0xGHIJKLMN, vi1xGHIJKLMN); in xnn_qu8_dwconv_minmax_rndnu_ukernel_up24x25__neon_mul8() local 229 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi2xGHIJKLMN); in xnn_qu8_dwconv_minmax_rndnu_ukernel_up24x25__neon_mul8() 249 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi3xGHIJKLMN); in xnn_qu8_dwconv_minmax_rndnu_ukernel_up24x25__neon_mul8() 269 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi4xGHIJKLMN); in xnn_qu8_dwconv_minmax_rndnu_ukernel_up24x25__neon_mul8() 289 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi5xGHIJKLMN); in xnn_qu8_dwconv_minmax_rndnu_ukernel_up24x25__neon_mul8() 309 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi6xGHIJKLMN); in xnn_qu8_dwconv_minmax_rndnu_ukernel_up24x25__neon_mul8() 329 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi7xGHIJKLMN); in xnn_qu8_dwconv_minmax_rndnu_ukernel_up24x25__neon_mul8() 349 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi8xGHIJKLMN); in xnn_qu8_dwconv_minmax_rndnu_ukernel_up24x25__neon_mul8() 369 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi9xGHIJKLMN); in xnn_qu8_dwconv_minmax_rndnu_ukernel_up24x25__neon_mul8() 389 vsumGHIJKLMN = vaddw_u8(vsumGHIJKLMN, vi10xGHIJKLMN); in xnn_qu8_dwconv_minmax_rndnu_ukernel_up24x25__neon_mul8() [all …]
|