/external/XNNPACK/src/qs8-gavgpool/gen/ |
D | 7p7x-minmax-neon-c32-acc2.c | 82 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi4x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c32_acc2() 83 vacc0x89ABCDEF = vaddw_s8(vacc0x89ABCDEF, vi4x89ABCDEF); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c32_acc2() 84 vacc0xGHIJKLMN = vaddw_s8(vacc0xGHIJKLMN, vi4xGHIJKLMN); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c32_acc2() 85 vacc0xOPQRSTUV = vaddw_s8(vacc0xOPQRSTUV, vi4xOPQRSTUV); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c32_acc2() 86 vacc1x01234567 = vaddw_s8(vacc1x01234567, vi5x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c32_acc2() 87 vacc1x89ABCDEF = vaddw_s8(vacc1x89ABCDEF, vi5x89ABCDEF); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c32_acc2() 88 vacc1xGHIJKLMN = vaddw_s8(vacc1xGHIJKLMN, vi5xGHIJKLMN); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c32_acc2() 89 vacc1xOPQRSTUV = vaddw_s8(vacc1xOPQRSTUV, vi5xOPQRSTUV); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c32_acc2() 90 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi6x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c32_acc2() 91 vacc0x89ABCDEF = vaddw_s8(vacc0x89ABCDEF, vi6x89ABCDEF); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c32_acc2() [all …]
|
D | 7p7x-minmax-neon-c24-acc2.c | 73 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi4x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c24_acc2() 74 vacc0x89ABCDEF = vaddw_s8(vacc0x89ABCDEF, vi4x89ABCDEF); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c24_acc2() 75 vacc0xGHIJKLMN = vaddw_s8(vacc0xGHIJKLMN, vi4xGHIJKLMN); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c24_acc2() 76 vacc1x01234567 = vaddw_s8(vacc1x01234567, vi5x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c24_acc2() 77 vacc1x89ABCDEF = vaddw_s8(vacc1x89ABCDEF, vi5x89ABCDEF); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c24_acc2() 78 vacc1xGHIJKLMN = vaddw_s8(vacc1xGHIJKLMN, vi5xGHIJKLMN); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c24_acc2() 79 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi6x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c24_acc2() 80 vacc0x89ABCDEF = vaddw_s8(vacc0x89ABCDEF, vi6x89ABCDEF); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c24_acc2() 81 vacc0xGHIJKLMN = vaddw_s8(vacc0xGHIJKLMN, vi6xGHIJKLMN); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c24_acc2() 115 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi4x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c24_acc2() [all …]
|
D | 7p7x-minmax-neon-c16-acc2.c | 64 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi4x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c16_acc2() 65 vacc0x89ABCDEF = vaddw_s8(vacc0x89ABCDEF, vi4x89ABCDEF); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c16_acc2() 66 vacc1x01234567 = vaddw_s8(vacc1x01234567, vi5x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c16_acc2() 67 vacc1x89ABCDEF = vaddw_s8(vacc1x89ABCDEF, vi5x89ABCDEF); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c16_acc2() 68 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi6x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c16_acc2() 69 vacc0x89ABCDEF = vaddw_s8(vacc0x89ABCDEF, vi6x89ABCDEF); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c16_acc2() 118 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi4x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c16_acc2() 119 vacc0x89ABCDEF = vaddw_s8(vacc0x89ABCDEF, vi4x89ABCDEF); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c16_acc2() 120 vacc1x01234567 = vaddw_s8(vacc1x01234567, vi5x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c16_acc2() 121 vacc1x89ABCDEF = vaddw_s8(vacc1x89ABCDEF, vi5x89ABCDEF); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c16_acc2() [all …]
|
D | 7p7x-minmax-neon-c8-acc2.c | 55 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi4x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c8_acc2() 56 vacc1x01234567 = vaddw_s8(vacc1x01234567, vi5x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c8_acc2() 57 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi6x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c8_acc2() 92 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi4x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c8_acc2() 93 vacc1x01234567 = vaddw_s8(vacc1x01234567, vi5x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c8_acc2() 94 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi6x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c8_acc2() 157 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi4x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c8_acc2() 158 vacc1x01234567 = vaddw_s8(vacc1x01234567, vi5x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c8_acc2() 159 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi6x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c8_acc2() 237 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi4x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7p7x__neon_c8_acc2() [all …]
|
D | 7x-minmax-neon-c32-acc2.c | 105 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi4x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2() 106 vacc0x89ABCDEF = vaddw_s8(vacc0x89ABCDEF, vi4x89ABCDEF); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2() 107 vacc0xGHIJKLMN = vaddw_s8(vacc0xGHIJKLMN, vi4xGHIJKLMN); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2() 108 vacc0xOPQRSTUV = vaddw_s8(vacc0xOPQRSTUV, vi4xOPQRSTUV); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2() 109 vacc1x01234567 = vaddw_s8(vacc1x01234567, vi5x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2() 110 vacc1x89ABCDEF = vaddw_s8(vacc1x89ABCDEF, vi5x89ABCDEF); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2() 111 vacc1xGHIJKLMN = vaddw_s8(vacc1xGHIJKLMN, vi5xGHIJKLMN); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2() 112 vacc1xOPQRSTUV = vaddw_s8(vacc1xOPQRSTUV, vi5xOPQRSTUV); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2() 113 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi6x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2() 114 vacc0x89ABCDEF = vaddw_s8(vacc0x89ABCDEF, vi6x89ABCDEF); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2() [all …]
|
D | 7x-minmax-neon-c24-acc2.c | 96 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi4x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c24_acc2() 97 vacc0x89ABCDEF = vaddw_s8(vacc0x89ABCDEF, vi4x89ABCDEF); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c24_acc2() 98 vacc0xGHIJKLMN = vaddw_s8(vacc0xGHIJKLMN, vi4xGHIJKLMN); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c24_acc2() 99 vacc1x01234567 = vaddw_s8(vacc1x01234567, vi5x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c24_acc2() 100 vacc1x89ABCDEF = vaddw_s8(vacc1x89ABCDEF, vi5x89ABCDEF); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c24_acc2() 101 vacc1xGHIJKLMN = vaddw_s8(vacc1xGHIJKLMN, vi5xGHIJKLMN); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c24_acc2() 102 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi6x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c24_acc2() 103 vacc0x89ABCDEF = vaddw_s8(vacc0x89ABCDEF, vi6x89ABCDEF); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c24_acc2() 104 vacc0xGHIJKLMN = vaddw_s8(vacc0xGHIJKLMN, vi6xGHIJKLMN); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c24_acc2() 246 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi4x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c24_acc2() [all …]
|
D | 7x-minmax-neon-c16-acc2.c | 87 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi4x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c16_acc2() 88 vacc0x89ABCDEF = vaddw_s8(vacc0x89ABCDEF, vi4x89ABCDEF); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c16_acc2() 89 vacc1x01234567 = vaddw_s8(vacc1x01234567, vi5x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c16_acc2() 90 vacc1x89ABCDEF = vaddw_s8(vacc1x89ABCDEF, vi5x89ABCDEF); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c16_acc2() 91 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi6x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c16_acc2() 92 vacc0x89ABCDEF = vaddw_s8(vacc0x89ABCDEF, vi6x89ABCDEF); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c16_acc2() 198 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi4x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c16_acc2() 199 vacc1x01234567 = vaddw_s8(vacc1x01234567, vi5x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c16_acc2() 200 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi6x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c16_acc2()
|
D | 7x-minmax-neon-c8-acc2.c | 78 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi4x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c8_acc2() 79 vacc1x01234567 = vaddw_s8(vacc1x01234567, vi5x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c8_acc2() 80 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi6x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c8_acc2() 155 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi4x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c8_acc2() 156 vacc1x01234567 = vaddw_s8(vacc1x01234567, vi5x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c8_acc2() 157 vacc0x01234567 = vaddw_s8(vacc0x01234567, vi6x01234567); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c8_acc2()
|
/external/libhevc/common/arm/ |
D | ihevc_sao_edge_offset_class1.s | 190 …VADDW.S8 Q10,Q10,D12 @pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0],… 201 …VADDW.S8 Q4,Q4,D13 @pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[1],… 209 …VADDW.S8 Q13,Q13,D24 @II pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[… 212 …VADDW.S8 Q14,Q14,D25 @II pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[… 248 …VADDW.S8 Q13,Q13,D24 @pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0],… 254 …VADDW.S8 Q14,Q14,D25 @pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[1],… 327 …VADDW.S8 Q10,Q10,D12 @pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0],… 335 …VADDW.S8 Q13,Q13,D24 @II pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[… 366 …VADDW.S8 Q13,Q13,D24 @pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0],…
|
D | ihevc_sao_edge_offset_class1_chroma.s | 196 …VADDW.S8 Q10,Q10,D12 @pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0],… 208 …VADDW.S8 Q14,Q14,D13 @pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[1],… 219 …VADDW.S8 Q13,Q13,D24 @II pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[… 224 …VADDW.S8 Q14,Q14,D25 @II pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[… 265 …VADDW.S8 Q13,Q13,D24 @pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0],… 271 …VADDW.S8 Q14,Q14,D25 @pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[1],… 351 …VADDW.S8 Q10,Q10,D12 @pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0],… 365 …VADDW.S8 Q13,Q13,D24 @II pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[… 401 …VADDW.S8 Q13,Q13,D24 @pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0],…
|
D | ihevc_sao_edge_offset_class0.s | 226 …VADDW.S8 Q9,Q9,D16 @pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0],… 238 …VADDW.S8 Q7,Q7,D17 @pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[1],… 245 …VADDW.S8 Q0,Q0,D30 @II pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[… 254 …VADDW.S8 Q14,Q14,D31 @II pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[… 332 …VADDW.S8 Q14,Q14,D26 @pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0],…
|
D | ihevc_sao_edge_offset_class0_chroma.s | 236 …VADDW.S8 Q9,Q9,D16 @pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0],… 244 …VADDW.S8 Q6,Q6,D17 @pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[1],… 260 …VADDW.S8 Q14,Q14,D26 @II pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[… 271 …VADDW.S8 Q15,Q15,D27 @II pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[… 399 …VADDW.S8 Q9,Q9,D16 @pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0],… 424 …VADDW.S8 Q12,Q12,D26 @II pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[…
|
D | ihevc_sao_edge_offset_class2.s | 320 …VADDW.S8 Q10,Q10,D10 @I pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0… 328 …VADDW.S8 Q11,Q11,D11 @I pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[1… 409 …VADDW.S8 Q13,Q13,D24 @II pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[… 418 …VADDW.S8 Q10,Q10,D10 @III pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val… 427 …VADDW.S8 Q14,Q14,D25 @II pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[… 438 …VADDW.S8 Q9,Q9,D11 @III pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val… 494 …VADDW.S8 Q10,Q10,D10 @pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0],… 500 …VADDW.S8 Q6,Q6,D11 @pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[1],… 632 …VADDW.S8 Q14,Q14,D24 @pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0],… 638 …VADDW.S8 Q15,Q15,D25 @pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[1],… [all …]
|
D | ihevc_sao_edge_offset_class3.s | 335 …VADDW.S8 Q10,Q10,D10 @I pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0… 342 …VADDW.S8 Q11,Q11,D11 @I pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[1… 436 …VADDW.S8 Q14,Q14,D24 @II pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[… 447 …VADDW.S8 Q13,Q13,D25 @II pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[… 454 …VADDW.S8 Q10,Q10,D10 @III pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val… 462 …VADDW.S8 Q11,Q11,D11 @III pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val… 529 …VADDW.S8 Q10,Q10,D24 @pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0],… 535 …VADDW.S8 Q11,Q11,D25 @pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[1],… 673 …VADDW.S8 Q14,Q14,D24 @pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0],… 679 …VADDW.S8 Q15,Q15,D25 @pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[1],… [all …]
|
D | ihevc_sao_edge_offset_class3_chroma.s | 426 …VADDW.S8 Q10,Q10,D22 @I pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0… 432 …VADDW.S8 Q9,Q9,D23 @I pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[1… 550 …VADDW.S8 Q14,Q14,D24 @II pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[… 564 …VADDW.S8 Q13,Q13,D25 @II pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[… 572 …VADDW.S8 Q10,Q10,D22 @III pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val… 584 …VADDW.S8 Q9,Q9,D23 @III pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val… 659 …VADDW.S8 Q10,Q10,D22 @pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0],… 663 …VADDW.S8 Q9,Q9,D23 @pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[1],… 834 …VADDW.S8 Q14,Q14,D24 @pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0],… 840 …VADDW.S8 Q15,Q15,D25 @pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[1],… [all …]
|
D | ihevc_sao_edge_offset_class2_chroma.s | 437 …VADDW.S8 Q10,Q10,D22 @I pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0… 442 …VADDW.S8 Q9,Q9,D23 @I pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[1… 550 …VADDW.S8 Q14,Q14,D24 @II pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[… 567 …VADDW.S8 Q13,Q13,D25 @II pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[… 580 …VADDW.S8 Q10,Q10,D22 @III pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val… 586 …VADDW.S8 Q9,Q9,D23 @III pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val… 657 …VADDW.S8 Q10,Q10,D24 @pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0],… 661 …VADDW.S8 Q9,Q9,D25 @pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[1],… 810 …VADDW.S8 Q14,Q14,D24 @pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0],… 815 …VADDW.S8 Q13,Q13,D25 @pi2_tmp_cur_row.val[1] = vaddw_s8(pi2_tmp_cur_row.val[1],… [all …]
|
/external/libvpx/libvpx/vp8/common/arm/neon/ |
D | loopfiltersimplehorizontaledge_neon.c | 65 q2s16 = vaddw_s8(q2s16, vget_low_s8(q4s8)); in vp8_loop_filter_simple_horizontal_edge_neon() 66 q3s16 = vaddw_s8(q3s16, vget_high_s8(q4s8)); in vp8_loop_filter_simple_horizontal_edge_neon()
|
D | loopfiltersimpleverticaledge_neon.c | 230 q2s16 = vaddw_s8(q2s16, vget_low_s8(q14s8)); in vp8_loop_filter_simple_vertical_edge_neon() 231 q13s16 = vaddw_s8(q13s16, vget_high_s8(q14s8)); in vp8_loop_filter_simple_vertical_edge_neon()
|
D | vp8_loopfilter_neon.c | 94 q2s16 = vaddw_s8(q2s16, vget_low_s8(q1s8)); in vp8_loop_filter_neon() 95 q11s16 = vaddw_s8(q11s16, vget_high_s8(q1s8)); in vp8_loop_filter_neon()
|
D | mbloopfilter_neon.c | 93 q2s16 = vaddw_s8(q2s16, vget_low_s8(q1s8)); in vp8_mbloop_filter_neon() 94 q13s16 = vaddw_s8(q13s16, vget_high_s8(q1s8)); in vp8_mbloop_filter_neon()
|
/external/XNNPACK/src/qs8-gavgpool/ |
D | multipass-neon.c.in | 58 …vacc${M % ACCUMULATORS}x${ABC[C:C+8]} = vaddw_s8(vacc${M % ACCUMULATORS}x${ABC[C:C+8]}, vi${M}x${A… 87 …vacc${M % ACCUMULATORS}x${ABC[0:8]} = vaddw_s8(vacc${M % ACCUMULATORS}x${ABC[0:8]}, vi${M}x${ABC[0… 125 …vacc${M % ACCUMULATORS}x${ABC[C:C+8]} = vaddw_s8(vacc${M % ACCUMULATORS}x${ABC[C:C+8]}, vi${M}x${A… 158 …vacc${M % ACCUMULATORS}x${ABC[0:8]} = vaddw_s8(vacc${M % ACCUMULATORS}x${ABC[0:8]}, vi${M}x${ABC[0… 219 …vacc${M % ACCUMULATORS}x${ABC[C:C+8]} = vaddw_s8(vacc${M % ACCUMULATORS}x${ABC[C:C+8]}, vi${M}x${A… 324 …vacc${M % ACCUMULATORS}x${ABC[0:8]} = vaddw_s8(vacc${M % ACCUMULATORS}x${ABC[0:8]}, vi${M}x${ABC[0…
|
D | unipass-neon.c.in | 69 …vacc${M % ACCUMULATORS}x${ABC[C:C+8]} = vaddw_s8(vacc${M % ACCUMULATORS}x${ABC[C:C+8]}, vi${M}x${A… 168 …vacc${M % ACCUMULATORS}x${ABC[0:8]} = vaddw_s8(vacc${M % ACCUMULATORS}x${ABC[0:8]}, vi${M}x${ABC[0…
|
/external/webp/src/dsp/ |
D | lossless_neon.c | 172 const uint8x8_t res = vqmovun_s16(vaddw_s8(avg_16, diff_avg)); in ClampedAddSubtractHalf_NEON() 459 const uint8x8_t delta = vqmovun_s16(vaddw_s8(avg_16, diff_avg)); \
|
/external/libgav1/libgav1/src/dsp/arm/ |
D | loop_filter_neon.cc | 97 vget_low_s16(vaddw_s8(vcombine_s16(q0mp0_3, zero), hev_option)); in Filter4() 118 const int16x8_t p1q1_a3 = vaddw_s8(p1q1_l, a3_ma3); in Filter4() 125 const int16x8_t p0q0_a = vaddw_s8(p0q0_l, a2_ma1); in Filter4()
|
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/ |
D | neon_tensor_utils.cc | 1232 temp = vaddw_s8(temp, vget_high_s8(input_value)); in NeonMatrixScalarMultiplyAccumulate()
|