/external/libhevc/common/arm64/ |
D | ihevc_intra_pred_luma_dc.s | 237 uxtl v20.8h, v0.8b 240 add v20.8h, v20.8h , v24.8h //col 1::7 add 3dc+2 (prol) 243 sqshrun v2.8b, v20.8h,#2 //columns shx2 movn (prol) 263 bsl v20.8b, v3.8b , v16.8b //row 1 (prol) 268 st1 {v20.8b},[x2], x3 //store row 1 (prol) 280 bsl v20.8b, v3.8b , v16.8b //row 3 (prol) 285 st1 {v20.8b},[x2], x3 //store row 3 (prol) 296 bsl v20.8b, v3.8b , v16.8b //row 5 (prol) 301 st1 {v20.8b},[x2], x3 //store row 5 (prol) 314 bsl v20.8b, v3.8b , v16.8b //row 7 (prol) [all …]
|
D | ihevc_intra_pred_chroma_ver.s | 117 ld2 {v20.8b, v21.8b}, [x6],#16 //16 loads (col 0:15) 127 st2 {v20.8b, v21.8b}, [x2],#16 128 st2 {v20.8b, v21.8b}, [x5],#16 129 st2 {v20.8b, v21.8b}, [x8],#16 130 st2 {v20.8b, v21.8b}, [x10],#16 140 st2 {v20.8b, v21.8b}, [x2],#16 141 st2 {v20.8b, v21.8b}, [x5],#16 142 st2 {v20.8b, v21.8b}, [x8],#16 143 st2 {v20.8b, v21.8b}, [x10],#16 153 st2 {v20.8b, v21.8b}, [x2],#16 [all …]
|
D | ihevc_sao_edge_offset_class1.s | 154 …SUB v20.16b, v17.16b , v5.16b //sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 160 ADD v5.16b, v5.16b , v20.16b //edge_idx = vaddq_s8(edge_idx, sign_down) 163 NEG v16.16b, v20.16b //sign_up = vnegq_s8(sign_down) 177 …Uxtl v20.8h, v3.8b //pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(… 179 …SADDW v20.8h, v20.8h , v5.8b //pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0]… 181 …SMAX v20.8h, v20.8h , v2.8h //pi2_tmp_cur_row.val[0] = vmaxq_s16(pi2_tmp_cur_row.val[0… 183 …UMIN v20.8h, v20.8h , v4.8h //pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vminq_u16… 197 xtn v20.8b, v20.8h //vmovn_s16(pi2_tmp_cur_row.val[0]) 200 xtn2 v20.16b, v1.8h //vmovn_s16(pi2_tmp_cur_row.val[1]) 209 ST1 { v20.16b},[x10],x1 //vst1q_u8(pu1_src_cpy, pu1_cur_row) [all …]
|
D | ihevc_inter_pred_luma_vert_w16inp_w16out.s | 182 smull v20.4s,v2.4h,v23.4h //mul_res2 = smull_u8(src_tmp3, coeffabs_1)// 185 smlal v20.4s,v1.4h,v22.4h //mul_res2 = smlal_u8(mul_res2, src_tmp2, coeffabs_0)// 187 smlal v20.4s,v3.4h,v24.4h //mul_res2 = smlal_u8(mul_res2, src_tmp4, coeffabs_2)// 189 smlal v20.4s,v4.4h,v25.4h //mul_res2 = smlal_u8(mul_res2, src_tmp1, coeffabs_3)// 191 smlal v20.4s,v5.4h,v26.4h //mul_res2 = smlal_u8(mul_res2, src_tmp2, coeffabs_4)// 193 smlal v20.4s,v6.4h,v27.4h //mul_res2 = smlal_u8(mul_res2, src_tmp3, coeffabs_5)// 194 smlal v20.4s,v7.4h,v28.4h //mul_res2 = smlal_u8(mul_res2, src_tmp4, coeffabs_6)// 195 smlal v20.4s,v16.4h,v29.4h //mul_res2 = smlal_u8(mul_res2, src_tmp1, coeffabs_7)// 210 sub v20.4s, v20.4s, v30.4s 230 shrn v20.4h, v20.4s, #6 [all …]
|
D | ihevc_inter_pred_filters_luma_vert_w16inp.s | 171 smull v20.4s, v2.4h, v23.4h //mul_res2 = vmull_u8(src_tmp3, coeffabs_1)// 174 smlal v20.4s, v1.4h, v22.4h //mul_res2 = vmlal_u8(mul_res2, src_tmp2, coeffabs_0)// 176 smlal v20.4s, v3.4h, v24.4h //mul_res2 = vmlal_u8(mul_res2, src_tmp4, coeffabs_2)// 178 smlal v20.4s, v4.4h, v25.4h //mul_res2 = vmlal_u8(mul_res2, src_tmp1, coeffabs_3)// 180 smlal v20.4s, v5.4h, v26.4h //mul_res2 = vmlal_u8(mul_res2, src_tmp2, coeffabs_4)// 182 smlal v20.4s, v6.4h, v27.4h //mul_res2 = vmlal_u8(mul_res2, src_tmp3, coeffabs_5)// 183 smlal v20.4s, v7.4h, v28.4h //mul_res2 = vmlal_u8(mul_res2, src_tmp4, coeffabs_6)// 184 smlal v20.4s, v16.4h, v29.4h //mul_res2 = vmlal_u8(mul_res2, src_tmp1, coeffabs_7)// 199 sqshrn v20.4h, v20.4s,#6 218 sqrshrun v20.8b, v20.8h,#6 //sto_res = vqmovun_s16(sto_res_tmp)// [all …]
|
D | ihevc_sao_edge_offset_class1_chroma.s | 183 …SUB v20.16b, v19.16b , v5.16b //sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 189 ADD v5.16b, v5.16b , v20.16b //edge_idx = vaddq_s8(edge_idx, sign_down) 193 NEG v16.16b, v20.16b //sign_up = vnegq_s8(sign_down) 209 …Uxtl v20.8h, v3.8b //pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(… 215 …SADDW v20.8h, v20.8h , v5.8b //pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0]… 217 …SMAX v20.8h, v20.8h , v2.8h //pi2_tmp_cur_row.val[0] = vmaxq_s16(pi2_tmp_cur_row.val[0… 219 …UMIN v20.8h, v20.8h , v4.8h //pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vminq_u16… 243 xtn v20.8b, v20.8h //vmovn_s16(pi2_tmp_cur_row.val[0]) 246 xtn2 v20.16b, v28.8h //vmovn_s16(pi2_tmp_cur_row.val[1]) 257 ST1 { v20.16b},[x10],x1 //vst1q_u8(pu1_src_cpy, pu1_cur_row) [all …]
|
D | ihevc_inter_pred_filters_luma_vert.s | 182 umull v20.8h, v2.8b, v23.8b //mul_res2 = vmull_u8(src_tmp3, coeffabs_1)// 186 umlsl v20.8h, v1.8b, v22.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp2, coeffabs_0)// 190 umlsl v20.8h, v3.8b, v24.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp4, coeffabs_2)// 193 umlal v20.8h, v4.8b, v25.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp1, coeffabs_3)// 196 umlal v20.8h, v5.8b, v26.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp2, coeffabs_4)// 199 umlsl v20.8h, v6.8b, v27.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp3, coeffabs_5)// 202 umlal v20.8h, v7.8b, v28.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp4, coeffabs_6)// 206 umlsl v20.8h, v16.8b, v29.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp1, coeffabs_7)// 224 sqrshrun v20.8b, v20.8h,#6 //sto_res = vqmovun_s16(sto_res_tmp)// 243 st1 {v20.8b},[x14],x6 //vst1_u8(pu1_dst_tmp,sto_res)// [all …]
|
D | ihevc_inter_pred_filters_luma_vert_w16out.s | 140 umull v20.8h, v2.8b, v23.8b //mul_res2 = vmull_u8(src_tmp3, coeffabs_1)// 144 umlsl v20.8h, v1.8b, v22.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp2, coeffabs_0)// 147 umlsl v20.8h, v3.8b, v24.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp4, coeffabs_2)// 151 umlal v20.8h, v4.8b, v25.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp1, coeffabs_3)// 154 umlal v20.8h, v5.8b, v26.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp2, coeffabs_4)// 157 umlsl v20.8h, v6.8b, v27.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp3, coeffabs_5)// 159 umlal v20.8h, v7.8b, v28.8b //mul_res2 = vmlal_u8(mul_res2, src_tmp4, coeffabs_6)// 162 umlsl v20.8h, v16.8b, v29.8b //mul_res2 = vmlsl_u8(mul_res2, src_tmp1, coeffabs_7)// 197 st1 {v20.16b},[x14],x6 //vst1_u8(pu1_dst_tmp,sto_res)// 234 umull v20.8h, v2.8b, v23.8b //mul_res2 = vmull_u8(src_tmp3, coeffabs_1)// [all …]
|
D | ihevc_sao_edge_offset_class2.s | 329 …Uxtl v20.8h, v5.8b //I pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u… 331 …SADDW v20.8h, v20.8h , v3.8b //I pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[… 333 …SMAX v20.8h, v20.8h , v2.8h //I pi2_tmp_cur_row.val[0] = vmaxq_s16(pi2_tmp_cur_row.val… 336 …UMIN v20.8h, v20.8h , v4.8h //I pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vminq_u… 340 xtn v20.8b, v20.8h //I vmovn_s16(pi2_tmp_cur_row.val[0]) 347 xtn2 v20.16b, v22.8h //I vmovn_s16(pi2_tmp_cur_row.val[1]) 351 ST1 { v20.16b},[x0],x1 //I vst1q_u8(pu1_src_cpy, pu1_cur_row) 427 …Uxtl v20.8h, v16.8b //III pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl… 431 …SADDW v20.8h, v20.8h , v3.8b //III pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.va… 435 …SMAX v20.8h, v20.8h , v2.8h //III pi2_tmp_cur_row.val[0] = vmaxq_s16(pi2_tmp_cur_row.v… [all …]
|
/external/libmpeg2/common/armv8/ |
D | icv_variance_av8.s | 86 umull v20.8h, v0.8b, v0.8b 91 uaddl v21.4s, v20.4h, v22.4h 93 uaddl2 v20.4s, v20.8h, v22.8h 96 add v20.4s, v20.4s, v21.4s 98 add v20.4s, v20.4s, v22.4s 99 addp v20.4s, v20.4s, v20.4s 100 addp v20.2s, v20.2s, v20.2s 106 smov x1, v20.2s[0]
|
D | ideint_spatial_filter_av8.s | 70 movi v20.8h, #0 115 uabal v20.8h, v2.8b, v4.8b 127 addp v20.8h, v20.8h, v20.8h 131 uaddlp v20.2s, v20.4h 140 mul v20.2s, v20.2s, v31.2s 149 smov x7, v20.2s[0] 177 smov x7, v20.2s[1]
|
D | ideint_cac_av8.s | 142 cmhs v20.8h, v16.8h, v18.8h 143 and v20.16b, v16.16b, v20.16b 148 mov v21.d[0], v20.d[1] 149 add v20.4h, v20.4h, v21.4h 186 add v20.4h, v0.4h, v20.4h 188 uaddlp v20.2s, v20.4h 218 cmhi v0.2s, v20.2s, v21.2s
|
/external/libhevc/decoder/arm64/ |
D | ihevcd_fmt_conv_420sp_to_rgba8888.s | 203 sMULL v20.4s, v6.4h, v0.h[0] ////(V-128)*C1 FOR R 217 sqshrn v7.4h, v20.4s,#13 ////D10 = (V-128)*C1>>13 4 16-BIT VALUES 230 UADDW v20.8h, v5.8h , v31.8b ////Q10 - HAS Y + B 239 sqxtun v20.8b, v20.8h 251 ZIP1 v27.8b, v20.8b, v21.8b 252 ZIP2 v21.8b, v20.8b, v21.8b 253 mov v20.d[0], v27.d[0] 259 mov v20.d[1], v21.d[0] 266 ZIP1 v25.8h, v20.8h, v22.8h 267 ZIP2 v19.8h, v20.8h, v22.8h [all …]
|
/external/libavc/common/armv8/ |
D | ih264_inter_pred_luma_horz_qpel_vert_hpel_av8.s | 172 uaddl v20.8h, v4.8b, v6.8b 175 mla v18.8h, v20.8h , v28.8h 177 uaddl v20.8h, v1.8b, v11.8b 179 mla v20.8h, v24.8h , v28.8h 183 mls v20.8h, v26.8h , v30.8h 188 st1 {v20.4s}, [x9], #16 189 ext v24.16b, v18.16b , v20.16b , #4 190 ext v26.16b, v18.16b , v20.16b , #6 192 ext v22.16b, v18.16b , v20.16b , #10 194 ext v24.16b, v18.16b , v20.16b , #2 [all …]
|
D | ih264_inter_pred_luma_horz_hpel_vert_hpel_av8.s | 110 uaddl v20.8h, v4.8b, v6.8b 113 mla v18.8h, v20.8h , v28.8h 115 uaddl v20.8h, v1.8b, v11.8b 117 mla v20.8h, v24.8h , v28.8h 121 mls v20.8h, v26.8h , v30.8h 126 ext v24.16b, v18.16b , v20.16b , #4 127 ext v26.16b, v18.16b , v20.16b , #6 129 ext v23.16b, v18.16b , v20.16b , #10 131 ext v24.16b, v18.16b , v20.16b , #2 132 ext v26.16b, v18.16b , v20.16b , #8 [all …]
|
D | ih264_inter_pred_chroma_av8.s | 153 umull v20.8h, v0.8b, v28.8b 155 umlal v20.8h, v3.8b, v29.8b 157 umlal v20.8h, v5.8b, v30.8b 159 umlal v20.8h, v8.8b, v31.8b 160 sqrshrun v26.8b, v20.8h, #6 184 umull v20.8h, v10.8b, v28.8b 185 umlal v20.8h, v13.8b, v29.8b 186 umlal v20.8h, v0.8b, v30.8b 187 umlal v20.8h, v3.8b, v31.8b 188 sqrshrun v26.8b, v20.8h, #6 [all …]
|
D | ih264_intra_pred_luma_16x16_av8.s | 217 dup v20.16b, v0.b[5] 224 st1 {v20.16b}, [x1], x3 320 dup v20.16b, w15 327 dup v20.8h, w11 334 uqshl v0.8h, v0.8h, v20.8h 336 dup v20.16b, v0.b[0] 340 st1 { v20.16b}, [x1], x3 341 st1 { v20.16b}, [x1], x3 342 st1 { v20.16b}, [x1], x3 343 st1 { v20.16b}, [x1], x3 [all …]
|
D | ih264_inter_pred_filters_luma_vert_av8.s | 139 uaddl v20.8h, v1.8b, v11.8b // temp4 = src[0_8] + src[5_8] 141 mla v20.8h, v18.8h , v22.8h // temp4 += temp3 * 20 149 mls v20.8h, v26.8h , v24.8h // temp4 -= temp5 * 5 157 sqrshrun v31.8b, v20.8h, #5 // dst[0_8] = CLIP_U8((temp4 +16) >> 5) 163 uaddl v20.8h, v6.8b, v0.8b 170 mls v18.8h, v20.8h , v24.8h 176 uaddl v20.8h, v8.8b, v2.8b 185 mls v14.8h, v20.8h , v24.8h 192 uaddl v20.8h, v9.8b, v7.8b // temp4 = src[0_8] + src[5_8] 201 mla v20.8h, v18.8h , v22.8h // temp4 += temp3 * 20 [all …]
|
D | ih264_resi_trans_quant_av8.s | 164 sub v20.4h, v15.4h , v16.4h //x2 = x5-x6 167 shl v22.4h, v20.4h, #1 //u_shift(x2,1,shft) 174 add v25.4h, v23.4h , v20.4h //x6 = u_shift(x3,1,shft) + x2; 198 add v20.4s, v0.4s, v23.4s 205 sshl v20.4s, v20.4s, v24.4s //shift row 1 210 xtn v20.4h, v20.4s //narrow row 1 215 neg v24.8h, v20.8h //get negative 221 cmeq v0.4h, v20.4h, #0 226 bsl v4.8b, v20.8b, v24.8b //restore sign of row 1 and 2 387 sub v20.4h, v15.4h , v16.4h //x2 = x5-x6 [all …]
|
D | ih264_inter_pred_luma_vert_qpel_av8.s | 146 uaddl v20.8h, v1.8b, v11.8b // temp4 = src[0_8] + src[5_8] 148 mla v20.8h, v18.8h , v22.8h // temp4 += temp3 * 20 156 mls v20.8h, v26.8h , v24.8h // temp4 -= temp5 * 5 164 sqrshrun v31.8b, v20.8h, #5 // dst[0_8] = CLIP_U8((temp4 +16) >> 5) 165 ld1 {v20.2s, v21.2s}, [x7], x2 // Load for interpolation row 0 166 urhadd v30.16b, v20.16b , v30.16b // Interpolation to obtain qpel value 172 uaddl v20.8h, v6.8b, v0.8b 179 mls v18.8h, v20.8h , v24.8h 187 uaddl v20.8h, v8.8b, v2.8b 196 mls v14.8h, v20.8h , v24.8h [all …]
|
/external/libavc/encoder/armv8/ |
D | ih264e_half_pel_av8.s | 172 …sqrshrun v20.8b, v8.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column1… 179 st1 {v20.8b, v21.8b}, [x1], #16 ////Store dest row0 303 uaddl v20.8h, v2.8b, v17.8b //// a0 + a5 (column1,row0) 305 umlal v20.8h, v8.8b, v1.8b //// a0 + a5 + 20a2 (column1,row0) 306 umlal v20.8h, v11.8b, v1.8b //// a0 + a5 + 20a2 + 20a3 (column1,row0) 307 umlsl v20.8h, v5.8b, v31.8b //// a0 + a5 + 20a2 + 20a3 - 5a1 (column1,row0) 308 umlsl v20.8h, v14.8b, v31.8b //// a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 (column1,row0) 309 mov v21.d[0], v20.d[1] 316 ext v30.8b, v20.8b , v21.8b , #4 321 ext v29.8b, v20.8b , v21.8b , #6 [all …]
|
D | ih264e_evaluate_intra16x16_modes_av8.s | 133 dup v20.8h, w11 140 uqshl v0.8h, v0.8h, v20.8h 166 dup v20.8b, v9.b[15] ///HORIZONTAL VALUE ROW=0// 176 uabdl v26.8h, v0.8b, v20.8b 187 dup v20.8b, v9.b[14] ///HORIZONTAL VALUE ROW=1// 197 uabal v26.8h, v2.8b, v20.8b 204 dup v20.8b, v9.b[13] ///HORIZONTAL VALUE ROW=2// 213 uabal v26.8h, v4.8b, v20.8b 220 dup v20.8b, v9.b[12] ///HORIZONTAL VALUE ROW=3// 229 uabal v26.8h, v6.8b, v20.8b [all …]
|
/external/llvm/test/MC/SystemZ/ |
D | insn-good-z13.s | 25 #CHECK: vab %v18, %v3, %v20 # encoding: [0xe7,0x23,0x40,0x00,0x0a,0xf3] 31 vab %v18, %v3, %v20 37 #CHECK: vaccb %v18, %v3, %v20 # encoding: [0xe7,0x23,0x40,0x00,0x0a,0xf1] 43 vaccb %v18, %v3, %v20 63 #CHECK: vaccf %v18, %v3, %v20 # encoding: [0xe7,0x23,0x40,0x00,0x2a,0xf1] 69 vaccf %v18, %v3, %v20 75 #CHECK: vaccg %v18, %v3, %v20 # encoding: [0xe7,0x23,0x40,0x00,0x3a,0xf1] 81 vaccg %v18, %v3, %v20 87 #CHECK: vacch %v18, %v3, %v20 # encoding: [0xe7,0x23,0x40,0x00,0x1a,0xf1] 93 vacch %v18, %v3, %v20 [all …]
|
/external/llvm/test/MC/Disassembler/SystemZ/ |
D | insns-z13.txt | 17 #CHECK: vab %v18, %v3, %v20 26 #CHECK: vaccb %v18, %v3, %v20 35 #CHECK: vacccq %v3, %v20, %v5, %v22 44 #CHECK: vaccf %v18, %v3, %v20 53 #CHECK: vaccg %v18, %v3, %v20 62 #CHECK: vacch %v18, %v3, %v20 71 #CHECK: vaccq %v18, %v3, %v20 80 #CHECK: vacq %v3, %v20, %v5, %v22 89 #CHECK: vaf %v18, %v3, %v20 98 #CHECK: vag %v18, %v3, %v20 [all …]
|
/external/boringssl/linux-aarch64/crypto/modes/ |
D | ghashv8-armx64.S | 26 eor v20.16b,v3.16b,v16.16b //twisted H 27 st1 {v20.2d},[x0],#16 //store Htable[0] 30 ext v16.16b,v20.16b,v20.16b,#8 //Karatsuba pre-processing 31 pmull v0.1q,v20.1d,v20.1d 32 eor v16.16b,v16.16b,v20.16b 33 pmull2 v2.1q,v20.2d,v20.2d 64 ld1 {v20.2d,v21.2d},[x1] //load twisted H, ... 71 pmull v0.1q,v20.1d,v3.1d //H.lo·Xi.lo 73 pmull2 v2.1q,v20.2d,v3.2d //H.hi·Xi.hi 119 ld1 {v20.2d,v21.2d},[x1],#32 //load twisted H, ..., H^2 [all …]
|