• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1///*****************************************************************************
2//*
3//* Copyright (C) 2012 Ittiam Systems Pvt Ltd, Bangalore
4//*
5//* Licensed under the Apache License, Version 2.0 (the "License");
6//* you may not use this file except in compliance with the License.
7//* You may obtain a copy of the License at:
8//*
9//* http://www.apache.org/licenses/LICENSE-2.0
10//*
11//* Unless required by applicable law or agreed to in writing, software
12//* distributed under the License is distributed on an "AS IS" BASIS,
13//* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
14//* See the License for the specific language governing permissions and
15//* limitations under the License.
16//*
17//*****************************************************************************/
18///**
19//*******************************************************************************
20//* //file
21//*  ihevc_inter_pred_chroma_horz_neon.s
22//*
23//* //brief
24//*  contains function definitions for inter prediction  interpolation.
25//* functions are coded using neon  intrinsics and can be compiled using
26
27//* rvct
28//*
29//* //author
30//*  yogeswaran rs / akshaya mukund
31//*
32//* //par list of functions:
33//*
34//*
35//* //remarks
36//*  none
37//*
38//*******************************************************************************
39//*/
40///**
41//*******************************************************************************
42//*
43//* //brief
44//*    chroma interprediction filter for horizontal input
45//*
46//* //par description:
47//*    applies a horizontal filter with coefficients pointed to  by 'pi1_coeff'
48//*    to the elements pointed by 'pu1_src' and  writes to the location pointed
49//*    by 'pu1_dst'  the output is downshifted by 6 and clipped to 8 bits
50//*    assumptions : the function is optimized considering the fact width is
51//*    multiple of 2,4 or 8. if width is 2, then height  should be multiple of 2.
52//*    width 4,8 is optimized further
53//*
54//* //param[in] pu1_src
55//*  uword8 pointer to the source
56//*
57//* //param[out] pu1_dst
58//*  uword8 pointer to the destination
59//*
60//* //param[in] src_strd
61//*  integer source stride
62//*
63//* //param[in] dst_strd
64//*  integer destination stride
65//*
66//* //param[in] pi1_coeff
67//*  word8 pointer to the filter coefficients
68//*
69//* //param[in] ht
70//*  integer height of the array
71//*
72//* //param[in] wd
73//*  integer width of the array
74//*
75//* //returns
76//*
77//* //remarks
78//*  none
79//*
80//*******************************************************************************
81//*/
82
83//void ihevc_inter_pred_chroma_horz(uword8 *pu1_src,
84//                                   uword8 *pu1_dst,
85//                                   word32 src_strd,
86//                                   word32 dst_strd,
87//                                   word8 *pi1_coeff,
88//                                   word32 ht,
89//                                   word32 wd)
90//**************variables vs registers*****************************************
91//x0 => *pu1_src
92//x1 => *pi2_dst
93//x2 =>  src_strd
94//x3 =>  dst_strd
95
96.text
97.align 4
98
99.include "ihevc_neon_macros.s"
100
101.globl ihevc_inter_pred_chroma_horz_av8
102
103.type ihevc_inter_pred_chroma_horz_av8, %function
104
105ihevc_inter_pred_chroma_horz_av8:
106
107    // stmfd sp!, {x4-x12, x14}                    //stack stores the values of the arguments
108
109    stp         d9,d10,[sp,#-16]!
110    stp         d11,d12,[sp,#-16]!
111    stp         d13,d14,[sp,#-16]!
112    stp         d8,d15,[sp,#-16]!           // Storing d15 using { sub sp,sp,#8; str d15,[sp] } is giving bus error.
113                                            // d8 is used as dummy register and stored along with d15 using stp. d8 is not used in the function.
114    stp         x19, x20,[sp,#-16]!
115
116    mov         x15,x4 // pi1_coeff
117    mov         x16,x5 // ht
118    mov         x17,x6 // wd
119
120
121    mov         x4,x15                      //loads pi1_coeff
122    mov         x7,x16                      //loads ht
123    mov         x10,x17                     //loads wd
124
125    ld1         {v0.8b},[x4]                //coeff = vld1_s8(pi1_coeff)
126    subs        x14,x7,#0                   //checks for ht == 0
127    abs         v2.8b, v0.8b                //vabs_s8(coeff)
128    mov         x11,#2
129    ble         end_loops
130
131    dup         v24.8b, v2.b[0]             //coeffabs_0 = vdup_lane_u8(coeffabs, 0)
132    sub         x12,x0,#2                   //pu1_src - 2
133    dup         v25.8b, v2.b[1]             //coeffabs_1 = vdup_lane_u8(coeffabs, 1)
134    add         x4,x12,x2                   //pu1_src_tmp2_8 = pu1_src + src_strd
135    dup         v26.8b, v2.b[2]             //coeffabs_2 = vdup_lane_u8(coeffabs, 2)
136
137    tst         x10,#3                      //checks wd for multiples
138    lsl         x5, x10, #1
139
140    dup         v27.8b, v2.b[3]             //coeffabs_3 = vdup_lane_u8(coeffabs, 3)
141
142    bne         outer_loop_4
143    cmp         x10,#12
144    beq         skip_16
145
146    cmp         x10,#8
147    bge         outer_loop_16
148skip_16:
149    tst         x7,#3
150
151    sub         x9,x0,#2
152    beq         outer_loop_ht_4             //jumps to else condition
153
154    b           outer_loop_8
155
156
157outer_loop_16:
158    mov         x10,x5                      //2wd
159    mul         x14, x14 , x10
160
161    sub         x20,x3,#16
162    neg         x6, x20
163
164    add         x4,x12,x2
165    mov         x9,#10
166    and         x0, x12, #31
167    sub         x20,x5,x3,lsl #1
168    neg         x8, x20
169    add         x20,x12, x2 , lsl #1
170    prfm        PLDL1KEEP,[x20]
171
172
173
174    add         x19,x12,#8
175    ld1         { v0.2s},[x12],x11          //vector load pu1_src
176    ld1         { v1.2s},[x19],x11          //vector load pu1_src
177    add         x20,x4, x2 , lsl #1
178    prfm        PLDL1KEEP,[x20]
179
180    ld1         { v2.2s},[x12],x11          //vector load pu1_src
181    ld1         { v3.2s},[x19],x11          //vector load pu1_src
182
183    ld1         { v4.2s},[x12],x11          //vector load pu1_src
184    ld1         { v5.2s},[x19],x11          //vector load pu1_src
185
186    ld1         { v6.2s},[x12],x9           //vector load pu1_src
187    ld1         { v7.2s},[x19],x9           //vector load pu1_src
188
189
190    add         x19,x4,#8
191    umull       v30.8h, v2.8b, v25.8b       //mul_res = vmull_u8(src[0_3], coeffabs_3)//
192    ld1         { v29.2s},[x4],x11          //vector load pu1_src
193    ld1         { v9.2s},[x19],x11          //vector load pu1_src
194
195    umlsl       v30.8h, v0.8b, v24.8b       //mul_res = vmlsl_u8(src[0_2], coeffabs_2)//
196
197    ld1         { v10.2s},[x4],x11          //vector load pu1_src
198    ld1         { v11.2s},[x19],x11         //vector load pu1_src
199
200    umlal       v30.8h, v4.8b, v26.8b       //mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
201
202    ld1         { v12.2s},[x4],x11          //vector load pu1_src
203    ld1         { v13.2s},[x19],x11         //vector load pu1_src
204
205    umlsl       v30.8h, v6.8b, v27.8b       //mul_res = vmlal_u8(src[0_1], coeffabs_1)//
206
207    ld1         { v14.2s},[x4],x9           //vector load pu1_src
208    ld1         { v15.2s},[x19],x9          //vector load pu1_src
209
210    umull       v28.8h, v3.8b, v25.8b
211
212    umlsl       v28.8h, v1.8b, v24.8b
213
214
215    umlal       v28.8h, v5.8b, v26.8b
216
217    umlsl       v28.8h, v7.8b, v27.8b
218
219
220    cmp         x14,#32
221    beq         epilog_end
222    sub         x14, x14,#64
223
224inner_loop_16:
225
226
227
228
229//     bgt            l_2
230
231//    add x20,x12, x2 , lsl #1
232    prfm        PLDL1KEEP,[x20]
233//    add x20,x4, x2 , lsl #1
234    prfm        PLDL1KEEP,[x20]
235
236
237
238    subs        x10,x10,#16
239
240    umull       v22.8h, v10.8b, v25.8b      //mul_res = vmull_u8(src[0_3], coeffabs_3)//
241
242
243    add         x20,x12,x8
244    csel        x12, x20, x12,eq
245    add         x20,x12,x2
246    csel        x4, x20, x4,eq
247    umlsl       v22.8h, v29.8b, v24.8b      //mul_res = vmlsl_u8(src[0_2], coeffabs_2)//
248
249
250
251    add         x20,x12, x2 , lsl #2
252    prfm        PLDL1KEEP,[x20]
253    sqrshrun    v30.8b, v30.8h,#6
254
255    add         x19,x12,#8
256    ld1         { v0.2s},[x12],x11          //vector load pu1_src
257    ld1         { v1.2s},[x19],x11          //vector load pu1_src
258
259    sqrshrun    v31.8b, v28.8h,#6
260
261
262
263    ld1         { v2.2s},[x12],x11          //vector load pu1_src
264    ld1         { v3.2s},[x19],x11          //vector load pu1_src
265    umlal       v22.8h, v12.8b, v26.8b      //mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
266
267
268
269
270    ld1         { v4.2s},[x12],x11          //vector load pu1_src
271    ld1         { v5.2s},[x19],x11          //vector load pu1_src
272    umlsl       v22.8h, v14.8b, v27.8b      //mul_res = vmlal_u8(src[0_1], coeffabs_1)//
273
274
275    ld1         { v6.2s},[x12],x9           //vector load pu1_src
276    ld1         { v7.2s},[x19],x9           //vector load pu1_src
277    umull       v20.8h, v11.8b, v25.8b      //mul_res = vmull_u8(src[0_3], coeffabs_3)//
278
279    add         x20,x4, x2 , lsl #2
280    prfm        PLDL1KEEP,[x20]
281    umlsl       v20.8h, v9.8b, v24.8b       //mul_res = vmlsl_u8(src[0_2], coeffabs_2)//
282
283    //mov       v30.s[1],v31.s[0]
284    add         x13,x1,#8
285    st1         { v30.4h}, [x1],x3
286    st1         { v31.4h}, [x13],x3
287    umlal       v20.8h, v13.8b, v26.8b      //mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
288
289    add         x19,x4,#8
290    ld1         { v29.2s},[x4],x11          //vector load pu1_src
291    ld1         { v9.2s},[x19],x11          //vector load pu1_src
292    umlsl       v20.8h, v15.8b, v27.8b      //mul_res = vmlal_u8(src[0_1], coeffabs_1)//
293
294
295    ld1         { v10.2s},[x4],x11          //vector load pu1_src
296    ld1         { v11.2s},[x19],x11         //vector load pu1_src
297    umull       v30.8h, v2.8b, v25.8b       //mul_res = vmull_u8(src[0_3], coeffabs_3)//
298
299    ld1         { v12.2s},[x4],x11          //vector load pu1_src
300    ld1         { v13.2s},[x19],x11         //vector load pu1_src
301    umlsl       v30.8h, v0.8b, v24.8b       //mul_res = vmlsl_u8(src[0_2], coeffabs_2)//
302
303    ld1         { v14.2s},[x4],x9           //vector load pu1_src
304    ld1         { v15.2s},[x19],x11         //vector load pu1_src
305    umlal       v30.8h, v4.8b, v26.8b       //mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
306
307    cmp         x10,#0
308    sqrshrun    v22.8b, v22.8h,#6
309    sqrshrun    v23.8b, v20.8h,#6
310
311
312
313    umlsl       v30.8h, v6.8b, v27.8b       //mul_res = vmlal_u8(src[0_1], coeffabs_1)//
314
315    csel        x10, x5, x10,eq             //2wd
316    umull       v28.8h, v3.8b, v25.8b
317
318
319    //add       x13,x1,#8
320    //mov       v22.s[1],v23.s[0]
321    st1         { v22.4h},[x1],x6           //store the result pu1_dst
322    st1         { v23.4h},[x13],x6          //store the result pu1_dst
323    umlsl       v28.8h, v1.8b, v24.8b
324
325
326    add         x20,x1,x8
327    csel        x1, x20, x1,eq
328    umlal       v28.8h, v5.8b, v26.8b
329
330    subs        x14,x14,#32                 //decrement the ht loop
331    umlsl       v28.8h, v7.8b, v27.8b
332
333//      mov            x0, x7
334
335    bgt         inner_loop_16
336
337
338
339    add         x14,x14,#64
340    cmp         x14,#32
341    beq         epilog_end
342
343epilog:
344    sqrshrun    v30.8b, v30.8h,#6
345    sqrshrun    v31.8b, v28.8h,#6
346
347
348
349    add         x13,x1,#8
350    //mov       v30.s[1],v31.s[0]
351    st1         { v30.4h}, [x1],x3
352    st1         { v31.4h}, [x13],x3
353
354    umull       v22.8h, v10.8b, v25.8b      //mul_res = vmull_u8(src[0_3], coeffabs_3)//
355
356
357
358
359    umlsl       v22.8h, v29.8b, v24.8b      //mul_res = vmlsl_u8(src[0_2], coeffabs_2)//
360    subs        x10,x10,#16                 //decrement the wd loop
361    umlal       v22.8h, v12.8b, v26.8b      //mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
362    add         x20,x12,x8
363    csel        x12, x20, x12,eq
364    umlsl       v22.8h, v14.8b, v27.8b      //mul_res = vmlal_u8(src[0_1], coeffabs_1)//
365    csel        x10, x5, x10,eq             //2wd
366
367
368    add         x20,x12,x2
369    csel        x4, x20, x4,eq
370    umull       v20.8h, v11.8b, v25.8b      //mul_res = vmull_u8(src[0_3], coeffabs_3)//
371
372    add         x19,x12,#8
373    ld1         { v0.2s},[x12],x11          //vector load pu1_src
374    ld1         { v1.2s},[x19],x11          //vector load pu1_src
375
376    umlsl       v20.8h, v9.8b, v24.8b       //mul_res = vmlsl_u8(src[0_2], coeffabs_2)//
377    ld1         { v2.2s},[x12],x11          //vector load pu1_src
378    ld1         { v3.2s},[x19],x11          //vector load pu1_src
379    umlal       v20.8h, v13.8b, v26.8b      //mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
380
381    ld1         { v4.2s},[x12],x11          //vector load pu1_src
382    ld1         { v5.2s},[x19],x11          //vector load pu1_src
383
384    umlsl       v20.8h, v15.8b, v27.8b      //mul_res = vmlal_u8(src[0_1], coeffabs_1)//
385    ld1         { v6.2s},[x12],x9           //vector load pu1_src
386    ld1         { v7.2s},[x19],x9           //vector load pu1_src
387    umull       v30.8h, v2.8b, v25.8b       //mul_res = vmull_u8(src[0_3], coeffabs_3)//
388
389
390    add         x19,x4,#8
391    ld1         { v29.2s},[x4],x11          //vector load pu1_src
392    ld1         { v9.2s},[x19],x11          //vector load pu1_src
393    umlsl       v30.8h, v0.8b, v24.8b       //mul_res = vmlsl_u8(src[0_2], coeffabs_2)//
394    ld1         { v10.2s},[x4],x11          //vector load pu1_src
395    ld1         { v11.2s},[x19],x11         //vector load pu1_src
396    umlal       v30.8h, v4.8b, v26.8b       //mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
397
398    umlsl       v30.8h, v6.8b, v27.8b       //mul_res = vmlal_u8(src[0_1], coeffabs_1)//
399
400    ld1         { v12.2s},[x4],x11          //vector load pu1_src
401    ld1         { v13.2s},[x19],x11         //vector load pu1_src
402    umull       v28.8h, v3.8b, v25.8b
403    ld1         { v14.2s},[x4],x9           //vector load pu1_src
404    ld1         { v15.2s},[x19],x9          //vector load pu1_src
405    umlsl       v28.8h, v1.8b, v24.8b
406    sqrshrun    v22.8b, v22.8h,#6
407    sqrshrun    v23.8b, v20.8h,#6
408
409    //mov       v22.s[1],v23.s[0]
410    st1         { v22.4h},[x1],x6           //store the result pu1_dst
411    st1         { v23.4h},[x13],x6          //store the result pu1_dst
412    umlal       v28.8h, v5.8b, v26.8b
413
414    umlsl       v28.8h, v7.8b, v27.8b
415    add         x20,x1,x8
416    csel        x1, x20, x1,eq
417
418
419
420epilog_end:
421    sqrshrun    v30.8b, v30.8h,#6
422    sqrshrun    v31.8b, v28.8h,#6
423
424
425    umull       v22.8h, v10.8b, v25.8b      //mul_res = vmull_u8(src[0_3], coeffabs_3)//
426    umlsl       v22.8h, v29.8b, v24.8b      //mul_res = vmlsl_u8(src[0_2], coeffabs_2)//
427    umlal       v22.8h, v12.8b, v26.8b      //mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
428    umlsl       v22.8h, v14.8b, v27.8b      //mul_res = vmlal_u8(src[0_1], coeffabs_1)//
429
430
431    umull       v20.8h, v11.8b, v25.8b      //mul_res = vmull_u8(src[0_3], coeffabs_3)//
432    umlsl       v20.8h, v9.8b, v24.8b       //mul_res = vmlsl_u8(src[0_2], coeffabs_2)//
433    umlal       v20.8h, v13.8b, v26.8b      //mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
434    umlsl       v20.8h, v15.8b, v27.8b      //mul_res = vmlal_u8(src[0_1], coeffabs_1)//
435    sqrshrun    v22.8b, v22.8h,#6
436    sqrshrun    v23.8b, v20.8h,#6
437
438    add         x13,x1,#8
439
440    //mov       v30.s[1],v31.s[0]
441    st1         { v30.4h}, [x1],x3
442    st1         { v31.4h}, [x13],x3
443
444    //mov       v22.s[1],v23.s[0]
445    st1         { v22.4h},[x1]              //store the result pu1_dst
446    st1         { v23.4h},[x13]             //store the result pu1_dst
447
448
449
450    b           end_loops
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470outer_loop_8:
471
472
473    add         x6,x1,x3                    //pu1_dst + dst_strd
474    mov         x7,x5
475    add         x4,x12,x2                   //pu1_src + src_strd
476
477
478inner_loop_8:
479    //ld1 {v0.2s, v1.2s},[x12],x11                //vector load pu1_src
480    ld1         {v0.2s},[x12],x11           //vector load pu1_src
481    ld1         {v1.2s},[x12],x11           //vector load pu1_src
482    ld1         {v2.2s},[x12],x11           //vector load pu1_src
483    ld1         {v3.2s},[x12],x11           //vector load pu1_src
484
485    //vext.u8    d2,d0,d1,#2                        //vector extract of src[0_2]
486    umull       v29.8h, v1.8b, v25.8b       //mul_res = vmull_u8(src[0_3], coeffabs_3)//
487    umlsl       v29.8h, v0.8b, v24.8b       //mul_res = vmlsl_u8(src[0_2], coeffabs_2)//
488    //vext.u8    d4,d0,d1,#4                        //vector extract of src[0_4]
489    //vext.u8    d6,d0,d1,#6                        //vector extract of src[0_6]
490    umlal       v29.8h, v2.8b, v26.8b       //mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
491    umlsl       v29.8h, v3.8b, v27.8b       //mul_res = vmlal_u8(src[0_1], coeffabs_1)//
492
493    ld1         {v4.2s},[x4],x11            //vector load pu1_src
494    ld1         {v5.2s},[x4],x11            //vector load pu1_src
495    ld1         {v6.2s},[x4],x11            //vector load pu1_src
496    ld1         {v7.2s},[x4],x11            //vector load pu1_src
497    //ld1 {v12.2s, v13.2s},[x4],x11                //vector load pu1_src + src_strd
498    //vext.u8    d14,d12,d13,#2                    //vector extract of src[0_2]
499    umull       v10.8h, v5.8b, v25.8b       //mul_res = vmull_u8(src[0_3], coeffabs_3)//
500    umlsl       v10.8h, v4.8b, v24.8b       //mul_res = vmlsl_u8(src[0_2], coeffabs_2)//
501    //vext.u8    d16,d12,d13,#4                    //vector extract of src[0_4]
502    //vext.u8    d18,d12,d13,#6                    //vector extract of src[0_6]
503    sqrshrun    v29.8b, v29.8h,#6           //right shift and saturating narrow result 1
504    umlal       v10.8h, v6.8b, v26.8b       //mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
505    umlsl       v10.8h, v7.8b, v27.8b       //mul_res = vmlal_u8(src[0_1], coeffabs_1)//
506
507    st1         {v29.8b},[x1],#8            //store the result pu1_dst
508
509    sqrshrun    v10.8b, v10.8h,#6           //right shift and saturating narrow result 2
510    subs        x7,x7,#8                    //decrement the wd loop
511    st1         {v10.8b},[x6],#8            //store the result pu1_dst
512    bgt         inner_loop_8
513
514    sub         x12,x12,x5
515    subs        x14,x14,#2                  //decrement the ht loop
516    sub         x1,x1,x5
517    add         x12,x12,x2,lsl #1
518    add         x1,x1,x3,lsl #1
519    bgt         outer_loop_8
520    b           end_loops
521
522//height if 4 comes
523outer_loop_ht_4:
524
525    mov         x7,x5
526
527prologue_ht_4:
528
529inner_loop_ht_4:
530
531    mov         x12,x9
532    mov         x4,x1
533
534    sub         x8, x2, #6
535
536    ld1         {v0.2s},[x12],x11           //(1)vector load pu1_src
537    ld1         {v1.2s},[x12],x11           //(1)vector load pu1_src
538    ld1         {v2.2s},[x12],x11           //(1)vector load pu1_src
539    //ld1 {v3.2s},[x12],x2                //(1)vector load pu1_src
540    ld1         {v3.2s},[x12],x8            //(1)vector load pu1_src
541
542    //sub        x12, x12, #6                //(1)
543
544    ld1         {v4.2s},[x12],x11           //(2)vector load pu1_src
545    ld1         {v5.2s},[x12],x11           //(2)vector load pu1_src
546    ld1         {v6.2s},[x12],x11           //(2)vector load pu1_src
547    //ld1 {v7.2s},[x12],x2                //(2)vector load pu1_src
548    ld1         {v7.2s},[x12],x8            //(2)vector load pu1_src
549
550    //sub        x12, x12, #6                //(2)
551
552    ld1         {v14.2s},[x12],x11          //(3)vector load pu1_src
553    umull       v29.8h, v1.8b, v25.8b       //(1)mul_res = vmull_u8(src[0_3], coeffabs_3)//
554
555    ld1         {v15.2s},[x12],x11          //(3)vector load pu1_src
556    umlsl       v29.8h, v0.8b, v24.8b       //(1)mul_res = vmlsl_u8(src[0_2], coeffabs_2)//
557
558    ld1         {v16.2s},[x12],x11          //(3)vector load pu1_src
559    umlal       v29.8h, v2.8b, v26.8b       //(1)mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
560
561    //ld1 {v17.2s},[x12],x2                //(3)vector load pu1_src
562    ld1         {v17.2s},[x12],x8           //(3)vector load pu1_src
563    umlsl       v29.8h, v3.8b, v27.8b       //(1)mul_res = vmlal_u8(src[0_1], coeffabs_1)//
564
565    //sub        x12, x12, #6                //(3)
566    umull       v10.8h, v5.8b, v25.8b       //(2)mul_res = vmull_u8(src[0_3], coeffabs_3)//
567
568    ld1         {v18.2s},[x12],x11          //(4)vector load pu1_src
569    umlsl       v10.8h, v4.8b, v24.8b       //(2)mul_res = vmlsl_u8(src[0_2], coeffabs_2)//
570
571    ld1         {v19.2s},[x12],x11          //(4)vector load pu1_src
572    umlal       v10.8h, v6.8b, v26.8b       //(2)mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
573
574    ld1         {v20.2s},[x12],x11          //(4)vector load pu1_src
575    umlsl       v10.8h, v7.8b, v27.8b       //(2)mul_res = vmlal_u8(src[0_1], coeffabs_1)//
576
577    ld1         {v21.2s},[x12],x2           //(4)vector load pu1_src
578    sqrshrun    v29.8b, v29.8h,#6           //(1)right shift and saturating narrow result 1
579
580    add         x9,x9,#8                    //(core loop)
581
582    subs        x7,x7,#8                    //(prologue)decrement the wd loop
583    beq         epilogue
584
585core_loop:
586    mov         x12,x9
587
588    ld1         {v0.2s},[x12],x11           //(1_1)vector load pu1_src
589    umull       v12.8h, v15.8b, v25.8b      //(3)mul_res = vmull_u8(src[0_3], coeffabs_3)//
590
591    ld1         {v1.2s},[x12],x11           //(1_1)vector load pu1_src
592    umlsl       v12.8h, v14.8b, v24.8b      //(3)mul_res = vmlsl_u8(src[0_2], coeffabs_2)//
593
594    ld1         {v2.2s},[x12],x11           //(1_1)vector load pu1_src
595    umlal       v12.8h, v16.8b, v26.8b      //(3)mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
596
597    //ld1 {v3.2s},[x12],x2                //(1_1)vector load pu1_src
598    ld1         {v3.2s},[x12],x8            //(1_1)vector load pu1_src
599    umlsl       v12.8h, v17.8b, v27.8b      //(3)mul_res = vmlal_u8(src[0_1], coeffabs_1)//
600
601    //sub        x12, x12, #6                //(1_1)
602
603    st1         {v29.8b},[x4],x3            //(1)store the result pu1_dst
604    sqrshrun    v10.8b, v10.8h,#6           //(2)right shift and saturating narrow result 2
605
606    ld1         {v4.2s},[x12],x11           //(2_1)vector load pu1_src
607    umull       v22.8h, v19.8b, v25.8b      //(4)mul_res = vmull_u8(src[0_3], coeffabs_3)//
608
609    ld1         {v5.2s},[x12],x11           //(2_1)vector load pu1_src
610    umlsl       v22.8h, v18.8b, v24.8b      //(4)mul_res = vmlsl_u8(src[0_2], coeffabs_2)//
611
612    ld1         {v6.2s},[x12],x11           //(2_1)vector load pu1_src
613    umlal       v22.8h, v20.8b, v26.8b      //(4)mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
614
615    //ld1 {v7.2s},[x12],x2                //(2_1)vector load pu1_src
616    ld1         {v7.2s},[x12],x8            //(2_1)vector load pu1_src
617    umlsl       v22.8h, v21.8b, v27.8b      //(4)mul_res = vmlal_u8(src[0_1], coeffabs_1)//
618
619    //sub        x12, x12, #6                //(2_1)
620
621    st1         {v10.8b},[x4],x3            //(2)store the result pu1_dst
622    sqrshrun    v12.8b, v12.8h,#6           //(3)right shift and saturating narrow result 1
623
624    ld1         {v14.2s},[x12],x11          //(3_1)vector load pu1_src
625    umull       v29.8h, v1.8b, v25.8b       //(1_1)mul_res = vmull_u8(src[0_3], coeffabs_3)//
626
627    ld1         {v15.2s},[x12],x11          //(3_1)vector load pu1_src
628    umlsl       v29.8h, v0.8b, v24.8b       //(1_1)mul_res = vmlsl_u8(src[0_2], coeffabs_2)//
629
630    ld1         {v16.2s},[x12],x11          //(3_1)vector load pu1_src
631    umlal       v29.8h, v2.8b, v26.8b       //(1_1)mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
632
633    //ld1 {v17.2s},[x12],x2                //(3_1)vector load pu1_src
634    ld1         {v17.2s},[x12],x8           //(3_1)vector load pu1_src
635    umlsl       v29.8h, v3.8b, v27.8b       //(1_1)mul_res = vmlal_u8(src[0_1], coeffabs_1)//
636
637    //sub        x12, x12, #6                //(3_1)
638
639    st1         {v12.8b},[x4],x3            //(3)store the result pu1_dst
640    sqrshrun    v22.8b, v22.8h,#6           //(4)right shift and saturating narrow result 2
641
642    add         x9,x9,#8                    //(core loop)
643
644    umull       v10.8h, v5.8b, v25.8b       //(2_1)mul_res = vmull_u8(src[0_3], coeffabs_3)//
645    ld1         {v18.2s},[x12],x11          //(4_1)vector load pu1_src
646
647    ld1         {v19.2s},[x12],x11          //(4_1)vector load pu1_src
648    umlsl       v10.8h, v4.8b, v24.8b       //(2_1)mul_res = vmlsl_u8(src[0_2], coeffabs_2)//
649
650    ld1         {v20.2s},[x12],x11          //(4_1)vector load pu1_src
651    umlal       v10.8h, v6.8b, v26.8b       //(2_1)mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
652
653    ld1         {v21.2s},[x12],x2           //(4_1)vector load pu1_src
654    umlsl       v10.8h, v7.8b, v27.8b       //(2_1)mul_res = vmlal_u8(src[0_1], coeffabs_1)//
655
656    add         x1,x1,#8                    //(core loop)
657
658    subs        x7,x7,#8                    //(core loop)
659
660    st1         {v22.8b},[x4], x3           //(4)store the result pu1_dst
661    sqrshrun    v29.8b, v29.8h,#6           //(1_1)right shift and saturating narrow result 1
662
663    mov         x4, x1                      //(core loop)
664
665    bgt         core_loop                   //loopback
666
667epilogue:
668    umull       v12.8h, v15.8b, v25.8b      //(3)mul_res = vmull_u8(src[0_3], coeffabs_3)//
669
670    umlsl       v12.8h, v14.8b, v24.8b      //(3)mul_res = vmlsl_u8(src[0_2], coeffabs_2)//
671
672    umlal       v12.8h, v16.8b, v26.8b      //(3)mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
673
674    umlsl       v12.8h, v17.8b, v27.8b      //(3)mul_res = vmlal_u8(src[0_1], coeffabs_1)//
675
676    st1         {v29.8b},[x4],x3            //(1)store the result pu1_dst
677    sqrshrun    v10.8b, v10.8h,#6           //(2)right shift and saturating narrow result 2
678
679    umull       v22.8h, v19.8b, v25.8b      //(4)mul_res = vmull_u8(src[0_3], coeffabs_3)//
680    umlsl       v22.8h, v18.8b, v24.8b      //(4)mul_res = vmlsl_u8(src[0_2], coeffabs_2)//
681
682    umlal       v22.8h, v20.8b, v26.8b      //(4)mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
683
684    umlsl       v22.8h, v21.8b, v27.8b      //(4)mul_res = vmlal_u8(src[0_1], coeffabs_1)//
685
686    st1         {v10.8b},[x4],x3            //(2)store the result pu1_dst
687    sqrshrun    v12.8b, v12.8h,#6           //(3)right shift and saturating narrow result 1
688
689    st1         {v12.8b},[x4],x3            //(3)store the result pu1_dst
690
691    add         x1,x1,#8                    //(core loop)
692
693    sqrshrun    v22.8b, v22.8h,#6           //(4)right shift and saturating narrow result 2
694
695
696    st1         {v22.8b},[x4], x3           //(4)store the result pu1_dst
697
698    sub         x9,x9,x5
699    subs        x14,x14,#4                  //decrement the ht loop
700    sub         x1,x1,x5
701    add         x9,x9,x2,lsl #2
702    add         x1,x1,x3,lsl #2
703    bgt         outer_loop_ht_4
704    b           end_loops
705
706outer_loop_4:
707    add         x6,x1,x3                    //pu1_dst + dst_strd
708    mov         x7,x5
709    add         x4,x12,x2                   //pu1_src + src_strd
710
711inner_loop_4:
712    //ld1 {v0.2s, v1.2s},[x12]                    //vector load pu1_src
713
714    ld1         {v20.2s},[x12],x11          //vector load pu1_src
715    ld1         {v21.2s},[x12],x11          //vector load pu1_src
716    ld1         {v22.2s},[x12],x11          //vector load pu1_src
717    ld1         {v23.2s},[x12]              //vector load pu1_src
718
719    sub         x12,x12,#2                  //increment the input pointer
720    ld1         {v16.2s},[x4],x11           //vector load pu1_src
721    ld1         {v17.2s},[x4],x11           //vector load pu1_src
722    ld1         {v18.2s},[x4],x11           //vector load pu1_src
723    ld1         {v19.2s},[x4]               //vector load pu1_src
724    //vext.u8    d2,d0,d1,#2                        //vector extract of src[0_2]
725    //vext.u8    d4,d0,d1,#4                        //vector extract of src[0_4]
726    //ld1 {v12.2s, v13.2s},[x4]                    //vector load pu1_src + src_strd
727    //vext.u8    d6,d0,d1,#6                        //vector extract of src[0_6]
728
729    sub         x4,x4,#2                    //increment the input pointer
730    //vext.u8    d14,d12,d13,#2                    //vector extract of src[0_2]
731    //vext.u8    d16,d12,d13,#4                    //vector extract of src[0_4]
732    //vext.u8    d18,d12,d13,#6                    //vector extract of src[0_6]
733
734    zip1        v0.2s, v20.2s, v16.2s
735    zip2        v4.2s, v20.2s, v16.2s       //vector zip the i iteration and ii interation in single register
736    zip1        v1.2s, v21.2s, v17.2s
737    zip2        v5.2s, v21.2s, v17.2s
738    zip1        v2.2s, v22.2s, v18.2s
739    zip2        v6.2s, v22.2s, v18.2s
740    zip1        v3.2s, v23.2s, v19.2s
741    zip2        v7.2s, v23.2s, v19.2s
742
743    umull       v29.8h, v1.8b, v25.8b       //arithmetic operations for ii iteration in the same time
744    umlsl       v29.8h, v0.8b, v24.8b
745    umlal       v29.8h, v2.8b, v26.8b
746    umlsl       v29.8h, v3.8b, v27.8b
747
748    sqrshrun    v29.8b, v29.8h,#6           //narrow right shift and saturating the result
749    st1         {v29.s}[0],[x1],#4          //store the i iteration result which is in upper part of the register
750    subs        x7,x7,#4                    //decrement the wd by 4
751
752    st1         {v29.s}[1],[x6],#4          //store the ii iteration result which is in lower part of the register
753
754    bgt         inner_loop_4
755
756    sub         x12,x12,x5
757    subs        x14,x14,#2                  //decrement the ht by 2
758    sub         x1,x1,x5
759    add         x12,x12,x2,lsl #1
760    add         x1,x1,x3,lsl #1
761    bgt         outer_loop_4
762
763end_loops:
764
765    // ldmfd sp!,{x4-x12,x15}                  //reload the registers from sp
766    ldp         x19, x20,[sp],#16
767    ldp         d8,d15,[sp],#16             // Loading d15 using { ldr d15,[sp]; add sp,sp,#8 } is giving bus error.
768                                            // d8 is used as dummy register and loaded along with d15 using ldp. d8 is not used in the function.
769    ldp         d13,d14,[sp],#16
770    ldp         d11,d12,[sp],#16
771    ldp         d9,d10,[sp],#16
772    ret
773
774
775
776
777
778
779
780
781