/external/libvpx/vpx_scale/arm/neon/ |
D | vp8_vpxyv12_extendframeborders_neon.asm | 44 add r6, r1, lr 45 sub r6, r6, r3, lsl #1 ;destptr2 46 sub r2, r6, #1 ;srcptr2 73 vst1.8 {q2, q3}, [r6], lr 75 vst1.8 {q6, q7}, [r6], lr 77 vst1.8 {q10, q11}, [r6], lr 79 vst1.8 {q14, q15}, [r6], lr 89 sub r6, r1, r3 ;destptr2 90 sub r2, r6, lr ;srcptr2 110 vst1.8 {q8, q9}, [r6]! [all …]
|
D | vp8_vpxyv12_copyframeyonly_neon.asm | 32 ldr r6, [r0, #yv12_buffer_config_y_stride] 43 add r10, r2, r6 71 add r2, r2, r6, lsl #1 102 add r6, r1, lr 103 sub r6, r6, r3, lsl #1 ;destptr2 104 sub r2, r6, #1 ;srcptr2 131 vst1.8 {q2, q3}, [r6], lr 133 vst1.8 {q6, q7}, [r6], lr 135 vst1.8 {q10, q11}, [r6], lr 137 vst1.8 {q14, q15}, [r6], lr [all …]
|
/external/tremolo/Tremolo/ |
D | mdctARM.s | 61 LDMDB r2!,{r5,r6,r7,r12} 64 MOV r6, r6, ASR #9 @ r6 = (*--r)>>9 78 MOV r14,r6, ASR #15 80 EORNE r6, r4, r14,ASR #31 81 STRH r6, [r0], r3 123 LDR r6, [r2],#8 128 RSB r6, r6, #0 133 MOV r6, r6, ASR #9 @ r6 = (-*l)>>9 146 MOV r14,r6, ASR #15 148 EORNE r6, r4, r14,ASR #31 [all …]
|
D | mdctLARM.s | 61 LDMDB r2!,{r5,r6,r7,r12} 64 MOV r6, r6, ASR #9 @ r6 = (*--r)>>9 78 MOV r14,r6, ASR #15 80 EORNE r6, r4, r14,ASR #31 81 STRH r6, [r0], r3 123 LDR r6, [r2],#8 128 RSB r6, r6, #0 133 MOV r6, r6, ASR #9 @ r6 = (-*l)>>9 146 MOV r14,r6, ASR #15 148 EORNE r6, r4, r14,ASR #31 [all …]
|
D | bitwiseARM.s | 67 STMFD r13!,{r5,r6} 79 LDRLT r6,[r3] @ r6 = ptr[1] 82 ORRLT r10,r10,r6,LSL r12 @ r10= first bitsLeftInSeg bits+crap 97 LDMIA r11,{r6,r12,r14} @ r6 = buffer 100 LDR r6,[r6] @ r6 = buffer->data 103 ADD r6,r6,r12 @ r6 = buffer->data+begin 105 LDRB r12,[r6],#1 @ r12= *buffer 116 LDMFD r13!,{r5,r6,r10,r11,PC} 125 LDMFD r13!,{r5,r6,r10,r11,PC} 130 LDMFD r13!,{r5,r6,r10,r11,PC} [all …]
|
/external/libvpx/vpx_scale/symbian/ |
D | gen_scalers_armv4.asm | 72 orr r6, r4, r5, lsl #16 ; b | a 74 mul r6, c51_205, r6 ; a * 51 + 205 * b 78 add r6, r6, #0x8000 80 mov r6, r6, lsr #24 81 strb r6, [dest], #1 108 orr r6, r4, r5, lsl #16 ; b | a 109 mul r6, c51_205, r6 114 add r6, r6, #0x8000 116 mov r6, r6, lsr #24 117 strb r6, [dest], #1 [all …]
|
D | gen_scalers_armv4.s | 99 orr r6, r4, r5, lsl #16 @ b | a 101 mul r6, c51_205, r6 @ a * 51 + 205 * b 105 add r6, r6, #0x8000 107 mov r6, r6, lsr #24 108 strb r6, [dest], #1 135 orr r6, r4, r5, lsl #16 @ b | a 136 mul r6, c51_205, r6 141 add r6, r6, #0x8000 143 mov r6, r6, lsr #24 144 strb r6, [dest], #1 [all …]
|
/external/libvpx/vpx_scale/arm/armv4/ |
D | gen_scalers_armv4.asm | 72 orr r6, r4, r5, lsl #16 ; b | a 74 mul r6, c51_205, r6 ; a * 51 + 205 * b 78 add r6, r6, #0x8000 80 mov r6, r6, lsr #24 81 strb r6, [dest], #1 108 orr r6, r4, r5, lsl #16 ; b | a 109 mul r6, c51_205, r6 114 add r6, r6, #0x8000 116 mov r6, r6, lsr #24 117 strb r6, [dest], #1 [all …]
|
/external/libvpx/vpx_scale/wce/ |
D | gen_scalers_armv4.asm | 72 orr r6, r4, r5, lsl #16 ; b | a 74 mul r6, c51_205, r6 ; a * 51 + 205 * b 78 add r6, r6, #0x8000 80 mov r6, r6, lsr #24 81 strb r6, [dest], #1 108 orr r6, r4, r5, lsl #16 ; b | a 109 mul r6, c51_205, r6 114 add r6, r6, #0x8000 116 mov r6, r6, lsr #24 117 strb r6, [dest], #1 [all …]
|
/external/libvpx/vp8/decoder/arm/armv6/ |
D | dequant_dc_idct_v6.asm | 29 ldr r6, [sp, #44] 42 strh r6, [r0], #2 45 smulbb r6, r4, r5 51 strh r6, [r0], #2 57 smulbb r6, r4, r5 63 strh r6, [r0], #2 66 smulbb r6, r4, r5 74 strh r6, [r0], #2 85 ldr r6, [r0, #8] 90 smulwt r9, r3, r6 [all …]
|
D | dequant_idct_v6.asm | 35 smulbb r6, r4, r5 41 strh r6, [r0], #2 44 smulbb r6, r4, r5 52 strh r6, [r0], #2 63 ldr r6, [r0, #8] 68 smulwt r9, r3, r6 69 smulwb r7, r3, r6 70 smulwt r10, r4, r6 71 smulwb r8, r4, r6 75 uadd16 r6, r6, r7 [all …]
|
D | dboolhuff_v6.asm | 38 stmdb sp!, {r4 - r6, lr} 41 mov r6, #0 50 orr r6, r6, r0, lsl r5 55 mov r0, r6 56 ldmia sp!, {r4 - r6, pc} 102 ; split RN r6 107 stmdb sp!, {r4 - r6, lr} 112 mov r6, r2, lsl #8 113 sub r6, r6, #256 ; split = 1 + (((range-1) * probability) >> 8) 115 smlawb r6, r6, prob, r12 [all …]
|
/external/libvpx/vp8/common/arm/armv6/ |
D | recon_v6.asm | 45 ldr r6, [dif, #0] ; 1 | 0 48 pkhbt r8, r6, r7, lsl #16 ; 2 | 0 49 pkhtb r9, r7, r6, asr #16 ; 3 | 1 63 ;; ldr r6, [dif, #8] ; 1 | 0 65 ldr r6, [dif, #0] ; 1 | 0 68 pkhbt r8, r6, r7, lsl #16 ; 2 | 0 69 pkhtb r9, r7, r6, asr #16 ; 3 | 1 83 ;; ldr r6, [dif, #16] ; 1 | 0 85 ldr r6, [dif, #0] ; 1 | 0 88 pkhbt r8, r6, r7, lsl #16 ; 2 | 0 [all …]
|
D | loopfilter_v6.asm | 71 ldr r6, [sp, #36] ; load thresh address 80 ldr r3, [r6], #4 ; thresh 89 uqsub8 r6, r9, r10 ; p3 - p2 94 orr r6, r6, r7 ; abs (p3-p2) 96 uqsub8 lr, r6, r2 ; compare to limit. lr: vp8_filter_mask 98 uqsub8 r6, r11, r12 ; p1 - p0 103 orr r6, r6, r7 ; abs (p1-p0) 104 uqsub8 r7, r6, r2 ; compare to limit 105 uqsub8 r8, r6, r3 ; compare to thresh -- save r8 for later 108 uqsub8 r6, r11, r10 ; p1 - q1 [all …]
|
D | idct_v6.asm | 12 ; r0 r1 r2 r3 r4 r5 r6 r7 r8 r9 r10 r11 r12 r14 65 mov r6, #4 ; i=4 1 i 90 subs r6, r6, #1 ; i-- 1 -- 96 mov r6, #4 ; i=4 1 i 126 subs r6, r6, #1 ; i-- 1 -- 164 smulwb r6, r4, r10 ; ((ip[4] * sinpi8sqrt2) >> 16) lt1 170 add r6, r6, r7 ; partial c1 lt1-lt2 182 pkhbt r6, r6, r7, lsl #16 ; partial c1_2 | partial c1_1 pack 185 usub16 r6, r6, r10 ; c1_2 | c1_1 c 242 ldr r6, [r0, #(4*2)] ; i5 | i4 5|4 [all …]
|
D | dc_only_idct_add_v6.asm | 28 ldr r6, [r1], r3 35 uxtab16 r7, r0, r6 36 uxtab16 r6, r0, r6, ror #8 40 usat16 r6, #8, r6 42 orr r7, r7, r6, lsl #8 44 ldr r6, [r1] 50 uxtab16 r7, r0, r6 51 uxtab16 r6, r0, r6, ror #8 55 usat16 r6, #8, r6 57 orr r7, r7, r6, lsl #8
|
D | bilinearfilter_v6.asm | 50 ldrb r6, [r0] ; load source data 59 pkhbt r6, r6, r7, lsl #16 ; src[1] | src[0] 62 smuad r6, r6, r5 ; apply the filter 73 add r6, r6, #0x40 ; round_shift_and_clamp 75 usat r6, #16, r6, asr #7 78 strh r6, [r1], r3 ; result is transposed and stored 88 ldrneb r6, [r0] ; load source data 115 ldrb r6, [r0] ; load data 120 strh r6, [r1], r3 ; store it to immediate buffer 163 ldr r6, [r0] ; load the data [all …]
|
/external/jpeg/ |
D | armv6_idct.S | 77 stmdb sp!, {r4, r5, r6, r7, r8, r9, r10, r11, r12, r14} 98 ldrsh r6, [r14, #96] 108 orreqs r8, r6, r7 118 mla r6, r11, r6, r2 126 rsb r2, r6, r2, lsl #1 163 add r0, r0, r6 164 sub r2, r2, r6 165 sub r6, r0, r6, lsl #1 197 add r6, r6, r5 199 sub r5, r6, r5, lsl #1 [all …]
|
/external/openssl/crypto/bn/asm/ |
D | s390x.S | 26 stmg %r6,%r10,48(%r15) 33 mlgr %r6,%r5 // *=w 35 alcgr %r6,zero 37 alcgr %r6,zero 42 algr %r9,%r6 49 mlgr %r6,%r5 51 alcgr %r6,zero 53 alcgr %r6,zero 58 algr %r9,%r6 73 mlgr %r6,%r5 // *=w [all …]
|
/external/libvpx/vp8/encoder/arm/armv5te/ |
D | vp8_packtokens_mbrow_armv5.asm | 34 ldr r6, _VP8_COMMON_MBrows_ 37 ldr r5, [r4, r6] ; load up mb_rows 65 ldr r6, [r1, #tokenextra_token] ; t 68 add r4, r4, r6, lsl #3 ; a = vp8_coef_encodings + t 73 ldr r6, [r4, #vp8_token_value] ; v 85 lsl r12, r6, r4 ; r12 = v << 32 - n 110 clz r6, r4 111 sub r6, r6, #24 ; shift 115 adds r3, r3, r6 ; count += shift 116 lsl r5, r4, r6 ; range <<= shift [all …]
|
D | vp8_packtokens_partitions_armv5.asm | 36 ldr r6, _VP8_COMMON_MBrows_ 39 ldr r5, [r4, r6] ; load up mb_rows 93 ldr r6, [r1, #tokenextra_token] ; t 96 add r4, r4, r6, lsl #3 ; a = vp8_coef_encodings + t 101 ldr r6, [r4, #vp8_token_value] ; v 113 lsl r12, r6, r4 ; r12 = v << 32 - n 138 clz r6, r4 139 sub r6, r6, #24 ; shift 143 adds r3, r3, r6 ; count += shift 144 lsl r5, r4, r6 ; range <<= shift [all …]
|
D | vp8_packtokens_armv5.asm | 44 ldr r6, [r1, #tokenextra_token] ; t 47 add r4, r4, r6, lsl #3 ; a = vp8_coef_encodings + t 52 ldr r6, [r4, #vp8_token_value] ; v 64 lsl r12, r6, r4 ; r12 = v << 32 - n 89 clz r6, r4 90 sub r6, r6, #24 ; shift 94 adds r3, r3, r6 ; count += shift 95 lsl r5, r4, r6 ; range <<= shift 98 sub r6, r6, r3 ; offset = shift - count 99 sub r4, r6, #1 ; offset-1 [all …]
|
D | boolhuff_armv5te.asm | 65 clz r6, r4 66 sub r6, r6, #24 ; shift 70 adds r3, r3, r6 ; count += shift 71 lsl r5, r4, r6 ; range <<= shift 74 sub r6, r6, r3 ; offset = shift - count 75 sub r4, r6, #1 ; offset-1 98 rsb r4, r6, #24 ; 24-offset 102 lsl r2, r2, r6 ; lowvalue <<= offset 103 mov r6, r3 ; shift = count 111 lsl r2, r2, r6 ; lowvalue <<= shift [all …]
|
/external/openssl/crypto/aes/asm/ |
D | aes-armv4.s | 125 ldrb r6,[r12,#0] 128 orr r0,r0,r6,lsl#24 132 ldrb r6,[r12,#4] 135 orr r1,r1,r6,lsl#24 139 ldrb r6,[r12,#8] 142 orr r2,r2,r6,lsl#24 146 ldrb r6,[r12,#12] 149 orr r3,r3,r6,lsl#24 156 mov r6,r0,lsr#8 159 strb r6,[r12,#2] [all …]
|
/external/openssl/crypto/sha/asm/ |
D | sha1-armv4-large.s | 10 ldmia r0,{r3,r4,r5,r6,r7} 16 mov r6,r6,ror#30 27 eor r10,r5,r6 @ F_xx_xx 32 eor r10,r10,r6,ror#2 @ F_00_19(B,C,D) 38 add r6,r8,r6,ror#2 @ E+=K_00_19 40 add r6,r6,r7,ror#27 @ E+=ROR(A,27) 44 add r6,r6,r9 @ E+=X[i] 48 add r6,r6,r10 @ E+=F_00_19(B,C,D) 55 add r5,r5,r6,ror#27 @ E+=ROR(A,27) 76 and r10,r6,r10,ror#2 [all …]
|