/external/libvpx/vp8/common/arm/armv6/ |
D | recon_v6.asm | 48 pkhbt r8, r6, r7, lsl #16 ; 2 | 0 51 uxtab16 r8, r8, r4 ; 2 | 0 + 3 | 2 | 2 | 0 54 usat16 r8, #8, r8 57 orr r8, r8, r9, lsl #8 59 str r8, [dst], stride 68 pkhbt r8, r6, r7, lsl #16 ; 2 | 0 71 uxtab16 r8, r8, r4 ; 2 | 0 + 3 | 2 | 2 | 0 74 usat16 r8, #8, r8 77 orr r8, r8, r9, lsl #8 79 str r8, [dst], stride [all …]
|
D | loopfilter_v6.asm | 91 uqsub8 r8, r10, r11 ; p2 - p1 95 orr r8, r8, r10 ; abs (p2-p1) 97 uqsub8 r8, r8, r2 ; compare to limit 99 orr lr, lr, r8 105 uqsub8 r8, r6, r3 ; compare to thresh -- save r8 for later 158 orr r10, r6, r8 ; calculate vp8_hevmask 166 ldr r8, [src], pstep ; p0 172 eor r8, r8, r12 ; p0 offset to convert to a signed value 177 str r8, [sp, #4] ; store ps0 temporarily 182 qsub8 r8, r9, r8 ; vp8_signed_char_clamp(vp8_filter + 3 * ( qs0 - ps0)) [all …]
|
D | filter_v6.asm | 57 ldrb r8, [r0, #-2] ; load source data 65 pkhbt lr, r8, r9, lsl #16 ; r9 | r8 66 pkhbt r8, r9, r10, lsl #16 ; r10 | r9 72 smuad r8, r8, r4 77 smlad r8, r11, r5, r8 86 smlad r11, r10, r6, r8 91 ldrneb r8, [r0, #-2] ; load data for next loop 148 ldr r8, [r0] ; load the data 153 smuad lr, r4, r8 ; apply filter 155 smulbt r8, r4, r8 [all …]
|
D | bilinearfilter_v6.asm | 52 ldrb r8, [r0, #2] 60 pkhbt r7, r7, r8, lsl #16 ; src[2] | src[1] 63 pkhbt r8, r8, r9, lsl #16 ; src[3] | src[2] 67 smuad r8, r8, r5 80 add r8, r8, #0x40 ; round_shift_and_clamp 83 usat r8, #16, r8, asr #7 86 strh r8, [r1], r3 ; result is transposed and stored 92 ldrneb r8, [r0, #2] 117 ldrb r8, [r0, #2] 124 strh r8, [r1], r3 [all …]
|
D | idct_v6.asm | 12 ; r0 r1 r2 r3 r4 r5 r6 r7 r8 r9 r10 r11 r12 r14 69 ldrsh r8, [r0, #16] ; input[8] 1, r8 unavail 2 [8] 73 add r9, r7, r8 ; a1 = [0] + [8] 1 a1 74 sub r7, r7, r8 ; b1 = [0] - [8] 1 b1 79 add r8, r7, r11 ; b1 + c1 1 b+c 80 strh r8, [r1, r2] ; out[pitch] = b1+c1 1 86 add r8, r2, r2 ; pitch * 2 1 p*2 87 strh r7, [r1, r8] ; out[pitch*2] = b1-c1 1 100 ldrsh r8, [r1, #6] ; input[3] 1, r8 un 2 [3] 104 …smulwb r10, r4, r8 ; ([3] * cospi8sqrt2minus1) >> 16 1, r10 un 2, r4/r8 ^1 … [all …]
|
/external/llvm/test/MC/ARM/ |
D | basic-thumb2-instructions.s | 24 adc r8, r12, #0xaa00aa00 34 @ CHECK: adc r8, r12, #2852170240 @ encoding: [0x4c,0xf1,0xaa,0x28] 70 add r2, r8, #0xff00 86 @ CHECK: add.w r2, r8, #65280 @ encoding: [0x08,0xf5,0x7f,0x42] 101 add r1, r2, r8 105 add.w r4, r8, r1, ror #12 106 add r10, r8 107 add r10, r10, r8 109 @ CHECK: add.w r1, r2, r8 @ encoding: [0x02,0xeb,0x08,0x01] 113 @ CHECK: add.w r4, r8, r1, ror #12 @ encoding: [0x08,0xeb,0x31,0x34] [all …]
|
D | basic-arm-instructions.s | 62 adc r6, r7, r8, lsl r9 63 adc r6, r7, r8, lsr r9 64 adc r6, r7, r8, asr r9 65 adc r6, r7, r8, ror r9 100 @ CHECK: adc r6, r7, r8, lsl r9 @ encoding: [0x18,0x69,0xa7,0xe0] 101 @ CHECK: adc r6, r7, r8, lsr r9 @ encoding: [0x38,0x69,0xa7,0xe0] 102 @ CHECK: adc r6, r7, r8, asr r9 @ encoding: [0x58,0x69,0xa7,0xe0] 103 @ CHECK: adc r6, r7, r8, ror r9 @ encoding: [0x78,0x69,0xa7,0xe0] 155 add r6, r7, r8, lsl r9 157 add r6, r7, r8, lsr r9 [all …]
|
D | diagnostics.s | 166 ssat r8, #0, r10, lsl #8 167 ssat r8, #33, r10, lsl #8 168 ssat r8, #1, r10, lsl #-1 169 ssat r8, #1, r10, lsl #32 170 ssat r8, #1, r10, asr #0 171 ssat r8, #1, r10, asr #33 172 ssat r8, #1, r10, lsr #5 173 ssat r8, #1, r10, lsl fred 174 ssat r8, #1, r10, lsl #fred 177 @ CHECK-ERRORS: ssat r8, #0, r10, lsl #8 [all …]
|
/external/libvpx/vp8/decoder/arm/armv6/ |
D | dequant_idct_v6.asm | 71 smulwb r8, r4, r6 74 pkhbt r8, r8, r10, lsl #16 84 usub16 r7, r8, r7 87 usub16 r8, r11, r14 90 uadd16 r6, r8, r7 91 usub16 r7, r8, r7 104 ldr r8, [r0], #4 108 smulwt lr, r3, r8 109 smulwt r10, r4, r8 110 pkhbt r11, r8, r6, lsl #16 [all …]
|
D | dequant_dc_idct_v6.asm | 93 smulwb r8, r4, r6 96 pkhbt r8, r8, r10, lsl #16 106 usub16 r7, r8, r7 109 usub16 r8, r11, r14 112 uadd16 r6, r8, r7 113 usub16 r7, r8, r7 126 ldr r8, [r0], #4 130 smulwt lr, r3, r8 131 smulwt r10, r4, r8 132 pkhbt r11, r8, r6, lsl #16 [all …]
|
/external/aac/libFDK/src/arm/ |
D | dct_arm.cpp | 117 LDR r8, [r1], #4 // val_tw = *twiddle++; in dct_IV_func1() 121 SMULWT r9, r5, r8 // accuX = accu2*val_tw.l in dct_IV_func1() 122 SMULWB r5, r5, r8 // accu2 = accu2*val_tw.h in dct_IV_func1() 124 SMLAWT r5, r4, r8, r5 // accu2 = accu2*val_tw.h + accu1*val_tw.l in dct_IV_func1() 125 SMLAWB r4, r4, r8, r9 // accu1 = accu1*val_tw.h - accu2*val_tw.l in dct_IV_func1() 127 LDR r8, [r1], #4 // val_tw = *twiddle++; in dct_IV_func1() 131 SMULWB r9, r7, r8 // accuX = accu4*val_tw.h in dct_IV_func1() 132 SMULWT r7, r7, r8 // accu4 = accu4*val_tw.l in dct_IV_func1() 134 SMLAWB r7, r6, r8, r7 // accu4 = accu4*val_tw.l+accu3*val_tw.h in dct_IV_func1() 135 SMLAWT r6, r6, r8, r9 // accu3 = accu3*val_tw.l-accu4*val_tw.h in dct_IV_func1() [all …]
|
/external/libvpx/vp8/encoder/arm/armv6/ |
D | vp8_mse16x16_armv6.asm | 41 usub8 r8, r5, r6 ; calculate difference 42 sel r7, r8, lr ; select bytes with positive difference 44 sel r8, r9, lr ; select bytes with negative difference 48 usad8 r6, r8, lr ; calculate sum of negative differences 49 orr r8, r8, r7 ; differences of all 4 pixels 54 uxtb16 r6, r8 ; byte (two pixels) to halfwords 55 uxtb16 r7, r8, ror #8 ; another two pixels to halfwords 62 usub8 r8, r5, r6 ; calculate difference 63 sel r7, r8, lr ; select bytes with positive difference 65 sel r8, r9, lr ; select bytes with negative difference [all …]
|
D | vp8_subtract_armv6.asm | 38 ldr r8, [r1, #vp8_blockd_predictor] 45 ldr r1, [r8], r2 ; pred 92 uxtb16 r8, r6 ; [s2 | s0] (A) 97 usub16 r6, r8, r9 ; [d2 | d0] (A) 103 pkhbt r8, r6, r7, lsl #16 ; [d1 | d0] (A) 106 str r8, [r0], #4 ; diff (A) 107 uxtb16 r8, r10 ; [s2 | s0] (B) 114 usub16 r6, r8, r9 ; [d2 | d0] (B) 119 pkhbt r8, r6, r7, lsl #16 ; [d1 | d0] (B) 122 str r8, [r0], #4 ; diff (B) [all …]
|
D | vp8_variance8x8_armv6.asm | 37 usub8 r8, r6, r7 ; calculate difference 38 sel r10, r8, lr ; select bytes with positive difference 40 sel r8, r9, lr ; select bytes with negative difference 44 usad8 r7, r8, lr ; calculate sum of negative differences 45 orr r8, r8, r10 ; differences of all 4 pixels 51 uxtb16 r7, r8 ; byte (two pixels) to halfwords 52 uxtb16 r10, r8, ror #8 ; another two pixels to halfwords 60 usub8 r8, r6, r7 ; calculate difference 62 sel r10, r8, lr ; select bytes with positive difference 65 sel r8, r9, lr ; select bytes with negative difference [all …]
|
/external/openssl/crypto/aes/asm/ |
D | aes-armv4.s | 231 and r8,lr,r0,lsr#8 237 ldr r5,[r10,r8,lsl#2] @ Te2[s0>>8] 238 and r8,lr,r1 245 ldr r8,[r10,r8,lsl#2] @ Te3[s1>>0] 250 eor r5,r5,r8,ror#8 251 and r8,lr,r2,lsr#16 @ i1 256 ldr r8,[r10,r8,lsl#2] @ Te1[s2>>16] 263 eor r1,r1,r8,ror#8 264 and r8,lr,r3,lsr#8 @ i1 269 ldr r8,[r10,r8,lsl#2] @ Te2[s3>>8] [all …]
|
/external/libffi/src/ia64/ |
D | unix.S | 60 mov r8 = in1 125 zxt1 r8 = r8 127 st8 [in1] = r8 131 sxt1 r8 = r8 133 st8 [in1] = r8 137 zxt2 r8 = r8 139 st8 [in1] = r8 143 sxt2 r8 = r8 145 st8 [in1] = r8 149 zxt4 r8 = r8 [all …]
|
/external/tremolo/Tremolo/ |
D | mdctARM.s | 187 LDMFD r12,{r8,r9,r10} @ r8 = step 191 MOV r8, r8, LSL #1 208 STRH r6, [r0], r8 226 LDMFD r12,{r8,r9,r10} @ r8 = step 230 MOV r8, r8, LSL #1 247 STRH r6, [r0], r8 272 LDR r8, [r1], #8 276 STMIA r2!,{r3,r4,r5,r6,r7,r8,r12,r14} 322 SMULL r8, r9, r7, r11 @ (r8, r9) = s2*T[1] 325 SMLAL r8, r9, r6, r10 @ (r8, r9) += s0*T[0] [all …]
|
D | mdctLARM.s | 187 LDMFD r12,{r8,r9,r10} @ r8 = step 191 MOV r8, r8, LSL #1 209 STRH r6, [r0], r8 227 LDMFD r12,{r8,r9,r10} @ r8 = step 231 MOV r8, r8, LSL #1 250 STRH r6, [r0], r8 275 LDR r8, [r1], #8 279 STMIA r2!,{r3,r4,r5,r6,r7,r8,r12,r14} 369 LDR r8, [r1],#16 @ r8 = ro0 = bX[0] 373 MOV r8, r8, ASR #8 [all …]
|
/external/qemu/distrib/jpeg-6b/ |
D | armv6_idct.S | 77 stmdb sp!, {r4, r5, r6, r7, r8, r9, r10, r11, r12, r14} 89 ldmdb r12!, {r8, r9, r10, r11} 102 mul r4, r8, r4 106 orreqs r8, r1, r2 107 orreqs r8, r3, r5 108 orreqs r8, r6, r7 121 ldmdb r12!, {r8, r9, r10, r11} 132 mul r7, r8, r7 151 add r8, r5, r7 157 smulwb r8, r8, r10 [all …]
|
/external/jpeg/ |
D | armv6_idct.S | 77 stmdb sp!, {r4, r5, r6, r7, r8, r9, r10, r11, r12, r14} 89 ldmdb r12!, {r8, r9, r10, r11} 102 mul r4, r8, r4 106 orreqs r8, r1, r2 107 orreqs r8, r3, r5 108 orreqs r8, r6, r7 121 ldmdb r12!, {r8, r9, r10, r11} 132 mul r7, r8, r7 151 add r8, r5, r7 157 smulwb r8, r8, r10 [all …]
|
/external/valgrind/main/none/tests/x86/ |
D | insn_basic.def | 27 adcb eflags[0x1,0x0] : r8.ub[12] r8.ub[34] => 1.ub[46] 28 adcb eflags[0x1,0x1] : r8.ub[12] r8.ub[34] => 1.ub[47] 29 adcb eflags[0x1,0x0] : r8.ub[12] m8.ub[34] => 1.ub[46] 30 adcb eflags[0x1,0x1] : r8.ub[12] m8.ub[34] => 1.ub[47] 31 adcb eflags[0x1,0x0] : m8.ub[12] r8.ub[34] => 1.ub[46] 32 adcb eflags[0x1,0x1] : m8.ub[12] r8.ub[34] => 1.ub[47] 64 addb r8.ub[12] r8.ub[34] => 1.ub[46] 65 addb r8.ub[12] m8.ub[34] => 1.ub[46] 66 addb m8.ub[12] r8.ub[34] => 1.ub[46] 84 andb r8.ub[0x34] r8.ub[0x56] => 1.ub[0x14] [all …]
|
/external/llvm/test/MC/Disassembler/ARM/ |
D | thumb2.txt | 10 # CHECK: adc r8, r12, #2852170240 56 # CHECK: add.w r2, r8, #65280 77 # CHECK: add.w r1, r2, r8 81 # CHECK: add.w r4, r8, r1, ror #12 114 # CHECK: and.w r4, r9, r8 115 # CHECK: and.w r1, r4, r8, asr #3 130 # CHECK: asrs.w r8, r3, #32 136 # CHECK: asrs.w r8, r8, #2 157 # CHECK: asrs.w r3, r4, r8 205 # CHECK: bic.w r8, r4, r1, lsr #11 [all …]
|
D | memory-arm-instructions.txt | 29 # CHECK: ldr r3, [r8, r1] 32 # CHECK: ldr r6, [r7, -r8]! 36 # CHECK: ldr r3, [r8, -r2, lsl #15 53 # CHECK: ldrb r3, [r8] 56 # CHECK: ldrb r8, [r1], #22 69 # CHECK: ldrb r9, [r8, r5] 74 # CHECK: ldrb r8, [r4], -r5 93 # CHECK: ldrbt r2, [r8], #-8 94 # CHECK: ldrbt r8, [r7], r6 108 # CHECK: ldrd r8, r9, [r2, #15] [all …]
|
/external/libvpx/vpx_scale/arm/armv4/ |
D | gen_scalers_armv4.asm | 79 and r8, mask, r3, lsr #24 ; d = src[3] 83 orr r7, r8, r7, lsl #16 ; c | d 92 orr r9, r9, r8, lsl #16 ; d | e 115 and r8, mask, r3, lsl #24 119 orr r7, r8, r7, lsl #16 ; c | d 176 ldrb r8, [r3], r1 ; d = des[dest_pitch*3] 180 orr r7, r8, r7, lsl #16 ; c | d 187 orr r9, r9, r8, lsl #16 ; d | e 301 stmdb sp!, {r4 - r8, lr} 314 mla r8, lr, r7, r5 ; c * 85 [all …]
|
/external/valgrind/main/none/tests/amd64/ |
D | insn_basic.def | 7 adcb eflags[0x1,0x0] : r8.ub[12] r8.ub[34] => 1.ub[46] 8 adcb eflags[0x1,0x1] : r8.ub[12] r8.ub[34] => 1.ub[47] 9 adcb eflags[0x1,0x0] : r8.ub[12] m8.ub[34] => 1.ub[46] 10 adcb eflags[0x1,0x1] : r8.ub[12] m8.ub[34] => 1.ub[47] 11 ###adcb eflags[0x1,0x0] : m8.ub[12] r8.ub[34] => 1.ub[46] 12 ###adcb eflags[0x1,0x1] : m8.ub[12] r8.ub[34] => 1.ub[47] 58 addb r8.ub[12] r8.ub[34] => 1.ub[46] 59 addb r8.ub[12] m8.ub[34] => 1.ub[46] 60 addb m8.ub[12] r8.ub[34] => 1.ub[46] 85 andb r8.ub[0x34] r8.ub[0x56] => 1.ub[0x14] [all …]
|