Home
last modified time | relevance | path

Searched refs:r8 (Results 1 – 25 of 856) sorted by relevance

12345678910>>...35

/external/linux-tools-perf/perf-3.12.0/arch/avr32/lib/
Dmemcpy.S31 3: ld.w r8, r11++
33 st.w r12++, r8
42 ld.ub r8, r11++
43 st.b r12++, r8
44 ld.ub r8, r11++
45 st.b r12++, r8
46 ld.ub r8, r11++
47 st.b r12++, r8
57 ld.ub r8, r11++
58 st.b r12++, r8
[all …]
/external/libhevc/common/arm/
Dihevc_intra_pred_luma_mode_18_34.s129 mov r8,r0
133 vld1.8 {d0},[r8],r6
135 vld1.8 {d1},[r8],r6
137 vld1.8 {d2},[r8],r6
138 vld1.8 {d3},[r8],r6
140 vld1.8 {d4},[r8],r6
141 vld1.8 {d5},[r8],r6
142 vld1.8 {d6},[r8],r6
144 vld1.8 {d7},[r8],r6
152 movne r8,r0
[all …]
Dihevc_sao_edge_offset_class3.s59 @r8=> ht
82 LDR r8,[sp,#0x40] @Loads ht
97 SUB r10,r8,#1 @ht-1
146 SUB r11,r8,#1 @ht - 1
189 MOV r12,r8 @Move ht
219 CMP r8,#4 @Compare ht with 4
227 LDREQB r8,[r5] @pu1_avail[0]
228 MOVNE r8,#-1
229 VMOV.8 d8[0],r8 @au1_mask = vsetq_lane_s8(-1, au1_mask, 0)
233 LDRB r8,[r5,#1] @pu1_avail[1]
[all …]
Dihevc_sao_edge_offset_class2_chroma.s61 @r8=> ht
90 LDR r8,[sp,#0x44] @Loads ht
107 SUB r10,r8,#1 @ht-1
192 SUB r11,r8,#1 @ht - 1
265 MOV r12,r8 @Move ht
296 CMP r8,#4 @Compare ht with 4
303 LDREQB r8,[r5] @pu1_avail[0]
305 MOVNE r8,#-1
306 VMOV.8 D8[0],r8 @au1_mask = vsetq_lane_s8(-1, au1_mask, 0)
309 VMOV.8 D8[1],r8 @au1_mask = vsetq_lane_s8(-1, au1_mask, 0)
[all …]
Dihevc_intra_pred_chroma_mode_18_34.s128 mov r8,r0
134 vld1.8 {d0,d1},[r8],r6
136 vld1.8 {d2,d3},[r8],r6
138 vld1.8 {d4,d5},[r8],r6
140 vld1.8 {d6,d7},[r8],r6
142 vld1.8 {d8,d9},[r8],r6
144 vld1.8 {d10,d11},[r8],r6
146 vld1.8 {d12,d13},[r8],r6
148 vld1.8 {d14,d15},[r8],r6
155 add r8,r0,#16
[all …]
/external/linux-tools-perf/perf-3.12.0/arch/arm/lib/
Dmemset.S36 stmfd sp!, {r8, lr}
37 mov r8, r1
41 stmgeia ip!, {r1, r3, r8, lr} @ 64 bytes at a time.
42 stmgeia ip!, {r1, r3, r8, lr}
43 stmgeia ip!, {r1, r3, r8, lr}
44 stmgeia ip!, {r1, r3, r8, lr}
46 ldmeqfd sp!, {r8, pc} @ Now <64 bytes to go.
51 stmneia ip!, {r1, r3, r8, lr}
52 stmneia ip!, {r1, r3, r8, lr}
54 stmneia ip!, {r1, r3, r8, lr}
[all …]
/external/openssl/crypto/modes/asm/
Dghash-x86_64.S12 movzbq 15(%rdi),%r8
20 movq 8(%rsi,%rax,1),%r8
23 movq %r8,%rdx
28 shrq $4,%r8
33 xorq 8(%rsi,%rbx,1),%r8
38 movq %r8,%rdx
40 xorq %r10,%r8
44 shrq $4,%r8
48 xorq 8(%rsi,%rax,1),%r8
53 movq %r8,%rdx
[all …]
/external/chromium_org/third_party/boringssl/src/crypto/bn/asm/
Drsaz-x86_64.pl163 movq %rax, %r8
203 addq %r8, %r8 #shlq \$1, %r8
205 adcq %r9, %r9 #shld \$1, %r8, %r9
209 addq %rdx, %r8
212 movq %r8, 8(%rsp)
216 movq 8($inp), %r8
218 mulq %r8
224 mulq %r8
232 mulq %r8
240 mulq %r8
[all …]
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/common/arm/armv6/
Ddequant_idct_v6.asm68 smulwb r8, r4, r6
71 pkhbt r8, r8, r10, lsl #16
81 usub16 r7, r8, r7
84 usub16 r8, r11, r14
87 uadd16 r6, r8, r7
88 usub16 r7, r8, r7
101 ldr r8, [r0], #4
105 smulwt lr, r3, r8
106 smulwt r10, r4, r8
107 pkhbt r11, r8, r6, lsl #16
[all …]
Dloopfilter_v6.asm92 uqsub8 r8, r10, r11 ; p2 - p1
96 orr r8, r8, r10 ; abs (p2-p1)
98 uqsub8 r8, r8, r2 ; compare to limit
100 orr lr, lr, r8
106 uqsub8 r8, r6, r3 ; compare to thresh -- save r8 for later
159 orr r10, r6, r8 ; calculate vp8_hevmask
167 ldr r8, [src], pstep ; p0
173 eor r8, r8, r12 ; p0 offset to convert to a signed value
178 str r8, [sp, #4] ; store ps0 temporarily
183 qsub8 r8, r9, r8 ; vp8_signed_char_clamp(vp8_filter + 3 * ( qs0 - ps0))
[all …]
/external/libvpx/libvpx/vp8/common/arm/armv6/
Ddequant_idct_v6.asm68 smulwb r8, r4, r6
71 pkhbt r8, r8, r10, lsl #16
81 usub16 r7, r8, r7
84 usub16 r8, r11, r14
87 uadd16 r6, r8, r7
88 usub16 r7, r8, r7
101 ldr r8, [r0], #4
105 smulwt lr, r3, r8
106 smulwt r10, r4, r8
107 pkhbt r11, r8, r6, lsl #16
[all …]
Dloopfilter_v6.asm92 uqsub8 r8, r10, r11 ; p2 - p1
96 orr r8, r8, r10 ; abs (p2-p1)
98 uqsub8 r8, r8, r2 ; compare to limit
100 orr lr, lr, r8
106 uqsub8 r8, r6, r3 ; compare to thresh -- save r8 for later
159 orr r10, r6, r8 ; calculate vp8_hevmask
167 ldr r8, [src], pstep ; p0
173 eor r8, r8, r12 ; p0 offset to convert to a signed value
178 str r8, [sp, #4] ; store ps0 temporarily
183 qsub8 r8, r9, r8 ; vp8_signed_char_clamp(vp8_filter + 3 * ( qs0 - ps0))
[all …]
/external/llvm/test/MC/ARM/
Dthumb2-narrow-dp.ll8 // - Rd, Rn and Rm are < r8
12 // - Rd, Rn and Rm are < r8
23 ANDS r8, r1, r8 // high registers so must use wide encoding
24 ANDS r8, r8, r1
25 ANDS r0, r8, r0
26 ANDS r1, r1, r8
37 // CHECK: ands.w r8, r1, r8 @ encoding: [0x11,0xea,0x08,0x08]
38 // CHECK: ands.w r8, r8, r1 @ encoding: [0x18,0xea,0x01,0x08]
39 // CHECK: ands.w r0, r8, r0 @ encoding: [0x18,0xea,0x00,0x00]
40 // CHECK: ands.w r1, r1, r8 @ encoding: [0x11,0xea,0x08,0x01]
[all …]
Dbasic-thumb2-instructions.s25 adc r8, r12, #0xaa00aa00
35 @ CHECK: adc r8, r12, #2852170240 @ encoding: [0x4c,0xf1,0xaa,0x28]
73 add r2, r8, #0xff00
97 @ CHECK: add.w r2, r8, #65280 @ encoding: [0x08,0xf5,0x7f,0x42]
120 add r1, r2, r8
124 add.w r4, r8, r1, ror #12
125 add r10, r8
126 add r10, r10, r8
128 @ CHECK: add.w r1, r2, r8 @ encoding: [0x02,0xeb,0x08,0x01]
132 @ CHECK: add.w r4, r8, r1, ror #12 @ encoding: [0x08,0xeb,0x31,0x34]
[all …]
/external/aac/libFDK/src/arm/
Ddct_arm.cpp117 LDR r8, [r1], #4 // val_tw = *twiddle++; in dct_IV_func1()
121 SMULWT r9, r5, r8 // accuX = accu2*val_tw.l in dct_IV_func1()
122 SMULWB r5, r5, r8 // accu2 = accu2*val_tw.h in dct_IV_func1()
124 SMLAWT r5, r4, r8, r5 // accu2 = accu2*val_tw.h + accu1*val_tw.l in dct_IV_func1()
125 SMLAWB r4, r4, r8, r9 // accu1 = accu1*val_tw.h - accu2*val_tw.l in dct_IV_func1()
127 LDR r8, [r1], #4 // val_tw = *twiddle++; in dct_IV_func1()
131 SMULWB r9, r7, r8 // accuX = accu4*val_tw.h in dct_IV_func1()
132 SMULWT r7, r7, r8 // accu4 = accu4*val_tw.l in dct_IV_func1()
134 SMLAWB r7, r6, r8, r7 // accu4 = accu4*val_tw.l+accu3*val_tw.h in dct_IV_func1()
135 SMLAWT r6, r6, r8, r9 // accu3 = accu3*val_tw.l-accu4*val_tw.h in dct_IV_func1()
[all …]
/external/chromium_org/third_party/boringssl/linux-x86_64/crypto/modes/
Dghash-x86_64.S15 movzbq 15(%rdi),%r8
23 movq 8(%rsi,%rax,1),%r8
26 movq %r8,%rdx
31 shrq $4,%r8
36 xorq 8(%rsi,%rbx,1),%r8
41 movq %r8,%rdx
43 xorq %r10,%r8
47 shrq $4,%r8
51 xorq 8(%rsi,%rax,1),%r8
56 movq %r8,%rdx
[all …]
/external/chromium_org/third_party/boringssl/mac-x86_64/crypto/modes/
Dghash-x86_64.S15 movzbq 15(%rdi),%r8
23 movq 8(%rsi,%rax,1),%r8
26 movq %r8,%rdx
31 shrq $4,%r8
36 xorq 8(%rsi,%rbx,1),%r8
41 movq %r8,%rdx
43 xorq %r10,%r8
47 shrq $4,%r8
51 xorq 8(%rsi,%rax,1),%r8
56 movq %r8,%rdx
[all …]
/external/libvpx/libvpx/vp8/encoder/arm/armv6/
Dvp8_mse16x16_armv6.asm44 usub8 r8, r5, r6 ; calculate difference
46 sel r7, r8, lr ; select bytes with positive difference
49 sel r8, r9, lr ; select bytes with negative difference
53 usad8 r6, r8, lr ; calculate sum of negative differences
54 orr r8, r8, r7 ; differences of all 4 pixels
59 uxtb16 r6, r8 ; byte (two pixels) to halfwords
60 uxtb16 r7, r8, ror #8 ; another two pixels to halfwords
67 usub8 r8, r5, r6 ; calculate difference
68 sel r7, r8, lr ; select bytes with positive difference
70 sel r8, r9, lr ; select bytes with negative difference
[all …]
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/encoder/arm/armv6/
Dvp8_mse16x16_armv6.asm44 usub8 r8, r5, r6 ; calculate difference
46 sel r7, r8, lr ; select bytes with positive difference
49 sel r8, r9, lr ; select bytes with negative difference
53 usad8 r6, r8, lr ; calculate sum of negative differences
54 orr r8, r8, r7 ; differences of all 4 pixels
59 uxtb16 r6, r8 ; byte (two pixels) to halfwords
60 uxtb16 r7, r8, ror #8 ; another two pixels to halfwords
67 usub8 r8, r5, r6 ; calculate difference
68 sel r7, r8, lr ; select bytes with positive difference
70 sel r8, r9, lr ; select bytes with negative difference
[all …]
/external/chromium_org/third_party/boringssl/win-x86_64/crypto/modes/
Dghash-x86_64.asm22 movzx r8,BYTE PTR[15+rdi]
30 mov r8,QWORD PTR[8+rax*1+rsi]
33 mov rdx,r8
38 shr r8,4
43 xor r8,QWORD PTR[8+rbx*1+rsi]
48 mov rdx,r8
50 xor r8,r10
54 shr r8,4
58 xor r8,QWORD PTR[8+rax*1+rsi]
63 mov rdx,r8
[all …]
/external/tremolo/Tremolo/
DmdctARM.s189 LDMFD r12,{r8,r9,r10} @ r8 = step
193 MOV r8, r8, LSL #1
210 STRH r6, [r0], r8
228 LDMFD r12,{r8,r9,r10} @ r8 = step
232 MOV r8, r8, LSL #1
249 STRH r6, [r0], r8
274 LDR r8, [r1], #8
278 STMIA r2!,{r3,r4,r5,r6,r7,r8,r12,r14}
326 SMULL r8, r9, r7, r11 @ (r8, r9) = s2*T[1]
329 SMLAL r8, r9, r6, r10 @ (r8, r9) += s0*T[0]
[all …]
DmdctLARM.s187 LDMFD r12,{r8,r9,r10} @ r8 = step
191 MOV r8, r8, LSL #1
209 STRH r6, [r0], r8
227 LDMFD r12,{r8,r9,r10} @ r8 = step
231 MOV r8, r8, LSL #1
250 STRH r6, [r0], r8
275 LDR r8, [r1], #8
279 STMIA r2!,{r3,r4,r5,r6,r7,r8,r12,r14}
369 LDR r8, [r1],#16 @ r8 = ro0 = bX[0]
373 MOV r8, r8, ASR #8
[all …]
/external/qemu/distrib/jpeg-6b/
Darmv6_idct.S77 stmdb sp!, {r4, r5, r6, r7, r8, r9, r10, r11, r12, r14}
89 ldmdb r12!, {r8, r9, r10, r11}
102 mul r4, r8, r4
106 orreqs r8, r1, r2
107 orreqs r8, r3, r5
108 orreqs r8, r6, r7
121 ldmdb r12!, {r8, r9, r10, r11}
132 mul r7, r8, r7
151 add r8, r5, r7
157 smulwb r8, r8, r10
[all …]
/external/jpeg/
Darmv6_idct.S77 stmdb sp!, {r4, r5, r6, r7, r8, r9, r10, r11, r12, r14}
89 ldmdb r12!, {r8, r9, r10, r11}
102 mul r4, r8, r4
106 orreqs r8, r1, r2
107 orreqs r8, r3, r5
108 orreqs r8, r6, r7
121 ldmdb r12!, {r8, r9, r10, r11}
132 mul r7, r8, r7
151 add r8, r5, r7
157 smulwb r8, r8, r10
[all …]
/external/libvpx/libvpx/vp9/common/arm/neon/
Dvp9_mb_lpf_neon.asm28 push {r4-r8, lr}
38 sub r8, r0, r1, lsl #3 ; move src pointer down by 8 lines
40 vld1.u8 {d0}, [r8@64], r1 ; p7
41 vld1.u8 {d1}, [r8@64], r1 ; p6
42 vld1.u8 {d2}, [r8@64], r1 ; p5
43 vld1.u8 {d3}, [r8@64], r1 ; p4
44 vld1.u8 {d4}, [r8@64], r1 ; p3
45 vld1.u8 {d5}, [r8@64], r1 ; p2
46 vld1.u8 {d6}, [r8@64], r1 ; p1
47 vld1.u8 {d7}, [r8@64], r1 ; p0
[all …]

12345678910>>...35