Home
last modified time | relevance | path

Searched refs:r10 (Results 1 – 25 of 525) sorted by relevance

12345678910>>...21

/third_party/libffi/src/tile/
Dtile.S83 #define FRAME_SIZE r10
253 SW sp, r10
254 .cfi_return_column r10
255 .cfi_offset r10, 0
258 addli r10, sp, -(CLOSURE_FRAME_SIZE - REG_SIZE)
262 SW r10, sp
271 addi r10, sp, LINKAGE_SIZE
275 STORE_REG(r0, r10)
276 STORE_REG(r1, r10)
277 STORE_REG(r2, r10)
[all …]
/third_party/skia/tests/sksl/runtime/
DSwitch.skvm12 10 r10 = trunc r1
13 11 r8 = eq_i32 r8 r10
21 19 r10 = eq_i32 r16 r10
22 20 r10 = bit_or r8 r10
23 21 r10 = bit_and r10 r15
24 22 r11 = select r10 r0 r11
25 23 r12 = select r10 r1 r12
26 24 r13 = select r10 r2 r13
27 25 r14 = select r10 r3 r14
28 26 r10 = bit_xor r9 r10
[all …]
DSwitchWithLoops.skvm12 10 r10 = splat 1 (1.4012985e-45)
13 11 r9 = eq_i32 r10 r9
14 12 r11 = bit_and r10 r9
16 14 r13 = add_i32 r11 r10
19 17 r13 = add_i32 r11 r10
23 21 r12 = add_i32 r11 r10
27 25 r15 = add_i32 r11 r10
31 29 r14 = add_i32 r11 r10
34 32 r14 = add_i32 r11 r10
38 36 r13 = add_i32 r11 r10
[all …]
/third_party/openssl/crypto/sha/asm/arm32/
Dsha1-armv4-large.S41 ldrb r10,[r1,#2]
46 orr r9,r9,r10,lsl#8
47 eor r10,r5,r6 @ F_xx_xx
54 eor r10,r5,r6 @ F_xx_xx
60 and r10,r4,r10,ror#2
62 eor r10,r10,r6,ror#2 @ F_00_19(B,C,D)
64 add r7,r7,r10 @ E+=F_00_19(B,C,D)
66 ldrb r10,[r1,#2]
71 orr r9,r9,r10,lsl#8
72 eor r10,r4,r5 @ F_xx_xx
[all …]
Dkeccak1600-armv4.S46 add r10,sp,#40
57 ldmia r10,{r10-r12,r14} @ A[1][0..1]
59 eor r0,r0,r10
62 ldrd r10,r11,[sp,#56]
65 eor r4,r4,r10
68 ldrd r10,r11,[sp,#72]
71 eor r8,r8,r10
74 ldrd r10,r11,[sp,#88]
77 eor r2,r2,r10
80 ldrd r10,r11,[sp,#104]
[all …]
/third_party/openssl/crypto/bn/asm/x86_64/
Dx86_64-mont5.s36 leaq -280(%rsp,%r9,8),%r10
38 andq $-1024,%r10
48 subq %r10,%r11
50 leaq (%r10,%r11,1),%rsp
52 cmpq %r10,%rsp
59 cmpq %r10,%rsp
63 leaq .Linc(%rip),%r10
69 movdqa 0(%r10),%xmm0
70 movdqa 16(%r10),%xmm1
71 leaq 24-112(%rsp,%r9,8),%r10
[all …]
Dx86_64-mont.s41 leaq -16(%rsp,%r9,8),%r10
43 andq $-1024,%r10
53 subq %r10,%r11
55 leaq (%r10,%r11,1),%rsp
57 cmpq %r10,%rsp
65 cmpq %r10,%rsp
82 movq %rax,%r10
85 imulq %r10,%rbp
89 addq %rax,%r10
103 movq %r10,%r11
[all …]
/third_party/openssl/crypto/bn/asm/
Dx86_64-mont5.pl79 $lo0="%r10";
127 lea -280(%rsp,$num,8),%r10 # future alloca(8*(num+2)+256+8)
129 and \$-1024,%r10 # minimize TLB usage
139 sub %r10,%r11
141 lea (%r10,%r11),%rsp
143 cmp %r10,%rsp
150 cmp %r10,%rsp
154 lea .Linc(%rip),%r10
165 movdqa 0(%r10),%xmm0 # 00000001000000010000000000000000
166 movdqa 16(%r10),%xmm1 # 00000002000000020000000200000002
[all …]
Drsaz-x86_64.pl147 movq %rdx, %r10
148 adcq \$0, %r10
151 addq %rax, %r10
194 addq %rax, %r10
241 adcq %r10, %r10
250 adcq %rdx, %r10
254 movq %r10, 24(%rsp)
303 movq 24($inp), %r10
314 mulq %r10
321 mulq %r10
[all …]
Dx86_64-mont.pl90 $lo0="%r10";
143 lea -16(%rsp,$num,8),%r10 # future alloca(8*(num+2))
145 and \$-1024,%r10 # minimize TLB usage
155 sub %r10,%r11
157 lea (%r10,%r11),%rsp
159 cmp %r10,%rsp
167 cmp %r10,%rsp
365 my @A=("%r10","%r11");
398 lea -32(%rsp,$num,8),%r10 # future alloca(8*(num+4))
400 and \$-1024,%r10 # minimize TLB usage
[all …]
/third_party/ffmpeg/libavcodec/arm/
Dvp8dsp_armv6.S61 push {r4-r10, lr}
64 mov r10, #0
68 stm r1!, {r10, lr}
71 stm r1!, {r10, lr}
74 stm r1!, {r10, lr}
77 stm r1!, {r10, lr}
125 sbfx r10, r4, #3, #13
130 sxth r10, r4
134 asr r10, #3 @ block[0][3]
143 strh r10, [r0], #32
[all …]
Dsbcdsp_armv6.S41 ldrd r10, r11, [r2, #16]
47 smlad r3, r8, r10, r3
50 ldrd r10, r11, [r2, #48]
55 smlad r3, r8, r10, r3
58 ldrd r10, r11, [r2, #8]
64 smlad r12, r8, r10, r14
67 ldrd r10, r11, [r2, #40]
72 smlad r12, r8, r10, r12
75 ldrd r10, r11, [r2, #72]
79 smlad r12, r8, r10, r12 @ t1[2] is done
[all …]
Dmpegaudiodsp_fixed_armv6.S75 sum8 r8, r9, r1, r0, r10, r11, r12, lr
76 sum8 r8, r9, r1, r2, r10, r11, r12, lr, rsb, 32
77 round r10, r8, r9
78 strh_post r10, r3, r4
84 ldr r10, [r1, #4]!
88 ldr r10, [r1, #4*64*\i]
91 smlal r8, r9, r10, r12
92 ldr r10, [r0, #4*64*(\i+1)]
101 smlal r8, r9, r12, r10
107 smlal r4, r7, r11, r10
[all …]
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/nnacl/assembly/avx/
DMatmulAvx.S89 movq 24(%rsp), %r10
102 imul %r10, %r12
105 imul %r10, %r13
108 imul %rax, %r10
282 addq %r10, %rdx
286 addq %r10, %rdx
290 addq %r10, %rdx
294 addq %r10, %rdx
298 addq %r10, %rdx
300 addq %r10, %rdx
[all …]
/third_party/openssl/crypto/chacha/asm/arm32/
Dchacha-armv4.S67 str r10,[sp,#4*(16+10)] @ off-load "rx"
80 ldr r10, [sp,#4*(13)]
92 mov r10,r10,ror#16
94 eor r10,r10,r1,ror#16
97 add r9,r9,r10
104 mov r10,r10,ror#24
106 eor r10,r10,r1,ror#24
109 add r9,r9,r10
111 str r10,[sp,#4*(16+13)]
112 ldr r10,[sp,#4*(16+15)]
[all …]
/third_party/python/Modules/_ctypes/libffi_osx/x86/
Ddarwin64.S47 movq (%rsp), %r10 /* Load return address. */
53 movq %r10, 24(%rax) /* Relocate return address. */
57 movq %r12, %r10
62 movq (%r10), %rdi
63 movq 8(%r10), %rsi
64 movq 16(%r10), %rdx
65 movq 24(%r10), %rcx
66 movq 32(%r10), %r8
67 movq 40(%r10), %r9
73 leaq 176(%r10), %rsp
[all …]
/third_party/gstreamer/gstplugins_good/gst/goom/
Dppc_zoom_ultimate.s40 ; r10 <=> int [16][16] precalccoeffs
63 mr r11,r10
82 lwz r10,4(r30) ; py2
90 sub r10,r10,r29
93 mullw r10,r10,r9
97 srawi r10,r10,16
99 add r29,r29,r10
103 rlwinm r10,r2,6,28-6,31-6 ; r10 <- (r2 << 2) & 0x000002D0 (r10=(r2%16)*4*16)
105 rlwimi r10, r29, 2, 28-2, 31-2 ; r10 <- ((r29 << 2) & 0x0000002D) | (r10 & !0x0000002D) (r10=…
116 lwzx r10,r11,r10 ; Loads coefs
[all …]
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/nnacl/assembly/arm32/
DConvDwFp32Row.S29 push {r4-r6, r8, r10, r11}
43 mov r10, r4 // input_channel
46 cmp r10, #16
48 sub r10, r10, #16
54 cmp r10, #16
72 sub r10, r10, #16
73 cmp r10, #16
89 cmp r10, #4
98 sub r10, r10, #4
99 cmp r10, #4
[all …]
DConvDw3x3Int8BorderPixel.S48 ldrb r10, [sp, #20] // in_zp
49 vdup.8 d18, r10
50 ldr r10, [sp, #24] // out_zp
51 vdup.32 q15, r10
52 ldr r10, [sp, #28] // out_multiplier
53 vdup.32 q14, r10
54 ldr r10, [sp, #32] // left_shift
55 vdup.32 q13, r10
56 ldr r10, [sp, #36] // right_shift
57 vdup.32 q12, r10
[all …]
DMatmulWinogradFp32.S59 ldr r10, [sp, #8] // in_channel
61 cmp r10, #16
63 cmp r10, #8
65 cmp r10, #4
67 cmp r10, #1
93 subs r10, r10, #16
97 cmp r10, #16
99 cmp r10, #8
101 cmp r10, #4
103 cmp r10, #1
[all …]
DMatmulInt8.S37 ldr r10, [sp, #44]
38 cmp r10, #0
47 ldr r10, [sp, #44]
48 cmp r10, #0
126 ldr r10, [sp, #44]
127 cmp r10, #0
143 ldr r10, [sp, #32]
144 ldr r11, [r10]!
150 ldr r10, [sp, #28]
151 ldr r11, [r10]
[all …]
/third_party/openssl/crypto/md5/asm/x86_64/
Dmd5-x86_64.s54 leal -680876936(%rax,%r10,1),%eax
63 leal -389564586(%rdx,%r10,1),%edx
72 leal 606105819(%rcx,%r10,1),%ecx
81 leal -1044525330(%rbx,%r10,1),%ebx
90 leal -176418897(%rax,%r10,1),%eax
99 leal 1200080426(%rdx,%r10,1),%edx
108 leal -1473231341(%rcx,%r10,1),%ecx
117 leal -45705983(%rbx,%r10,1),%ebx
126 leal 1770035416(%rax,%r10,1),%eax
135 leal -1958414417(%rdx,%r10,1),%edx
[all …]
/third_party/openssl/crypto/aes/asm/
Dvpaes-x86_64.pl87 ## Clobbers %xmm1-%xmm5, %r9, %r10, %r11, %rax
110 lea .Lk_mc_backward(%rip),%r10
123 movdqa -0x40(%r11,%r10), %xmm1 # .Lk_mc_forward[]
125 movdqa (%r11,%r10), %xmm4 # .Lk_mc_backward[]
166 movdqa -0x60(%r10), %xmm4 # 3 : sbou .Lk_sbo
167 movdqa -0x50(%r10), %xmm0 # 0 : sbot .Lk_sbo+16
171 movdqa 0x40(%r11,%r10), %xmm1 # .Lk_sr[]
200 lea .Lk_dsbd(%rip),%r10
207 add %r10, %r11
215 movdqa -0x20(%r10),%xmm4 # 4 : sb9u
[all …]
/third_party/openssl/crypto/ec/asm/x86_64/
Decp_nistz256-x86_64.s2419 movq 16(%rsi),%r10
2424 adcq %r10,%r10
2430 movq %r10,%rcx
2432 sbbq 16(%rsi),%r10
2440 cmovcq %rcx,%r10
2443 movq %r10,16(%rdi)
2474 movq 16(%rsi),%r10
2482 movq %r10,%rcx
2484 adcq 16(%rsi),%r10
2493 cmovzq %rcx,%r10
[all …]
/third_party/openssl/crypto/camellia/asm/
Dcmll-x86_64.pl525 shl \$32,%r10 # @S[2]||
527 or %r11,%r10 # ||@S[3]
530 &_saveround (2,$out,-128,"%r8","%r10"); # KA<<<0
533 &_rotl128 ("%r8","%r10",15);
534 &_saveround (6,$out,-128,"%r8","%r10"); # KA<<<15
535 &_rotl128 ("%r8","%r10",15); # 15+15=30
536 &_saveround (8,$out,-128,"%r8","%r10"); # KA<<<30
539 &_rotl128 ("%r8","%r10",15); # 30+15=45
543 &_rotl128 ("%r8","%r10",15); # 45+15=60
544 &_saveround (14,$out,-128,"%r8","%r10"); # KA<<<60
[all …]

12345678910>>...21