/third_party/libffi/src/tile/ |
D | tile.S | 83 #define FRAME_SIZE r10 253 SW sp, r10 254 .cfi_return_column r10 255 .cfi_offset r10, 0 258 addli r10, sp, -(CLOSURE_FRAME_SIZE - REG_SIZE) 262 SW r10, sp 271 addi r10, sp, LINKAGE_SIZE 275 STORE_REG(r0, r10) 276 STORE_REG(r1, r10) 277 STORE_REG(r2, r10) [all …]
|
/third_party/skia/tests/sksl/runtime/ |
D | Switch.skvm | 12 10 r10 = trunc r1 13 11 r8 = eq_i32 r8 r10 21 19 r10 = eq_i32 r16 r10 22 20 r10 = bit_or r8 r10 23 21 r10 = bit_and r10 r15 24 22 r11 = select r10 r0 r11 25 23 r12 = select r10 r1 r12 26 24 r13 = select r10 r2 r13 27 25 r14 = select r10 r3 r14 28 26 r10 = bit_xor r9 r10 [all …]
|
D | SwitchWithLoops.skvm | 12 10 r10 = splat 1 (1.4012985e-45) 13 11 r9 = eq_i32 r10 r9 14 12 r11 = bit_and r10 r9 16 14 r13 = add_i32 r11 r10 19 17 r13 = add_i32 r11 r10 23 21 r12 = add_i32 r11 r10 27 25 r15 = add_i32 r11 r10 31 29 r14 = add_i32 r11 r10 34 32 r14 = add_i32 r11 r10 38 36 r13 = add_i32 r11 r10 [all …]
|
/third_party/openssl/crypto/sha/asm/arm32/ |
D | sha1-armv4-large.S | 41 ldrb r10,[r1,#2] 46 orr r9,r9,r10,lsl#8 47 eor r10,r5,r6 @ F_xx_xx 54 eor r10,r5,r6 @ F_xx_xx 60 and r10,r4,r10,ror#2 62 eor r10,r10,r6,ror#2 @ F_00_19(B,C,D) 64 add r7,r7,r10 @ E+=F_00_19(B,C,D) 66 ldrb r10,[r1,#2] 71 orr r9,r9,r10,lsl#8 72 eor r10,r4,r5 @ F_xx_xx [all …]
|
D | keccak1600-armv4.S | 46 add r10,sp,#40 57 ldmia r10,{r10-r12,r14} @ A[1][0..1] 59 eor r0,r0,r10 62 ldrd r10,r11,[sp,#56] 65 eor r4,r4,r10 68 ldrd r10,r11,[sp,#72] 71 eor r8,r8,r10 74 ldrd r10,r11,[sp,#88] 77 eor r2,r2,r10 80 ldrd r10,r11,[sp,#104] [all …]
|
/third_party/openssl/crypto/bn/asm/x86_64/ |
D | x86_64-mont5.s | 36 leaq -280(%rsp,%r9,8),%r10 38 andq $-1024,%r10 48 subq %r10,%r11 50 leaq (%r10,%r11,1),%rsp 52 cmpq %r10,%rsp 59 cmpq %r10,%rsp 63 leaq .Linc(%rip),%r10 69 movdqa 0(%r10),%xmm0 70 movdqa 16(%r10),%xmm1 71 leaq 24-112(%rsp,%r9,8),%r10 [all …]
|
D | x86_64-mont.s | 41 leaq -16(%rsp,%r9,8),%r10 43 andq $-1024,%r10 53 subq %r10,%r11 55 leaq (%r10,%r11,1),%rsp 57 cmpq %r10,%rsp 65 cmpq %r10,%rsp 82 movq %rax,%r10 85 imulq %r10,%rbp 89 addq %rax,%r10 103 movq %r10,%r11 [all …]
|
/third_party/openssl/crypto/bn/asm/ |
D | x86_64-mont5.pl | 79 $lo0="%r10"; 127 lea -280(%rsp,$num,8),%r10 # future alloca(8*(num+2)+256+8) 129 and \$-1024,%r10 # minimize TLB usage 139 sub %r10,%r11 141 lea (%r10,%r11),%rsp 143 cmp %r10,%rsp 150 cmp %r10,%rsp 154 lea .Linc(%rip),%r10 165 movdqa 0(%r10),%xmm0 # 00000001000000010000000000000000 166 movdqa 16(%r10),%xmm1 # 00000002000000020000000200000002 [all …]
|
D | rsaz-x86_64.pl | 147 movq %rdx, %r10 148 adcq \$0, %r10 151 addq %rax, %r10 194 addq %rax, %r10 241 adcq %r10, %r10 250 adcq %rdx, %r10 254 movq %r10, 24(%rsp) 303 movq 24($inp), %r10 314 mulq %r10 321 mulq %r10 [all …]
|
D | x86_64-mont.pl | 90 $lo0="%r10"; 143 lea -16(%rsp,$num,8),%r10 # future alloca(8*(num+2)) 145 and \$-1024,%r10 # minimize TLB usage 155 sub %r10,%r11 157 lea (%r10,%r11),%rsp 159 cmp %r10,%rsp 167 cmp %r10,%rsp 365 my @A=("%r10","%r11"); 398 lea -32(%rsp,$num,8),%r10 # future alloca(8*(num+4)) 400 and \$-1024,%r10 # minimize TLB usage [all …]
|
/third_party/ffmpeg/libavcodec/arm/ |
D | vp8dsp_armv6.S | 61 push {r4-r10, lr} 64 mov r10, #0 68 stm r1!, {r10, lr} 71 stm r1!, {r10, lr} 74 stm r1!, {r10, lr} 77 stm r1!, {r10, lr} 125 sbfx r10, r4, #3, #13 130 sxth r10, r4 134 asr r10, #3 @ block[0][3] 143 strh r10, [r0], #32 [all …]
|
D | sbcdsp_armv6.S | 41 ldrd r10, r11, [r2, #16] 47 smlad r3, r8, r10, r3 50 ldrd r10, r11, [r2, #48] 55 smlad r3, r8, r10, r3 58 ldrd r10, r11, [r2, #8] 64 smlad r12, r8, r10, r14 67 ldrd r10, r11, [r2, #40] 72 smlad r12, r8, r10, r12 75 ldrd r10, r11, [r2, #72] 79 smlad r12, r8, r10, r12 @ t1[2] is done [all …]
|
D | mpegaudiodsp_fixed_armv6.S | 75 sum8 r8, r9, r1, r0, r10, r11, r12, lr 76 sum8 r8, r9, r1, r2, r10, r11, r12, lr, rsb, 32 77 round r10, r8, r9 78 strh_post r10, r3, r4 84 ldr r10, [r1, #4]! 88 ldr r10, [r1, #4*64*\i] 91 smlal r8, r9, r10, r12 92 ldr r10, [r0, #4*64*(\i+1)] 101 smlal r8, r9, r12, r10 107 smlal r4, r7, r11, r10 [all …]
|
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/nnacl/assembly/avx/ |
D | MatmulAvx.S | 89 movq 24(%rsp), %r10 102 imul %r10, %r12 105 imul %r10, %r13 108 imul %rax, %r10 282 addq %r10, %rdx 286 addq %r10, %rdx 290 addq %r10, %rdx 294 addq %r10, %rdx 298 addq %r10, %rdx 300 addq %r10, %rdx [all …]
|
/third_party/openssl/crypto/chacha/asm/arm32/ |
D | chacha-armv4.S | 67 str r10,[sp,#4*(16+10)] @ off-load "rx" 80 ldr r10, [sp,#4*(13)] 92 mov r10,r10,ror#16 94 eor r10,r10,r1,ror#16 97 add r9,r9,r10 104 mov r10,r10,ror#24 106 eor r10,r10,r1,ror#24 109 add r9,r9,r10 111 str r10,[sp,#4*(16+13)] 112 ldr r10,[sp,#4*(16+15)] [all …]
|
/third_party/python/Modules/_ctypes/libffi_osx/x86/ |
D | darwin64.S | 47 movq (%rsp), %r10 /* Load return address. */ 53 movq %r10, 24(%rax) /* Relocate return address. */ 57 movq %r12, %r10 62 movq (%r10), %rdi 63 movq 8(%r10), %rsi 64 movq 16(%r10), %rdx 65 movq 24(%r10), %rcx 66 movq 32(%r10), %r8 67 movq 40(%r10), %r9 73 leaq 176(%r10), %rsp [all …]
|
/third_party/gstreamer/gstplugins_good/gst/goom/ |
D | ppc_zoom_ultimate.s | 40 ; r10 <=> int [16][16] precalccoeffs 63 mr r11,r10 82 lwz r10,4(r30) ; py2 90 sub r10,r10,r29 93 mullw r10,r10,r9 97 srawi r10,r10,16 99 add r29,r29,r10 103 rlwinm r10,r2,6,28-6,31-6 ; r10 <- (r2 << 2) & 0x000002D0 (r10=(r2%16)*4*16) 105 rlwimi r10, r29, 2, 28-2, 31-2 ; r10 <- ((r29 << 2) & 0x0000002D) | (r10 & !0x0000002D) (r10=… 116 lwzx r10,r11,r10 ; Loads coefs [all …]
|
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/nnacl/assembly/arm32/ |
D | ConvDwFp32Row.S | 29 push {r4-r6, r8, r10, r11} 43 mov r10, r4 // input_channel 46 cmp r10, #16 48 sub r10, r10, #16 54 cmp r10, #16 72 sub r10, r10, #16 73 cmp r10, #16 89 cmp r10, #4 98 sub r10, r10, #4 99 cmp r10, #4 [all …]
|
D | ConvDw3x3Int8BorderPixel.S | 48 ldrb r10, [sp, #20] // in_zp 49 vdup.8 d18, r10 50 ldr r10, [sp, #24] // out_zp 51 vdup.32 q15, r10 52 ldr r10, [sp, #28] // out_multiplier 53 vdup.32 q14, r10 54 ldr r10, [sp, #32] // left_shift 55 vdup.32 q13, r10 56 ldr r10, [sp, #36] // right_shift 57 vdup.32 q12, r10 [all …]
|
D | MatmulWinogradFp32.S | 59 ldr r10, [sp, #8] // in_channel 61 cmp r10, #16 63 cmp r10, #8 65 cmp r10, #4 67 cmp r10, #1 93 subs r10, r10, #16 97 cmp r10, #16 99 cmp r10, #8 101 cmp r10, #4 103 cmp r10, #1 [all …]
|
D | MatmulInt8.S | 37 ldr r10, [sp, #44] 38 cmp r10, #0 47 ldr r10, [sp, #44] 48 cmp r10, #0 126 ldr r10, [sp, #44] 127 cmp r10, #0 143 ldr r10, [sp, #32] 144 ldr r11, [r10]! 150 ldr r10, [sp, #28] 151 ldr r11, [r10] [all …]
|
/third_party/openssl/crypto/md5/asm/x86_64/ |
D | md5-x86_64.s | 54 leal -680876936(%rax,%r10,1),%eax 63 leal -389564586(%rdx,%r10,1),%edx 72 leal 606105819(%rcx,%r10,1),%ecx 81 leal -1044525330(%rbx,%r10,1),%ebx 90 leal -176418897(%rax,%r10,1),%eax 99 leal 1200080426(%rdx,%r10,1),%edx 108 leal -1473231341(%rcx,%r10,1),%ecx 117 leal -45705983(%rbx,%r10,1),%ebx 126 leal 1770035416(%rax,%r10,1),%eax 135 leal -1958414417(%rdx,%r10,1),%edx [all …]
|
/third_party/openssl/crypto/aes/asm/ |
D | vpaes-x86_64.pl | 87 ## Clobbers %xmm1-%xmm5, %r9, %r10, %r11, %rax 110 lea .Lk_mc_backward(%rip),%r10 123 movdqa -0x40(%r11,%r10), %xmm1 # .Lk_mc_forward[] 125 movdqa (%r11,%r10), %xmm4 # .Lk_mc_backward[] 166 movdqa -0x60(%r10), %xmm4 # 3 : sbou .Lk_sbo 167 movdqa -0x50(%r10), %xmm0 # 0 : sbot .Lk_sbo+16 171 movdqa 0x40(%r11,%r10), %xmm1 # .Lk_sr[] 200 lea .Lk_dsbd(%rip),%r10 207 add %r10, %r11 215 movdqa -0x20(%r10),%xmm4 # 4 : sb9u [all …]
|
/third_party/openssl/crypto/ec/asm/x86_64/ |
D | ecp_nistz256-x86_64.s | 2419 movq 16(%rsi),%r10 2424 adcq %r10,%r10 2430 movq %r10,%rcx 2432 sbbq 16(%rsi),%r10 2440 cmovcq %rcx,%r10 2443 movq %r10,16(%rdi) 2474 movq 16(%rsi),%r10 2482 movq %r10,%rcx 2484 adcq 16(%rsi),%r10 2493 cmovzq %rcx,%r10 [all …]
|
/third_party/openssl/crypto/camellia/asm/ |
D | cmll-x86_64.pl | 525 shl \$32,%r10 # @S[2]|| 527 or %r11,%r10 # ||@S[3] 530 &_saveround (2,$out,-128,"%r8","%r10"); # KA<<<0 533 &_rotl128 ("%r8","%r10",15); 534 &_saveround (6,$out,-128,"%r8","%r10"); # KA<<<15 535 &_rotl128 ("%r8","%r10",15); # 15+15=30 536 &_saveround (8,$out,-128,"%r8","%r10"); # KA<<<30 539 &_rotl128 ("%r8","%r10",15); # 30+15=45 543 &_rotl128 ("%r8","%r10",15); # 45+15=60 544 &_saveround (14,$out,-128,"%r8","%r10"); # KA<<<60 [all …]
|