/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/nnacl/assembly/fp16/ |
D | MatmulFp16.S | 49 mov w10, w6 // reload lhs row 331 cmp w10, #1 334 cmp w10, #2 337 cmp w10, #3 340 cmp w10, #4 343 cmp w10, #5 346 cmp w10, #6 349 cmp w10, #7 352 cmp w10, #8 355 cmp w10, #9 [all …]
|
D | PostFuncBiasReluC4Fp16.S | 38 mov w10, #0 41 cmp w10, w3 46 add w10, w10, #4
|
D | PostFuncBiasReluC8Fp16.S | 38 mov w10, #0 41 cmp w10, w3 46 add w10, w10, #8
|
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/nnacl/assembly/arm64/ |
D | MatmulFp32.S | 58 mov w10, w6 // reload lhs row 271 cmp w10, #1 275 cmp w10, #2 279 cmp w10, #3 283 cmp w10, #4 287 cmp w10, #5 291 cmp w10, #6 295 cmp w10, #7 299 cmp w10, #8 303 cmp w10, #9 [all …]
|
D | DeconvDwInt8Post.S | 43 ldr w10, [sp, #16] 44 dup v31.4s, w10 // acc_max
|
D | PostFuncBiasReluC4.S | 47 mov w10, #0 50 cmp w10, w3 55 add w10, w10, #4
|
D | PostFuncInt8C4Neon64.S | 59 ldr w10, [sp, #16] 67 dup v29.4s, w10
|
D | PostFuncBiasReluC8.S | 43 mov w10, #0 46 cmp w10, w3 51 add w10, w10, #8
|
D | MatmulInt8Opt.S | 57 ldr w10, [sp, #16] 239 dup v10.4s, w10
|
/third_party/openssl/crypto/sha/asm/arm64/ |
D | sha256-armv8.S | 208 eor w10,w20,w20,ror#14 214 eor w16,w16,w10,ror#11 // Sigma1(e) 215 ror w10,w24,#2 222 eor w17,w10,w17,ror#13 // Sigma0(a) 229 ldp w9,w10,[x1],#2*4 276 rev w10,w10 // 7 285 add w20,w20,w10 // h+=X[i] 481 str w10,[sp,#12] 488 ror w10,w21,#2 495 eor w10,w10,w21,ror#13 [all …]
|
D | sha1-armv8.S | 130 add w22,w22,w10 // future e+=X[i] 289 eor w8,w8,w10 314 eor w10,w10,w12 318 eor w10,w10,w19 322 eor w10,w10,w7 325 ror w10,w10,#31 335 add w21,w21,w10 // future e+=X[i] 358 eor w13,w13,w10 414 eor w19,w19,w10 482 eor w8,w8,w10 [all …]
|
/third_party/openssl/crypto/chacha/asm/arm64/ |
D | chacha-armv8.S | 88 add w6,w6,w10 104 eor w10,w10,w14 108 ror w10,w10,#20 112 add w6,w6,w10 128 eor w10,w10,w14 132 ror w10,w10,#25 135 add w5,w5,w10 151 eor w10,w10,w15 155 ror w10,w10,#20 159 add w5,w5,w10 [all …]
|
/third_party/ffmpeg/libavcodec/aarch64/ |
D | hevcdsp_sao_neon.S | 41 add w10, w8, w5 // x10 = k + sao_left_class 42 and w10, w10, #0x1F
|
D | h264cmc_neon.S | 35 lsr w10, w4, #1 37 lsl w10, w10, #1 38 add w9, w9, w10 205 lsr w10, w4, #1 207 lsl w10, w10, #1 208 add w9, w9, w10
|
D | h264idct_neon.S | 381 mov w10, #16 394 2: subs w10, w10, #4
|
/third_party/openh264/codec/encoder/core/arm64/ |
D | svc_motion_estimation_aarch64_neon.S | 123 ldrh w10, [x9] // sum of last line of pFeatureOfBlock[i] 136 subs w10, w10, w12 138 add w0, w10, w11 202 ldrh w10, [x9] // sum of last line of pFeatureOfBlock[i] 214 subs w10, w10, w12 216 add w0, w10, w11
|
D | intra_pred_sad_3_opt_aarch64_neon.S | 405 mov w10, w0 406 SELECT_BEST_COST_PREFER_HIGHER w10 410 sub w9, w10, w2 417 sub w8, w10, w1 432 mov w0, w10
|
/third_party/ffmpeg/libavfilter/aarch64/ |
D | vf_nlmeans_neon.S | 43 1: mov w10, w6 // width copy for each line 71 subs w10, w10, #16 // width dec
|
/third_party/ffmpeg/libswscale/aarch64/ |
D | yuv2rgb_neon.S | 26 ldp w9, w10, [sp, #\yoff] 29 ldr w10, [sp, #\ycoeff] 37 dup v0.8H, w10 55 dup v0.8H, w10 71 dup v0.8H, w10
|
D | rgb2rgb_neon.S | 63 ldrb w10, [x1], #1 65 bfi w9, w10, #8, #8
|
/third_party/openssl/crypto/aes/asm/arm64/ |
D | aesv8-armx.S | 602 add w10, w8, #1 604 rev w10, w10 605 mov v6.s[3],w10 657 add w10,w8,#2 671 rev w10,w10 674 mov v6.s[3], w10
|
/third_party/openh264/codec/processing/src/arm64/ |
D | down_sample_aarch64_neon.S | 229 mov w10, #32767 230 and w8, w6, w10 238 and w9, w7, w10
|
/third_party/glslang/Test/ |
D | stringToDouble.vert | 14 double w10 = 1234567890123456789012345.0;
|
/third_party/vk-gl-cts/framework/randomshaders/ |
D | rsgProgramExecutor.cpp | 114 float w10 = x*(1.0f-y); in interpolateVertexQuad() local 116 return quad.x()*w00 + quad.y()*w10 + quad.z()*w01 + quad.w()*w11; in interpolateVertexQuad()
|
/third_party/optimized-routines/string/aarch64/ |
D | memcpy-advsimd.S | 28 #define C_lw w10
|