| /kernel/linux/linux-5.10/arch/arm64/crypto/ |
| D | chacha-neon-core.S | 4 * Copyright (C) 2016-2018 Linaro, Ltd. <ard.biesheuvel@linaro.org> 11 * ChaCha20 256-bit cipher algorithm, RFC7539, x64 SSSE3 functions 29 * chacha_permute - permute one block 31 * Permute one 64-byte block where the state matrix is stored in the four NEON 32 * registers v0-v3. It performs matrix operations on four words in parallel, 42 ld1 {v12.4s}, [x10] 45 // x0 += x1, x3 = rotl32(x3 ^ x0, 16) 46 add v0.4s, v0.4s, v1.4s 47 eor v3.16b, v3.16b, v0.16b 51 add v2.4s, v2.4s, v3.4s [all …]
|
| D | sm3-ce-core.S | 1 /* SPDX-License-Identifier: GPL-2.0-only */ 3 * sm3-ce-core.S - SM3 secure hash using ARMv8.2 Crypto Extensions 11 .irp b, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12 12 .set .Lv\b\().4s, \b 16 .inst 0xce60c000 | .L\rd | (.L\rn << 5) | (.L\rm << 16) 20 .inst 0xce60c400 | .L\rd | (.L\rn << 5) | (.L\rm << 16) 24 .inst 0xce400000 | .L\rd | (.L\rn << 5) | (.L\ra << 10) | (.L\rm << 16) 28 .inst 0xce408000 | .L\rd | (.L\rn << 5) | ((\imm2) << 12) | (.L\rm << 16) 32 .inst 0xce408400 | .L\rd | (.L\rn << 5) | ((\imm2) << 12) | (.L\rm << 16) 36 .inst 0xce408800 | .L\rd | (.L\rn << 5) | ((\imm2) << 12) | (.L\rm << 16) [all …]
|
| D | sha512-ce-core.S | 1 /* SPDX-License-Identifier: GPL-2.0 */ 3 * sha512-ce-core.S - core SHA-384/SHA-512 transform using v8 Crypto Extensions 15 .irp b,0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19 21 .inst 0xce608000 | .L\rd | (.L\rn << 5) | (.L\rm << 16) 25 .inst 0xce608400 | .L\rd | (.L\rn << 5) | (.L\rm << 16) 33 .inst 0xce608800 | .L\rd | (.L\rn << 5) | (.L\rm << 16) 37 * The SHA-512 round constants 40 .align 4 85 ld1 {v\rc1\().2d}, [x4], #16 88 ext v6.16b, v\i2\().16b, v\i3\().16b, #8 [all …]
|
| D | sha256-core.S_shipped | 1 // SPDX-License-Identifier: GPL-2.0 11 // Copyright 2014-2016 The OpenSSL Project Authors. All Rights Reserved. 30 // SHA256-hw SHA256(*) SHA512 31 // Apple A7 1.97 10.5 (+33%) 6.73 (-1%(**)) 32 // Cortex-A53 2.38 15.5 (+115%) 10.0 (+150%(***)) 33 // Cortex-A57 2.31 11.6 (+86%) 7.51 (+260%(***)) 35 // X-Gene 20.0 (+100%) 12.8 (+300%(***)) 40 // (**) The result is a trade-off: it's possible to improve it by 42 // on Cortex-A53 (or by 4 cycles per round). 43 // (***) Super-impressive coefficients over gcc-generated code are [all …]
|
| D | aes-ce-core.S | 1 /* SPDX-License-Identifier: GPL-2.0-only */ 3 * Copyright (C) 2013 - 2017 Linaro Ltd <ard.biesheuvel@linaro.org> 9 .arch armv8-a+crypto 13 ld1 {v0.16b}, [x2] 14 ld1 {v1.4s}, [x0], #16 18 mov v3.16b, v1.16b 20 0: mov v2.16b, v1.16b 21 ld1 {v3.4s}, [x0], #16 22 1: aese v0.16b, v2.16b 23 aesmc v0.16b, v0.16b [all …]
|
| D | sha2-ce-core.S | 1 /* SPDX-License-Identifier: GPL-2.0-only */ 3 * sha2-ce-core.S - core SHA-224/SHA-256 transform using v8 Crypto Extensions 12 .arch armv8-a+crypto 30 mov dg2v.16b, dg0v.16b 32 add t1.4s, v\s0\().4s, \rc\().4s 33 sha256h dg0q, dg1q, t0.4s 34 sha256h2 dg1q, dg2q, t0.4s 37 add t0.4s, v\s0\().4s, \rc\().4s 39 sha256h dg0q, dg1q, t1.4s 40 sha256h2 dg1q, dg2q, t1.4s [all …]
|
| /kernel/linux/linux-6.6/arch/arm64/crypto/ |
| D | chacha-neon-core.S | 4 * Copyright (C) 2016-2018 Linaro, Ltd. <ard.biesheuvel@linaro.org> 11 * ChaCha20 256-bit cipher algorithm, RFC7539, x64 SSSE3 functions 29 * chacha_permute - permute one block 31 * Permute one 64-byte block where the state matrix is stored in the four NEON 32 * registers v0-v3. It performs matrix operations on four words in parallel, 42 ld1 {v12.4s}, [x10] 45 // x0 += x1, x3 = rotl32(x3 ^ x0, 16) 46 add v0.4s, v0.4s, v1.4s 47 eor v3.16b, v3.16b, v0.16b 51 add v2.4s, v2.4s, v3.4s [all …]
|
| D | sm4-ce-asm.h | 1 /* SPDX-License-Identifier: GPL-2.0-or-later */ 8 ld1 {v24.16b-v27.16b}, [ptr], #64; \ 9 ld1 {v28.16b-v31.16b}, [ptr]; 12 sm4e b0.4s, v24.4s; \ 13 sm4e b0.4s, v25.4s; \ 14 sm4e b0.4s, v26.4s; \ 15 sm4e b0.4s, v27.4s; \ 16 sm4e b0.4s, v28.4s; \ 17 sm4e b0.4s, v29.4s; \ 18 sm4e b0.4s, v30.4s; \ [all …]
|
| D | sm4-neon-core.S | 1 /* SPDX-License-Identifier: GPL-2.0-or-later */ 5 * https://tools.ietf.org/id/draft-ribose-cfrg-sm4-10.html 35 ld1 {v16.16b-v19.16b}, [x5], #64; \ 36 ld1 {v20.16b-v23.16b}, [x5], #64; \ 37 ld1 {v24.16b-v27.16b}, [x5], #64; \ 38 ld1 {v28.16b-v31.16b}, [x5]; 41 zip1 RTMP0.4s, s0.4s, s1.4s; \ 42 zip1 RTMP1.4s, s2.4s, s3.4s; \ 43 zip2 RTMP2.4s, s0.4s, s1.4s; \ 44 zip2 RTMP3.4s, s2.4s, s3.4s; \ [all …]
|
| D | sm4-ce-gcm-core.S | 1 /* SPDX-License-Identifier: GPL-2.0-or-later */ 3 * SM4-GCM AEAD Algorithm using ARMv8 Crypto Extensions 14 #include "sm4-ce-asm.h" 16 .arch armv8-a+crypto 19 .set .Lv\b\().4s, \b 37 * output: r0:r1 (low 128-bits in r0, high in r1) 40 ext T0.16b, m1.16b, m1.16b, #8; \ 45 eor T0.16b, T0.16b, T1.16b; \ 46 ext T1.16b, RZERO.16b, T0.16b, #8; \ 47 ext T0.16b, T0.16b, RZERO.16b, #8; \ [all …]
|
| D | sm3-ce-core.S | 1 /* SPDX-License-Identifier: GPL-2.0-only */ 3 * sm3-ce-core.S - SM3 secure hash using ARMv8.2 Crypto Extensions 12 .irp b, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12 13 .set .Lv\b\().4s, \b 17 .inst 0xce60c000 | .L\rd | (.L\rn << 5) | (.L\rm << 16) 21 .inst 0xce60c400 | .L\rd | (.L\rn << 5) | (.L\rm << 16) 25 .inst 0xce400000 | .L\rd | (.L\rn << 5) | (.L\ra << 10) | (.L\rm << 16) 29 .inst 0xce408000 | .L\rd | (.L\rn << 5) | ((\imm2) << 12) | (.L\rm << 16) 33 .inst 0xce408400 | .L\rd | (.L\rn << 5) | ((\imm2) << 12) | (.L\rm << 16) 37 .inst 0xce408800 | .L\rd | (.L\rn << 5) | ((\imm2) << 12) | (.L\rm << 16) [all …]
|
| D | sha512-ce-core.S | 1 /* SPDX-License-Identifier: GPL-2.0 */ 3 * sha512-ce-core.S - core SHA-384/SHA-512 transform using v8 Crypto Extensions 15 .irp b,0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19 21 .inst 0xce608000 | .L\rd | (.L\rn << 5) | (.L\rm << 16) 25 .inst 0xce608400 | .L\rd | (.L\rn << 5) | (.L\rm << 16) 33 .inst 0xce608800 | .L\rd | (.L\rn << 5) | (.L\rm << 16) 37 * The SHA-512 round constants 40 .align 4 85 ld1 {v\rc1\().2d}, [x4], #16 88 ext v6.16b, v\i2\().16b, v\i3\().16b, #8 [all …]
|
| D | aes-ce-core.S | 1 /* SPDX-License-Identifier: GPL-2.0-only */ 3 * Copyright (C) 2013 - 2017 Linaro Ltd <ard.biesheuvel@linaro.org> 9 .arch armv8-a+crypto 13 ld1 {v0.16b}, [x2] 14 ld1 {v1.4s}, [x0], #16 18 mov v3.16b, v1.16b 20 0: mov v2.16b, v1.16b 21 ld1 {v3.4s}, [x0], #16 22 1: aese v0.16b, v2.16b 23 aesmc v0.16b, v0.16b [all …]
|
| D | sm4-ce-core.S | 1 /* SPDX-License-Identifier: GPL-2.0-or-later */ 5 * https://tools.ietf.org/id/draft-ribose-cfrg-sm4-10.html 13 #include "sm4-ce-asm.h" 15 .arch armv8-a+crypto 17 .irp b, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, \ 19 .set .Lv\b\().4s, \b 27 .inst 0xce60c800 | (.L\vm << 16) | (.L\vn << 5) | .L\vd 45 * x0: 128-bit key 51 ld1 {v0.16b}, [x0]; 52 rev32 v0.16b, v0.16b; [all …]
|
| D | sha2-ce-core.S | 1 /* SPDX-License-Identifier: GPL-2.0-only */ 3 * sha2-ce-core.S - core SHA-224/SHA-256 transform using v8 Crypto Extensions 12 .arch armv8-a+crypto 30 mov dg2v.16b, dg0v.16b 32 add t1.4s, v\s0\().4s, \rc\().4s 33 sha256h dg0q, dg1q, t0.4s 34 sha256h2 dg1q, dg2q, t0.4s 37 add t0.4s, v\s0\().4s, \rc\().4s 39 sha256h dg0q, dg1q, t1.4s 40 sha256h2 dg1q, dg2q, t1.4s [all …]
|
| D | aes-ce-ccm-core.S | 1 /* SPDX-License-Identifier: GPL-2.0-only */ 3 * aesce-ccm-core.S - AES-CCM transform for ARMv8 with Crypto Extensions 5 * Copyright (C) 2013 - 2017 Linaro Ltd <ard.biesheuvel@linaro.org> 12 .arch armv8-a+crypto 19 ld1 {v0.16b}, [x0] /* load mac */ 21 sub w3, w3, #16 22 eor v1.16b, v1.16b, v1.16b 27 ext v1.16b, v1.16b, v1.16b, #1 /* rotate in the input bytes */ 30 eor v0.16b, v0.16b, v1.16b 31 1: ld1 {v3.4s}, [x4] /* load first round key */ [all …]
|
| /kernel/linux/linux-6.6/tools/testing/selftests/powerpc/lib/ |
| D | reg.S | 1 /* SPDX-License-Identifier: GPL-2.0-or-later */ 7 #include <ppc-asm.h> 11 /* Non volatile GPR - unsigned long buf[18] */ 15 ld 16, 2*8(3) 17 ld 18, 4*8(3) 29 ld 30, 16*8(3) 37 std 16, 2*8(3) 39 std 18, 4*8(3) 51 std 30, 16*8(3) 56 /* Double Precision Float - double buf[32] */ [all …]
|
| /kernel/linux/linux-5.10/tools/testing/selftests/powerpc/lib/ |
| D | reg.S | 1 /* SPDX-License-Identifier: GPL-2.0-or-later */ 7 #include <ppc-asm.h> 11 /* Non volatile GPR - unsigned long buf[18] */ 15 ld 16, 2*8(3) 17 ld 18, 4*8(3) 29 ld 30, 16*8(3) 37 std 16, 2*8(3) 39 std 18, 4*8(3) 51 std 30, 16*8(3) 56 /* Single Precision Float - float buf[32] */ [all …]
|
| /kernel/linux/linux-6.6/Documentation/driver-api/media/drivers/ccs/ |
| D | ccs-regs.asc | 1 # SPDX-License-Identifier: GPL-2.0-only OR BSD-3-Clause 2 # Copyright (C) 2019--2020 Intel Corporation 5 # - f field LSB MSB rflags 6 # - e enum value # after a field 7 # - e enum value [LSB MSB] 8 # - b bool bit 9 # - l arg name min max elsize [discontig...] 12 # 8, 16, 32 register bits (default is 8) 19 module_model_id 0x0000 16 23 - e GRBG 0 [all …]
|
| /kernel/linux/linux-5.10/arch/x86/crypto/ |
| D | cast5-avx-x86_64-asm_64.S | 1 /* SPDX-License-Identifier: GPL-2.0-or-later */ 3 * Cast5 Cipher 16-way parallel algorithm (AVX/x86_64) 6 * <Johannes.Goetzfried@informatik.stud.uni-erlangen.de> 14 .file "cast5-avx-x86_64-asm_64.S" 23 #define kr (16*4) 24 #define rr ((16*4)+16) 26 /* s-boxes */ 33 16-way AVX cast5 88 shrq $16, src; \ 89 movl s1(, RID1, 4), dst ## d; \ [all …]
|
| /kernel/linux/linux-6.6/arch/x86/crypto/ |
| D | cast5-avx-x86_64-asm_64.S | 1 /* SPDX-License-Identifier: GPL-2.0-or-later */ 3 * Cast5 Cipher 16-way parallel algorithm (AVX/x86_64) 6 * <Johannes.Goetzfried@informatik.stud.uni-erlangen.de> 14 .file "cast5-avx-x86_64-asm_64.S" 23 #define kr (16*4) 24 #define rr ((16*4)+16) 26 /* s-boxes */ 33 16-way AVX cast5 88 movl (RID2,RID1,4), dst ## d; \ 91 op1 (RID1,RID2,4), dst ## d; \ [all …]
|
| /kernel/linux/linux-5.10/arch/csky/abiv1/ |
| D | memcpy.S | 1 /* SPDX-License-Identifier: GPL-2.0 */ 2 // Copyright (C) 2018 Hangzhou C-SKY Microsystems co.,ltd. 25 cmplti r4, 4 36 cmplti r4, 16 42 ldw r5, (r3, 4) 46 stw r5, (r7, 4) 49 subi r4, 16 50 addi r3, 16 51 addi r7, 16 52 cmplti r4, 16 [all …]
|
| /kernel/linux/linux-5.10/arch/alpha/lib/ |
| D | ev6-memset.S | 1 /* SPDX-License-Identifier: GPL-2.0 */ 3 * arch/alpha/lib/ev6-memset.S 8 * 21264 version contributed by Rick Gorton <rick.gorton@alpha-processor.com> 13 * ftp.digital.com/pub/Digital/info/semiconductor/literature/dsc-library.html 15 * E - either cluster 16 * U - upper subcluster; U0 - subcluster U0; U1 - subcluster U1 17 * L - lower subcluster; L0 - subcluster L0; L1 - subcluster L1 48 * undertake a major re-write to interleave the constant materialization 49 * with other parts of the fall-through code. This is important, even 55 bis $16,$16,$0 # E : return value [all …]
|
| /kernel/linux/linux-6.6/arch/alpha/lib/ |
| D | ev6-memset.S | 1 /* SPDX-License-Identifier: GPL-2.0 */ 3 * arch/alpha/lib/ev6-memset.S 8 * 21264 version contributed by Rick Gorton <rick.gorton@alpha-processor.com> 13 * ftp.digital.com/pub/Digital/info/semiconductor/literature/dsc-library.html 15 * E - either cluster 16 * U - upper subcluster; U0 - subcluster U0; U1 - subcluster U1 17 * L - lower subcluster; L0 - subcluster L0; L1 - subcluster L1 48 * undertake a major re-write to interleave the constant materialization 49 * with other parts of the fall-through code. This is important, even 55 bis $16,$16,$0 # E : return value [all …]
|
| /kernel/linux/linux-6.6/drivers/gpu/drm/vmwgfx/device_include/ |
| D | svga3d_surfacedefs.h | 2 * Copyright 2008-2021 VMware, Inc. 3 * SPDX-License-Identifier: GPL-2.0 OR MIT 28 * svga3d_surfacedefs.h -- 71 SVGA3DBLOCKDESC_BUFFER = 1 << 4, 93 SVGA3DBLOCKDESC_SRGB = 1 << 16, 335 4, 336 4, 338 { { 0 }, { 8 }, { 16 }, { 24 } } }, 343 4, 344 4, [all …]
|