| /kernel/linux/linux-6.6/arch/arm64/crypto/ |
| D | sm4-ce-asm.h | 11 #define SM4_CRYPT_BLK_BE(b0) \ argument 12 sm4e b0.4s, v24.4s; \ 13 sm4e b0.4s, v25.4s; \ 14 sm4e b0.4s, v26.4s; \ 15 sm4e b0.4s, v27.4s; \ 16 sm4e b0.4s, v28.4s; \ 17 sm4e b0.4s, v29.4s; \ 18 sm4e b0.4s, v30.4s; \ 19 sm4e b0.4s, v31.4s; \ 20 rev64 b0.4s, b0.4s; \ [all …]
|
| D | sm4-ce-gcm-core.S | 109 #define SM4_CRYPT_PMUL_128x128_BLK(b0, r0, r1, m0, m1, T0, T1) \ argument 110 rev32 b0.16b, b0.16b; \ 112 sm4e b0.4s, v24.4s; \ 114 sm4e b0.4s, v25.4s; \ 116 sm4e b0.4s, v26.4s; \ 118 sm4e b0.4s, v27.4s; \ 120 sm4e b0.4s, v28.4s; \ 122 sm4e b0.4s, v29.4s; \ 124 sm4e b0.4s, v30.4s; \ 126 sm4e b0.4s, v31.4s; \ [all …]
|
| D | sm4-neon-core.S | 131 #define SM4_CRYPT_BLK4_BE(b0, b1, b2, b3) \ argument 137 ROUND4(0, b0, b1, b2, b3); \ 138 ROUND4(1, b1, b2, b3, b0); \ 139 ROUND4(2, b2, b3, b0, b1); \ 140 ROUND4(3, b3, b0, b1, b2); \ 144 rev32 b0.16b, b0.16b; \ 149 rotate_clockwise_4x4(b0, b1, b2, b3); \ 154 #define SM4_CRYPT_BLK4(b0, b1, b2, b3) \ argument 155 rev32 b0.16b, b0.16b; \ 159 SM4_CRYPT_BLK4_BE(b0, b1, b2, b3); [all …]
|
| /kernel/linux/linux-5.10/arch/c6x/lib/ |
| D | strasgi.S | 24 || cmpltu .l2 B2, B7, B0 28 || [B0] ldw .d2t1 *B4++, A0 32 [B0] sub .d2 B6, 24, B7 33 || [B0] b .s2 _strasgi_loop 34 || cmpltu .l2 B1, B6, B0 36 [B0] ldw .d2t1 *B4++, A1 39 || cmpltu .l2 12, B6, B0 41 [B0] ldw .d2t1 *B4++, A5 44 || cmpltu .l2 8, B6, B0 46 [B0] ldw .d2t1 *B4++, A7 [all …]
|
| D | csum_64plus.S | 33 || AND .S2 3,B4,B0 34 OR .L2X B0,A1,B0 ; non aligned condition 39 [B0] BNOP .S1 L6,5 79 CMPGT .L2 B5,0,B0 80 [!B0] BNOP .S1 L82,4 112 L82: AND .S2X 1,A6,B0 113 [!B0] BNOP .S1 L9,5 136 L9: SHRU .S2X A9,16,B0 137 [!B0] BNOP .S1 L10,5 187 SHL .S2 B4,2,B0 [all …]
|
| D | remi.S | 10 ;; __c6xabi_divi A0,A1,A2,A4,A6,B0,B1,B2,B4,B5 11 ;; __c6xabi_divu A0,A1,A2,A4,A6,B0,B1,B2,B4 12 ;; __c6xabi_remi A1,A2,A4,A5,A6,B0,B1,B2,B4 13 ;; __c6xabi_remu A1,A4,A5,A7,B0,B1,B2,B4 36 || xor .s2x B2, A1, B0 39 [B0] addkpc .s2 _divu_ret_1, B3, 1 40 [!B0] addkpc .s2 _divu_ret_2, B3, 1
|
| D | remu.S | 10 ;; __c6xabi_divi A0,A1,A2,A4,A6,B0,B1,B2,B4,B5 11 ;; __c6xabi_divu A0,A1,A2,A4,A6,B0,B1,B2,B4 12 ;; __c6xabi_remi A1,A2,A4,A5,A6,B0,B1,B2,B4 13 ;; __c6xabi_remu A1,A4,A5,A7,B0,B1,B2,B4 44 cmpgt .l2 B1, 7, B0 50 || [B0] b .s1 _remu_loop
|
| D | divu.S | 10 ;; __c6xabi_divi A0,A1,A2,A4,A6,B0,B1,B2,B4,B5 11 ;; __c6xabi_divu A0,A1,A2,A4,A6,B0,B1,B2,B4 12 ;; __c6xabi_remi A1,A2,A4,A5,A6,B0,B1,B2,B4 13 ;; __c6xabi_remu A1,A4,A5,A7,B0,B1,B2,B4 59 cmpgt .l2 B1, 7, B0 64 || [B0] b .s1 _divu_loop
|
| /kernel/linux/linux-6.6/arch/s390/crypto/ |
| D | chacha-s390.S | 441 #define B0 %v1 macro 495 VLR B0,K1 526 VAF A0,A0,B0 551 VX B0,B0,C0 557 VERLLF B0,B0,12 564 VAF A0,A0,B0 589 VX B0,B0,C0 595 VERLLF B0,B0,7 608 VSLDB B0,B0,B0,4 621 VAF A0,A0,B0 [all …]
|
| /kernel/linux/linux-5.10/arch/c6x/kernel/ |
| D | entry.S | 48 STW .D2T2 B0,*SP--[2] ; save original B0 49 MVKL .S2 current_ksp,B0 50 MVKH .S2 current_ksp,B0 51 LDW .D2T2 *B0,B1 ; KSP 55 XOR .D2 SP,B1,B0 ; (SP ^ KSP) 56 LDW .D2T2 *+SP[1],B1 ; restore B0/B1 57 LDW .D2T2 *++SP[2],B0 58 SHR .S2 B0,THREAD_SHIFT,B0 ; 0 if already using kstack 59 [B0] STDW .D2T2 SP:DP,*--B1[1] ; user: save user sp/dp kstack 60 [B0] MV .S2 B1,SP ; and switch to kstack [all …]
|
| D | head.S | 24 MVKL .S2 current_ksp,B0 25 MVKH .S2 current_ksp,B0 26 LDW .D2T2 *B0,B15 29 SHR .S2 B6,3,B0 ; number of dwords to clear 33 BDEC .S2 bss_loop,B0 35 CMPLT .L2 B0,0,B1
|
| /kernel/linux/linux-6.6/include/media/ |
| D | v4l2-h264.h | 26 * ordered P/B0/B1 lists 29 * This object stores the context of the P/B0/B1 reference list builder. 55 * v4l2_h264_build_b_ref_lists() - Build the B0/B1 reference lists 58 * @b0_reflist: 32 sized array used to store the B0 reference list. Each entry 63 * This functions builds the B0/B1 reference lists. This procedure is described 66 * need to pass B0/B1 reference lists to the hardware.
|
| /kernel/linux/linux-5.10/include/media/ |
| D | v4l2-h264.h | 25 * ordered P/B0/B1 lists 28 * This object stores the context of the P/B0/B1 reference list builder. 51 * v4l2_h264_build_b_ref_lists() - Build the B0/B1 reference lists 54 * @b0_reflist: 16-bytes array used to store the B0 reference list. Each entry 59 * This functions builds the B0/B1 reference lists. This procedure is described 62 * need to pass B0/B1 reference lists to the hardware.
|
| /kernel/linux/linux-6.6/arch/arm/nwfpe/ |
| D | softfloat-macros | 339 value formed by concatenating `b0' and `b1'. Addition is modulo 2^128, so 346 bits64 a0, bits64 a1, bits64 b0, bits64 b1, bits64 *z0Ptr, bits64 *z1Ptr ) 352 *z0Ptr = a0 + b0 + ( z1 < a1 ); 359 192-bit value formed by concatenating `b0', `b1', and `b2'. Addition is 370 bits64 b0, 385 z0 = a0 + b0; 397 Subtracts the 128-bit value formed by concatenating `b0' and `b1' from the 406 bits64 a0, bits64 a1, bits64 b0, bits64 b1, bits64 *z0Ptr, bits64 *z1Ptr ) 410 *z0Ptr = a0 - b0 - ( a1 < b1 ); 416 Subtracts the 192-bit value formed by concatenating `b0', `b1', and `b2' [all …]
|
| /kernel/linux/linux-5.10/arch/arm/nwfpe/ |
| D | softfloat-macros | 339 value formed by concatenating `b0' and `b1'. Addition is modulo 2^128, so 346 bits64 a0, bits64 a1, bits64 b0, bits64 b1, bits64 *z0Ptr, bits64 *z1Ptr ) 352 *z0Ptr = a0 + b0 + ( z1 < a1 ); 359 192-bit value formed by concatenating `b0', `b1', and `b2'. Addition is 370 bits64 b0, 385 z0 = a0 + b0; 397 Subtracts the 128-bit value formed by concatenating `b0' and `b1' from the 406 bits64 a0, bits64 a1, bits64 b0, bits64 b1, bits64 *z0Ptr, bits64 *z1Ptr ) 410 *z0Ptr = a0 - b0 - ( a1 < b1 ); 416 Subtracts the 192-bit value formed by concatenating `b0', `b1', and `b2' [all …]
|
| /kernel/linux/linux-6.6/lib/crypto/ |
| D | curve25519-hacl64.c | 44 u64 b0 = b[0]; in modulo_carry_top() local 46 u64 b0_ = b0 + 19 * (b4 >> 51); in modulo_carry_top() 129 u64 b0; in fmul_shift_reduce() local 151 b0 = output[0]; in fmul_shift_reduce() 152 output[0] = 19 * b0; in fmul_shift_reduce() 190 u128 b0; in fmul_fmul() local 201 b0 = t[0]; in fmul_fmul() 203 b0_ = ((b0) + (((u128)(19) * (((u64)(((b4) >> (51)))))))); in fmul_fmul() 248 u128 b0; in fsquare_fsquare_() local 258 b0 = tmp[0]; in fsquare_fsquare_() [all …]
|
| /kernel/linux/linux-5.10/lib/crypto/ |
| D | curve25519-hacl64.c | 46 u64 b0 = b[0]; in modulo_carry_top() local 48 u64 b0_ = b0 + 19 * (b4 >> 51); in modulo_carry_top() 131 u64 b0; in fmul_shift_reduce() local 153 b0 = output[0]; in fmul_shift_reduce() 154 output[0] = 19 * b0; in fmul_shift_reduce() 192 u128 b0; in fmul_fmul() local 203 b0 = t[0]; in fmul_fmul() 205 b0_ = ((b0) + (((u128)(19) * (((u64)(((b4) >> (51)))))))); in fmul_fmul() 250 u128 b0; in fsquare_fsquare_() local 260 b0 = tmp[0]; in fsquare_fsquare_() [all …]
|
| /kernel/linux/linux-6.6/drivers/crypto/nx/ |
| D | nx-aes-ccm.c | 134 unsigned int cryptlen, u8 *b0) in generate_b0() argument 138 memcpy(b0, iv, 16); in generate_b0() 140 lp = b0[0]; in generate_b0() 144 *b0 |= (8 * ((m - 2) / 2)); in generate_b0() 148 *b0 |= 64; in generate_b0() 150 return set_msg_len(b0 + 16 - l, cryptlen, l); in generate_b0() 164 u8 tmp[16], *b1 = NULL, *b0 = NULL, *result = NULL; in generate_pat() local 186 b0 = nx_ctx->csbcpb->cpb.aes_ccm.in_pat_or_b0; in generate_pat() 189 * operation on 2 AES blocks, B0 (stored in the csbcpb) and B1, in generate_pat() 191 b0 = nx_ctx->csbcpb->cpb.aes_ccm.in_pat_or_b0; in generate_pat() [all …]
|
| /kernel/linux/linux-5.10/drivers/crypto/nx/ |
| D | nx-aes-ccm.c | 134 unsigned int cryptlen, u8 *b0) in generate_b0() argument 139 memcpy(b0, iv, 16); in generate_b0() 141 lp = b0[0]; in generate_b0() 145 *b0 |= (8 * ((m - 2) / 2)); in generate_b0() 149 *b0 |= 64; in generate_b0() 151 rc = set_msg_len(b0 + 16 - l, cryptlen, l); in generate_b0() 167 u8 tmp[16], *b1 = NULL, *b0 = NULL, *result = NULL; in generate_pat() local 189 b0 = nx_ctx->csbcpb->cpb.aes_ccm.in_pat_or_b0; in generate_pat() 192 * operation on 2 AES blocks, B0 (stored in the csbcpb) and B1, in generate_pat() 194 b0 = nx_ctx->csbcpb->cpb.aes_ccm.in_pat_or_b0; in generate_pat() [all …]
|
| /kernel/linux/linux-6.6/drivers/net/wireless/intel/iwlwifi/cfg/ |
| D | ax210.c | 29 #define IWL_SO_A_JF_B_FW_PRE "iwlwifi-so-a0-jf-b0" 30 #define IWL_SO_A_HR_B_FW_PRE "iwlwifi-so-a0-hr-b0" 35 #define IWL_MA_A_HR_B_FW_PRE "iwlwifi-ma-a0-hr-b0" 39 #define IWL_MA_B_HR_B_FW_PRE "iwlwifi-ma-b0-hr-b0" 40 #define IWL_MA_B_GF_A_FW_PRE "iwlwifi-ma-b0-gf-a0" 41 #define IWL_MA_B_GF4_A_FW_PRE "iwlwifi-ma-b0-gf4-a0" 42 #define IWL_MA_B_MR_A_FW_PRE "iwlwifi-ma-b0-mr-a0" 306 MODULE_FIRMWARE("iwlwifi-ma-b0-gf-a0.pnvm"); 307 MODULE_FIRMWARE("iwlwifi-ma-b0-gf4-a0.pnvm");
|
| /kernel/linux/linux-6.6/drivers/gpu/drm/i915/ |
| D | intel_step.c | 37 [1] = { COMMON_GT_MEDIA_STEP(B0), .display_step = STEP_B0 }, 54 [3] = { COMMON_STEP(B0) }, 63 [1] = { COMMON_STEP(B0) }, 68 [1] = { COMMON_GT_MEDIA_STEP(B0), .display_step = STEP_C0 }, 76 [1] = { COMMON_GT_MEDIA_STEP(B0), .display_step = STEP_D0 }, 81 [1] = { COMMON_STEP(B0) }, 87 [1] = { COMMON_STEP(B0) }, 93 [0x4] = { COMMON_GT_MEDIA_STEP(B0), .display_step = STEP_B0 }, 100 [0x4] = { COMMON_GT_MEDIA_STEP(B0), .display_step = STEP_B0 }, 108 [0x4] = { COMMON_GT_MEDIA_STEP(B0) }, [all …]
|
| /kernel/linux/linux-6.6/arch/ia64/lib/ |
| D | ip_fast_csum.S | 77 br.ret.sptk.many b0 85 mov r34=b0 90 br.call.sptk.many b0=do_csum 94 mov b0=r34 95 br.ret.sptk.many b0 146 br.ret.sptk.many b0
|
| /kernel/linux/linux-5.10/arch/ia64/lib/ |
| D | ip_fast_csum.S | 77 br.ret.sptk.many b0 85 mov r34=b0 90 br.call.sptk.many b0=do_csum 94 mov b0=r34 95 br.ret.sptk.many b0 146 br.ret.sptk.many b0
|
| /kernel/linux/linux-5.10/fs/f2fs/ |
| D | hash.c | 28 __u32 b0 = buf[0], b1 = buf[1]; in TEA_transform() local 34 b0 += ((b1 << 4)+a) ^ (b1+sum) ^ ((b1 >> 5)+b); in TEA_transform() 35 b1 += ((b0 << 4)+c) ^ (b0+sum) ^ ((b0 >> 5)+d); in TEA_transform() 38 buf[0] += b0; in TEA_transform()
|
| /kernel/linux/linux-5.10/arch/arm/crypto/ |
| D | ghash-ce-core.S | 226 vmull.p64 XL, HH4_L, XL_L // a0 * b0 227 vmull.p64 XM, HH34_H, XL2_H // (a1 + a0)(b1 + b0) 231 vmull.p64 XL2, HH3_L, XM2_H // a0 * b0 232 vmull.p64 XM2, HH34_L, XM2_L // (a1 + a0)(b1 + b0) 240 vmull.p64 XL2, HH_L, T3_H // a0 * b0 241 vmull.p64 XM2, SHASH2_H, T3_L // (a1 + a0)(b1 + b0) 249 vmull.p64 XL2, SHASH_L, T1_H // a0 * b0 250 vmull.p64 XM2, SHASH2_p64, T1_L // (a1 + a0)(b1 + b0) 284 __pmull_\pn XL, XL_L, SHASH_L, s1l, s2l, s3l, s4l @ a0 * b0 285 __pmull_\pn XM, T1_L, SHASH2_\pn @ (a1+a0)(b1+b0)
|