/arch/x86/crypto/ |
D | sha512-avx2-asm.S | 203 xor T1, y1 # y1 = (a>>39) ^ (a>>34) # S0 204 rorx $28, a, T1 # T1 = (a >> 28) # S0 207 xor T1, y1 # y1 = (a>>39) ^ (a>>34) ^ (a>>28) # S0 213 add y1, h # h = k + w + h + S0 # -- 217 add y2, h # h = k + w + h + S0 + S1 + CH = t1 + S0# -- 218 add y3, h # h = t1 + S0 + MAJ # -- 267 xor T1, y1 # y1 = (a>>39) ^ (a>>34) # S0 269 rorx $28, a, T1 # T1 = (a >> 28) # S0 272 xor T1, y1 # y1 = (a>>39) ^ (a>>34) ^ (a>>28) # S0 278 add y1, h # h = k + w + h + S0 # -- [all …]
|
D | sha256-avx2-asm.S | 183 xor T1, y1 # y1 = (a>>22) ^ (a>>13) # S0 184 rorx $2, a, T1 # T1 = (a >> 2) # S0 188 xor T1, y1 # y1 = (a>>22) ^ (a>>13) ^ (a>>2) # S0 195 add y1, h # h = k + w + h + S0 # -- 201 add y2, h # h = k + w + h + S0 + S1 + CH = t1 + S0# -- 202 add y3, h # h = t1 + S0 + MAJ # -- 232 xor T1, y1 # y1 = (a>>22) ^ (a>>13) # S0 235 rorx $2, a, T1 # T1 = (a >> 2) # S0 239 xor T1, y1 # y1 = (a>>22) ^ (a>>13) ^ (a>>2) # S0 247 add y1, h # h = k + w + h + S0 # -- [all …]
|
D | sha256-avx-asm.S | 177 MY_ROR 2, y1 # y1 = S0 = (a>>2) ^ (a>>13) ^ (a>>22) 189 add y1, h # h = h + S1 + CH + k + w + S0 192 add y0, h # h = h + S1 + CH + k + w + S0 + MAJ 215 MY_ROR 2, y1 # y1 = S0 = (a>>2) ^ (a>>13) ^ (a>>22) 227 add y1, h # h = h + S1 + CH + k + w + S0 230 add y0, h # h = h + S1 + CH + k + w + S0 + MAJ 252 MY_ROR 2, y1 # y1 = S0 = (a>>2) ^ (a>>13) ^ (a>>22) 264 add y1, h # h = h + S1 + CH + k + w + S0 268 add y0, h # h = h + S1 + CH + k + w + S0 + MAJ 289 MY_ROR 2, y1 # y1 = S0 = (a>>2) ^ (a>>13) ^ (a>>22) [all …]
|
D | sha256-ssse3-asm.S | 172 ror $2, y1 # y1 = S0 = (a>>2) ^ (a>>13) ^ (a>>22) 185 add y1, h # h = h + S1 + CH + k + w + S0 188 add y0, h # h = h + S1 + CH + k + w + S0 + MAJ 214 ror $2, y1 # y1 = S0 = (a>>2) ^ (a>>13) ^ (a>>22) 226 add y1, h # h = h + S1 + CH + k + w + S0 229 add y0, h # h = h + S1 + CH + k + w + S0 + MAJ 254 ror $2, y1 # y1 = S0 = (a>>2) ^ (a>>13) ^ (a>>22) 266 add y1, h # h = h + S1 + CH + k + w + S0 270 add y0, h # h = h + S1 + CH + k + w + S0 + MAJ 294 ror $2, y1 # y1 = S0 = (a>>2) ^ (a>>13) ^ (a>>2 [all …]
|
D | serpent-sse2-i586-asm_32.S | 152 #define S0(x0, x1, x2, x3, x4) \ macro 531 S0(RA, RB, RC, RD, RE); LK(RC, RB, RD, RA, RE, 1); 539 S0(RC, RA, RE, RD, RB); LK(RE, RA, RD, RC, RB, 9); 547 S0(RE, RC, RB, RD, RA); LK(RB, RC, RD, RE, RA, 17); 555 S0(RB, RE, RA, RD, RC); LK(RA, RE, RD, RB, RC, 25);
|
D | serpent-sse2-x86_64-asm_64.S | 652 S(S0, RA, RB, RC, RD, RE); LK2(RC, RB, RD, RA, RE, 1); 660 S(S0, RC, RA, RE, RD, RB); LK2(RE, RA, RD, RC, RB, 9); 668 S(S0, RE, RC, RB, RD, RA); LK2(RB, RC, RD, RE, RA, 17); 676 S(S0, RB, RE, RA, RD, RC); LK2(RA, RE, RD, RB, RC, 25);
|
D | serpent-avx-x86_64-asm_64.S | 585 S(S0, RA, RB, RC, RD, RE); LK2(RC, RB, RD, RA, RE, 1); 593 S(S0, RC, RA, RE, RD, RB); LK2(RE, RA, RD, RC, RB, 9); 601 S(S0, RE, RC, RB, RD, RA); LK2(RB, RC, RD, RE, RA, 17); 609 S(S0, RB, RE, RA, RD, RC); LK2(RA, RE, RD, RB, RC, 25);
|
D | serpent-avx2-asm_64.S | 577 S(S0, RA, RB, RC, RD, RE); LK2(RC, RB, RD, RA, RE, 1); 585 S(S0, RC, RA, RE, RD, RB); LK2(RE, RA, RD, RC, RB, 9); 593 S(S0, RE, RC, RB, RD, RA); LK2(RB, RC, RD, RE, RA, 17); 601 S(S0, RB, RE, RA, RD, RC); LK2(RA, RE, RD, RB, RC, 25);
|
D | sha512-avx-asm.S | 155 RORQ tmp0, 28 # 28 # tmp = ((((a ror5)^a)ror6)^a)ror28 = S0(a) 156 add tmp0, h_64 # a(next_state) = T1 + Maj(a,b,c) S0(a)
|
D | sha512-ssse3-asm.S | 148 ror $28, tmp0 # 28 # tmp = ((((a ror5)^a)ror6)^a)ror28 = S0(a) 149 add tmp0, h_64 # a(next_state) = T1 + Maj(a,b,c) S0(a)
|
/arch/h8300/lib/ |
D | libgcc.h | 21 #define S0 r4
|
/arch/cris/arch-v32/kernel/ |
D | kgdb.c | 322 S0, S1, S2, S3, enumerator 573 } else if (regno >= S0 && regno <= S15) { in write_register() 575 …igned char *)&sreg.s0_0 + (reg.srs * 16 * sizeof(unsigned int)) + (regno - S0) * sizeof(unsigned i… in write_register() 620 } else if (regno >= S0 && regno <= S15) { in read_register() 622 …t *)((char *)&sreg.s0_0 + (reg.srs * 16 * sizeof(unsigned int)) + (regno - S0) * sizeof(unsigned i… in read_register()
|