• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /* Copyright (C) 1995-1998 Eric Young (eay@cryptsoft.com)
2  * All rights reserved.
3  *
4  * This package is an SSL implementation written
5  * by Eric Young (eay@cryptsoft.com).
6  * The implementation was written so as to conform with Netscapes SSL.
7  *
8  * This library is free for commercial and non-commercial use as long as
9  * the following conditions are aheared to.  The following conditions
10  * apply to all code found in this distribution, be it the RC4, RSA,
11  * lhash, DES, etc., code; not just the SSL code.  The SSL documentation
12  * included with this distribution is covered by the same copyright terms
13  * except that the holder is Tim Hudson (tjh@cryptsoft.com).
14  *
15  * Copyright remains Eric Young's, and as such any Copyright notices in
16  * the code are not to be removed.
17  * If this package is used in a product, Eric Young should be given attribution
18  * as the author of the parts of the library used.
19  * This can be in the form of a textual message at program startup or
20  * in documentation (online or textual) provided with the package.
21  *
22  * Redistribution and use in source and binary forms, with or without
23  * modification, are permitted provided that the following conditions
24  * are met:
25  * 1. Redistributions of source code must retain the copyright
26  *    notice, this list of conditions and the following disclaimer.
27  * 2. Redistributions in binary form must reproduce the above copyright
28  *    notice, this list of conditions and the following disclaimer in the
29  *    documentation and/or other materials provided with the distribution.
30  * 3. All advertising materials mentioning features or use of this software
31  *    must display the following acknowledgement:
32  *    "This product includes cryptographic software written by
33  *     Eric Young (eay@cryptsoft.com)"
34  *    The word 'cryptographic' can be left out if the rouines from the library
35  *    being used are not cryptographic related :-).
36  * 4. If you include any Windows specific code (or a derivative thereof) from
37  *    the apps directory (application code) you must include an acknowledgement:
38  *    "This product includes software written by Tim Hudson (tjh@cryptsoft.com)"
39  *
40  * THIS SOFTWARE IS PROVIDED BY ERIC YOUNG ``AS IS'' AND
41  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
42  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
43  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
44  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
45  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
46  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
47  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
48  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
49  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
50  * SUCH DAMAGE.
51  *
52  * The licence and distribution terms for any publically available version or
53  * derivative of this code cannot be changed.  i.e. this code cannot simply be
54  * copied and put under another distribution licence
55  * [including the GNU Public Licence.] */
56 
57 #include <openssl/sha.h>
58 
59 #include <string.h>
60 
61 #include <openssl/mem.h>
62 
63 #include "../../internal.h"
64 #include "../digest/md32_common.h"
65 #include "internal.h"
66 
67 
SHA1_Init(SHA_CTX * sha)68 int SHA1_Init(SHA_CTX *sha) {
69   OPENSSL_memset(sha, 0, sizeof(SHA_CTX));
70   sha->h[0] = 0x67452301UL;
71   sha->h[1] = 0xefcdab89UL;
72   sha->h[2] = 0x98badcfeUL;
73   sha->h[3] = 0x10325476UL;
74   sha->h[4] = 0xc3d2e1f0UL;
75   return 1;
76 }
77 
SHA1(const uint8_t * data,size_t len,uint8_t out[SHA_DIGEST_LENGTH])78 uint8_t *SHA1(const uint8_t *data, size_t len, uint8_t out[SHA_DIGEST_LENGTH]) {
79   SHA_CTX ctx;
80   SHA1_Init(&ctx);
81   SHA1_Update(&ctx, data, len);
82   SHA1_Final(out, &ctx);
83   OPENSSL_cleanse(&ctx, sizeof(ctx));
84   return out;
85 }
86 
87 #if !defined(SHA1_ASM)
88 static void sha1_block_data_order(uint32_t *state, const uint8_t *data,
89                                   size_t num);
90 #endif
91 
SHA1_Transform(SHA_CTX * c,const uint8_t data[SHA_CBLOCK])92 void SHA1_Transform(SHA_CTX *c, const uint8_t data[SHA_CBLOCK]) {
93   sha1_block_data_order(c->h, data, 1);
94 }
95 
SHA1_Update(SHA_CTX * c,const void * data,size_t len)96 int SHA1_Update(SHA_CTX *c, const void *data, size_t len) {
97   crypto_md32_update(&sha1_block_data_order, c->h, c->data, SHA_CBLOCK, &c->num,
98                      &c->Nh, &c->Nl, data, len);
99   return 1;
100 }
101 
SHA1_Final(uint8_t out[SHA_DIGEST_LENGTH],SHA_CTX * c)102 int SHA1_Final(uint8_t out[SHA_DIGEST_LENGTH], SHA_CTX *c) {
103   crypto_md32_final(&sha1_block_data_order, c->h, c->data, SHA_CBLOCK, &c->num,
104                     c->Nh, c->Nl, /*is_big_endian=*/1);
105 
106   CRYPTO_store_u32_be(out, c->h[0]);
107   CRYPTO_store_u32_be(out + 4, c->h[1]);
108   CRYPTO_store_u32_be(out + 8, c->h[2]);
109   CRYPTO_store_u32_be(out + 12, c->h[3]);
110   CRYPTO_store_u32_be(out + 16, c->h[4]);
111   return 1;
112 }
113 
114 #define ROTATE(a, n) (((a) << (n)) | ((a) >> (32 - (n))))
115 #define Xupdate(a, ix, ia, ib, ic, id) \
116   do {                                 \
117     (a) = ((ia) ^ (ib) ^ (ic) ^ (id)); \
118     (ix) = (a) = ROTATE((a), 1);       \
119   } while (0)
120 
121 #define K_00_19 0x5a827999UL
122 #define K_20_39 0x6ed9eba1UL
123 #define K_40_59 0x8f1bbcdcUL
124 #define K_60_79 0xca62c1d6UL
125 
126 // As  pointed out by Wei Dai <weidai@eskimo.com>, F() below can be simplified
127 // to the code in F_00_19.  Wei attributes these optimisations to Peter
128 // Gutmann's SHS code, and he attributes it to Rich Schroeppel. #define
129 // F(x,y,z) (((x) & (y))  |  ((~(x)) & (z))) I've just become aware of another
130 // tweak to be made, again from Wei Dai, in F_40_59, (x&a)|(y&a) -> (x|y)&a
131 #define F_00_19(b, c, d) ((((c) ^ (d)) & (b)) ^ (d))
132 #define F_20_39(b, c, d) ((b) ^ (c) ^ (d))
133 #define F_40_59(b, c, d) (((b) & (c)) | (((b) | (c)) & (d)))
134 #define F_60_79(b, c, d) F_20_39(b, c, d)
135 
136 #define BODY_00_15(i, a, b, c, d, e, f, xi)                               \
137   do {                                                                    \
138     (f) = (xi) + (e) + K_00_19 + ROTATE((a), 5) + F_00_19((b), (c), (d)); \
139     (b) = ROTATE((b), 30);                                                \
140   } while (0)
141 
142 #define BODY_16_19(i, a, b, c, d, e, f, xi, xa, xb, xc, xd)         \
143   do {                                                              \
144     Xupdate(f, xi, xa, xb, xc, xd);                                 \
145     (f) += (e) + K_00_19 + ROTATE((a), 5) + F_00_19((b), (c), (d)); \
146     (b) = ROTATE((b), 30);                                          \
147   } while (0)
148 
149 #define BODY_20_31(i, a, b, c, d, e, f, xi, xa, xb, xc, xd)         \
150   do {                                                              \
151     Xupdate(f, xi, xa, xb, xc, xd);                                 \
152     (f) += (e) + K_20_39 + ROTATE((a), 5) + F_20_39((b), (c), (d)); \
153     (b) = ROTATE((b), 30);                                          \
154   } while (0)
155 
156 #define BODY_32_39(i, a, b, c, d, e, f, xa, xb, xc, xd)             \
157   do {                                                              \
158     Xupdate(f, xa, xa, xb, xc, xd);                                 \
159     (f) += (e) + K_20_39 + ROTATE((a), 5) + F_20_39((b), (c), (d)); \
160     (b) = ROTATE((b), 30);                                          \
161   } while (0)
162 
163 #define BODY_40_59(i, a, b, c, d, e, f, xa, xb, xc, xd)             \
164   do {                                                              \
165     Xupdate(f, xa, xa, xb, xc, xd);                                 \
166     (f) += (e) + K_40_59 + ROTATE((a), 5) + F_40_59((b), (c), (d)); \
167     (b) = ROTATE((b), 30);                                          \
168   } while (0)
169 
170 #define BODY_60_79(i, a, b, c, d, e, f, xa, xb, xc, xd)                   \
171   do {                                                                    \
172     Xupdate(f, xa, xa, xb, xc, xd);                                       \
173     (f) = (xa) + (e) + K_60_79 + ROTATE((a), 5) + F_60_79((b), (c), (d)); \
174     (b) = ROTATE((b), 30);                                                \
175   } while (0)
176 
177 #ifdef X
178 #undef X
179 #endif
180 
181 /* Originally X was an array. As it's automatic it's natural
182 * to expect RISC compiler to accomodate at least part of it in
183 * the register bank, isn't it? Unfortunately not all compilers
184 * "find" this expectation reasonable:-( On order to make such
185 * compilers generate better code I replace X[] with a bunch of
186 * X0, X1, etc. See the function body below...
187 *         <appro@fy.chalmers.se> */
188 #define X(i)  XX##i
189 
190 #if !defined(SHA1_ASM)
sha1_block_data_order(uint32_t * state,const uint8_t * data,size_t num)191 static void sha1_block_data_order(uint32_t *state, const uint8_t *data,
192                                   size_t num) {
193   register uint32_t A, B, C, D, E, T;
194   uint32_t XX0, XX1, XX2, XX3, XX4, XX5, XX6, XX7, XX8, XX9, XX10,
195       XX11, XX12, XX13, XX14, XX15;
196 
197   A = state[0];
198   B = state[1];
199   C = state[2];
200   D = state[3];
201   E = state[4];
202 
203   for (;;) {
204     X(0) = CRYPTO_load_u32_be(data);
205     data += 4;
206     X(1) = CRYPTO_load_u32_be(data);
207     data += 4;
208     BODY_00_15(0, A, B, C, D, E, T, X(0));
209     X(2) = CRYPTO_load_u32_be(data);
210     data += 4;
211     BODY_00_15(1, T, A, B, C, D, E, X(1));
212     X(3) = CRYPTO_load_u32_be(data);
213     data += 4;
214     BODY_00_15(2, E, T, A, B, C, D, X(2));
215     X(4) = CRYPTO_load_u32_be(data);
216     data += 4;
217     BODY_00_15(3, D, E, T, A, B, C, X(3));
218     X(5) = CRYPTO_load_u32_be(data);
219     data += 4;
220     BODY_00_15(4, C, D, E, T, A, B, X(4));
221     X(6) = CRYPTO_load_u32_be(data);
222     data += 4;
223     BODY_00_15(5, B, C, D, E, T, A, X(5));
224     X(7) = CRYPTO_load_u32_be(data);
225     data += 4;
226     BODY_00_15(6, A, B, C, D, E, T, X(6));
227     X(8) = CRYPTO_load_u32_be(data);
228     data += 4;
229     BODY_00_15(7, T, A, B, C, D, E, X(7));
230     X(9) = CRYPTO_load_u32_be(data);
231     data += 4;
232     BODY_00_15(8, E, T, A, B, C, D, X(8));
233     X(10) = CRYPTO_load_u32_be(data);
234     data += 4;
235     BODY_00_15(9, D, E, T, A, B, C, X(9));
236     X(11) = CRYPTO_load_u32_be(data);
237     data += 4;
238     BODY_00_15(10, C, D, E, T, A, B, X(10));
239     X(12) = CRYPTO_load_u32_be(data);
240     data += 4;
241     BODY_00_15(11, B, C, D, E, T, A, X(11));
242     X(13) = CRYPTO_load_u32_be(data);
243     data += 4;
244     BODY_00_15(12, A, B, C, D, E, T, X(12));
245     X(14) = CRYPTO_load_u32_be(data);
246     data += 4;
247     BODY_00_15(13, T, A, B, C, D, E, X(13));
248     X(15) = CRYPTO_load_u32_be(data);
249     data += 4;
250     BODY_00_15(14, E, T, A, B, C, D, X(14));
251     BODY_00_15(15, D, E, T, A, B, C, X(15));
252 
253     BODY_16_19(16, C, D, E, T, A, B, X(0), X(0), X(2), X(8), X(13));
254     BODY_16_19(17, B, C, D, E, T, A, X(1), X(1), X(3), X(9), X(14));
255     BODY_16_19(18, A, B, C, D, E, T, X(2), X(2), X(4), X(10), X(15));
256     BODY_16_19(19, T, A, B, C, D, E, X(3), X(3), X(5), X(11), X(0));
257 
258     BODY_20_31(20, E, T, A, B, C, D, X(4), X(4), X(6), X(12), X(1));
259     BODY_20_31(21, D, E, T, A, B, C, X(5), X(5), X(7), X(13), X(2));
260     BODY_20_31(22, C, D, E, T, A, B, X(6), X(6), X(8), X(14), X(3));
261     BODY_20_31(23, B, C, D, E, T, A, X(7), X(7), X(9), X(15), X(4));
262     BODY_20_31(24, A, B, C, D, E, T, X(8), X(8), X(10), X(0), X(5));
263     BODY_20_31(25, T, A, B, C, D, E, X(9), X(9), X(11), X(1), X(6));
264     BODY_20_31(26, E, T, A, B, C, D, X(10), X(10), X(12), X(2), X(7));
265     BODY_20_31(27, D, E, T, A, B, C, X(11), X(11), X(13), X(3), X(8));
266     BODY_20_31(28, C, D, E, T, A, B, X(12), X(12), X(14), X(4), X(9));
267     BODY_20_31(29, B, C, D, E, T, A, X(13), X(13), X(15), X(5), X(10));
268     BODY_20_31(30, A, B, C, D, E, T, X(14), X(14), X(0), X(6), X(11));
269     BODY_20_31(31, T, A, B, C, D, E, X(15), X(15), X(1), X(7), X(12));
270 
271     BODY_32_39(32, E, T, A, B, C, D, X(0), X(2), X(8), X(13));
272     BODY_32_39(33, D, E, T, A, B, C, X(1), X(3), X(9), X(14));
273     BODY_32_39(34, C, D, E, T, A, B, X(2), X(4), X(10), X(15));
274     BODY_32_39(35, B, C, D, E, T, A, X(3), X(5), X(11), X(0));
275     BODY_32_39(36, A, B, C, D, E, T, X(4), X(6), X(12), X(1));
276     BODY_32_39(37, T, A, B, C, D, E, X(5), X(7), X(13), X(2));
277     BODY_32_39(38, E, T, A, B, C, D, X(6), X(8), X(14), X(3));
278     BODY_32_39(39, D, E, T, A, B, C, X(7), X(9), X(15), X(4));
279 
280     BODY_40_59(40, C, D, E, T, A, B, X(8), X(10), X(0), X(5));
281     BODY_40_59(41, B, C, D, E, T, A, X(9), X(11), X(1), X(6));
282     BODY_40_59(42, A, B, C, D, E, T, X(10), X(12), X(2), X(7));
283     BODY_40_59(43, T, A, B, C, D, E, X(11), X(13), X(3), X(8));
284     BODY_40_59(44, E, T, A, B, C, D, X(12), X(14), X(4), X(9));
285     BODY_40_59(45, D, E, T, A, B, C, X(13), X(15), X(5), X(10));
286     BODY_40_59(46, C, D, E, T, A, B, X(14), X(0), X(6), X(11));
287     BODY_40_59(47, B, C, D, E, T, A, X(15), X(1), X(7), X(12));
288     BODY_40_59(48, A, B, C, D, E, T, X(0), X(2), X(8), X(13));
289     BODY_40_59(49, T, A, B, C, D, E, X(1), X(3), X(9), X(14));
290     BODY_40_59(50, E, T, A, B, C, D, X(2), X(4), X(10), X(15));
291     BODY_40_59(51, D, E, T, A, B, C, X(3), X(5), X(11), X(0));
292     BODY_40_59(52, C, D, E, T, A, B, X(4), X(6), X(12), X(1));
293     BODY_40_59(53, B, C, D, E, T, A, X(5), X(7), X(13), X(2));
294     BODY_40_59(54, A, B, C, D, E, T, X(6), X(8), X(14), X(3));
295     BODY_40_59(55, T, A, B, C, D, E, X(7), X(9), X(15), X(4));
296     BODY_40_59(56, E, T, A, B, C, D, X(8), X(10), X(0), X(5));
297     BODY_40_59(57, D, E, T, A, B, C, X(9), X(11), X(1), X(6));
298     BODY_40_59(58, C, D, E, T, A, B, X(10), X(12), X(2), X(7));
299     BODY_40_59(59, B, C, D, E, T, A, X(11), X(13), X(3), X(8));
300 
301     BODY_60_79(60, A, B, C, D, E, T, X(12), X(14), X(4), X(9));
302     BODY_60_79(61, T, A, B, C, D, E, X(13), X(15), X(5), X(10));
303     BODY_60_79(62, E, T, A, B, C, D, X(14), X(0), X(6), X(11));
304     BODY_60_79(63, D, E, T, A, B, C, X(15), X(1), X(7), X(12));
305     BODY_60_79(64, C, D, E, T, A, B, X(0), X(2), X(8), X(13));
306     BODY_60_79(65, B, C, D, E, T, A, X(1), X(3), X(9), X(14));
307     BODY_60_79(66, A, B, C, D, E, T, X(2), X(4), X(10), X(15));
308     BODY_60_79(67, T, A, B, C, D, E, X(3), X(5), X(11), X(0));
309     BODY_60_79(68, E, T, A, B, C, D, X(4), X(6), X(12), X(1));
310     BODY_60_79(69, D, E, T, A, B, C, X(5), X(7), X(13), X(2));
311     BODY_60_79(70, C, D, E, T, A, B, X(6), X(8), X(14), X(3));
312     BODY_60_79(71, B, C, D, E, T, A, X(7), X(9), X(15), X(4));
313     BODY_60_79(72, A, B, C, D, E, T, X(8), X(10), X(0), X(5));
314     BODY_60_79(73, T, A, B, C, D, E, X(9), X(11), X(1), X(6));
315     BODY_60_79(74, E, T, A, B, C, D, X(10), X(12), X(2), X(7));
316     BODY_60_79(75, D, E, T, A, B, C, X(11), X(13), X(3), X(8));
317     BODY_60_79(76, C, D, E, T, A, B, X(12), X(14), X(4), X(9));
318     BODY_60_79(77, B, C, D, E, T, A, X(13), X(15), X(5), X(10));
319     BODY_60_79(78, A, B, C, D, E, T, X(14), X(0), X(6), X(11));
320     BODY_60_79(79, T, A, B, C, D, E, X(15), X(1), X(7), X(12));
321 
322     state[0] = (state[0] + E) & 0xffffffffL;
323     state[1] = (state[1] + T) & 0xffffffffL;
324     state[2] = (state[2] + A) & 0xffffffffL;
325     state[3] = (state[3] + B) & 0xffffffffL;
326     state[4] = (state[4] + C) & 0xffffffffL;
327 
328     if (--num == 0) {
329       break;
330     }
331 
332     A = state[0];
333     B = state[1];
334     C = state[2];
335     D = state[3];
336     E = state[4];
337   }
338 }
339 #endif
340 
341 #undef ROTATE
342 #undef Xupdate
343 #undef K_00_19
344 #undef K_20_39
345 #undef K_40_59
346 #undef K_60_79
347 #undef F_00_19
348 #undef F_20_39
349 #undef F_40_59
350 #undef F_60_79
351 #undef BODY_00_15
352 #undef BODY_16_19
353 #undef BODY_20_31
354 #undef BODY_32_39
355 #undef BODY_40_59
356 #undef BODY_60_79
357 #undef X
358