summaryrefslogtreecommitdiff
path: root/src/lib/libcrypto/sha/sha_locl.h
diff options
context:
space:
mode:
Diffstat (limited to 'src/lib/libcrypto/sha/sha_locl.h')
-rw-r--r--src/lib/libcrypto/sha/sha_locl.h563
1 files changed, 375 insertions, 188 deletions
diff --git a/src/lib/libcrypto/sha/sha_locl.h b/src/lib/libcrypto/sha/sha_locl.h
index 6646a8915b..3e6f489b87 100644
--- a/src/lib/libcrypto/sha/sha_locl.h
+++ b/src/lib/libcrypto/sha/sha_locl.h
@@ -60,180 +60,105 @@
60#include <string.h> 60#include <string.h>
61 61
62#include <openssl/opensslconf.h> 62#include <openssl/opensslconf.h>
63 63#include <openssl/sha.h>
64#ifdef undef
65/* one or the other needs to be defined */
66#ifndef SHA_1 /* FIPE 180-1 */
67#define SHA_0 /* FIPS 180 */
68#endif
69#endif
70
71#undef c2nl
72#define c2nl(c,l) (l =(((unsigned long)(*((c)++)))<<24), \
73 l|=(((unsigned long)(*((c)++)))<<16), \
74 l|=(((unsigned long)(*((c)++)))<< 8), \
75 l|=(((unsigned long)(*((c)++))) ))
76
77#undef p_c2nl
78#define p_c2nl(c,l,n) { \
79 switch (n) { \
80 case 0: l =((unsigned long)(*((c)++)))<<24; \
81 case 1: l|=((unsigned long)(*((c)++)))<<16; \
82 case 2: l|=((unsigned long)(*((c)++)))<< 8; \
83 case 3: l|=((unsigned long)(*((c)++))); \
84 } \
85 }
86
87#undef c2nl_p
88/* NOTE the pointer is not incremented at the end of this */
89#define c2nl_p(c,l,n) { \
90 l=0; \
91 (c)+=n; \
92 switch (n) { \
93 case 3: l =((unsigned long)(*(--(c))))<< 8; \
94 case 2: l|=((unsigned long)(*(--(c))))<<16; \
95 case 1: l|=((unsigned long)(*(--(c))))<<24; \
96 } \
97 }
98
99#undef p_c2nl_p
100#define p_c2nl_p(c,l,sc,len) { \
101 switch (sc) \
102 { \
103 case 0: l =((unsigned long)(*((c)++)))<<24; \
104 if (--len == 0) break; \
105 case 1: l|=((unsigned long)(*((c)++)))<<16; \
106 if (--len == 0) break; \
107 case 2: l|=((unsigned long)(*((c)++)))<< 8; \
108 } \
109 }
110
111#undef nl2c
112#define nl2c(l,c) (*((c)++)=(unsigned char)(((l)>>24)&0xff), \
113 *((c)++)=(unsigned char)(((l)>>16)&0xff), \
114 *((c)++)=(unsigned char)(((l)>> 8)&0xff), \
115 *((c)++)=(unsigned char)(((l) )&0xff))
116
117#undef c2l
118#define c2l(c,l) (l =(((unsigned long)(*((c)++))) ), \
119 l|=(((unsigned long)(*((c)++)))<< 8), \
120 l|=(((unsigned long)(*((c)++)))<<16), \
121 l|=(((unsigned long)(*((c)++)))<<24))
122
123#undef p_c2l
124#define p_c2l(c,l,n) { \
125 switch (n) { \
126 case 0: l =((unsigned long)(*((c)++))); \
127 case 1: l|=((unsigned long)(*((c)++)))<< 8; \
128 case 2: l|=((unsigned long)(*((c)++)))<<16; \
129 case 3: l|=((unsigned long)(*((c)++)))<<24; \
130 } \
131 }
132
133#undef c2l_p
134/* NOTE the pointer is not incremented at the end of this */
135#define c2l_p(c,l,n) { \
136 l=0; \
137 (c)+=n; \
138 switch (n) { \
139 case 3: l =((unsigned long)(*(--(c))))<<16; \
140 case 2: l|=((unsigned long)(*(--(c))))<< 8; \
141 case 1: l|=((unsigned long)(*(--(c)))); \
142 } \
143 }
144
145#undef p_c2l_p
146#define p_c2l_p(c,l,sc,len) { \
147 switch (sc) \
148 { \
149 case 0: l =((unsigned long)(*((c)++))); \
150 if (--len == 0) break; \
151 case 1: l|=((unsigned long)(*((c)++)))<< 8; \
152 if (--len == 0) break; \
153 case 2: l|=((unsigned long)(*((c)++)))<<16; \
154 } \
155 }
156
157#undef l2c
158#define l2c(l,c) (*((c)++)=(unsigned char)(((l) )&0xff), \
159 *((c)++)=(unsigned char)(((l)>> 8)&0xff), \
160 *((c)++)=(unsigned char)(((l)>>16)&0xff), \
161 *((c)++)=(unsigned char)(((l)>>24)&0xff))
162 64
163#ifndef SHA_LONG_LOG2 65#ifndef SHA_LONG_LOG2
164#define SHA_LONG_LOG2 2 /* default to 32 bits */ 66#define SHA_LONG_LOG2 2 /* default to 32 bits */
165#endif 67#endif
166 68
167#undef ROTATE 69#define DATA_ORDER_IS_BIG_ENDIAN
168#undef Endian_Reverse32 70
169#if defined(WIN32) 71#define HASH_LONG SHA_LONG
170#define ROTATE(a,n) _lrotl(a,n) 72#define HASH_LONG_LOG2 SHA_LONG_LOG2
171#elif defined(__GNUC__) && !defined(PEDANTIC) 73#define HASH_CTX SHA_CTX
172/* some inline assembler templates by <appro@fy.chalmers.se> */ 74#define HASH_CBLOCK SHA_CBLOCK
173#if defined(__i386) && !defined(NO_ASM) 75#define HASH_LBLOCK SHA_LBLOCK
174#define ROTATE(a,n) ({ register unsigned int ret; \ 76#define HASH_MAKE_STRING(c,s) do { \
175 asm ("roll %1,%0" \ 77 unsigned long ll; \
176 : "=r"(ret) \ 78 ll=(c)->h0; HOST_l2c(ll,(s)); \
177 : "I"(n), "0"(a) \ 79 ll=(c)->h1; HOST_l2c(ll,(s)); \
178 : "cc"); \ 80 ll=(c)->h2; HOST_l2c(ll,(s)); \
179 ret; \ 81 ll=(c)->h3; HOST_l2c(ll,(s)); \
180 }) 82 ll=(c)->h4; HOST_l2c(ll,(s)); \
181#ifndef I386_ONLY 83 } while (0)
182#define Endian_Reverse32(a) \ 84
183 { register unsigned int ltmp=(a); \ 85#if defined(SHA_0)
184 asm ("bswapl %0" \ 86
185 : "=r"(ltmp) : "0"(ltmp)); \ 87# define HASH_UPDATE SHA_Update
186 (a)=ltmp; \ 88# define HASH_TRANSFORM SHA_Transform
187 } 89# define HASH_FINAL SHA_Final
188#endif 90# define HASH_INIT SHA_Init
189#elif defined(__powerpc) 91# define HASH_BLOCK_HOST_ORDER sha_block_host_order
190#define ROTATE(a,n) ({ register unsigned int ret; \ 92# define HASH_BLOCK_DATA_ORDER sha_block_data_order
191 asm ("rlwinm %0,%1,%2,0,31" \ 93# define Xupdate(a,ix,ia,ib,ic,id) (ix=(a)=(ia^ib^ic^id))
192 : "=r"(ret) \ 94
193 : "r"(a), "I"(n)); \ 95 void sha_block_host_order (SHA_CTX *c, const void *p,int num);
194 ret; \ 96 void sha_block_data_order (SHA_CTX *c, const void *p,int num);
195 }) 97
196/* Endian_Reverse32 is not needed for PowerPC */ 98#elif defined(SHA_1)
197#endif 99
198#endif 100# define HASH_UPDATE SHA1_Update
101# define HASH_TRANSFORM SHA1_Transform
102# define HASH_FINAL SHA1_Final
103# define HASH_INIT SHA1_Init
104# define HASH_BLOCK_HOST_ORDER sha1_block_host_order
105# define HASH_BLOCK_DATA_ORDER sha1_block_data_order
106# if defined(__MWERKS__) && defined(__MC68K__)
107 /* Metrowerks for Motorola fails otherwise:-( <appro@fy.chalmers.se> */
108# define Xupdate(a,ix,ia,ib,ic,id) do { (a)=(ia^ib^ic^id); \
109 ix=(a)=ROTATE((a),1); \
110 } while (0)
111# else
112# define Xupdate(a,ix,ia,ib,ic,id) ( (a)=(ia^ib^ic^id), \
113 ix=(a)=ROTATE((a),1) \
114 )
115# endif
116
117# ifdef SHA1_ASM
118# if defined(__i386) || defined(_M_IX86) || defined(__INTEL__)
119# define sha1_block_host_order sha1_block_asm_host_order
120# define DONT_IMPLEMENT_BLOCK_HOST_ORDER
121# define sha1_block_data_order sha1_block_asm_data_order
122# define DONT_IMPLEMENT_BLOCK_DATA_ORDER
123# define HASH_BLOCK_DATA_ORDER_ALIGNED sha1_block_asm_data_order
124# endif
125# endif
126 void sha1_block_host_order (SHA_CTX *c, const void *p,int num);
127 void sha1_block_data_order (SHA_CTX *c, const void *p,int num);
199 128
200/* A nice byte order reversal from Wei Dai <weidai@eskimo.com> */
201#ifdef ROTATE
202#ifndef Endian_Reverse32
203/* 5 instructions with rotate instruction, else 9 */
204#define Endian_Reverse32(a) \
205 { \
206 unsigned long t=(a); \
207 (a)=((ROTATE(t,8)&0x00FF00FF)|(ROTATE((t&0x00FF00FF),24))); \
208 }
209#endif
210#else 129#else
211#define ROTATE(a,n) (((a)<<(n))|(((a)&0xffffffff)>>(32-(n)))) 130# error "Either SHA_0 or SHA_1 must be defined."
212#ifndef Endian_Reverse32
213/* 6 instructions with rotate instruction, else 8 */
214#define Endian_Reverse32(a) \
215 { \
216 unsigned long t=(a); \
217 t=(((t>>8)&0x00FF00FF)|((t&0x00FF00FF)<<8)); \
218 (a)=ROTATE(t,16); \
219 }
220#endif 131#endif
221/* 132
222 * Originally the middle line started with l=(((l&0xFF00FF00)>>8)|... 133#ifndef FLAT_INC
223 * It's rewritten as above for two reasons: 134#include "../md32_common.h"
224 * - RISCs aren't good at long constants and have to explicitely 135#else
225 * compose 'em with several (well, usually 2) instructions in a 136#include "md32_common.h"
226 * register before performing the actual operation and (as you
227 * already realized:-) having same constant should inspire the
228 * compiler to permanently allocate the only register for it;
229 * - most modern CPUs have two ALUs, but usually only one has
230 * circuitry for shifts:-( this minor tweak inspires compiler
231 * to schedule shift instructions in a better way...
232 *
233 * <appro@fy.chalmers.se>
234 */
235#endif 137#endif
236 138
139#define INIT_DATA_h0 0x67452301UL
140#define INIT_DATA_h1 0xefcdab89UL
141#define INIT_DATA_h2 0x98badcfeUL
142#define INIT_DATA_h3 0x10325476UL
143#define INIT_DATA_h4 0xc3d2e1f0UL
144
145void HASH_INIT (SHA_CTX *c)
146 {
147 c->h0=INIT_DATA_h0;
148 c->h1=INIT_DATA_h1;
149 c->h2=INIT_DATA_h2;
150 c->h3=INIT_DATA_h3;
151 c->h4=INIT_DATA_h4;
152 c->Nl=0;
153 c->Nh=0;
154 c->num=0;
155 }
156
157#define K_00_19 0x5a827999UL
158#define K_20_39 0x6ed9eba1UL
159#define K_40_59 0x8f1bbcdcUL
160#define K_60_79 0xca62c1d6UL
161
237/* As pointed out by Wei Dai <weidai@eskimo.com>, F() below can be 162/* As pointed out by Wei Dai <weidai@eskimo.com>, F() below can be
238 * simplified to the code in F_00_19. Wei attributes these optimisations 163 * simplified to the code in F_00_19. Wei attributes these optimisations
239 * to Peter Gutmann's SHS code, and he attributes it to Rich Schroeppel. 164 * to Peter Gutmann's SHS code, and he attributes it to Rich Schroeppel.
@@ -246,43 +171,305 @@
246#define F_40_59(b,c,d) (((b) & (c)) | (((b)|(c)) & (d))) 171#define F_40_59(b,c,d) (((b) & (c)) | (((b)|(c)) & (d)))
247#define F_60_79(b,c,d) F_20_39(b,c,d) 172#define F_60_79(b,c,d) F_20_39(b,c,d)
248 173
249#undef Xupdate 174#define BODY_00_15(i,a,b,c,d,e,f,xi) \
250#ifdef SHA_0 175 (f)=xi+(e)+K_00_19+ROTATE((a),5)+F_00_19((b),(c),(d)); \
251#define Xupdate(a,i,ia,ib,ic,id) X[(i)&0x0f]=(a)=\
252 (ia[(i)&0x0f]^ib[((i)+2)&0x0f]^ic[((i)+8)&0x0f]^id[((i)+13)&0x0f]);
253#endif
254#ifdef SHA_1
255#define Xupdate(a,i,ia,ib,ic,id) (a)=\
256 (ia[(i)&0x0f]^ib[((i)+2)&0x0f]^ic[((i)+8)&0x0f]^id[((i)+13)&0x0f]);\
257 X[(i)&0x0f]=(a)=ROTATE((a),1);
258#endif
259
260#define BODY_00_15(i,a,b,c,d,e,f,xa) \
261 (f)=xa[i]+(e)+K_00_19+ROTATE((a),5)+F_00_19((b),(c),(d)); \
262 (b)=ROTATE((b),30); 176 (b)=ROTATE((b),30);
263 177
264#define BODY_16_19(i,a,b,c,d,e,f,xa,xb,xc,xd) \ 178#define BODY_16_19(i,a,b,c,d,e,f,xi,xa,xb,xc,xd) \
265 Xupdate(f,i,xa,xb,xc,xd); \ 179 Xupdate(f,xi,xa,xb,xc,xd); \
266 (f)+=(e)+K_00_19+ROTATE((a),5)+F_00_19((b),(c),(d)); \ 180 (f)+=(e)+K_00_19+ROTATE((a),5)+F_00_19((b),(c),(d)); \
267 (b)=ROTATE((b),30); 181 (b)=ROTATE((b),30);
268 182
269#define BODY_20_31(i,a,b,c,d,e,f,xa,xb,xc,xd) \ 183#define BODY_20_31(i,a,b,c,d,e,f,xi,xa,xb,xc,xd) \
270 Xupdate(f,i,xa,xb,xc,xd); \ 184 Xupdate(f,xi,xa,xb,xc,xd); \
271 (f)+=(e)+K_20_39+ROTATE((a),5)+F_20_39((b),(c),(d)); \ 185 (f)+=(e)+K_20_39+ROTATE((a),5)+F_20_39((b),(c),(d)); \
272 (b)=ROTATE((b),30); 186 (b)=ROTATE((b),30);
273 187
274#define BODY_32_39(i,a,b,c,d,e,f,xa) \ 188#define BODY_32_39(i,a,b,c,d,e,f,xa,xb,xc,xd) \
275 Xupdate(f,i,xa,xa,xa,xa); \ 189 Xupdate(f,xa,xa,xb,xc,xd); \
276 (f)+=(e)+K_20_39+ROTATE((a),5)+F_20_39((b),(c),(d)); \ 190 (f)+=(e)+K_20_39+ROTATE((a),5)+F_20_39((b),(c),(d)); \
277 (b)=ROTATE((b),30); 191 (b)=ROTATE((b),30);
278 192
279#define BODY_40_59(i,a,b,c,d,e,f,xa) \ 193#define BODY_40_59(i,a,b,c,d,e,f,xa,xb,xc,xd) \
280 Xupdate(f,i,xa,xa,xa,xa); \ 194 Xupdate(f,xa,xa,xb,xc,xd); \
281 (f)+=(e)+K_40_59+ROTATE((a),5)+F_40_59((b),(c),(d)); \ 195 (f)+=(e)+K_40_59+ROTATE((a),5)+F_40_59((b),(c),(d)); \
282 (b)=ROTATE((b),30); 196 (b)=ROTATE((b),30);
283 197
284#define BODY_60_79(i,a,b,c,d,e,f,xa) \ 198#define BODY_60_79(i,a,b,c,d,e,f,xa,xb,xc,xd) \
285 Xupdate(f,i,xa,xa,xa,xa); \ 199 Xupdate(f,xa,xa,xb,xc,xd); \
286 (f)=X[(i)&0x0f]+(e)+K_60_79+ROTATE((a),5)+F_60_79((b),(c),(d)); \ 200 (f)=xa+(e)+K_60_79+ROTATE((a),5)+F_60_79((b),(c),(d)); \
287 (b)=ROTATE((b),30); 201 (b)=ROTATE((b),30);
288 202
203#ifdef X
204#undef X
205#endif
206#ifndef MD32_XARRAY
207 /*
208 * Originally X was an array. As it's automatic it's natural
209 * to expect RISC compiler to accomodate at least part of it in
210 * the register bank, isn't it? Unfortunately not all compilers
211 * "find" this expectation reasonable:-( On order to make such
212 * compilers generate better code I replace X[] with a bunch of
213 * X0, X1, etc. See the function body below...
214 * <appro@fy.chalmers.se>
215 */
216# define X(i) XX##i
217#else
218 /*
219 * However! Some compilers (most notably HP C) get overwhelmed by
220 * that many local variables so that we have to have the way to
221 * fall down to the original behavior.
222 */
223# define X(i) XX[i]
224#endif
225
226#ifndef DONT_IMPLEMENT_BLOCK_HOST_ORDER
227void HASH_BLOCK_HOST_ORDER (SHA_CTX *c, const void *d, int num)
228 {
229 const SHA_LONG *W=d;
230 register unsigned long A,B,C,D,E,T;
231#ifndef MD32_XARRAY
232 unsigned long XX0, XX1, XX2, XX3, XX4, XX5, XX6, XX7,
233 XX8, XX9,XX10,XX11,XX12,XX13,XX14,XX15;
234#else
235 SHA_LONG XX[16];
236#endif
237
238 A=c->h0;
239 B=c->h1;
240 C=c->h2;
241 D=c->h3;
242 E=c->h4;
243
244 for (;;)
245 {
246 BODY_00_15( 0,A,B,C,D,E,T,W[ 0]);
247 BODY_00_15( 1,T,A,B,C,D,E,W[ 1]);
248 BODY_00_15( 2,E,T,A,B,C,D,W[ 2]);
249 BODY_00_15( 3,D,E,T,A,B,C,W[ 3]);
250 BODY_00_15( 4,C,D,E,T,A,B,W[ 4]);
251 BODY_00_15( 5,B,C,D,E,T,A,W[ 5]);
252 BODY_00_15( 6,A,B,C,D,E,T,W[ 6]);
253 BODY_00_15( 7,T,A,B,C,D,E,W[ 7]);
254 BODY_00_15( 8,E,T,A,B,C,D,W[ 8]);
255 BODY_00_15( 9,D,E,T,A,B,C,W[ 9]);
256 BODY_00_15(10,C,D,E,T,A,B,W[10]);
257 BODY_00_15(11,B,C,D,E,T,A,W[11]);
258 BODY_00_15(12,A,B,C,D,E,T,W[12]);
259 BODY_00_15(13,T,A,B,C,D,E,W[13]);
260 BODY_00_15(14,E,T,A,B,C,D,W[14]);
261 BODY_00_15(15,D,E,T,A,B,C,W[15]);
262
263 BODY_16_19(16,C,D,E,T,A,B,X( 0),W[ 0],W[ 2],W[ 8],W[13]);
264 BODY_16_19(17,B,C,D,E,T,A,X( 1),W[ 1],W[ 3],W[ 9],W[14]);
265 BODY_16_19(18,A,B,C,D,E,T,X( 2),W[ 2],W[ 4],W[10],W[15]);
266 BODY_16_19(19,T,A,B,C,D,E,X( 3),W[ 3],W[ 5],W[11],X( 0));
267
268 BODY_20_31(20,E,T,A,B,C,D,X( 4),W[ 4],W[ 6],W[12],X( 1));
269 BODY_20_31(21,D,E,T,A,B,C,X( 5),W[ 5],W[ 7],W[13],X( 2));
270 BODY_20_31(22,C,D,E,T,A,B,X( 6),W[ 6],W[ 8],W[14],X( 3));
271 BODY_20_31(23,B,C,D,E,T,A,X( 7),W[ 7],W[ 9],W[15],X( 4));
272 BODY_20_31(24,A,B,C,D,E,T,X( 8),W[ 8],W[10],X( 0),X( 5));
273 BODY_20_31(25,T,A,B,C,D,E,X( 9),W[ 9],W[11],X( 1),X( 6));
274 BODY_20_31(26,E,T,A,B,C,D,X(10),W[10],W[12],X( 2),X( 7));
275 BODY_20_31(27,D,E,T,A,B,C,X(11),W[11],W[13],X( 3),X( 8));
276 BODY_20_31(28,C,D,E,T,A,B,X(12),W[12],W[14],X( 4),X( 9));
277 BODY_20_31(29,B,C,D,E,T,A,X(13),W[13],W[15],X( 5),X(10));
278 BODY_20_31(30,A,B,C,D,E,T,X(14),W[14],X( 0),X( 6),X(11));
279 BODY_20_31(31,T,A,B,C,D,E,X(15),W[15],X( 1),X( 7),X(12));
280
281 BODY_32_39(32,E,T,A,B,C,D,X( 0),X( 2),X( 8),X(13));
282 BODY_32_39(33,D,E,T,A,B,C,X( 1),X( 3),X( 9),X(14));
283 BODY_32_39(34,C,D,E,T,A,B,X( 2),X( 4),X(10),X(15));
284 BODY_32_39(35,B,C,D,E,T,A,X( 3),X( 5),X(11),X( 0));
285 BODY_32_39(36,A,B,C,D,E,T,X( 4),X( 6),X(12),X( 1));
286 BODY_32_39(37,T,A,B,C,D,E,X( 5),X( 7),X(13),X( 2));
287 BODY_32_39(38,E,T,A,B,C,D,X( 6),X( 8),X(14),X( 3));
288 BODY_32_39(39,D,E,T,A,B,C,X( 7),X( 9),X(15),X( 4));
289
290 BODY_40_59(40,C,D,E,T,A,B,X( 8),X(10),X( 0),X( 5));
291 BODY_40_59(41,B,C,D,E,T,A,X( 9),X(11),X( 1),X( 6));
292 BODY_40_59(42,A,B,C,D,E,T,X(10),X(12),X( 2),X( 7));
293 BODY_40_59(43,T,A,B,C,D,E,X(11),X(13),X( 3),X( 8));
294 BODY_40_59(44,E,T,A,B,C,D,X(12),X(14),X( 4),X( 9));
295 BODY_40_59(45,D,E,T,A,B,C,X(13),X(15),X( 5),X(10));
296 BODY_40_59(46,C,D,E,T,A,B,X(14),X( 0),X( 6),X(11));
297 BODY_40_59(47,B,C,D,E,T,A,X(15),X( 1),X( 7),X(12));
298 BODY_40_59(48,A,B,C,D,E,T,X( 0),X( 2),X( 8),X(13));
299 BODY_40_59(49,T,A,B,C,D,E,X( 1),X( 3),X( 9),X(14));
300 BODY_40_59(50,E,T,A,B,C,D,X( 2),X( 4),X(10),X(15));
301 BODY_40_59(51,D,E,T,A,B,C,X( 3),X( 5),X(11),X( 0));
302 BODY_40_59(52,C,D,E,T,A,B,X( 4),X( 6),X(12),X( 1));
303 BODY_40_59(53,B,C,D,E,T,A,X( 5),X( 7),X(13),X( 2));
304 BODY_40_59(54,A,B,C,D,E,T,X( 6),X( 8),X(14),X( 3));
305 BODY_40_59(55,T,A,B,C,D,E,X( 7),X( 9),X(15),X( 4));
306 BODY_40_59(56,E,T,A,B,C,D,X( 8),X(10),X( 0),X( 5));
307 BODY_40_59(57,D,E,T,A,B,C,X( 9),X(11),X( 1),X( 6));
308 BODY_40_59(58,C,D,E,T,A,B,X(10),X(12),X( 2),X( 7));
309 BODY_40_59(59,B,C,D,E,T,A,X(11),X(13),X( 3),X( 8));
310
311 BODY_60_79(60,A,B,C,D,E,T,X(12),X(14),X( 4),X( 9));
312 BODY_60_79(61,T,A,B,C,D,E,X(13),X(15),X( 5),X(10));
313 BODY_60_79(62,E,T,A,B,C,D,X(14),X( 0),X( 6),X(11));
314 BODY_60_79(63,D,E,T,A,B,C,X(15),X( 1),X( 7),X(12));
315 BODY_60_79(64,C,D,E,T,A,B,X( 0),X( 2),X( 8),X(13));
316 BODY_60_79(65,B,C,D,E,T,A,X( 1),X( 3),X( 9),X(14));
317 BODY_60_79(66,A,B,C,D,E,T,X( 2),X( 4),X(10),X(15));
318 BODY_60_79(67,T,A,B,C,D,E,X( 3),X( 5),X(11),X( 0));
319 BODY_60_79(68,E,T,A,B,C,D,X( 4),X( 6),X(12),X( 1));
320 BODY_60_79(69,D,E,T,A,B,C,X( 5),X( 7),X(13),X( 2));
321 BODY_60_79(70,C,D,E,T,A,B,X( 6),X( 8),X(14),X( 3));
322 BODY_60_79(71,B,C,D,E,T,A,X( 7),X( 9),X(15),X( 4));
323 BODY_60_79(72,A,B,C,D,E,T,X( 8),X(10),X( 0),X( 5));
324 BODY_60_79(73,T,A,B,C,D,E,X( 9),X(11),X( 1),X( 6));
325 BODY_60_79(74,E,T,A,B,C,D,X(10),X(12),X( 2),X( 7));
326 BODY_60_79(75,D,E,T,A,B,C,X(11),X(13),X( 3),X( 8));
327 BODY_60_79(76,C,D,E,T,A,B,X(12),X(14),X( 4),X( 9));
328 BODY_60_79(77,B,C,D,E,T,A,X(13),X(15),X( 5),X(10));
329 BODY_60_79(78,A,B,C,D,E,T,X(14),X( 0),X( 6),X(11));
330 BODY_60_79(79,T,A,B,C,D,E,X(15),X( 1),X( 7),X(12));
331
332 c->h0=(c->h0+E)&0xffffffffL;
333 c->h1=(c->h1+T)&0xffffffffL;
334 c->h2=(c->h2+A)&0xffffffffL;
335 c->h3=(c->h3+B)&0xffffffffL;
336 c->h4=(c->h4+C)&0xffffffffL;
337
338 if (--num <= 0) break;
339
340 A=c->h0;
341 B=c->h1;
342 C=c->h2;
343 D=c->h3;
344 E=c->h4;
345
346 W+=SHA_LBLOCK;
347 }
348 }
349#endif
350
351#ifndef DONT_IMPLEMENT_BLOCK_DATA_ORDER
352void HASH_BLOCK_DATA_ORDER (SHA_CTX *c, const void *p, int num)
353 {
354 const unsigned char *data=p;
355 register unsigned long A,B,C,D,E,T,l;
356#ifndef MD32_XARRAY
357 unsigned long XX0, XX1, XX2, XX3, XX4, XX5, XX6, XX7,
358 XX8, XX9,XX10,XX11,XX12,XX13,XX14,XX15;
359#else
360 SHA_LONG XX[16];
361#endif
362
363 A=c->h0;
364 B=c->h1;
365 C=c->h2;
366 D=c->h3;
367 E=c->h4;
368
369 for (;;)
370 {
371
372 HOST_c2l(data,l); X( 0)=l; HOST_c2l(data,l); X( 1)=l;
373 BODY_00_15( 0,A,B,C,D,E,T,X( 0)); HOST_c2l(data,l); X( 2)=l;
374 BODY_00_15( 1,T,A,B,C,D,E,X( 1)); HOST_c2l(data,l); X( 3)=l;
375 BODY_00_15( 2,E,T,A,B,C,D,X( 2)); HOST_c2l(data,l); X( 4)=l;
376 BODY_00_15( 3,D,E,T,A,B,C,X( 3)); HOST_c2l(data,l); X( 5)=l;
377 BODY_00_15( 4,C,D,E,T,A,B,X( 4)); HOST_c2l(data,l); X( 6)=l;
378 BODY_00_15( 5,B,C,D,E,T,A,X( 5)); HOST_c2l(data,l); X( 7)=l;
379 BODY_00_15( 6,A,B,C,D,E,T,X( 6)); HOST_c2l(data,l); X( 8)=l;
380 BODY_00_15( 7,T,A,B,C,D,E,X( 7)); HOST_c2l(data,l); X( 9)=l;
381 BODY_00_15( 8,E,T,A,B,C,D,X( 8)); HOST_c2l(data,l); X(10)=l;
382 BODY_00_15( 9,D,E,T,A,B,C,X( 9)); HOST_c2l(data,l); X(11)=l;
383 BODY_00_15(10,C,D,E,T,A,B,X(10)); HOST_c2l(data,l); X(12)=l;
384 BODY_00_15(11,B,C,D,E,T,A,X(11)); HOST_c2l(data,l); X(13)=l;
385 BODY_00_15(12,A,B,C,D,E,T,X(12)); HOST_c2l(data,l); X(14)=l;
386 BODY_00_15(13,T,A,B,C,D,E,X(13)); HOST_c2l(data,l); X(15)=l;
387 BODY_00_15(14,E,T,A,B,C,D,X(14));
388 BODY_00_15(15,D,E,T,A,B,C,X(15));
389
390 BODY_16_19(16,C,D,E,T,A,B,X( 0),X( 0),X( 2),X( 8),X(13));
391 BODY_16_19(17,B,C,D,E,T,A,X( 1),X( 1),X( 3),X( 9),X(14));
392 BODY_16_19(18,A,B,C,D,E,T,X( 2),X( 2),X( 4),X(10),X(15));
393 BODY_16_19(19,T,A,B,C,D,E,X( 3),X( 3),X( 5),X(11),X( 0));
394
395 BODY_20_31(20,E,T,A,B,C,D,X( 4),X( 4),X( 6),X(12),X( 1));
396 BODY_20_31(21,D,E,T,A,B,C,X( 5),X( 5),X( 7),X(13),X( 2));
397 BODY_20_31(22,C,D,E,T,A,B,X( 6),X( 6),X( 8),X(14),X( 3));
398 BODY_20_31(23,B,C,D,E,T,A,X( 7),X( 7),X( 9),X(15),X( 4));
399 BODY_20_31(24,A,B,C,D,E,T,X( 8),X( 8),X(10),X( 0),X( 5));
400 BODY_20_31(25,T,A,B,C,D,E,X( 9),X( 9),X(11),X( 1),X( 6));
401 BODY_20_31(26,E,T,A,B,C,D,X(10),X(10),X(12),X( 2),X( 7));
402 BODY_20_31(27,D,E,T,A,B,C,X(11),X(11),X(13),X( 3),X( 8));
403 BODY_20_31(28,C,D,E,T,A,B,X(12),X(12),X(14),X( 4),X( 9));
404 BODY_20_31(29,B,C,D,E,T,A,X(13),X(13),X(15),X( 5),X(10));
405 BODY_20_31(30,A,B,C,D,E,T,X(14),X(14),X( 0),X( 6),X(11));
406 BODY_20_31(31,T,A,B,C,D,E,X(15),X(15),X( 1),X( 7),X(12));
407
408 BODY_32_39(32,E,T,A,B,C,D,X( 0),X( 2),X( 8),X(13));
409 BODY_32_39(33,D,E,T,A,B,C,X( 1),X( 3),X( 9),X(14));
410 BODY_32_39(34,C,D,E,T,A,B,X( 2),X( 4),X(10),X(15));
411 BODY_32_39(35,B,C,D,E,T,A,X( 3),X( 5),X(11),X( 0));
412 BODY_32_39(36,A,B,C,D,E,T,X( 4),X( 6),X(12),X( 1));
413 BODY_32_39(37,T,A,B,C,D,E,X( 5),X( 7),X(13),X( 2));
414 BODY_32_39(38,E,T,A,B,C,D,X( 6),X( 8),X(14),X( 3));
415 BODY_32_39(39,D,E,T,A,B,C,X( 7),X( 9),X(15),X( 4));
416
417 BODY_40_59(40,C,D,E,T,A,B,X( 8),X(10),X( 0),X( 5));
418 BODY_40_59(41,B,C,D,E,T,A,X( 9),X(11),X( 1),X( 6));
419 BODY_40_59(42,A,B,C,D,E,T,X(10),X(12),X( 2),X( 7));
420 BODY_40_59(43,T,A,B,C,D,E,X(11),X(13),X( 3),X( 8));
421 BODY_40_59(44,E,T,A,B,C,D,X(12),X(14),X( 4),X( 9));
422 BODY_40_59(45,D,E,T,A,B,C,X(13),X(15),X( 5),X(10));
423 BODY_40_59(46,C,D,E,T,A,B,X(14),X( 0),X( 6),X(11));
424 BODY_40_59(47,B,C,D,E,T,A,X(15),X( 1),X( 7),X(12));
425 BODY_40_59(48,A,B,C,D,E,T,X( 0),X( 2),X( 8),X(13));
426 BODY_40_59(49,T,A,B,C,D,E,X( 1),X( 3),X( 9),X(14));
427 BODY_40_59(50,E,T,A,B,C,D,X( 2),X( 4),X(10),X(15));
428 BODY_40_59(51,D,E,T,A,B,C,X( 3),X( 5),X(11),X( 0));
429 BODY_40_59(52,C,D,E,T,A,B,X( 4),X( 6),X(12),X( 1));
430 BODY_40_59(53,B,C,D,E,T,A,X( 5),X( 7),X(13),X( 2));
431 BODY_40_59(54,A,B,C,D,E,T,X( 6),X( 8),X(14),X( 3));
432 BODY_40_59(55,T,A,B,C,D,E,X( 7),X( 9),X(15),X( 4));
433 BODY_40_59(56,E,T,A,B,C,D,X( 8),X(10),X( 0),X( 5));
434 BODY_40_59(57,D,E,T,A,B,C,X( 9),X(11),X( 1),X( 6));
435 BODY_40_59(58,C,D,E,T,A,B,X(10),X(12),X( 2),X( 7));
436 BODY_40_59(59,B,C,D,E,T,A,X(11),X(13),X( 3),X( 8));
437
438 BODY_60_79(60,A,B,C,D,E,T,X(12),X(14),X( 4),X( 9));
439 BODY_60_79(61,T,A,B,C,D,E,X(13),X(15),X( 5),X(10));
440 BODY_60_79(62,E,T,A,B,C,D,X(14),X( 0),X( 6),X(11));
441 BODY_60_79(63,D,E,T,A,B,C,X(15),X( 1),X( 7),X(12));
442 BODY_60_79(64,C,D,E,T,A,B,X( 0),X( 2),X( 8),X(13));
443 BODY_60_79(65,B,C,D,E,T,A,X( 1),X( 3),X( 9),X(14));
444 BODY_60_79(66,A,B,C,D,E,T,X( 2),X( 4),X(10),X(15));
445 BODY_60_79(67,T,A,B,C,D,E,X( 3),X( 5),X(11),X( 0));
446 BODY_60_79(68,E,T,A,B,C,D,X( 4),X( 6),X(12),X( 1));
447 BODY_60_79(69,D,E,T,A,B,C,X( 5),X( 7),X(13),X( 2));
448 BODY_60_79(70,C,D,E,T,A,B,X( 6),X( 8),X(14),X( 3));
449 BODY_60_79(71,B,C,D,E,T,A,X( 7),X( 9),X(15),X( 4));
450 BODY_60_79(72,A,B,C,D,E,T,X( 8),X(10),X( 0),X( 5));
451 BODY_60_79(73,T,A,B,C,D,E,X( 9),X(11),X( 1),X( 6));
452 BODY_60_79(74,E,T,A,B,C,D,X(10),X(12),X( 2),X( 7));
453 BODY_60_79(75,D,E,T,A,B,C,X(11),X(13),X( 3),X( 8));
454 BODY_60_79(76,C,D,E,T,A,B,X(12),X(14),X( 4),X( 9));
455 BODY_60_79(77,B,C,D,E,T,A,X(13),X(15),X( 5),X(10));
456 BODY_60_79(78,A,B,C,D,E,T,X(14),X( 0),X( 6),X(11));
457 BODY_60_79(79,T,A,B,C,D,E,X(15),X( 1),X( 7),X(12));
458
459 c->h0=(c->h0+E)&0xffffffffL;
460 c->h1=(c->h1+T)&0xffffffffL;
461 c->h2=(c->h2+A)&0xffffffffL;
462 c->h3=(c->h3+B)&0xffffffffL;
463 c->h4=(c->h4+C)&0xffffffffL;
464
465 if (--num <= 0) break;
466
467 A=c->h0;
468 B=c->h1;
469 C=c->h2;
470 D=c->h3;
471 E=c->h4;
472
473 }
474 }
475#endif