diff options
| author | jsing <> | 2024-03-26 05:21:35 +0000 |
|---|---|---|
| committer | jsing <> | 2024-03-26 05:21:35 +0000 |
| commit | 60ada654237dd4cad2245c11f7a9ecf4f6091bc5 (patch) | |
| tree | d200f9ecb8d57b94c38e6f06e679fc6035b87aa7 /src | |
| parent | 0d30941fc41ea9b32baa169075a9593164087f09 (diff) | |
| download | openbsd-60ada654237dd4cad2245c11f7a9ecf4f6091bc5.tar.gz openbsd-60ada654237dd4cad2245c11f7a9ecf4f6091bc5.tar.bz2 openbsd-60ada654237dd4cad2245c11f7a9ecf4f6091bc5.zip | |
Demacro MD5 and improve data loading.
Use static inline functions instead of macros and improve handling of
aligned data. Also number rounds as per RFC 1321.
ok tb@
Diffstat (limited to 'src')
| -rw-r--r-- | src/lib/libcrypto/md5/md5.c | 270 |
1 files changed, 154 insertions, 116 deletions
diff --git a/src/lib/libcrypto/md5/md5.c b/src/lib/libcrypto/md5/md5.c index c2ee2958df..cb1a9a3a09 100644 --- a/src/lib/libcrypto/md5/md5.c +++ b/src/lib/libcrypto/md5/md5.c | |||
| @@ -1,4 +1,4 @@ | |||
| 1 | /* $OpenBSD: md5.c,v 1.18 2023/08/15 08:39:27 jsing Exp $ */ | 1 | /* $OpenBSD: md5.c,v 1.19 2024/03/26 05:21:35 jsing Exp $ */ |
| 2 | /* Copyright (C) 1995-1998 Eric Young (eay@cryptsoft.com) | 2 | /* Copyright (C) 1995-1998 Eric Young (eay@cryptsoft.com) |
| 3 | * All rights reserved. | 3 | * All rights reserved. |
| 4 | * | 4 | * |
| @@ -90,47 +90,64 @@ void md5_block_asm_data_order(MD5_CTX *c, const void *p, size_t num); | |||
| 90 | 90 | ||
| 91 | #include "md32_common.h" | 91 | #include "md32_common.h" |
| 92 | 92 | ||
| 93 | /* | 93 | #ifndef MD5_ASM |
| 94 | #define F(x,y,z) (((x) & (y)) | ((~(x)) & (z))) | 94 | static inline uint32_t |
| 95 | #define G(x,y,z) (((x) & (z)) | ((y) & (~(z)))) | 95 | md5_F(uint32_t x, uint32_t y, uint32_t z) |
| 96 | */ | 96 | { |
| 97 | return (x & y) | (~x & z); | ||
| 98 | } | ||
| 97 | 99 | ||
| 98 | /* As pointed out by Wei Dai <weidai@eskimo.com>, the above can be | 100 | static inline uint32_t |
| 99 | * simplified to the code below. Wei attributes these optimizations | 101 | md5_G(uint32_t x, uint32_t y, uint32_t z) |
| 100 | * to Peter Gutmann's SHS code, and he attributes it to Rich Schroeppel. | 102 | { |
| 101 | */ | 103 | return (x & z) | (y & ~z); |
| 102 | #define F(b,c,d) ((((c) ^ (d)) & (b)) ^ (d)) | 104 | } |
| 103 | #define G(b,c,d) ((((b) ^ (c)) & (d)) ^ (c)) | ||
| 104 | #define H(b,c,d) ((b) ^ (c) ^ (d)) | ||
| 105 | #define I(b,c,d) (((~(d)) | (b)) ^ (c)) | ||
| 106 | 105 | ||
| 107 | #define R0(a,b,c,d,k,s,t) { \ | 106 | static inline uint32_t |
| 108 | a+=((k)+(t)+F((b),(c),(d))); \ | 107 | md5_H(uint32_t x, uint32_t y, uint32_t z) |
| 109 | a=ROTATE(a,s); \ | 108 | { |
| 110 | a+=b; };\ | 109 | return x ^ y ^ z; |
| 110 | } | ||
| 111 | |||
| 112 | static inline uint32_t | ||
| 113 | md5_I(uint32_t x, uint32_t y, uint32_t z) | ||
| 114 | { | ||
| 115 | return y ^ (x | ~z); | ||
| 116 | } | ||
| 111 | 117 | ||
| 112 | #define R1(a,b,c,d,k,s,t) { \ | 118 | static inline void |
| 113 | a+=((k)+(t)+G((b),(c),(d))); \ | 119 | md5_round1(uint32_t *a, uint32_t b, uint32_t c, uint32_t d, uint32_t x, |
| 114 | a=ROTATE(a,s); \ | 120 | uint32_t t, uint32_t s) |
| 115 | a+=b; }; | 121 | { |
| 122 | *a = b + crypto_rol_u32(*a + md5_F(b, c, d) + x + t, s); | ||
| 123 | } | ||
| 116 | 124 | ||
| 117 | #define R2(a,b,c,d,k,s,t) { \ | 125 | static inline void |
| 118 | a+=((k)+(t)+H((b),(c),(d))); \ | 126 | md5_round2(uint32_t *a, uint32_t b, uint32_t c, uint32_t d, uint32_t x, |
| 119 | a=ROTATE(a,s); \ | 127 | uint32_t t, uint32_t s) |
| 120 | a+=b; }; | 128 | { |
| 129 | *a = b + crypto_rol_u32(*a + md5_G(b, c, d) + x + t, s); | ||
| 130 | } | ||
| 121 | 131 | ||
| 122 | #define R3(a,b,c,d,k,s,t) { \ | 132 | static inline void |
| 123 | a+=((k)+(t)+I((b),(c),(d))); \ | 133 | md5_round3(uint32_t *a, uint32_t b, uint32_t c, uint32_t d, uint32_t x, |
| 124 | a=ROTATE(a,s); \ | 134 | uint32_t t, uint32_t s) |
| 125 | a+=b; }; | 135 | { |
| 136 | *a = b + crypto_rol_u32(*a + md5_H(b, c, d) + x + t, s); | ||
| 137 | } | ||
| 126 | 138 | ||
| 127 | /* Implemented from RFC1321 The MD5 Message-Digest Algorithm. */ | 139 | static inline void |
| 140 | md5_round4(uint32_t *a, uint32_t b, uint32_t c, uint32_t d, uint32_t x, | ||
| 141 | uint32_t t, uint32_t s) | ||
| 142 | { | ||
| 143 | *a = b + crypto_rol_u32(*a + md5_I(b, c, d) + x + t, s); | ||
| 144 | } | ||
| 128 | 145 | ||
| 129 | #ifndef MD5_ASM | ||
| 130 | static void | 146 | static void |
| 131 | md5_block_data_order(MD5_CTX *c, const void *_in, size_t num) | 147 | md5_block_data_order(MD5_CTX *c, const void *_in, size_t num) |
| 132 | { | 148 | { |
| 133 | const uint8_t *in = _in; | 149 | const uint8_t *in = _in; |
| 150 | const MD5_LONG *in32; | ||
| 134 | MD5_LONG A, B, C, D; | 151 | MD5_LONG A, B, C, D; |
| 135 | MD5_LONG X0, X1, X2, X3, X4, X5, X6, X7, | 152 | MD5_LONG X0, X1, X2, X3, X4, X5, X6, X7, |
| 136 | X8, X9, X10, X11, X12, X13, X14, X15; | 153 | X8, X9, X10, X11, X12, X13, X14, X15; |
| @@ -140,93 +157,114 @@ md5_block_data_order(MD5_CTX *c, const void *_in, size_t num) | |||
| 140 | C = c->C; | 157 | C = c->C; |
| 141 | D = c->D; | 158 | D = c->D; |
| 142 | 159 | ||
| 143 | for (; num--; ) { | 160 | while (num-- > 0) { |
| 144 | X0 = crypto_load_le32toh(&in[0 * 4]); | 161 | if ((uintptr_t)in % 4 == 0) { |
| 145 | X1 = crypto_load_le32toh(&in[1 * 4]); | 162 | /* Input is 32 bit aligned. */ |
| 146 | X2 = crypto_load_le32toh(&in[2 * 4]); | 163 | in32 = (const MD5_LONG *)in; |
| 147 | X3 = crypto_load_le32toh(&in[3 * 4]); | 164 | X0 = le32toh(in32[0]); |
| 148 | X4 = crypto_load_le32toh(&in[4 * 4]); | 165 | X1 = le32toh(in32[1]); |
| 149 | X5 = crypto_load_le32toh(&in[5 * 4]); | 166 | X2 = le32toh(in32[2]); |
| 150 | X6 = crypto_load_le32toh(&in[6 * 4]); | 167 | X3 = le32toh(in32[3]); |
| 151 | X7 = crypto_load_le32toh(&in[7 * 4]); | 168 | X4 = le32toh(in32[4]); |
| 152 | X8 = crypto_load_le32toh(&in[8 * 4]); | 169 | X5 = le32toh(in32[5]); |
| 153 | X9 = crypto_load_le32toh(&in[9 * 4]); | 170 | X6 = le32toh(in32[6]); |
| 154 | X10 = crypto_load_le32toh(&in[10 * 4]); | 171 | X7 = le32toh(in32[7]); |
| 155 | X11 = crypto_load_le32toh(&in[11 * 4]); | 172 | X8 = le32toh(in32[8]); |
| 156 | X12 = crypto_load_le32toh(&in[12 * 4]); | 173 | X9 = le32toh(in32[9]); |
| 157 | X13 = crypto_load_le32toh(&in[13 * 4]); | 174 | X10 = le32toh(in32[10]); |
| 158 | X14 = crypto_load_le32toh(&in[14 * 4]); | 175 | X11 = le32toh(in32[11]); |
| 159 | X15 = crypto_load_le32toh(&in[15 * 4]); | 176 | X12 = le32toh(in32[12]); |
| 177 | X13 = le32toh(in32[13]); | ||
| 178 | X14 = le32toh(in32[14]); | ||
| 179 | X15 = le32toh(in32[15]); | ||
| 180 | } else { | ||
| 181 | /* Input is not 32 bit aligned. */ | ||
| 182 | X0 = crypto_load_le32toh(&in[0 * 4]); | ||
| 183 | X1 = crypto_load_le32toh(&in[1 * 4]); | ||
| 184 | X2 = crypto_load_le32toh(&in[2 * 4]); | ||
| 185 | X3 = crypto_load_le32toh(&in[3 * 4]); | ||
| 186 | X4 = crypto_load_le32toh(&in[4 * 4]); | ||
| 187 | X5 = crypto_load_le32toh(&in[5 * 4]); | ||
| 188 | X6 = crypto_load_le32toh(&in[6 * 4]); | ||
| 189 | X7 = crypto_load_le32toh(&in[7 * 4]); | ||
| 190 | X8 = crypto_load_le32toh(&in[8 * 4]); | ||
| 191 | X9 = crypto_load_le32toh(&in[9 * 4]); | ||
| 192 | X10 = crypto_load_le32toh(&in[10 * 4]); | ||
| 193 | X11 = crypto_load_le32toh(&in[11 * 4]); | ||
| 194 | X12 = crypto_load_le32toh(&in[12 * 4]); | ||
| 195 | X13 = crypto_load_le32toh(&in[13 * 4]); | ||
| 196 | X14 = crypto_load_le32toh(&in[14 * 4]); | ||
| 197 | X15 = crypto_load_le32toh(&in[15 * 4]); | ||
| 198 | } | ||
| 160 | in += MD5_CBLOCK; | 199 | in += MD5_CBLOCK; |
| 161 | 200 | ||
| 162 | /* Round 0 */ | 201 | md5_round1(&A, B, C, D, X0, 0xd76aa478L, 7); |
| 163 | R0(A, B, C, D, X0, 7, 0xd76aa478L); | 202 | md5_round1(&D, A, B, C, X1, 0xe8c7b756L, 12); |
| 164 | R0(D, A, B, C, X1, 12, 0xe8c7b756L); | 203 | md5_round1(&C, D, A, B, X2, 0x242070dbL, 17); |
| 165 | R0(C, D, A, B, X2, 17, 0x242070dbL); | 204 | md5_round1(&B, C, D, A, X3, 0xc1bdceeeL, 22); |
| 166 | R0(B, C, D, A, X3, 22, 0xc1bdceeeL); | 205 | md5_round1(&A, B, C, D, X4, 0xf57c0fafL, 7); |
| 167 | R0(A, B, C, D, X4, 7, 0xf57c0fafL); | 206 | md5_round1(&D, A, B, C, X5, 0x4787c62aL, 12); |
| 168 | R0(D, A, B, C, X5, 12, 0x4787c62aL); | 207 | md5_round1(&C, D, A, B, X6, 0xa8304613L, 17); |
| 169 | R0(C, D, A, B, X6, 17, 0xa8304613L); | 208 | md5_round1(&B, C, D, A, X7, 0xfd469501L, 22); |
| 170 | R0(B, C, D, A, X7, 22, 0xfd469501L); | 209 | md5_round1(&A, B, C, D, X8, 0x698098d8L, 7); |
| 171 | R0(A, B, C, D, X8, 7, 0x698098d8L); | 210 | md5_round1(&D, A, B, C, X9, 0x8b44f7afL, 12); |
| 172 | R0(D, A, B, C, X9, 12, 0x8b44f7afL); | 211 | md5_round1(&C, D, A, B, X10, 0xffff5bb1L, 17); |
| 173 | R0(C, D, A, B, X10, 17, 0xffff5bb1L); | 212 | md5_round1(&B, C, D, A, X11, 0x895cd7beL, 22); |
| 174 | R0(B, C, D, A, X11, 22, 0x895cd7beL); | 213 | md5_round1(&A, B, C, D, X12, 0x6b901122L, 7); |
| 175 | R0(A, B, C, D, X12, 7, 0x6b901122L); | 214 | md5_round1(&D, A, B, C, X13, 0xfd987193L, 12); |
| 176 | R0(D, A, B, C, X13, 12, 0xfd987193L); | 215 | md5_round1(&C, D, A, B, X14, 0xa679438eL, 17); |
| 177 | R0(C, D, A, B, X14, 17, 0xa679438eL); | 216 | md5_round1(&B, C, D, A, X15, 0x49b40821L, 22); |
| 178 | R0(B, C, D, A, X15, 22, 0x49b40821L); | 217 | |
| 179 | /* Round 1 */ | 218 | md5_round2(&A, B, C, D, X1, 0xf61e2562L, 5); |
| 180 | R1(A, B, C, D, X1, 5, 0xf61e2562L); | 219 | md5_round2(&D, A, B, C, X6, 0xc040b340L, 9); |
| 181 | R1(D, A, B, C, X6, 9, 0xc040b340L); | 220 | md5_round2(&C, D, A, B, X11, 0x265e5a51L, 14); |
| 182 | R1(C, D, A, B, X11, 14, 0x265e5a51L); | 221 | md5_round2(&B, C, D, A, X0, 0xe9b6c7aaL, 20); |
| 183 | R1(B, C, D, A, X0, 20, 0xe9b6c7aaL); | 222 | md5_round2(&A, B, C, D, X5, 0xd62f105dL, 5); |
| 184 | R1(A, B, C, D, X5, 5, 0xd62f105dL); | 223 | md5_round2(&D, A, B, C, X10, 0x02441453L, 9); |
| 185 | R1(D, A, B, C, X10, 9, 0x02441453L); | 224 | md5_round2(&C, D, A, B, X15, 0xd8a1e681L, 14); |
| 186 | R1(C, D, A, B, X15, 14, 0xd8a1e681L); | 225 | md5_round2(&B, C, D, A, X4, 0xe7d3fbc8L, 20); |
| 187 | R1(B, C, D, A, X4, 20, 0xe7d3fbc8L); | 226 | md5_round2(&A, B, C, D, X9, 0x21e1cde6L, 5); |
| 188 | R1(A, B, C, D, X9, 5, 0x21e1cde6L); | 227 | md5_round2(&D, A, B, C, X14, 0xc33707d6L, 9); |
| 189 | R1(D, A, B, C, X14, 9, 0xc33707d6L); | 228 | md5_round2(&C, D, A, B, X3, 0xf4d50d87L, 14); |
| 190 | R1(C, D, A, B, X3, 14, 0xf4d50d87L); | 229 | md5_round2(&B, C, D, A, X8, 0x455a14edL, 20); |
| 191 | R1(B, C, D, A, X8, 20, 0x455a14edL); | 230 | md5_round2(&A, B, C, D, X13, 0xa9e3e905L, 5); |
| 192 | R1(A, B, C, D, X13, 5, 0xa9e3e905L); | 231 | md5_round2(&D, A, B, C, X2, 0xfcefa3f8L, 9); |
| 193 | R1(D, A, B, C, X2, 9, 0xfcefa3f8L); | 232 | md5_round2(&C, D, A, B, X7, 0x676f02d9L, 14); |
| 194 | R1(C, D, A, B, X7, 14, 0x676f02d9L); | 233 | md5_round2(&B, C, D, A, X12, 0x8d2a4c8aL, 20); |
| 195 | R1(B, C, D, A, X12, 20, 0x8d2a4c8aL); | 234 | |
| 196 | /* Round 2 */ | 235 | md5_round3(&A, B, C, D, X5, 0xfffa3942L, 4); |
| 197 | R2(A, B, C, D, X5, 4, 0xfffa3942L); | 236 | md5_round3(&D, A, B, C, X8, 0x8771f681L, 11); |
| 198 | R2(D, A, B, C, X8, 11, 0x8771f681L); | 237 | md5_round3(&C, D, A, B, X11, 0x6d9d6122L, 16); |
| 199 | R2(C, D, A, B, X11, 16, 0x6d9d6122L); | 238 | md5_round3(&B, C, D, A, X14, 0xfde5380cL, 23); |
| 200 | R2(B, C, D, A, X14, 23, 0xfde5380cL); | 239 | md5_round3(&A, B, C, D, X1, 0xa4beea44L, 4); |
| 201 | R2(A, B, C, D, X1, 4, 0xa4beea44L); | 240 | md5_round3(&D, A, B, C, X4, 0x4bdecfa9L, 11); |
| 202 | R2(D, A, B, C, X4, 11, 0x4bdecfa9L); | 241 | md5_round3(&C, D, A, B, X7, 0xf6bb4b60L, 16); |
| 203 | R2(C, D, A, B, X7, 16, 0xf6bb4b60L); | 242 | md5_round3(&B, C, D, A, X10, 0xbebfbc70L, 23); |
| 204 | R2(B, C, D, A, X10, 23, 0xbebfbc70L); | 243 | md5_round3(&A, B, C, D, X13, 0x289b7ec6L, 4); |
| 205 | R2(A, B, C, D, X13, 4, 0x289b7ec6L); | 244 | md5_round3(&D, A, B, C, X0, 0xeaa127faL, 11); |
| 206 | R2(D, A, B, C, X0, 11, 0xeaa127faL); | 245 | md5_round3(&C, D, A, B, X3, 0xd4ef3085L, 16); |
| 207 | R2(C, D, A, B, X3, 16, 0xd4ef3085L); | 246 | md5_round3(&B, C, D, A, X6, 0x04881d05L, 23); |
| 208 | R2(B, C, D, A, X6, 23, 0x04881d05L); | 247 | md5_round3(&A, B, C, D, X9, 0xd9d4d039L, 4); |
| 209 | R2(A, B, C, D, X9, 4, 0xd9d4d039L); | 248 | md5_round3(&D, A, B, C, X12, 0xe6db99e5L, 11); |
| 210 | R2(D, A, B, C, X12, 11, 0xe6db99e5L); | 249 | md5_round3(&C, D, A, B, X15, 0x1fa27cf8L, 16); |
| 211 | R2(C, D, A, B, X15, 16, 0x1fa27cf8L); | 250 | md5_round3(&B, C, D, A, X2, 0xc4ac5665L, 23); |
| 212 | R2(B, C, D, A, X2, 23, 0xc4ac5665L); | 251 | |
| 213 | /* Round 3 */ | 252 | md5_round4(&A, B, C, D, X0, 0xf4292244L, 6); |
| 214 | R3(A, B, C, D, X0, 6, 0xf4292244L); | 253 | md5_round4(&D, A, B, C, X7, 0x432aff97L, 10); |
| 215 | R3(D, A, B, C, X7, 10, 0x432aff97L); | 254 | md5_round4(&C, D, A, B, X14, 0xab9423a7L, 15); |
| 216 | R3(C, D, A, B, X14, 15, 0xab9423a7L); | 255 | md5_round4(&B, C, D, A, X5, 0xfc93a039L, 21); |
| 217 | R3(B, C, D, A, X5, 21, 0xfc93a039L); | 256 | md5_round4(&A, B, C, D, X12, 0x655b59c3L, 6); |
| 218 | R3(A, B, C, D, X12, 6, 0x655b59c3L); | 257 | md5_round4(&D, A, B, C, X3, 0x8f0ccc92L, 10); |
| 219 | R3(D, A, B, C, X3, 10, 0x8f0ccc92L); | 258 | md5_round4(&C, D, A, B, X10, 0xffeff47dL, 15); |
| 220 | R3(C, D, A, B, X10, 15, 0xffeff47dL); | 259 | md5_round4(&B, C, D, A, X1, 0x85845dd1L, 21); |
| 221 | R3(B, C, D, A, X1, 21, 0x85845dd1L); | 260 | md5_round4(&A, B, C, D, X8, 0x6fa87e4fL, 6); |
| 222 | R3(A, B, C, D, X8, 6, 0x6fa87e4fL); | 261 | md5_round4(&D, A, B, C, X15, 0xfe2ce6e0L, 10); |
| 223 | R3(D, A, B, C, X15, 10, 0xfe2ce6e0L); | 262 | md5_round4(&C, D, A, B, X6, 0xa3014314L, 15); |
| 224 | R3(C, D, A, B, X6, 15, 0xa3014314L); | 263 | md5_round4(&B, C, D, A, X13, 0x4e0811a1L, 21); |
| 225 | R3(B, C, D, A, X13, 21, 0x4e0811a1L); | 264 | md5_round4(&A, B, C, D, X4, 0xf7537e82L, 6); |
| 226 | R3(A, B, C, D, X4, 6, 0xf7537e82L); | 265 | md5_round4(&D, A, B, C, X11, 0xbd3af235L, 10); |
| 227 | R3(D, A, B, C, X11, 10, 0xbd3af235L); | 266 | md5_round4(&C, D, A, B, X2, 0x2ad7d2bbL, 15); |
| 228 | R3(C, D, A, B, X2, 15, 0x2ad7d2bbL); | 267 | md5_round4(&B, C, D, A, X9, 0xeb86d391L, 21); |
| 229 | R3(B, C, D, A, X9, 21, 0xeb86d391L); | ||
| 230 | 268 | ||
| 231 | A = c->A += A; | 269 | A = c->A += A; |
| 232 | B = c->B += B; | 270 | B = c->B += B; |
