1 /* SHA-256 and SHA-512 implementation based on code by Oliver Gay 2 * <olivier.gay (at) a3.epfl.ch> under a BSD-style license. See below. 3 */ 4 5 /* 6 * FIPS 180-2 SHA-224/256/384/512 implementation 7 * Last update: 02/02/2007 8 * Issue date: 04/30/2005 9 * 10 * Copyright (C) 2005, 2007 Olivier Gay <olivier.gay (at) a3.epfl.ch> 11 * All rights reserved. 12 * 13 * Redistribution and use in source and binary forms, with or without 14 * modification, are permitted provided that the following conditions 15 * are met: 16 * 1. Redistributions of source code must retain the above copyright 17 * notice, this list of conditions and the following disclaimer. 18 * 2. Redistributions in binary form must reproduce the above copyright 19 * notice, this list of conditions and the following disclaimer in the 20 * documentation and/or other materials provided with the distribution. 21 * 3. Neither the name of the project nor the names of its contributors 22 * may be used to endorse or promote products derived from this software 23 * without specific prior written permission. 24 * 25 * THIS SOFTWARE IS PROVIDED BY THE PROJECT AND CONTRIBUTORS ``AS IS'' AND 26 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 27 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 28 * ARE DISCLAIMED. IN NO EVENT SHALL THE PROJECT OR CONTRIBUTORS BE LIABLE 29 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 30 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 31 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 32 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 33 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 34 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 35 * SUCH DAMAGE. 36 */ 37 38 #include "avb_sha.h" 39 40 #define SHFR(x, n) (x >> n) 41 #define ROTR(x, n) ((x >> n) | (x << ((sizeof(x) << 3) - n))) 42 #define ROTL(x, n) ((x << n) | (x >> ((sizeof(x) << 3) - n))) 43 #define CH(x, y, z) ((x & y) ^ (~x & z)) 44 #define MAJ(x, y, z) ((x & y) ^ (x & z) ^ (y & z)) 45 46 #define SHA256_F1(x) (ROTR(x, 2) ^ ROTR(x, 13) ^ ROTR(x, 22)) 47 #define SHA256_F2(x) (ROTR(x, 6) ^ ROTR(x, 11) ^ ROTR(x, 25)) 48 #define SHA256_F3(x) (ROTR(x, 7) ^ ROTR(x, 18) ^ SHFR(x, 3)) 49 #define SHA256_F4(x) (ROTR(x, 17) ^ ROTR(x, 19) ^ SHFR(x, 10)) 50 51 #define UNPACK32(x, str) \ 52 { \ 53 *((str) + 3) = (uint8_t)((x)); \ 54 *((str) + 2) = (uint8_t)((x) >> 8); \ 55 *((str) + 1) = (uint8_t)((x) >> 16); \ 56 *((str) + 0) = (uint8_t)((x) >> 24); \ 57 } 58 59 #define PACK32(str, x) \ 60 { \ 61 *(x) = ((uint32_t) * ((str) + 3)) | ((uint32_t) * ((str) + 2) << 8) | \ 62 ((uint32_t) * ((str) + 1) << 16) | \ 63 ((uint32_t) * ((str) + 0) << 24); \ 64 } 65 66 /* Macros used for loops unrolling */ 67 68 #define SHA256_SCR(i) \ 69 { w[i] = SHA256_F4(w[i - 2]) + w[i - 7] + SHA256_F3(w[i - 15]) + w[i - 16]; } 70 71 #define SHA256_EXP(a, b, c, d, e, f, g, h, j) \ 72 { \ 73 t1 = wv[h] + SHA256_F2(wv[e]) + CH(wv[e], wv[f], wv[g]) + sha256_k[j] + \ 74 w[j]; \ 75 t2 = SHA256_F1(wv[a]) + MAJ(wv[a], wv[b], wv[c]); \ 76 wv[d] += t1; \ 77 wv[h] = t1 + t2; \ 78 } 79 80 static const uint32_t sha256_h0[8] = {0x6a09e667, 81 0xbb67ae85, 82 0x3c6ef372, 83 0xa54ff53a, 84 0x510e527f, 85 0x9b05688c, 86 0x1f83d9ab, 87 0x5be0cd19}; 88 89 static const uint32_t sha256_k[64] = { 90 0x428a2f98, 0x71374491, 0xb5c0fbcf, 0xe9b5dba5, 0x3956c25b, 0x59f111f1, 91 0x923f82a4, 0xab1c5ed5, 0xd807aa98, 0x12835b01, 0x243185be, 0x550c7dc3, 92 0x72be5d74, 0x80deb1fe, 0x9bdc06a7, 0xc19bf174, 0xe49b69c1, 0xefbe4786, 93 0x0fc19dc6, 0x240ca1cc, 0x2de92c6f, 0x4a7484aa, 0x5cb0a9dc, 0x76f988da, 94 0x983e5152, 0xa831c66d, 0xb00327c8, 0xbf597fc7, 0xc6e00bf3, 0xd5a79147, 95 0x06ca6351, 0x14292967, 0x27b70a85, 0x2e1b2138, 0x4d2c6dfc, 0x53380d13, 96 0x650a7354, 0x766a0abb, 0x81c2c92e, 0x92722c85, 0xa2bfe8a1, 0xa81a664b, 97 0xc24b8b70, 0xc76c51a3, 0xd192e819, 0xd6990624, 0xf40e3585, 0x106aa070, 98 0x19a4c116, 0x1e376c08, 0x2748774c, 0x34b0bcb5, 0x391c0cb3, 0x4ed8aa4a, 99 0x5b9cca4f, 0x682e6ff3, 0x748f82ee, 0x78a5636f, 0x84c87814, 0x8cc70208, 100 0x90befffa, 0xa4506ceb, 0xbef9a3f7, 0xc67178f2}; 101 102 /* SHA-256 implementation */ 103 void avb_sha256_init(AvbSHA256Ctx* ctx) { 104 #ifndef UNROLL_LOOPS 105 int i; 106 for (i = 0; i < 8; i++) { 107 ctx->h[i] = sha256_h0[i]; 108 } 109 #else 110 ctx->h[0] = sha256_h0[0]; 111 ctx->h[1] = sha256_h0[1]; 112 ctx->h[2] = sha256_h0[2]; 113 ctx->h[3] = sha256_h0[3]; 114 ctx->h[4] = sha256_h0[4]; 115 ctx->h[5] = sha256_h0[5]; 116 ctx->h[6] = sha256_h0[6]; 117 ctx->h[7] = sha256_h0[7]; 118 #endif /* !UNROLL_LOOPS */ 119 120 ctx->len = 0; 121 ctx->tot_len = 0; 122 } 123 124 static void SHA256_transform(AvbSHA256Ctx* ctx, 125 const uint8_t* message, 126 unsigned int block_nb) { 127 uint32_t w[64]; 128 uint32_t wv[8]; 129 uint32_t t1, t2; 130 const unsigned char* sub_block; 131 int i; 132 133 #ifndef UNROLL_LOOPS 134 int j; 135 #endif 136 137 for (i = 0; i < (int)block_nb; i++) { 138 sub_block = message + (i << 6); 139 140 #ifndef UNROLL_LOOPS 141 for (j = 0; j < 16; j++) { 142 PACK32(&sub_block[j << 2], &w[j]); 143 } 144 145 for (j = 16; j < 64; j++) { 146 SHA256_SCR(j); 147 } 148 149 for (j = 0; j < 8; j++) { 150 wv[j] = ctx->h[j]; 151 } 152 153 for (j = 0; j < 64; j++) { 154 t1 = wv[7] + SHA256_F2(wv[4]) + CH(wv[4], wv[5], wv[6]) + sha256_k[j] + 155 w[j]; 156 t2 = SHA256_F1(wv[0]) + MAJ(wv[0], wv[1], wv[2]); 157 wv[7] = wv[6]; 158 wv[6] = wv[5]; 159 wv[5] = wv[4]; 160 wv[4] = wv[3] + t1; 161 wv[3] = wv[2]; 162 wv[2] = wv[1]; 163 wv[1] = wv[0]; 164 wv[0] = t1 + t2; 165 } 166 167 for (j = 0; j < 8; j++) { 168 ctx->h[j] += wv[j]; 169 } 170 #else 171 PACK32(&sub_block[0], &w[0]); 172 PACK32(&sub_block[4], &w[1]); 173 PACK32(&sub_block[8], &w[2]); 174 PACK32(&sub_block[12], &w[3]); 175 PACK32(&sub_block[16], &w[4]); 176 PACK32(&sub_block[20], &w[5]); 177 PACK32(&sub_block[24], &w[6]); 178 PACK32(&sub_block[28], &w[7]); 179 PACK32(&sub_block[32], &w[8]); 180 PACK32(&sub_block[36], &w[9]); 181 PACK32(&sub_block[40], &w[10]); 182 PACK32(&sub_block[44], &w[11]); 183 PACK32(&sub_block[48], &w[12]); 184 PACK32(&sub_block[52], &w[13]); 185 PACK32(&sub_block[56], &w[14]); 186 PACK32(&sub_block[60], &w[15]); 187 188 SHA256_SCR(16); 189 SHA256_SCR(17); 190 SHA256_SCR(18); 191 SHA256_SCR(19); 192 SHA256_SCR(20); 193 SHA256_SCR(21); 194 SHA256_SCR(22); 195 SHA256_SCR(23); 196 SHA256_SCR(24); 197 SHA256_SCR(25); 198 SHA256_SCR(26); 199 SHA256_SCR(27); 200 SHA256_SCR(28); 201 SHA256_SCR(29); 202 SHA256_SCR(30); 203 SHA256_SCR(31); 204 SHA256_SCR(32); 205 SHA256_SCR(33); 206 SHA256_SCR(34); 207 SHA256_SCR(35); 208 SHA256_SCR(36); 209 SHA256_SCR(37); 210 SHA256_SCR(38); 211 SHA256_SCR(39); 212 SHA256_SCR(40); 213 SHA256_SCR(41); 214 SHA256_SCR(42); 215 SHA256_SCR(43); 216 SHA256_SCR(44); 217 SHA256_SCR(45); 218 SHA256_SCR(46); 219 SHA256_SCR(47); 220 SHA256_SCR(48); 221 SHA256_SCR(49); 222 SHA256_SCR(50); 223 SHA256_SCR(51); 224 SHA256_SCR(52); 225 SHA256_SCR(53); 226 SHA256_SCR(54); 227 SHA256_SCR(55); 228 SHA256_SCR(56); 229 SHA256_SCR(57); 230 SHA256_SCR(58); 231 SHA256_SCR(59); 232 SHA256_SCR(60); 233 SHA256_SCR(61); 234 SHA256_SCR(62); 235 SHA256_SCR(63); 236 237 wv[0] = ctx->h[0]; 238 wv[1] = ctx->h[1]; 239 wv[2] = ctx->h[2]; 240 wv[3] = ctx->h[3]; 241 wv[4] = ctx->h[4]; 242 wv[5] = ctx->h[5]; 243 wv[6] = ctx->h[6]; 244 wv[7] = ctx->h[7]; 245 246 SHA256_EXP(0, 1, 2, 3, 4, 5, 6, 7, 0); 247 SHA256_EXP(7, 0, 1, 2, 3, 4, 5, 6, 1); 248 SHA256_EXP(6, 7, 0, 1, 2, 3, 4, 5, 2); 249 SHA256_EXP(5, 6, 7, 0, 1, 2, 3, 4, 3); 250 SHA256_EXP(4, 5, 6, 7, 0, 1, 2, 3, 4); 251 SHA256_EXP(3, 4, 5, 6, 7, 0, 1, 2, 5); 252 SHA256_EXP(2, 3, 4, 5, 6, 7, 0, 1, 6); 253 SHA256_EXP(1, 2, 3, 4, 5, 6, 7, 0, 7); 254 SHA256_EXP(0, 1, 2, 3, 4, 5, 6, 7, 8); 255 SHA256_EXP(7, 0, 1, 2, 3, 4, 5, 6, 9); 256 SHA256_EXP(6, 7, 0, 1, 2, 3, 4, 5, 10); 257 SHA256_EXP(5, 6, 7, 0, 1, 2, 3, 4, 11); 258 SHA256_EXP(4, 5, 6, 7, 0, 1, 2, 3, 12); 259 SHA256_EXP(3, 4, 5, 6, 7, 0, 1, 2, 13); 260 SHA256_EXP(2, 3, 4, 5, 6, 7, 0, 1, 14); 261 SHA256_EXP(1, 2, 3, 4, 5, 6, 7, 0, 15); 262 SHA256_EXP(0, 1, 2, 3, 4, 5, 6, 7, 16); 263 SHA256_EXP(7, 0, 1, 2, 3, 4, 5, 6, 17); 264 SHA256_EXP(6, 7, 0, 1, 2, 3, 4, 5, 18); 265 SHA256_EXP(5, 6, 7, 0, 1, 2, 3, 4, 19); 266 SHA256_EXP(4, 5, 6, 7, 0, 1, 2, 3, 20); 267 SHA256_EXP(3, 4, 5, 6, 7, 0, 1, 2, 21); 268 SHA256_EXP(2, 3, 4, 5, 6, 7, 0, 1, 22); 269 SHA256_EXP(1, 2, 3, 4, 5, 6, 7, 0, 23); 270 SHA256_EXP(0, 1, 2, 3, 4, 5, 6, 7, 24); 271 SHA256_EXP(7, 0, 1, 2, 3, 4, 5, 6, 25); 272 SHA256_EXP(6, 7, 0, 1, 2, 3, 4, 5, 26); 273 SHA256_EXP(5, 6, 7, 0, 1, 2, 3, 4, 27); 274 SHA256_EXP(4, 5, 6, 7, 0, 1, 2, 3, 28); 275 SHA256_EXP(3, 4, 5, 6, 7, 0, 1, 2, 29); 276 SHA256_EXP(2, 3, 4, 5, 6, 7, 0, 1, 30); 277 SHA256_EXP(1, 2, 3, 4, 5, 6, 7, 0, 31); 278 SHA256_EXP(0, 1, 2, 3, 4, 5, 6, 7, 32); 279 SHA256_EXP(7, 0, 1, 2, 3, 4, 5, 6, 33); 280 SHA256_EXP(6, 7, 0, 1, 2, 3, 4, 5, 34); 281 SHA256_EXP(5, 6, 7, 0, 1, 2, 3, 4, 35); 282 SHA256_EXP(4, 5, 6, 7, 0, 1, 2, 3, 36); 283 SHA256_EXP(3, 4, 5, 6, 7, 0, 1, 2, 37); 284 SHA256_EXP(2, 3, 4, 5, 6, 7, 0, 1, 38); 285 SHA256_EXP(1, 2, 3, 4, 5, 6, 7, 0, 39); 286 SHA256_EXP(0, 1, 2, 3, 4, 5, 6, 7, 40); 287 SHA256_EXP(7, 0, 1, 2, 3, 4, 5, 6, 41); 288 SHA256_EXP(6, 7, 0, 1, 2, 3, 4, 5, 42); 289 SHA256_EXP(5, 6, 7, 0, 1, 2, 3, 4, 43); 290 SHA256_EXP(4, 5, 6, 7, 0, 1, 2, 3, 44); 291 SHA256_EXP(3, 4, 5, 6, 7, 0, 1, 2, 45); 292 SHA256_EXP(2, 3, 4, 5, 6, 7, 0, 1, 46); 293 SHA256_EXP(1, 2, 3, 4, 5, 6, 7, 0, 47); 294 SHA256_EXP(0, 1, 2, 3, 4, 5, 6, 7, 48); 295 SHA256_EXP(7, 0, 1, 2, 3, 4, 5, 6, 49); 296 SHA256_EXP(6, 7, 0, 1, 2, 3, 4, 5, 50); 297 SHA256_EXP(5, 6, 7, 0, 1, 2, 3, 4, 51); 298 SHA256_EXP(4, 5, 6, 7, 0, 1, 2, 3, 52); 299 SHA256_EXP(3, 4, 5, 6, 7, 0, 1, 2, 53); 300 SHA256_EXP(2, 3, 4, 5, 6, 7, 0, 1, 54); 301 SHA256_EXP(1, 2, 3, 4, 5, 6, 7, 0, 55); 302 SHA256_EXP(0, 1, 2, 3, 4, 5, 6, 7, 56); 303 SHA256_EXP(7, 0, 1, 2, 3, 4, 5, 6, 57); 304 SHA256_EXP(6, 7, 0, 1, 2, 3, 4, 5, 58); 305 SHA256_EXP(5, 6, 7, 0, 1, 2, 3, 4, 59); 306 SHA256_EXP(4, 5, 6, 7, 0, 1, 2, 3, 60); 307 SHA256_EXP(3, 4, 5, 6, 7, 0, 1, 2, 61); 308 SHA256_EXP(2, 3, 4, 5, 6, 7, 0, 1, 62); 309 SHA256_EXP(1, 2, 3, 4, 5, 6, 7, 0, 63); 310 311 ctx->h[0] += wv[0]; 312 ctx->h[1] += wv[1]; 313 ctx->h[2] += wv[2]; 314 ctx->h[3] += wv[3]; 315 ctx->h[4] += wv[4]; 316 ctx->h[5] += wv[5]; 317 ctx->h[6] += wv[6]; 318 ctx->h[7] += wv[7]; 319 #endif /* !UNROLL_LOOPS */ 320 } 321 } 322 323 void avb_sha256_update(AvbSHA256Ctx* ctx, const uint8_t* data, uint32_t len) { 324 unsigned int block_nb; 325 unsigned int new_len, rem_len, tmp_len; 326 const uint8_t* shifted_data; 327 328 tmp_len = AVB_SHA256_BLOCK_SIZE - ctx->len; 329 rem_len = len < tmp_len ? len : tmp_len; 330 331 avb_memcpy(&ctx->block[ctx->len], data, rem_len); 332 333 if (ctx->len + len < AVB_SHA256_BLOCK_SIZE) { 334 ctx->len += len; 335 return; 336 } 337 338 new_len = len - rem_len; 339 block_nb = new_len / AVB_SHA256_BLOCK_SIZE; 340 341 shifted_data = data + rem_len; 342 343 SHA256_transform(ctx, ctx->block, 1); 344 SHA256_transform(ctx, shifted_data, block_nb); 345 346 rem_len = new_len % AVB_SHA256_BLOCK_SIZE; 347 348 avb_memcpy(ctx->block, &shifted_data[block_nb << 6], rem_len); 349 350 ctx->len = rem_len; 351 ctx->tot_len += (block_nb + 1) << 6; 352 } 353 354 uint8_t* avb_sha256_final(AvbSHA256Ctx* ctx) { 355 unsigned int block_nb; 356 unsigned int pm_len; 357 unsigned int len_b; 358 #ifndef UNROLL_LOOPS 359 int i; 360 #endif 361 362 block_nb = 363 (1 + ((AVB_SHA256_BLOCK_SIZE - 9) < (ctx->len % AVB_SHA256_BLOCK_SIZE))); 364 365 len_b = (ctx->tot_len + ctx->len) << 3; 366 pm_len = block_nb << 6; 367 368 avb_memset(ctx->block + ctx->len, 0, pm_len - ctx->len); 369 ctx->block[ctx->len] = 0x80; 370 UNPACK32(len_b, ctx->block + pm_len - 4); 371 372 SHA256_transform(ctx, ctx->block, block_nb); 373 374 #ifndef UNROLL_LOOPS 375 for (i = 0; i < 8; i++) { 376 UNPACK32(ctx->h[i], &ctx->buf[i << 2]); 377 } 378 #else 379 UNPACK32(ctx->h[0], &ctx->buf[0]); 380 UNPACK32(ctx->h[1], &ctx->buf[4]); 381 UNPACK32(ctx->h[2], &ctx->buf[8]); 382 UNPACK32(ctx->h[3], &ctx->buf[12]); 383 UNPACK32(ctx->h[4], &ctx->buf[16]); 384 UNPACK32(ctx->h[5], &ctx->buf[20]); 385 UNPACK32(ctx->h[6], &ctx->buf[24]); 386 UNPACK32(ctx->h[7], &ctx->buf[28]); 387 #endif /* !UNROLL_LOOPS */ 388 389 return ctx->buf; 390 } 391