// Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved. // SPDX-License-Identifier: Apache-2.0 // // An implementation of the compress function of SHA256 using avx // The implementation is based on: // Gueron, S., Krasnov, V. Parallelizing message schedules to accelerate the // computations of hash functions. J Cryptogr Eng 2, 241–253 (2012). // https://doi.org/10.1007/s13389-012-0037-z // // Written by Nir Drucker and Shay Gueron // AWS Cryptographic Algorithms Group. // (ndrucker@amazon.com, gueron@amazon.com) #include "internal/avx_defs.h" #include "sha256_defs.h" // This file depends on vec_t and on the macros LOAD, ADD32, ALIGNR8, SRL32, // SLL32, SRL64 that are defined in avx_defs.h #include "sha256_compress_x86_64_avx_helper.c" #define MS_VEC_NUM (SHA256_BLOCK_BYTE_LEN / sizeof(vec_t)) #define WORDS_IN_VEC (sizeof(vec_t) / sizeof(sha256_word_t)) _INLINE_ void load_data(OUT vec_t x[MS_VEC_NUM], IN OUT sha256_msg_schedule_t *ms, IN const uint8_t *data) { // 32 bits (4 bytes) swap masks const vec_t shuf_mask = _mm_setr_epi32(0x00010203, 0x04050607, 0x08090a0b, 0x0c0d0e0f); PRAGMA_LOOP_UNROLL_4 for(size_t i = 0; i < MS_VEC_NUM; i++) { x[i] = LOAD(&data[sizeof(vec_t) * i]); x[i] = SHUF8(x[i], shuf_mask); STORE(&ms->w[WORDS_IN_VEC * i], ADD32(x[i], LOAD(&K256[WORDS_IN_VEC * i]))); } } _INLINE_ void rounds_0_47(sha256_state_t * cur_state, vec_t x[MS_VEC_NUM], sha256_msg_schedule_t *ms) { const vec_t lo_mask = _mm_setr_epi32(0x03020100, 0x0b0a0908, -1, -1); const vec_t hi_mask = _mm_setr_epi32(-1, -1, 0x03020100, 0x0b0a0908); // The first SHA256_BLOCK_WORDS_NUM entries of K256 were loaded in // load_data(...). size_t k256_idx = SHA256_BLOCK_WORDS_NUM; // Rounds 0-47 (0-15, 16-31, 32-47) for(size_t i = 0; i < 3; i++) { PRAGMA_LOOP_UNROLL_4 for(size_t j = 0; j < MS_VEC_NUM; j++) { const size_t pos = WORDS_IN_VEC * j; const vec_t y = sha256_update_x_avx(x, &K256[k256_idx], lo_mask, hi_mask); sha_round(cur_state, ms->w[pos + 0], 0); sha_round(cur_state, ms->w[pos + 1], 0); sha_round(cur_state, ms->w[pos + 2], 0); sha_round(cur_state, ms->w[pos + 3], 0); STORE(&ms->w[pos], y); k256_idx += WORDS_IN_VEC; } } } _INLINE_ void rounds_48_63(sha256_state_t * cur_state, const sha256_msg_schedule_t *ms) { PRAGMA_LOOP_UNROLL_16 for(size_t i = SHA256_FINAL_ROUND_START_IDX; i < SHA256_ROUNDS_NUM; i++) { sha_round(cur_state, ms->w[LSB4(i)], 0); } } void sha256_compress_x86_64_avx(sha256_state_t *state, const uint8_t * data, size_t blocks_num) { sha256_state_t cur_state; sha256_msg_schedule_t ms; vec_t x[MS_VEC_NUM]; while(blocks_num--) { my_memcpy(cur_state.w, state->w, sizeof(cur_state.w)); load_data(x, &ms, data); data += SHA256_BLOCK_BYTE_LEN; rounds_0_47(&cur_state, x, &ms); rounds_48_63(&cur_state, &ms); accumulate_state(state, &cur_state); } secure_clean(&cur_state, sizeof(cur_state)); secure_clean(&ms, sizeof(ms)); }