diff options
Diffstat (limited to 'src/plugins/crypto_native/aes_cbc.c')
-rw-r--r-- | src/plugins/crypto_native/aes_cbc.c | 459 |
1 files changed, 28 insertions, 431 deletions
diff --git a/src/plugins/crypto_native/aes_cbc.c b/src/plugins/crypto_native/aes_cbc.c index 02d96b31c79..1f21dc149fa 100644 --- a/src/plugins/crypto_native/aes_cbc.c +++ b/src/plugins/crypto_native/aes_cbc.c @@ -19,7 +19,7 @@ #include <vnet/plugin/plugin.h> #include <vnet/crypto/crypto.h> #include <crypto_native/crypto_native.h> -#include <crypto_native/aes.h> +#include <vppinfra/crypto/aes_cbc.h> #if __GNUC__ > 4 && !__clang__ && CLIB_DEBUG == 0 #pragma GCC optimize ("O3") @@ -48,412 +48,6 @@ #define u32xN_splat u32x4_splat #endif -typedef struct -{ - u8x16 encrypt_key[15]; - u8xN decrypt_key[15]; -} aes_cbc_key_data_t; - -static_always_inline void __clib_unused -aes_cbc_dec (u8x16 * k, u8x16u * src, u8x16u * dst, u8x16u * iv, int count, - int rounds) -{ - u8x16 r[4], c[4], f; - - f = iv[0]; - while (count >= 64) - { - clib_prefetch_load (src + 8); - clib_prefetch_load (dst + 8); - - c[0] = r[0] = src[0]; - c[1] = r[1] = src[1]; - c[2] = r[2] = src[2]; - c[3] = r[3] = src[3]; - -#if __x86_64__ - r[0] ^= k[0]; - r[1] ^= k[0]; - r[2] ^= k[0]; - r[3] ^= k[0]; - - for (int i = 1; i < rounds; i++) - { - r[0] = aes_dec_round (r[0], k[i]); - r[1] = aes_dec_round (r[1], k[i]); - r[2] = aes_dec_round (r[2], k[i]); - r[3] = aes_dec_round (r[3], k[i]); - } - - r[0] = aes_dec_last_round (r[0], k[rounds]); - r[1] = aes_dec_last_round (r[1], k[rounds]); - r[2] = aes_dec_last_round (r[2], k[rounds]); - r[3] = aes_dec_last_round (r[3], k[rounds]); -#else - for (int i = 0; i < rounds - 1; i++) - { - r[0] = vaesimcq_u8 (vaesdq_u8 (r[0], k[i])); - r[1] = vaesimcq_u8 (vaesdq_u8 (r[1], k[i])); - r[2] = vaesimcq_u8 (vaesdq_u8 (r[2], k[i])); - r[3] = vaesimcq_u8 (vaesdq_u8 (r[3], k[i])); - } - r[0] = vaesdq_u8 (r[0], k[rounds - 1]) ^ k[rounds]; - r[1] = vaesdq_u8 (r[1], k[rounds - 1]) ^ k[rounds]; - r[2] = vaesdq_u8 (r[2], k[rounds - 1]) ^ k[rounds]; - r[3] = vaesdq_u8 (r[3], k[rounds - 1]) ^ k[rounds]; -#endif - dst[0] = r[0] ^ f; - dst[1] = r[1] ^ c[0]; - dst[2] = r[2] ^ c[1]; - dst[3] = r[3] ^ c[2]; - f = c[3]; - - count -= 64; - src += 4; - dst += 4; - } - - while (count > 0) - { - c[0] = r[0] = src[0]; -#if __x86_64__ - r[0] ^= k[0]; - for (int i = 1; i < rounds; i++) - r[0] = aes_dec_round (r[0], k[i]); - r[0] = aes_dec_last_round (r[0], k[rounds]); -#else - c[0] = r[0] = src[0]; - for (int i = 0; i < rounds - 1; i++) - r[0] = vaesimcq_u8 (vaesdq_u8 (r[0], k[i])); - r[0] = vaesdq_u8 (r[0], k[rounds - 1]) ^ k[rounds]; -#endif - dst[0] = r[0] ^ f; - f = c[0]; - - count -= 16; - src += 1; - dst += 1; - } -} - -#if __x86_64__ -#if defined(__VAES__) && defined(__AVX512F__) - -static_always_inline u8x64 -aes_block_load_x4 (u8 * src[], int i) -{ - u8x64 r = { }; - r = u8x64_insert_u8x16 (r, aes_block_load (src[0] + i), 0); - r = u8x64_insert_u8x16 (r, aes_block_load (src[1] + i), 1); - r = u8x64_insert_u8x16 (r, aes_block_load (src[2] + i), 2); - r = u8x64_insert_u8x16 (r, aes_block_load (src[3] + i), 3); - return r; -} - -static_always_inline void -aes_block_store_x4 (u8 * dst[], int i, u8x64 r) -{ - aes_block_store (dst[0] + i, u8x64_extract_u8x16 (r, 0)); - aes_block_store (dst[1] + i, u8x64_extract_u8x16 (r, 1)); - aes_block_store (dst[2] + i, u8x64_extract_u8x16 (r, 2)); - aes_block_store (dst[3] + i, u8x64_extract_u8x16 (r, 3)); -} - -static_always_inline u8x64 -aes4_cbc_dec_permute (u8x64 a, u8x64 b) -{ - return (u8x64) u64x8_shuffle2 (a, b, 6, 7, 8, 9, 10, 11, 12, 13); -} - -static_always_inline void -aes4_cbc_dec (u8x64 *k, u8x64u *src, u8x64u *dst, u8x16u *iv, int count, - aes_key_size_t rounds) -{ - u8x64 f, r[4], c[4] = { }; - __mmask8 m; - int i, n_blocks = count >> 4; - - f = (u8x64) _mm512_mask_loadu_epi64 (_mm512_setzero_si512 (), 0xc0, - (__m512i *) (iv - 3)); - - while (n_blocks >= 16) - { - c[0] = src[0]; - c[1] = src[1]; - c[2] = src[2]; - c[3] = src[3]; - - r[0] = c[0] ^ k[0]; - r[1] = c[1] ^ k[0]; - r[2] = c[2] ^ k[0]; - r[3] = c[3] ^ k[0]; - - for (i = 1; i < rounds; i++) - { - r[0] = aes_dec_round_x4 (r[0], k[i]); - r[1] = aes_dec_round_x4 (r[1], k[i]); - r[2] = aes_dec_round_x4 (r[2], k[i]); - r[3] = aes_dec_round_x4 (r[3], k[i]); - } - - r[0] = aes_dec_last_round_x4 (r[0], k[i]); - r[1] = aes_dec_last_round_x4 (r[1], k[i]); - r[2] = aes_dec_last_round_x4 (r[2], k[i]); - r[3] = aes_dec_last_round_x4 (r[3], k[i]); - - dst[0] = r[0] ^= aes4_cbc_dec_permute (f, c[0]); - dst[1] = r[1] ^= aes4_cbc_dec_permute (c[0], c[1]); - dst[2] = r[2] ^= aes4_cbc_dec_permute (c[1], c[2]); - dst[3] = r[3] ^= aes4_cbc_dec_permute (c[2], c[3]); - f = c[3]; - - n_blocks -= 16; - src += 4; - dst += 4; - } - - if (n_blocks >= 12) - { - c[0] = src[0]; - c[1] = src[1]; - c[2] = src[2]; - - r[0] = c[0] ^ k[0]; - r[1] = c[1] ^ k[0]; - r[2] = c[2] ^ k[0]; - - for (i = 1; i < rounds; i++) - { - r[0] = aes_dec_round_x4 (r[0], k[i]); - r[1] = aes_dec_round_x4 (r[1], k[i]); - r[2] = aes_dec_round_x4 (r[2], k[i]); - } - - r[0] = aes_dec_last_round_x4 (r[0], k[i]); - r[1] = aes_dec_last_round_x4 (r[1], k[i]); - r[2] = aes_dec_last_round_x4 (r[2], k[i]); - - dst[0] = r[0] ^= aes4_cbc_dec_permute (f, c[0]); - dst[1] = r[1] ^= aes4_cbc_dec_permute (c[0], c[1]); - dst[2] = r[2] ^= aes4_cbc_dec_permute (c[1], c[2]); - f = c[2]; - - n_blocks -= 12; - src += 3; - dst += 3; - } - else if (n_blocks >= 8) - { - c[0] = src[0]; - c[1] = src[1]; - - r[0] = c[0] ^ k[0]; - r[1] = c[1] ^ k[0]; - - for (i = 1; i < rounds; i++) - { - r[0] = aes_dec_round_x4 (r[0], k[i]); - r[1] = aes_dec_round_x4 (r[1], k[i]); - } - - r[0] = aes_dec_last_round_x4 (r[0], k[i]); - r[1] = aes_dec_last_round_x4 (r[1], k[i]); - - dst[0] = r[0] ^= aes4_cbc_dec_permute (f, c[0]); - dst[1] = r[1] ^= aes4_cbc_dec_permute (c[0], c[1]); - f = c[1]; - - n_blocks -= 8; - src += 2; - dst += 2; - } - else if (n_blocks >= 4) - { - c[0] = src[0]; - - r[0] = c[0] ^ k[0]; - - for (i = 1; i < rounds; i++) - { - r[0] = aes_dec_round_x4 (r[0], k[i]); - } - - r[0] = aes_dec_last_round_x4 (r[0], k[i]); - - dst[0] = r[0] ^= aes4_cbc_dec_permute (f, c[0]); - f = c[0]; - - n_blocks -= 4; - src += 1; - dst += 1; - } - - if (n_blocks > 0) - { - m = (1 << (n_blocks * 2)) - 1; - c[0] = (u8x64) _mm512_mask_loadu_epi64 ((__m512i) c[0], m, - (__m512i *) src); - f = aes4_cbc_dec_permute (f, c[0]); - r[0] = c[0] ^ k[0]; - for (i = 1; i < rounds; i++) - r[0] = aes_dec_round_x4 (r[0], k[i]); - r[0] = aes_dec_last_round_x4 (r[0], k[i]); - _mm512_mask_storeu_epi64 ((__m512i *) dst, m, (__m512i) (r[0] ^ f)); - } -} -#elif defined(__VAES__) - -static_always_inline u8x32 -aes_block_load_x2 (u8 *src[], int i) -{ - u8x32 r = {}; - r = u8x32_insert_lo (r, aes_block_load (src[0] + i)); - r = u8x32_insert_hi (r, aes_block_load (src[1] + i)); - return r; -} - -static_always_inline void -aes_block_store_x2 (u8 *dst[], int i, u8x32 r) -{ - aes_block_store (dst[0] + i, u8x32_extract_lo (r)); - aes_block_store (dst[1] + i, u8x32_extract_hi (r)); -} - -static_always_inline u8x32 -aes2_cbc_dec_permute (u8x32 a, u8x32 b) -{ - return (u8x32) u64x4_shuffle2 ((u64x4) a, (u64x4) b, 2, 3, 4, 5); -} - -static_always_inline void -aes2_cbc_dec (u8x32 *k, u8x32u *src, u8x32u *dst, u8x16u *iv, int count, - aes_key_size_t rounds) -{ - u8x32 f = {}, r[4], c[4] = {}; - int i, n_blocks = count >> 4; - - f = u8x32_insert_hi (f, *iv); - - while (n_blocks >= 8) - { - c[0] = src[0]; - c[1] = src[1]; - c[2] = src[2]; - c[3] = src[3]; - - r[0] = c[0] ^ k[0]; - r[1] = c[1] ^ k[0]; - r[2] = c[2] ^ k[0]; - r[3] = c[3] ^ k[0]; - - for (i = 1; i < rounds; i++) - { - r[0] = aes_dec_round_x2 (r[0], k[i]); - r[1] = aes_dec_round_x2 (r[1], k[i]); - r[2] = aes_dec_round_x2 (r[2], k[i]); - r[3] = aes_dec_round_x2 (r[3], k[i]); - } - - r[0] = aes_dec_last_round_x2 (r[0], k[i]); - r[1] = aes_dec_last_round_x2 (r[1], k[i]); - r[2] = aes_dec_last_round_x2 (r[2], k[i]); - r[3] = aes_dec_last_round_x2 (r[3], k[i]); - - dst[0] = r[0] ^= aes2_cbc_dec_permute (f, c[0]); - dst[1] = r[1] ^= aes2_cbc_dec_permute (c[0], c[1]); - dst[2] = r[2] ^= aes2_cbc_dec_permute (c[1], c[2]); - dst[3] = r[3] ^= aes2_cbc_dec_permute (c[2], c[3]); - f = c[3]; - - n_blocks -= 8; - src += 4; - dst += 4; - } - - if (n_blocks >= 6) - { - c[0] = src[0]; - c[1] = src[1]; - c[2] = src[2]; - - r[0] = c[0] ^ k[0]; - r[1] = c[1] ^ k[0]; - r[2] = c[2] ^ k[0]; - - for (i = 1; i < rounds; i++) - { - r[0] = aes_dec_round_x2 (r[0], k[i]); - r[1] = aes_dec_round_x2 (r[1], k[i]); - r[2] = aes_dec_round_x2 (r[2], k[i]); - } - - r[0] = aes_dec_last_round_x2 (r[0], k[i]); - r[1] = aes_dec_last_round_x2 (r[1], k[i]); - r[2] = aes_dec_last_round_x2 (r[2], k[i]); - - dst[0] = r[0] ^= aes2_cbc_dec_permute (f, c[0]); - dst[1] = r[1] ^= aes2_cbc_dec_permute (c[0], c[1]); - dst[2] = r[2] ^= aes2_cbc_dec_permute (c[1], c[2]); - f = c[2]; - - n_blocks -= 6; - src += 3; - dst += 3; - } - else if (n_blocks >= 4) - { - c[0] = src[0]; - c[1] = src[1]; - - r[0] = c[0] ^ k[0]; - r[1] = c[1] ^ k[0]; - - for (i = 1; i < rounds; i++) - { - r[0] = aes_dec_round_x2 (r[0], k[i]); - r[1] = aes_dec_round_x2 (r[1], k[i]); - } - - r[0] = aes_dec_last_round_x2 (r[0], k[i]); - r[1] = aes_dec_last_round_x2 (r[1], k[i]); - - dst[0] = r[0] ^= aes2_cbc_dec_permute (f, c[0]); - dst[1] = r[1] ^= aes2_cbc_dec_permute (c[0], c[1]); - f = c[1]; - - n_blocks -= 4; - src += 2; - dst += 2; - } - else if (n_blocks >= 2) - { - c[0] = src[0]; - r[0] = c[0] ^ k[0]; - - for (i = 1; i < rounds; i++) - r[0] = aes_dec_round_x2 (r[0], k[i]); - - r[0] = aes_dec_last_round_x2 (r[0], k[i]); - dst[0] = r[0] ^= aes2_cbc_dec_permute (f, c[0]); - f = c[0]; - - n_blocks -= 2; - src += 1; - dst += 1; - } - - if (n_blocks > 0) - { - u8x16 rl = *(u8x16u *) src ^ u8x32_extract_lo (k[0]); - for (i = 1; i < rounds; i++) - rl = aes_dec_round (rl, u8x32_extract_lo (k[i])); - rl = aes_dec_last_round (rl, u8x32_extract_lo (k[i])); - *(u8x16 *) dst = rl ^ u8x32_extract_hi (f); - } -} -#endif -#endif - static_always_inline u32 aes_ops_enc_aes_cbc (vlib_main_t * vm, vnet_crypto_op_t * ops[], u32 n_ops, aes_key_size_t ks) @@ -658,28 +252,6 @@ decrypt: return n_ops; } -static_always_inline void * -aes_cbc_key_exp (vnet_crypto_key_t * key, aes_key_size_t ks) -{ - u8x16 e[15], d[15]; - aes_cbc_key_data_t *kd; - kd = clib_mem_alloc_aligned (sizeof (*kd), CLIB_CACHE_LINE_BYTES); - aes_key_expand (e, key->data, ks); - aes_key_enc_to_dec (e, d, ks); - for (int i = 0; i < AES_KEY_ROUNDS (ks) + 1; i++) - { -#if defined(__VAES__) && defined(__AVX512F__) - kd->decrypt_key[i] = u8x64_splat_u8x16 (d[i]); -#elif defined(__VAES__) - kd->decrypt_key[i] = u8x32_splat_u8x16 (d[i]); -#else - kd->decrypt_key[i] = d[i]; -#endif - kd->encrypt_key[i] = e[i]; - } - return kd; -} - #define foreach_aes_cbc_handler_type _(128) _(192) _(256) #define _(x) \ @@ -689,12 +261,37 @@ static u32 aes_ops_dec_aes_cbc_##x \ static u32 aes_ops_enc_aes_cbc_##x \ (vlib_main_t * vm, vnet_crypto_op_t * ops[], u32 n_ops) \ { return aes_ops_enc_aes_cbc (vm, ops, n_ops, AES_KEY_##x); } \ -static void * aes_cbc_key_exp_##x (vnet_crypto_key_t *key) \ -{ return aes_cbc_key_exp (key, AES_KEY_##x); } foreach_aes_cbc_handler_type; #undef _ +static void * +aes_cbc_key_exp_128 (vnet_crypto_key_t *key) +{ + aes_cbc_key_data_t *kd; + kd = clib_mem_alloc_aligned (sizeof (*kd), CLIB_CACHE_LINE_BYTES); + clib_aes128_cbc_key_expand (kd, key->data); + return kd; +} + +static void * +aes_cbc_key_exp_192 (vnet_crypto_key_t *key) +{ + aes_cbc_key_data_t *kd; + kd = clib_mem_alloc_aligned (sizeof (*kd), CLIB_CACHE_LINE_BYTES); + clib_aes192_cbc_key_expand (kd, key->data); + return kd; +} + +static void * +aes_cbc_key_exp_256 (vnet_crypto_key_t *key) +{ + aes_cbc_key_data_t *kd; + kd = clib_mem_alloc_aligned (sizeof (*kd), CLIB_CACHE_LINE_BYTES); + clib_aes256_cbc_key_expand (kd, key->data); + return kd; +} + #include <fcntl.h> clib_error_t * |