summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--MAINTAINERS6
-rw-r--r--src/plugins/crypto_native/CMakeLists.txt (renamed from src/plugins/crypto_ia32/CMakeLists.txt)6
-rw-r--r--src/plugins/crypto_native/FEATURE.yaml (renamed from src/plugins/crypto_ia32/FEATURE.yaml)0
-rw-r--r--src/plugins/crypto_native/aes.h (renamed from src/plugins/crypto_ia32/aesni.h)24
-rw-r--r--src/plugins/crypto_native/aes_cbc.c (renamed from src/plugins/crypto_ia32/aes_cbc.c)46
-rw-r--r--src/plugins/crypto_native/aes_gcm.c (renamed from src/plugins/crypto_ia32/aes_gcm.c)38
-rw-r--r--src/plugins/crypto_native/crypto_native.h (renamed from src/plugins/crypto_ia32/crypto_ia32.h)34
-rw-r--r--src/plugins/crypto_native/ghash.h (renamed from src/plugins/crypto_ia32/ghash.h)0
-rw-r--r--src/plugins/crypto_native/main.c (renamed from src/plugins/crypto_ia32/main.c)38
9 files changed, 96 insertions, 96 deletions
diff --git a/MAINTAINERS b/MAINTAINERS
index c0af54cc35b..e0afbaa2792 100644
--- a/MAINTAINERS
+++ b/MAINTAINERS
@@ -186,10 +186,10 @@ I: nhrp
M: Neale Ranns <nranns@cisco.com>
F: src/vnet/nhrp/
-Crypto IA32 Plugin
-I: crypto-ia32
+Crypto native Plugin
+I: crypto-native
M: Damjan Marion <damarion@cisco.com>
-F: src/plugins/crypto_ia32/
+F: src/plugins/crypto_native/
Crypto openssl Plugin
I: crypto-openssl
diff --git a/src/plugins/crypto_ia32/CMakeLists.txt b/src/plugins/crypto_native/CMakeLists.txt
index 0a8b560731a..cd701ec7d55 100644
--- a/src/plugins/crypto_ia32/CMakeLists.txt
+++ b/src/plugins/crypto_native/CMakeLists.txt
@@ -15,7 +15,7 @@ if(NOT CMAKE_SYSTEM_PROCESSOR MATCHES "amd64.*|x86_64.*|AMD64.*")
return()
endif()
-add_vpp_plugin(crypto_ia32 SOURCES main.c)
+add_vpp_plugin(crypto_native SOURCES main.c)
list(APPEND VARIANTS "sse42\;-march=silvermont")
list(APPEND VARIANTS "avx2\;-march=core-avx2")
@@ -29,9 +29,9 @@ endif()
foreach(VARIANT ${VARIANTS})
list(GET VARIANT 0 v)
list(GET VARIANT 1 f)
- set(l crypto_ia32_${v})
+ set(l crypto_native_${v})
add_library(${l} OBJECT aes_cbc.c aes_gcm.c)
set_target_properties(${l} PROPERTIES POSITION_INDEPENDENT_CODE ON)
target_compile_options(${l} PUBLIC ${f} -Wall -fno-common -maes)
- target_sources(crypto_ia32_plugin PRIVATE $<TARGET_OBJECTS:${l}>)
+ target_sources(crypto_native_plugin PRIVATE $<TARGET_OBJECTS:${l}>)
endforeach()
diff --git a/src/plugins/crypto_ia32/FEATURE.yaml b/src/plugins/crypto_native/FEATURE.yaml
index 206caceb2d4..206caceb2d4 100644
--- a/src/plugins/crypto_ia32/FEATURE.yaml
+++ b/src/plugins/crypto_native/FEATURE.yaml
diff --git a/src/plugins/crypto_ia32/aesni.h b/src/plugins/crypto_native/aes.h
index ece61c13cf3..d0923913034 100644
--- a/src/plugins/crypto_ia32/aesni.h
+++ b/src/plugins/crypto_native/aes.h
@@ -20,13 +20,13 @@
typedef enum
{
- AESNI_KEY_128 = 0,
- AESNI_KEY_192 = 1,
- AESNI_KEY_256 = 2,
-} aesni_key_size_t;
+ AES_KEY_128 = 0,
+ AES_KEY_192 = 1,
+ AES_KEY_256 = 2,
+} aes_key_size_t;
-#define AESNI_KEY_ROUNDS(x) (10 + x *2)
-#define AESNI_KEY_BYTES(x) (16 + x * 8)
+#define AES_KEY_ROUNDS(x) (10 + x * 2)
+#define AES_KEY_BYTES(x) (16 + x * 8)
/* AES-NI based AES key expansion based on code samples from
@@ -178,17 +178,17 @@ aes256_key_expand (__m128i * k, u8 * key)
}
static_always_inline void
-aes_key_expand (__m128i * k, u8 * key, aesni_key_size_t ks)
+aes_key_expand (__m128i * k, u8 * key, aes_key_size_t ks)
{
switch (ks)
{
- case AESNI_KEY_128:
+ case AES_KEY_128:
aes128_key_expand (k, key);
break;
- case AESNI_KEY_192:
+ case AES_KEY_192:
aes192_key_expand (k, key);
break;
- case AESNI_KEY_256:
+ case AES_KEY_256:
aes256_key_expand (k, key);
break;
}
@@ -196,9 +196,9 @@ aes_key_expand (__m128i * k, u8 * key, aesni_key_size_t ks)
static_always_inline void
-aes_key_enc_to_dec (__m128i * k, aesni_key_size_t ks)
+aes_key_enc_to_dec (__m128i * k, aes_key_size_t ks)
{
- int rounds = AESNI_KEY_ROUNDS (ks);
+ int rounds = AES_KEY_ROUNDS (ks);
__m128i r;
r = k[rounds];
diff --git a/src/plugins/crypto_ia32/aes_cbc.c b/src/plugins/crypto_native/aes_cbc.c
index 910f8ca7413..c814b13725f 100644
--- a/src/plugins/crypto_ia32/aes_cbc.c
+++ b/src/plugins/crypto_native/aes_cbc.c
@@ -19,8 +19,8 @@
#include <vnet/plugin/plugin.h>
#include <vnet/crypto/crypto.h>
#include <x86intrin.h>
-#include <crypto_ia32/crypto_ia32.h>
-#include <crypto_ia32/aesni.h>
+#include <crypto_native/crypto_native.h>
+#include <crypto_native/aes.h>
#if __GNUC__ > 4 && !__clang__ && CLIB_DEBUG == 0
#pragma GCC optimize ("O3")
@@ -87,7 +87,7 @@ aes_block_store_x4 (u8 * dst[], int i, __m512i r)
static_always_inline void __clib_unused
aes_cbc_dec (__m128i * k, u8 * src, u8 * dst, u8 * iv, int count,
- aesni_key_size_t rounds)
+ aes_key_size_t rounds)
{
__m128i r0, r1, r2, r3, c0, c1, c2, c3, f;
int i;
@@ -152,7 +152,7 @@ aes_cbc_dec (__m128i * k, u8 * src, u8 * dst, u8 * iv, int count,
#ifdef __VAES__
static_always_inline void
vaes_cbc_dec (__m512i * k, u8 * src, u8 * dst, u8 * iv, int count,
- aesni_key_size_t rounds)
+ aes_key_size_t rounds)
{
__m512i permute = { 6, 7, 8, 9, 10, 11, 12, 13 };
__m512i r0, r1, r2, r3, c0, c1, c2, c3, f = { };
@@ -236,12 +236,12 @@ vaes_cbc_dec (__m512i * k, u8 * src, u8 * dst, u8 * iv, int count,
static_always_inline u32
aesni_ops_enc_aes_cbc (vlib_main_t * vm, vnet_crypto_op_t * ops[],
- u32 n_ops, aesni_key_size_t ks)
+ u32 n_ops, aes_key_size_t ks)
{
- crypto_ia32_main_t *cm = &crypto_ia32_main;
- crypto_ia32_per_thread_data_t *ptd = vec_elt_at_index (cm->per_thread_data,
- vm->thread_index);
- int rounds = AESNI_KEY_ROUNDS (ks);
+ crypto_native_main_t *cm = &crypto_native_main;
+ crypto_native_per_thread_data_t *ptd =
+ vec_elt_at_index (cm->per_thread_data, vm->thread_index);
+ int rounds = AES_KEY_ROUNDS (ks);
u8 dummy[8192];
u32 i, j, count, n_left = n_ops;
u32xN dummy_mask = { };
@@ -372,10 +372,10 @@ more:
static_always_inline u32
aesni_ops_dec_aes_cbc (vlib_main_t * vm, vnet_crypto_op_t * ops[],
- u32 n_ops, aesni_key_size_t ks)
+ u32 n_ops, aes_key_size_t ks)
{
- crypto_ia32_main_t *cm = &crypto_ia32_main;
- int rounds = AESNI_KEY_ROUNDS (ks);
+ crypto_native_main_t *cm = &crypto_native_main;
+ int rounds = AES_KEY_ROUNDS (ks);
vnet_crypto_op_t *op = ops[0];
aes_cbc_key_data_t *kd = (aes_cbc_key_data_t *) cm->key_data[op->key_index];
u32 n_left = n_ops;
@@ -401,7 +401,7 @@ decrypt:
}
static_always_inline void *
-aesni_cbc_key_exp (vnet_crypto_key_t * key, aesni_key_size_t ks)
+aesni_cbc_key_exp (vnet_crypto_key_t * key, aes_key_size_t ks)
{
__m128i e[15], d[15];
aes_cbc_key_data_t *kd;
@@ -409,7 +409,7 @@ aesni_cbc_key_exp (vnet_crypto_key_t * key, aesni_key_size_t ks)
aes_key_expand (e, key->data, ks);
aes_key_expand (d, key->data, ks);
aes_key_enc_to_dec (d, ks);
- for (int i = 0; i < AESNI_KEY_ROUNDS (ks) + 1; i++)
+ for (int i = 0; i < AES_KEY_ROUNDS (ks) + 1; i++)
{
#if __VAES__
kd->decrypt_key[i] = _mm512_broadcast_i64x2 (d[i]);
@@ -426,12 +426,12 @@ aesni_cbc_key_exp (vnet_crypto_key_t * key, aesni_key_size_t ks)
#define _(x) \
static u32 aesni_ops_dec_aes_cbc_##x \
(vlib_main_t * vm, vnet_crypto_op_t * ops[], u32 n_ops) \
-{ return aesni_ops_dec_aes_cbc (vm, ops, n_ops, AESNI_KEY_##x); } \
+{ return aesni_ops_dec_aes_cbc (vm, ops, n_ops, AES_KEY_##x); } \
static u32 aesni_ops_enc_aes_cbc_##x \
(vlib_main_t * vm, vnet_crypto_op_t * ops[], u32 n_ops) \
-{ return aesni_ops_enc_aes_cbc (vm, ops, n_ops, AESNI_KEY_##x); } \
+{ return aesni_ops_enc_aes_cbc (vm, ops, n_ops, AES_KEY_##x); } \
static void * aesni_cbc_key_exp_##x (vnet_crypto_key_t *key) \
-{ return aesni_cbc_key_exp (key, AESNI_KEY_##x); }
+{ return aesni_cbc_key_exp (key, AES_KEY_##x); }
foreach_aesni_cbc_handler_type;
#undef _
@@ -440,17 +440,17 @@ foreach_aesni_cbc_handler_type;
clib_error_t *
#ifdef __VAES__
-crypto_ia32_aesni_cbc_init_vaes (vlib_main_t * vm)
+crypto_native_aes_cbc_init_vaes (vlib_main_t * vm)
#elif __AVX512F__
-crypto_ia32_aesni_cbc_init_avx512 (vlib_main_t * vm)
+crypto_native_aes_cbc_init_avx512 (vlib_main_t * vm)
#elif __AVX2__
-crypto_ia32_aesni_cbc_init_avx2 (vlib_main_t * vm)
+crypto_native_aes_cbc_init_avx2 (vlib_main_t * vm)
#else
-crypto_ia32_aesni_cbc_init_sse42 (vlib_main_t * vm)
+crypto_native_aes_cbc_init_sse42 (vlib_main_t * vm)
#endif
{
- crypto_ia32_main_t *cm = &crypto_ia32_main;
- crypto_ia32_per_thread_data_t *ptd;
+ crypto_native_main_t *cm = &crypto_native_main;
+ crypto_native_per_thread_data_t *ptd;
clib_error_t *err = 0;
int fd;
diff --git a/src/plugins/crypto_ia32/aes_gcm.c b/src/plugins/crypto_native/aes_gcm.c
index e45dda79faf..3eb7ae84e19 100644
--- a/src/plugins/crypto_ia32/aes_gcm.c
+++ b/src/plugins/crypto_native/aes_gcm.c
@@ -19,9 +19,9 @@
#include <vnet/plugin/plugin.h>
#include <vnet/crypto/crypto.h>
#include <x86intrin.h>
-#include <crypto_ia32/crypto_ia32.h>
-#include <crypto_ia32/aesni.h>
-#include <crypto_ia32/ghash.h>
+#include <crypto_native/crypto_native.h>
+#include <crypto_native/aes.h>
+#include <crypto_native/ghash.h>
#if __GNUC__ > 4 && !__clang__ && CLIB_DEBUG == 0
#pragma GCC optimize ("O3")
@@ -650,9 +650,9 @@ aes_gcm (const u8 * in, u8 * out, const u8 * addt, const u8 * iv, u8 * tag,
static_always_inline u32
aesni_ops_enc_aes_gcm (vlib_main_t * vm, vnet_crypto_op_t * ops[],
- u32 n_ops, aesni_key_size_t ks)
+ u32 n_ops, aes_key_size_t ks)
{
- crypto_ia32_main_t *cm = &crypto_ia32_main;
+ crypto_native_main_t *cm = &crypto_native_main;
vnet_crypto_op_t *op = ops[0];
aes_gcm_key_data_t *kd;
u32 n_left = n_ops;
@@ -661,7 +661,7 @@ aesni_ops_enc_aes_gcm (vlib_main_t * vm, vnet_crypto_op_t * ops[],
next:
kd = (aes_gcm_key_data_t *) cm->key_data[op->key_index];
aes_gcm (op->src, op->dst, op->aad, op->iv, op->tag, op->len, op->aad_len,
- op->tag_len, kd, AESNI_KEY_ROUNDS (ks), /* is_encrypt */ 1);
+ op->tag_len, kd, AES_KEY_ROUNDS (ks), /* is_encrypt */ 1);
op->status = VNET_CRYPTO_OP_STATUS_COMPLETED;
if (--n_left)
@@ -675,9 +675,9 @@ next:
static_always_inline u32
aesni_ops_dec_aes_gcm (vlib_main_t * vm, vnet_crypto_op_t * ops[],
- u32 n_ops, aesni_key_size_t ks)
+ u32 n_ops, aes_key_size_t ks)
{
- crypto_ia32_main_t *cm = &crypto_ia32_main;
+ crypto_native_main_t *cm = &crypto_native_main;
vnet_crypto_op_t *op = ops[0];
aes_gcm_key_data_t *kd;
u32 n_left = n_ops;
@@ -686,7 +686,7 @@ aesni_ops_dec_aes_gcm (vlib_main_t * vm, vnet_crypto_op_t * ops[],
next:
kd = (aes_gcm_key_data_t *) cm->key_data[op->key_index];
rv = aes_gcm (op->src, op->dst, op->aad, op->iv, op->tag, op->len,
- op->aad_len, op->tag_len, kd, AESNI_KEY_ROUNDS (ks),
+ op->aad_len, op->tag_len, kd, AES_KEY_ROUNDS (ks),
/* is_encrypt */ 0);
if (rv)
@@ -709,7 +709,7 @@ next:
}
static_always_inline void *
-aesni_gcm_key_exp (vnet_crypto_key_t * key, aesni_key_size_t ks)
+aesni_gcm_key_exp (vnet_crypto_key_t * key, aes_key_size_t ks)
{
aes_gcm_key_data_t *kd;
__m128i H;
@@ -722,7 +722,7 @@ aesni_gcm_key_exp (vnet_crypto_key_t * key, aesni_key_size_t ks)
/* pre-calculate H */
H = kd->Ke[0];
- for (i = 1; i < AESNI_KEY_ROUNDS (ks); i += 1)
+ for (i = 1; i < AES_KEY_ROUNDS (ks); i += 1)
H = _mm_aesenc_si128 (H, kd->Ke[i]);
H = _mm_aesenclast_si128 (H, kd->Ke[i]);
H = aesni_gcm_bswap (H);
@@ -735,28 +735,28 @@ aesni_gcm_key_exp (vnet_crypto_key_t * key, aesni_key_size_t ks)
#define _(x) \
static u32 aesni_ops_dec_aes_gcm_##x \
(vlib_main_t * vm, vnet_crypto_op_t * ops[], u32 n_ops) \
-{ return aesni_ops_dec_aes_gcm (vm, ops, n_ops, AESNI_KEY_##x); } \
+{ return aesni_ops_dec_aes_gcm (vm, ops, n_ops, AES_KEY_##x); } \
static u32 aesni_ops_enc_aes_gcm_##x \
(vlib_main_t * vm, vnet_crypto_op_t * ops[], u32 n_ops) \
-{ return aesni_ops_enc_aes_gcm (vm, ops, n_ops, AESNI_KEY_##x); } \
+{ return aesni_ops_enc_aes_gcm (vm, ops, n_ops, AES_KEY_##x); } \
static void * aesni_gcm_key_exp_##x (vnet_crypto_key_t *key) \
-{ return aesni_gcm_key_exp (key, AESNI_KEY_##x); }
+{ return aesni_gcm_key_exp (key, AES_KEY_##x); }
foreach_aesni_gcm_handler_type;
#undef _
clib_error_t *
#ifdef __VAES__
-crypto_ia32_aesni_gcm_init_vaes (vlib_main_t * vm)
+crypto_native_aes_gcm_init_vaes (vlib_main_t * vm)
#elif __AVX512F__
-crypto_ia32_aesni_gcm_init_avx512 (vlib_main_t * vm)
+crypto_native_aes_gcm_init_avx512 (vlib_main_t * vm)
#elif __AVX2__
-crypto_ia32_aesni_gcm_init_avx2 (vlib_main_t * vm)
+crypto_native_aes_gcm_init_avx2 (vlib_main_t * vm)
#else
-crypto_ia32_aesni_gcm_init_sse42 (vlib_main_t * vm)
+crypto_native_aes_gcm_init_sse42 (vlib_main_t * vm)
#endif
{
- crypto_ia32_main_t *cm = &crypto_ia32_main;
+ crypto_native_main_t *cm = &crypto_native_main;
#define _(x) \
vnet_crypto_register_ops_handler (vm, cm->crypto_engine_index, \
diff --git a/src/plugins/crypto_ia32/crypto_ia32.h b/src/plugins/crypto_native/crypto_native.h
index e2b30071854..83362102f77 100644
--- a/src/plugins/crypto_ia32/crypto_ia32.h
+++ b/src/plugins/crypto_native/crypto_native.h
@@ -15,36 +15,36 @@
*------------------------------------------------------------------
*/
-#ifndef __crypto_ia32_h__
-#define __crypto_ia32_h__
+#ifndef __crypto_native_h__
+#define __crypto_native_h__
-typedef void *(crypto_ia32_key_fn_t) (vnet_crypto_key_t * key);
+typedef void *(crypto_native_key_fn_t) (vnet_crypto_key_t * key);
typedef struct
{
__m128i cbc_iv[4];
-} crypto_ia32_per_thread_data_t;
+} crypto_native_per_thread_data_t;
typedef struct
{
u32 crypto_engine_index;
- crypto_ia32_per_thread_data_t *per_thread_data;
- crypto_ia32_key_fn_t *key_fn[VNET_CRYPTO_N_ALGS];
+ crypto_native_per_thread_data_t *per_thread_data;
+ crypto_native_key_fn_t *key_fn[VNET_CRYPTO_N_ALGS];
void **key_data;
-} crypto_ia32_main_t;
+} crypto_native_main_t;
-extern crypto_ia32_main_t crypto_ia32_main;
+extern crypto_native_main_t crypto_native_main;
-clib_error_t *crypto_ia32_aesni_cbc_init_sse42 (vlib_main_t * vm);
-clib_error_t *crypto_ia32_aesni_cbc_init_avx2 (vlib_main_t * vm);
-clib_error_t *crypto_ia32_aesni_cbc_init_avx512 (vlib_main_t * vm);
-clib_error_t *crypto_ia32_aesni_cbc_init_vaes (vlib_main_t * vm);
+clib_error_t *crypto_native_aes_cbc_init_sse42 (vlib_main_t * vm);
+clib_error_t *crypto_native_aes_cbc_init_avx2 (vlib_main_t * vm);
+clib_error_t *crypto_native_aes_cbc_init_avx512 (vlib_main_t * vm);
+clib_error_t *crypto_native_aes_cbc_init_vaes (vlib_main_t * vm);
-clib_error_t *crypto_ia32_aesni_gcm_init_sse42 (vlib_main_t * vm);
-clib_error_t *crypto_ia32_aesni_gcm_init_avx2 (vlib_main_t * vm);
-clib_error_t *crypto_ia32_aesni_gcm_init_avx512 (vlib_main_t * vm);
-clib_error_t *crypto_ia32_aesni_gcm_init_vaes (vlib_main_t * vm);
-#endif /* __crypto_ia32_h__ */
+clib_error_t *crypto_native_aes_gcm_init_sse42 (vlib_main_t * vm);
+clib_error_t *crypto_native_aes_gcm_init_avx2 (vlib_main_t * vm);
+clib_error_t *crypto_native_aes_gcm_init_avx512 (vlib_main_t * vm);
+clib_error_t *crypto_native_aes_gcm_init_vaes (vlib_main_t * vm);
+#endif /* __crypto_native_h__ */
/*
* fd.io coding-style-patch-verification: ON
diff --git a/src/plugins/crypto_ia32/ghash.h b/src/plugins/crypto_native/ghash.h
index 0b2f629e28a..0b2f629e28a 100644
--- a/src/plugins/crypto_ia32/ghash.h
+++ b/src/plugins/crypto_native/ghash.h
diff --git a/src/plugins/crypto_ia32/main.c b/src/plugins/crypto_native/main.c
index bcfd7fa6ffd..2c03ee20309 100644
--- a/src/plugins/crypto_ia32/main.c
+++ b/src/plugins/crypto_native/main.c
@@ -18,16 +18,16 @@
#include <vlib/vlib.h>
#include <vnet/plugin/plugin.h>
#include <vnet/crypto/crypto.h>
-#include <crypto_ia32/crypto_ia32.h>
+#include <crypto_native/crypto_native.h>
-crypto_ia32_main_t crypto_ia32_main;
+crypto_native_main_t crypto_native_main;
static void
-crypto_ia32_key_handler (vlib_main_t * vm, vnet_crypto_key_op_t kop,
- vnet_crypto_key_index_t idx)
+crypto_native_key_handler (vlib_main_t * vm, vnet_crypto_key_op_t kop,
+ vnet_crypto_key_index_t idx)
{
vnet_crypto_key_t *key = vnet_crypto_get_key (idx);
- crypto_ia32_main_t *cm = &crypto_ia32_main;
+ crypto_native_main_t *cm = &crypto_native_main;
if (cm->key_fn[key->alg] == 0)
return;
@@ -56,9 +56,9 @@ crypto_ia32_key_handler (vlib_main_t * vm, vnet_crypto_key_op_t kop,
}
clib_error_t *
-crypto_ia32_init (vlib_main_t * vm)
+crypto_native_init (vlib_main_t * vm)
{
- crypto_ia32_main_t *cm = &crypto_ia32_main;
+ crypto_native_main_t *cm = &crypto_native_main;
vlib_thread_main_t *tm = vlib_get_thread_main ();
clib_error_t *error = 0;
@@ -69,17 +69,17 @@ crypto_ia32_init (vlib_main_t * vm)
CLIB_CACHE_LINE_BYTES);
cm->crypto_engine_index =
- vnet_crypto_register_engine (vm, "ia32", 100,
- "Intel IA32 ISA Optimized Crypto");
+ vnet_crypto_register_engine (vm, "native", 100,
+ "Native ISA Optimized Crypto");
if (clib_cpu_supports_vaes ())
- error = crypto_ia32_aesni_cbc_init_vaes (vm);
+ error = crypto_native_aes_cbc_init_vaes (vm);
else if (clib_cpu_supports_avx512f ())
- error = crypto_ia32_aesni_cbc_init_avx512 (vm);
+ error = crypto_native_aes_cbc_init_avx512 (vm);
else if (clib_cpu_supports_avx2 ())
- error = crypto_ia32_aesni_cbc_init_avx2 (vm);
+ error = crypto_native_aes_cbc_init_avx2 (vm);
else
- error = crypto_ia32_aesni_cbc_init_sse42 (vm);
+ error = crypto_native_aes_cbc_init_sse42 (vm);
if (error)
goto error;
@@ -87,20 +87,20 @@ crypto_ia32_init (vlib_main_t * vm)
if (clib_cpu_supports_pclmulqdq ())
{
if (clib_cpu_supports_vaes ())
- error = crypto_ia32_aesni_gcm_init_vaes (vm);
+ error = crypto_native_aes_gcm_init_vaes (vm);
else if (clib_cpu_supports_avx512f ())
- error = crypto_ia32_aesni_gcm_init_avx512 (vm);
+ error = crypto_native_aes_gcm_init_avx512 (vm);
else if (clib_cpu_supports_avx2 ())
- error = crypto_ia32_aesni_gcm_init_avx2 (vm);
+ error = crypto_native_aes_gcm_init_avx2 (vm);
else
- error = crypto_ia32_aesni_gcm_init_sse42 (vm);
+ error = crypto_native_aes_gcm_init_sse42 (vm);
if (error)
goto error;
}
vnet_crypto_register_key_handler (vm, cm->crypto_engine_index,
- crypto_ia32_key_handler);
+ crypto_native_key_handler);
error:
@@ -111,7 +111,7 @@ error:
}
/* *INDENT-OFF* */
-VLIB_INIT_FUNCTION (crypto_ia32_init) =
+VLIB_INIT_FUNCTION (crypto_native_init) =
{
.runs_after = VLIB_INITS ("vnet_crypto_init"),
};