aboutsummaryrefslogtreecommitdiffstats
path: root/src/plugins
diff options
context:
space:
mode:
authorNeale Ranns <nranns@cisco.com>2019-04-16 21:39:37 +0000
committerDamjan Marion <dmarion@me.com>2019-04-17 13:03:45 +0000
commitd8cfbebce78e26a6ef7f6693e7c90dc3c6435d51 (patch)
tree1f5553c691b72e9b158e3a9d57cc443e5a477868 /src/plugins
parent732a0f103315559aef9c2872cb485270c6133268 (diff)
crypto-ipsecmb: enable GCM
Change-Id: I670d7899bcc63a419daf481167dc445a6386cce8 Signed-off-by: Neale Ranns <nranns@cisco.com>
Diffstat (limited to 'src/plugins')
-rw-r--r--src/plugins/crypto_ipsecmb/ipsecmb.c244
1 files changed, 220 insertions, 24 deletions
diff --git a/src/plugins/crypto_ipsecmb/ipsecmb.c b/src/plugins/crypto_ipsecmb/ipsecmb.c
index d579c0d74a6..9b3cf3b7afd 100644
--- a/src/plugins/crypto_ipsecmb/ipsecmb.c
+++ b/src/plugins/crypto_ipsecmb/ipsecmb.c
@@ -36,6 +36,26 @@ typedef struct ipsecmb_main_t_
ipsecmb_per_thread_data_t *per_thread_data;
} ipsecmb_main_t;
+/**
+ * AES GCM key=expansion VFT
+ */
+typedef void (*ase_gcm_pre_t) (const void *key,
+ struct gcm_key_data * key_data);
+
+typedef struct ipsecmb_gcm_pre_vft_t_
+{
+ ase_gcm_pre_t ase_gcm_pre_128;
+ ase_gcm_pre_t ase_gcm_pre_192;
+ ase_gcm_pre_t ase_gcm_pre_256;
+} ipsecmb_gcm_pre_vft_t;
+
+static ipsecmb_gcm_pre_vft_t ipsecmb_gcm_pre_vft;
+
+#define INIT_IPSEC_MB_GCM_PRE(_arch) \
+ ipsecmb_gcm_pre_vft.ase_gcm_pre_128 = aes_gcm_pre_128_##_arch; \
+ ipsecmb_gcm_pre_vft.ase_gcm_pre_192 = aes_gcm_pre_192_##_arch; \
+ ipsecmb_gcm_pre_vft.ase_gcm_pre_256 = aes_gcm_pre_256_##_arch;
+
static ipsecmb_main_t ipsecmb_main;
#define foreach_ipsecmb_hmac_op \
@@ -44,11 +64,22 @@ static ipsecmb_main_t ipsecmb_main;
_(SHA384, SHA_384, sha384) \
_(SHA512, SHA_512, sha512)
-#define foreach_ipsecmb_cipher_op \
+/*
+ * (Alg, key-len-bits, key-len-bytes, iv-len-bytes)
+ */
+#define foreach_ipsecmb_cbc_cipher_op \
_(AES_128_CBC, 128, 16, 16) \
_(AES_192_CBC, 192, 24, 16) \
_(AES_256_CBC, 256, 32, 16)
+/*
+ * (Alg, key-len-bits, key-len-bytes, iv-len-bytes)
+ */
+#define foreach_ipsecmb_gcm_cipher_op \
+ _(AES_128_GCM, 128, 16, 12) \
+ _(AES_192_GCM, 192, 24, 12) \
+ _(AES_256_GCM, 256, 32, 12)
+
always_inline void
hash_expand_keys (const MB_MGR * mgr,
const u8 * key,
@@ -202,11 +233,11 @@ ipsecmb_retire_cipher_job (JOB_AES_HMAC * job, u32 * n_fail)
}
static_always_inline u32
-ipsecmb_ops_cipher_inline (vlib_main_t * vm,
- ipsecmb_per_thread_data_t * ptd,
- vnet_crypto_op_t * ops[],
- u32 n_ops, u32 key_len, u32 iv_len,
- keyexp_t fn, JOB_CIPHER_DIRECTION direction)
+ipsecmb_ops_cbc_cipher_inline (vlib_main_t * vm,
+ ipsecmb_per_thread_data_t * ptd,
+ vnet_crypto_op_t * ops[],
+ u32 n_ops, u32 key_len, u32 iv_len,
+ keyexp_t fn, JOB_CIPHER_DIRECTION direction)
{
JOB_AES_HMAC *job;
u32 i, n_fail = 0;
@@ -270,9 +301,9 @@ ipsecmb_ops_cipher_inline (vlib_main_t * vm,
#define _(a, b, c, d) \
static_always_inline u32 \
-ipsecmb_ops_cipher_enc_##a (vlib_main_t * vm, \
- vnet_crypto_op_t * ops[], \
- u32 n_ops) \
+ipsecmb_ops_cbc_cipher_enc_##a (vlib_main_t * vm, \
+ vnet_crypto_op_t * ops[], \
+ u32 n_ops) \
{ \
ipsecmb_per_thread_data_t *ptd; \
ipsecmb_main_t *imbm; \
@@ -280,18 +311,18 @@ ipsecmb_ops_cipher_enc_##a (vlib_main_t * vm, \
imbm = &ipsecmb_main; \
ptd = vec_elt_at_index (imbm->per_thread_data, vm->thread_index); \
\
- return ipsecmb_ops_cipher_inline (vm, ptd, ops, n_ops, c, d, \
- ptd->mgr->keyexp_##b, \
- ENCRYPT); \
+ return ipsecmb_ops_cbc_cipher_inline (vm, ptd, ops, n_ops, c, d, \
+ ptd->mgr->keyexp_##b, \
+ ENCRYPT); \
}
-foreach_ipsecmb_cipher_op;
+foreach_ipsecmb_cbc_cipher_op;
#undef _
#define _(a, b, c, d) \
static_always_inline u32 \
-ipsecmb_ops_cipher_dec_##a (vlib_main_t * vm, \
- vnet_crypto_op_t * ops[], \
- u32 n_ops) \
+ipsecmb_ops_cbc_cipher_dec_##a (vlib_main_t * vm, \
+ vnet_crypto_op_t * ops[], \
+ u32 n_ops) \
{ \
ipsecmb_per_thread_data_t *ptd; \
ipsecmb_main_t *imbm; \
@@ -299,11 +330,161 @@ ipsecmb_ops_cipher_dec_##a (vlib_main_t * vm, \
imbm = &ipsecmb_main; \
ptd = vec_elt_at_index (imbm->per_thread_data, vm->thread_index); \
\
- return ipsecmb_ops_cipher_inline (vm, ptd, ops, n_ops, c, d, \
- ptd->mgr->keyexp_##b, \
- DECRYPT); \
+ return ipsecmb_ops_cbc_cipher_inline (vm, ptd, ops, n_ops, c, d, \
+ ptd->mgr->keyexp_##b, \
+ DECRYPT); \
+ }
+foreach_ipsecmb_cbc_cipher_op;
+#undef _
+
+always_inline void
+ipsecmb_retire_gcm_cipher_job (JOB_AES_HMAC * job,
+ u32 * n_fail, JOB_CIPHER_DIRECTION direction)
+{
+ vnet_crypto_op_t *op = job->user_data;
+
+ if (STS_COMPLETED != job->status)
+ {
+ op->status = VNET_CRYPTO_OP_STATUS_FAIL_BAD_HMAC;
+ *n_fail = *n_fail + 1;
+ }
+ else
+ op->status = VNET_CRYPTO_OP_STATUS_COMPLETED;
+
+ if (DECRYPT == direction)
+ {
+ if ((memcmp (op->tag, job->auth_tag_output, op->tag_len)))
+ {
+ *n_fail = *n_fail + 1;
+ op->status = VNET_CRYPTO_OP_STATUS_FAIL_BAD_HMAC;
+ }
+ }
+}
+
+static_always_inline u32
+ipsecmb_ops_gcm_cipher_inline (vlib_main_t * vm,
+ ipsecmb_per_thread_data_t * ptd,
+ vnet_crypto_op_t * ops[],
+ u32 n_ops, u32 key_len, u32 iv_len,
+ ase_gcm_pre_t fn,
+ JOB_CIPHER_DIRECTION direction)
+{
+ JOB_AES_HMAC *job;
+ u32 i, n_fail = 0;
+ u8 scratch[n_ops][64];
+
+ /*
+ * queue all the jobs first ...
+ */
+ for (i = 0; i < n_ops; i++)
+ {
+ struct gcm_key_data key_data;
+ vnet_crypto_op_t *op = ops[i];
+ u32 nonce[3];
+ __m128i iv;
+
+ fn (op->key, &key_data);
+
+ job = IMB_GET_NEXT_JOB (ptd->mgr);
+
+ job->src = op->src;
+ job->dst = op->dst;
+ job->msg_len_to_cipher_in_bytes = op->len;
+ job->cipher_start_src_offset_in_bytes = 0;
+
+ job->hash_alg = AES_GMAC;
+ job->cipher_mode = GCM;
+ job->cipher_direction = direction;
+ job->chain_order = (direction == ENCRYPT ? CIPHER_HASH : HASH_CIPHER);
+
+ if (direction == ENCRYPT)
+ {
+ if (op->flags & VNET_CRYPTO_OP_FLAG_INIT_IV)
+ {
+ iv = ptd->cbc_iv;
+ // only use 8 bytes of the IV
+ clib_memcpy_fast (op->iv, &iv, 8);
+ ptd->cbc_iv = _mm_aesenc_si128 (iv, iv);
+ }
+ nonce[0] = op->salt;
+ clib_memcpy_fast (nonce + 1, op->iv, 8);
+ job->iv = (u8 *) nonce;
+ }
+ else
+ {
+ nonce[0] = op->salt;
+ clib_memcpy_fast (nonce + 1, op->iv, 8);
+ job->iv = op->iv;
+ }
+
+ job->aes_key_len_in_bytes = key_len;
+ job->aes_enc_key_expanded = &key_data;
+ job->aes_dec_key_expanded = &key_data;
+ job->iv_len_in_bytes = iv_len;
+
+ job->u.GCM.aad = op->aad;
+ job->u.GCM.aad_len_in_bytes = op->aad_len;
+ job->auth_tag_output_len_in_bytes = op->tag_len;
+ if (DECRYPT == direction)
+ job->auth_tag_output = scratch[i];
+ else
+ job->auth_tag_output = op->tag;
+ job->user_data = op;
+
+ job = IMB_SUBMIT_JOB (ptd->mgr);
+
+ if (job)
+ ipsecmb_retire_gcm_cipher_job (job, &n_fail, direction);
+ }
+
+ /*
+ * .. then flush (i.e. complete) them
+ * We will have queued enough to satisfy the 'multi' buffer
+ */
+ while ((job = IMB_FLUSH_JOB (ptd->mgr)))
+ {
+ ipsecmb_retire_gcm_cipher_job (job, &n_fail, direction);
+ }
+
+ return n_ops - n_fail;
+}
+
+#define _(a, b, c, d) \
+static_always_inline u32 \
+ipsecmb_ops_gcm_cipher_enc_##a (vlib_main_t * vm, \
+ vnet_crypto_op_t * ops[], \
+ u32 n_ops) \
+{ \
+ ipsecmb_per_thread_data_t *ptd; \
+ ipsecmb_main_t *imbm; \
+ \
+ imbm = &ipsecmb_main; \
+ ptd = vec_elt_at_index (imbm->per_thread_data, vm->thread_index); \
+ \
+ return ipsecmb_ops_gcm_cipher_inline (vm, ptd, ops, n_ops, c, d, \
+ ipsecmb_gcm_pre_vft.ase_gcm_pre_##b, \
+ ENCRYPT); \
+ }
+foreach_ipsecmb_gcm_cipher_op;
+#undef _
+
+#define _(a, b, c, d) \
+static_always_inline u32 \
+ipsecmb_ops_gcm_cipher_dec_##a (vlib_main_t * vm, \
+ vnet_crypto_op_t * ops[], \
+ u32 n_ops) \
+{ \
+ ipsecmb_per_thread_data_t *ptd; \
+ ipsecmb_main_t *imbm; \
+ \
+ imbm = &ipsecmb_main; \
+ ptd = vec_elt_at_index (imbm->per_thread_data, vm->thread_index); \
+ \
+ return ipsecmb_ops_gcm_cipher_inline (vm, ptd, ops, n_ops, c, d, \
+ ipsecmb_gcm_pre_vft.ase_gcm_pre_##b, \
+ DECRYPT); \
}
-foreach_ipsecmb_cipher_op;
+foreach_ipsecmb_gcm_cipher_op;
#undef _
clib_error_t *
@@ -356,6 +537,7 @@ crypto_ipsecmb_init (vlib_main_t * vm)
{
ptd->mgr = alloc_mb_mgr (0);
init_mb_mgr_avx512 (ptd->mgr);
+ INIT_IPSEC_MB_GCM_PRE (avx_gen4);
}
}
else if (clib_cpu_supports_avx2 ())
@@ -364,6 +546,7 @@ crypto_ipsecmb_init (vlib_main_t * vm)
{
ptd->mgr = alloc_mb_mgr (0);
init_mb_mgr_avx2 (ptd->mgr);
+ INIT_IPSEC_MB_GCM_PRE (avx_gen2);
}
}
else
@@ -372,6 +555,7 @@ crypto_ipsecmb_init (vlib_main_t * vm)
{
ptd->mgr = alloc_mb_mgr (0);
init_mb_mgr_sse (ptd->mgr);
+ INIT_IPSEC_MB_GCM_PRE (sse);
}
}
@@ -387,15 +571,27 @@ crypto_ipsecmb_init (vlib_main_t * vm)
#undef _
#define _(a, b, c, d) \
vnet_crypto_register_ops_handler (vm, eidx, VNET_CRYPTO_OP_##a##_ENC, \
- ipsecmb_ops_cipher_enc_##a); \
+ ipsecmb_ops_cbc_cipher_enc_##a); \
+
+ foreach_ipsecmb_cbc_cipher_op;
+#undef _
+#define _(a, b, c, d) \
+ vnet_crypto_register_ops_handler (vm, eidx, VNET_CRYPTO_OP_##a##_DEC, \
+ ipsecmb_ops_cbc_cipher_dec_##a); \
+
+ foreach_ipsecmb_cbc_cipher_op;
+#undef _
+#define _(a, b, c, d) \
+ vnet_crypto_register_ops_handler (vm, eidx, VNET_CRYPTO_OP_##a##_ENC, \
+ ipsecmb_ops_gcm_cipher_enc_##a); \
- foreach_ipsecmb_cipher_op;
+ foreach_ipsecmb_gcm_cipher_op;
#undef _
#define _(a, b, c, d) \
vnet_crypto_register_ops_handler (vm, eidx, VNET_CRYPTO_OP_##a##_DEC, \
- ipsecmb_ops_cipher_dec_##a); \
+ ipsecmb_ops_gcm_cipher_dec_##a); \
- foreach_ipsecmb_cipher_op;
+ foreach_ipsecmb_gcm_cipher_op;
#undef _
return (NULL);