aboutsummaryrefslogtreecommitdiffstats
path: root/src/plugins/crypto_ipsecmb/ipsecmb.c
diff options
context:
space:
mode:
Diffstat (limited to 'src/plugins/crypto_ipsecmb/ipsecmb.c')
-rw-r--r--src/plugins/crypto_ipsecmb/ipsecmb.c381
1 files changed, 381 insertions, 0 deletions
diff --git a/src/plugins/crypto_ipsecmb/ipsecmb.c b/src/plugins/crypto_ipsecmb/ipsecmb.c
new file mode 100644
index 00000000000..309623ec5e4
--- /dev/null
+++ b/src/plugins/crypto_ipsecmb/ipsecmb.c
@@ -0,0 +1,381 @@
+/*
+ * ipsecmb.c - Intel IPSec Multi-buffer library Crypto Engine
+ *
+ * Copyright (c) 2019 Cisco Systemss
+ * Licensed under the Apache License, Version 2.0 (the "License");
+ * you may not use this file except in compliance with the License.
+ * You may obtain a copy of the License at:
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+#include <intel-ipsec-mb.h>
+
+#include <vnet/vnet.h>
+#include <vnet/plugin/plugin.h>
+#include <vpp/app/version.h>
+#include <vnet/crypto/crypto.h>
+#include <vppinfra/cpu.h>
+
+typedef struct
+{
+ MB_MGR *mgr;
+} ipsecmb_per_thread_data_t;
+
+typedef struct ipsecmb_main_t_
+{
+ ipsecmb_per_thread_data_t *per_thread_data;
+} ipsecmb_main_t;
+
+static ipsecmb_main_t ipsecmb_main;
+
+#define foreach_ipsecmb_hmac_op \
+ _(SHA1, SHA1, sha1) \
+ _(SHA256, SHA_256, sha256) \
+ _(SHA384, SHA_384, sha384) \
+ _(SHA512, SHA_512, sha512)
+
+#define foreach_ipsecmb_cipher_op \
+ _(AES_128_CBC, 128) \
+ _(AES_192_CBC, 192) \
+ _(AES_256_CBC, 256)
+
+always_inline void
+hash_expand_keys (const MB_MGR * mgr,
+ const u8 * key,
+ u32 length,
+ u8 block_size,
+ u8 ipad[256], u8 opad[256], hash_one_block_t fn)
+{
+ u8 buf[block_size];
+ int i = 0;
+
+ if (length > block_size)
+ {
+ return;
+ }
+
+ memset (buf, 0x36, sizeof (buf));
+ for (i = 0; i < length; i++)
+ {
+ buf[i] ^= key[i];
+ }
+ fn (buf, ipad);
+
+ memset (buf, 0x5c, sizeof (buf));
+
+ for (i = 0; i < length; i++)
+ {
+ buf[i] ^= key[i];
+ }
+ fn (buf, opad);
+}
+
+always_inline void
+ipsecmb_retire_hmac_job (JOB_AES_HMAC * job, u32 * n_fail)
+{
+ vnet_crypto_op_t *op = job->user_data;
+
+ if (STS_COMPLETED != job->status)
+ {
+ op->status = VNET_CRYPTO_OP_STATUS_FAIL_BAD_HMAC;
+ *n_fail = *n_fail + 1;
+ }
+ else
+ op->status = VNET_CRYPTO_OP_STATUS_COMPLETED;
+
+ if (op->flags & VNET_CRYPTO_OP_FLAG_HMAC_CHECK)
+ {
+ if ((memcmp (op->digest, job->auth_tag_output, op->digest_len)))
+ {
+ *n_fail = *n_fail + 1;
+ op->status = VNET_CRYPTO_OP_STATUS_FAIL_BAD_HMAC;
+ }
+ }
+ else
+ clib_memcpy_fast (op->digest, job->auth_tag_output, op->digest_len);
+}
+
+static_always_inline u32
+ipsecmb_ops_hmac_inline (vlib_main_t * vm,
+ const ipsecmb_per_thread_data_t * ptd,
+ vnet_crypto_op_t * ops[],
+ u32 n_ops,
+ u32 block_size,
+ hash_one_block_t fn, JOB_HASH_ALG alg)
+{
+ JOB_AES_HMAC *job;
+ u32 i, n_fail = 0;
+ u8 scratch[n_ops][64];
+
+ /*
+ * queue all the jobs first ...
+ */
+ for (i = 0; i < n_ops; i++)
+ {
+ vnet_crypto_op_t *op = ops[i];
+ u8 ipad[256], opad[256];
+
+ hash_expand_keys (ptd->mgr, op->key, op->key_len,
+ block_size, ipad, opad, fn);
+
+ job = IMB_GET_NEXT_JOB (ptd->mgr);
+
+ job->src = op->src;
+ job->hash_start_src_offset_in_bytes = 0;
+ job->msg_len_to_hash_in_bytes = op->len;
+ job->hash_alg = alg;
+ job->auth_tag_output_len_in_bytes = op->digest_len;
+ job->auth_tag_output = scratch[i];
+
+ job->cipher_mode = NULL_CIPHER;
+ job->cipher_direction = DECRYPT;
+ job->chain_order = HASH_CIPHER;
+
+ job->aes_key_len_in_bytes = op->key_len;
+
+ job->u.HMAC._hashed_auth_key_xor_ipad = ipad;
+ job->u.HMAC._hashed_auth_key_xor_opad = opad;
+ job->user_data = op;
+
+ job = IMB_SUBMIT_JOB (ptd->mgr);
+
+ if (job)
+ ipsecmb_retire_hmac_job (job, &n_fail);
+ }
+
+ /*
+ * .. then flush (i.e. complete) them
+ * We will have queued enough to satisfy the 'multi' buffer
+ */
+ while ((job = IMB_FLUSH_JOB (ptd->mgr)))
+ {
+ ipsecmb_retire_hmac_job (job, &n_fail);
+ }
+
+ return n_ops - n_fail;
+}
+
+#define _(a, b, c) \
+static_always_inline u32 \
+ipsecmb_ops_hmac_##a (vlib_main_t * vm, \
+ vnet_crypto_op_t * ops[], \
+ u32 n_ops) \
+{ \
+ ipsecmb_per_thread_data_t *ptd; \
+ ipsecmb_main_t *imbm; \
+ \
+ imbm = &ipsecmb_main; \
+ ptd = vec_elt_at_index (imbm->per_thread_data, vm->thread_index); \
+ \
+ return ipsecmb_ops_hmac_inline (vm, ptd, ops, n_ops, \
+ b##_BLOCK_SIZE, \
+ ptd->mgr->c##_one_block, \
+ b); \
+ }
+foreach_ipsecmb_hmac_op;
+#undef _
+
+#define EXPANDED_KEY_N_BYTES (16 * 15)
+
+always_inline void
+ipsecmb_retire_cipher_job (JOB_AES_HMAC * job, u32 * n_fail)
+{
+ vnet_crypto_op_t *op = job->user_data;
+
+ if (STS_COMPLETED != job->status)
+ {
+ op->status = VNET_CRYPTO_OP_STATUS_FAIL_BAD_HMAC;
+ *n_fail = *n_fail + 1;
+ }
+ else
+ op->status = VNET_CRYPTO_OP_STATUS_COMPLETED;
+}
+
+static_always_inline u32
+ipsecmb_ops_cipher_inline (vlib_main_t * vm,
+ const ipsecmb_per_thread_data_t * ptd,
+ vnet_crypto_op_t * ops[],
+ u32 n_ops,
+ keyexp_t fn, JOB_CIPHER_DIRECTION direction)
+{
+ JOB_AES_HMAC *job;
+ u32 i, n_fail = 0;
+
+ /*
+ * queue all the jobs first ...
+ */
+ for (i = 0; i < n_ops; i++)
+ {
+ u8 aes_enc_key_expanded[EXPANDED_KEY_N_BYTES];
+ u8 aes_dec_key_expanded[EXPANDED_KEY_N_BYTES];
+ vnet_crypto_op_t *op = ops[i];
+
+ fn (op->key, aes_enc_key_expanded, aes_dec_key_expanded);
+
+ job = IMB_GET_NEXT_JOB (ptd->mgr);
+
+ job->src = op->src;
+ job->dst = op->dst;
+ job->msg_len_to_cipher_in_bytes = op->len;
+ job->cipher_start_src_offset_in_bytes = 0;
+
+ job->hash_alg = NULL_HASH;
+ job->cipher_mode = CBC;
+ job->cipher_direction = direction;
+ job->chain_order = (direction == ENCRYPT ? CIPHER_HASH : HASH_CIPHER);
+
+ job->aes_key_len_in_bytes = op->key_len;
+ job->aes_enc_key_expanded = aes_enc_key_expanded;
+ job->aes_dec_key_expanded = aes_dec_key_expanded;
+ job->iv = op->iv;
+ job->iv_len_in_bytes = op->iv_len;
+
+ job->user_data = op;
+
+ job = IMB_SUBMIT_JOB (ptd->mgr);
+
+ if (job)
+ ipsecmb_retire_cipher_job (job, &n_fail);
+ }
+
+ /*
+ * .. then flush (i.e. complete) them
+ * We will have queued enough to satisfy the 'multi' buffer
+ */
+ while ((job = IMB_FLUSH_JOB (ptd->mgr)))
+ {
+ ipsecmb_retire_cipher_job (job, &n_fail);
+ }
+
+ return n_ops - n_fail;
+}
+
+#define _(a, b) \
+static_always_inline u32 \
+ipsecmb_ops_cipher_enc_##a (vlib_main_t * vm, \
+ vnet_crypto_op_t * ops[], \
+ u32 n_ops) \
+{ \
+ ipsecmb_per_thread_data_t *ptd; \
+ ipsecmb_main_t *imbm; \
+ \
+ imbm = &ipsecmb_main; \
+ ptd = vec_elt_at_index (imbm->per_thread_data, vm->thread_index); \
+ \
+ return ipsecmb_ops_cipher_inline (vm, ptd, ops, n_ops, \
+ ptd->mgr->keyexp_##b, \
+ ENCRYPT); \
+ }
+foreach_ipsecmb_cipher_op;
+#undef _
+
+#define _(a, b) \
+static_always_inline u32 \
+ipsecmb_ops_cipher_dec_##a (vlib_main_t * vm, \
+ vnet_crypto_op_t * ops[], \
+ u32 n_ops) \
+{ \
+ ipsecmb_per_thread_data_t *ptd; \
+ ipsecmb_main_t *imbm; \
+ \
+ imbm = &ipsecmb_main; \
+ ptd = vec_elt_at_index (imbm->per_thread_data, vm->thread_index); \
+ \
+ return ipsecmb_ops_cipher_inline (vm, ptd, ops, n_ops, \
+ ptd->mgr->keyexp_##b, \
+ DECRYPT); \
+ }
+foreach_ipsecmb_cipher_op;
+#undef _
+
+static clib_error_t *
+crypto_ipsecmb_init (vlib_main_t * vm)
+{
+ ipsecmb_main_t *imbm = &ipsecmb_main;
+ ipsecmb_per_thread_data_t *ptd;
+ vlib_thread_main_t *tm = vlib_get_thread_main ();
+ clib_error_t *error;
+ u32 eidx;
+
+ if ((error = vlib_call_init_function (vm, vnet_crypto_init)))
+ return error;
+
+ /*
+ * A priority that is better than OpenSSL but worse than VPP natvie
+ */
+ eidx = vnet_crypto_register_engine (vm, "ipsecmb", 80,
+ "Intel IPSEC multi-buffer");
+
+ vec_validate (imbm->per_thread_data, tm->n_vlib_mains - 1);
+
+ if (clib_cpu_supports_avx512f ())
+ {
+ vec_foreach (ptd, imbm->per_thread_data)
+ {
+ ptd->mgr = alloc_mb_mgr (0);
+ init_mb_mgr_avx512 (ptd->mgr);
+ }
+ }
+ else if (clib_cpu_supports_avx2 ())
+ {
+ vec_foreach (ptd, imbm->per_thread_data)
+ {
+ ptd->mgr = alloc_mb_mgr (0);
+ init_mb_mgr_avx2 (ptd->mgr);
+ }
+ }
+ else
+ {
+ vec_foreach (ptd, imbm->per_thread_data)
+ {
+ ptd->mgr = alloc_mb_mgr (0);
+ init_mb_mgr_sse (ptd->mgr);
+ }
+ }
+
+#define _(a, b, c) \
+ vnet_crypto_register_ops_handler (vm, eidx, VNET_CRYPTO_OP_##a##_HMAC, \
+ ipsecmb_ops_hmac_##a); \
+
+ foreach_ipsecmb_hmac_op;
+#undef _
+#define _(a, b) \
+ vnet_crypto_register_ops_handler (vm, eidx, VNET_CRYPTO_OP_##a##_ENC, \
+ ipsecmb_ops_cipher_enc_##a); \
+
+ foreach_ipsecmb_cipher_op;
+#undef _
+#define _(a, b) \
+ vnet_crypto_register_ops_handler (vm, eidx, VNET_CRYPTO_OP_##a##_DEC, \
+ ipsecmb_ops_cipher_dec_##a); \
+
+ foreach_ipsecmb_cipher_op;
+#undef _
+
+ return 0;
+}
+
+VLIB_INIT_FUNCTION (crypto_ipsecmb_init);
+
+/* *INDENT-OFF* */
+VLIB_PLUGIN_REGISTER () =
+{
+ .version = VPP_BUILD_VER,
+ .description = "Intel IPSEC multi-buffer",
+};
+/* *INDENT-ON* */
+
+/*
+ * fd.io coding-style-patch-verification: ON
+ *
+ * Local Variables:
+ * eval: (c-set-style "gnu")
+ * End:
+ */