aboutsummaryrefslogtreecommitdiffstats
path: root/src/vnet/crypto/node.c
diff options
context:
space:
mode:
authorFan Zhang <roy.fan.zhang@intel.com>2020-04-29 14:00:03 +0100
committerDamjan Marion <dmarion@me.com>2020-04-30 14:38:33 +0000
commitf539578bac8b64886b57c460c9d74273e6613f8b (patch)
tree190e09705fe1ebb46ca86a48c19de49fcaf0cbb0 /src/vnet/crypto/node.c
parent162330f25aeec09694fffaaa31ba9b318620eb9c (diff)
crypto: introduce async crypto infra
Type: feature Signed-off-by: Damjan Marion <damarion@cisco.com> Signed-off-by: Filip Tehlar <ftehlar@cisco.com> Signed-off-by: Fan Zhang <roy.fan.zhang@intel.com> Signed-off-by: Piotr Bronowski <piotrx.bronowski@intel.com> Signed-off-by: Dariusz Kazimierski <dariuszx.kazimierski@intel.com> Signed-off-by: Piotr Kleski <piotrx.kleski@intel.com> Change-Id: I4c3fcccf55c36842b7b48aed260fef2802b5c54b
Diffstat (limited to 'src/vnet/crypto/node.c')
-rw-r--r--src/vnet/crypto/node.c184
1 files changed, 184 insertions, 0 deletions
diff --git a/src/vnet/crypto/node.c b/src/vnet/crypto/node.c
new file mode 100644
index 00000000000..51ee63d1d62
--- /dev/null
+++ b/src/vnet/crypto/node.c
@@ -0,0 +1,184 @@
+/*
+ * Copyright (c) 2020 Cisco and/or its affiliates.
+ * Licensed under the Apache License, Version 2.0 (the "License");
+ * you may not use this file except in compliance with the License.
+ * You may obtain a copy of the License at:
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+#include <stdbool.h>
+#include <vlib/vlib.h>
+#include <vnet/crypto/crypto.h>
+
+typedef enum
+{
+#define _(sym,str) VNET_CRYPTO_ASYNC_ERROR_##sym,
+ foreach_crypto_op_status
+#undef _
+ VNET_CRYPTO_ASYNC_N_ERROR,
+} vnet_crypto_async_error_t;
+
+static char *vnet_crypto_async_error_strings[] = {
+#define _(sym,string) string,
+ foreach_crypto_op_status
+#undef _
+};
+
+#define foreach_crypto_dispatch_next \
+ _(ERR_DROP, "error-drop")
+
+typedef enum
+{
+#define _(n, s) CRYPTO_DISPATCH_NEXT_##n,
+ foreach_crypto_dispatch_next
+#undef _
+ CRYPTO_DISPATCH_N_NEXT,
+} crypto_dispatch_next_t;
+
+typedef struct
+{
+ vnet_crypto_op_status_t op_status;
+ vnet_crypto_async_op_id_t op;
+} crypto_dispatch_trace_t;
+
+static u8 *
+format_crypto_dispatch_trace (u8 * s, va_list * args)
+{
+ CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
+ CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
+ crypto_dispatch_trace_t *t = va_arg (*args, crypto_dispatch_trace_t *);
+
+ s = format (s, "%U: %U", format_vnet_crypto_async_op, t->op,
+ format_vnet_crypto_op_status, t->op_status);
+ return s;
+}
+
+static void
+vnet_crypto_async_add_trace (vlib_main_t * vm, vlib_node_runtime_t * node,
+ vlib_buffer_t * b,
+ vnet_crypto_async_op_id_t op_id,
+ vnet_crypto_op_status_t status)
+{
+ crypto_dispatch_trace_t *tr = vlib_add_trace (vm, node, b, sizeof (*tr));
+ tr->op_status = status;
+ tr->op = op_id;
+}
+
+static_always_inline u32
+crypto_dequeue_frame (vlib_main_t * vm, vlib_node_runtime_t * node,
+ vnet_crypto_thread_t * ct,
+ vnet_crypto_frame_dequeue_t * hdl,
+ u32 n_cache, u32 * n_total)
+{
+ vnet_crypto_async_frame_t *cf = (hdl) (vm);
+
+ while (cf)
+ {
+ vec_validate (ct->buffer_indice, n_cache + cf->n_elts);
+ vec_validate (ct->nexts, n_cache + cf->n_elts);
+ clib_memcpy_fast (ct->buffer_indice + n_cache, cf->buffer_indices,
+ sizeof (u32) * cf->n_elts);
+ if (cf->state == VNET_CRYPTO_FRAME_STATE_SUCCESS)
+ {
+ clib_memcpy_fast (ct->nexts + n_cache, cf->next_node_index,
+ sizeof (u16) * cf->n_elts);
+ }
+ else
+ {
+ u32 i;
+ for (i = 0; i < cf->n_elts; i++)
+ {
+ if (cf->elts[i].status != VNET_CRYPTO_OP_STATUS_COMPLETED)
+ {
+ ct->nexts[i + n_cache] = CRYPTO_DISPATCH_NEXT_ERR_DROP;
+ vlib_node_increment_counter (vm, node->node_index,
+ cf->elts[i].status, 1);
+ }
+ else
+ ct->nexts[i + n_cache] = cf->next_node_index[i];
+ }
+ }
+ n_cache += cf->n_elts;
+ *n_total += cf->n_elts;
+ if (n_cache >= VLIB_FRAME_SIZE)
+ {
+ vlib_buffer_enqueue_to_next (vm, node, ct->buffer_indice, ct->nexts,
+ n_cache);
+ n_cache = 0;
+ }
+
+ if (PREDICT_FALSE (node->flags & VLIB_NODE_FLAG_TRACE))
+ {
+ u32 i;
+
+ for (i = 0; i < cf->n_elts; i++)
+ {
+ vlib_buffer_t *b = vlib_get_buffer (vm, cf->buffer_indices[i]);
+ if (b->flags & VLIB_BUFFER_IS_TRACED)
+ vnet_crypto_async_add_trace (vm, node, b, cf->op,
+ cf->elts[i].status);
+ }
+ }
+ vnet_crypto_async_free_frame (vm, cf);
+ cf = (hdl) (vm);
+ }
+
+ return n_cache;
+}
+
+VLIB_NODE_FN (crypto_dispatch_node) (vlib_main_t * vm,
+ vlib_node_runtime_t * node,
+ vlib_frame_t * frame)
+{
+ vnet_crypto_main_t *cm = &crypto_main;
+ vnet_crypto_thread_t *ct = cm->threads + vm->thread_index;
+ u32 n_dispatched = 0, n_cache = 0;
+ u32 index;
+
+ /* *INDENT-OFF* */
+ clib_bitmap_foreach (index, cm->async_active_ids, ({
+ n_cache = crypto_dequeue_frame (vm, node, ct, cm->dequeue_handlers[index],
+ n_cache, &n_dispatched);
+ }));
+ /* *INDENT-ON* */
+ if (n_cache)
+ vlib_buffer_enqueue_to_next (vm, node, ct->buffer_indice, ct->nexts,
+ n_cache);
+
+ return n_dispatched;
+}
+
+/* *INDENT-OFF* */
+VLIB_REGISTER_NODE (crypto_dispatch_node) = {
+ .name = "crypto-dispatch",
+ .type = VLIB_NODE_TYPE_INPUT,
+ .state = VLIB_NODE_STATE_DISABLED,
+ .format_trace = format_crypto_dispatch_trace,
+
+ .n_errors = ARRAY_LEN(vnet_crypto_async_error_strings),
+ .error_strings = vnet_crypto_async_error_strings,
+
+ .n_next_nodes = CRYPTO_DISPATCH_N_NEXT,
+ .next_nodes = {
+#define _(n, s) \
+ [CRYPTO_DISPATCH_NEXT_##n] = s,
+ foreach_crypto_dispatch_next
+#undef _
+ },
+};
+/* *INDENT-ON* */
+
+/*
+ * fd.io coding-style-patch-verification: ON
+ *
+ * Local Variables:
+ * eval: (c-set-style "gnu")
+ * End:
+ */