diff options
author | Fan Zhang <roy.fan.zhang@intel.com> | 2020-04-29 14:00:03 +0100 |
---|---|---|
committer | Damjan Marion <dmarion@me.com> | 2020-04-30 14:38:33 +0000 |
commit | f539578bac8b64886b57c460c9d74273e6613f8b (patch) | |
tree | 190e09705fe1ebb46ca86a48c19de49fcaf0cbb0 /src/vnet/crypto/node.c | |
parent | 162330f25aeec09694fffaaa31ba9b318620eb9c (diff) |
crypto: introduce async crypto infra
Type: feature
Signed-off-by: Damjan Marion <damarion@cisco.com>
Signed-off-by: Filip Tehlar <ftehlar@cisco.com>
Signed-off-by: Fan Zhang <roy.fan.zhang@intel.com>
Signed-off-by: Piotr Bronowski <piotrx.bronowski@intel.com>
Signed-off-by: Dariusz Kazimierski <dariuszx.kazimierski@intel.com>
Signed-off-by: Piotr Kleski <piotrx.kleski@intel.com>
Change-Id: I4c3fcccf55c36842b7b48aed260fef2802b5c54b
Diffstat (limited to 'src/vnet/crypto/node.c')
-rw-r--r-- | src/vnet/crypto/node.c | 184 |
1 files changed, 184 insertions, 0 deletions
diff --git a/src/vnet/crypto/node.c b/src/vnet/crypto/node.c new file mode 100644 index 00000000000..51ee63d1d62 --- /dev/null +++ b/src/vnet/crypto/node.c @@ -0,0 +1,184 @@ +/* + * Copyright (c) 2020 Cisco and/or its affiliates. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at: + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +#include <stdbool.h> +#include <vlib/vlib.h> +#include <vnet/crypto/crypto.h> + +typedef enum +{ +#define _(sym,str) VNET_CRYPTO_ASYNC_ERROR_##sym, + foreach_crypto_op_status +#undef _ + VNET_CRYPTO_ASYNC_N_ERROR, +} vnet_crypto_async_error_t; + +static char *vnet_crypto_async_error_strings[] = { +#define _(sym,string) string, + foreach_crypto_op_status +#undef _ +}; + +#define foreach_crypto_dispatch_next \ + _(ERR_DROP, "error-drop") + +typedef enum +{ +#define _(n, s) CRYPTO_DISPATCH_NEXT_##n, + foreach_crypto_dispatch_next +#undef _ + CRYPTO_DISPATCH_N_NEXT, +} crypto_dispatch_next_t; + +typedef struct +{ + vnet_crypto_op_status_t op_status; + vnet_crypto_async_op_id_t op; +} crypto_dispatch_trace_t; + +static u8 * +format_crypto_dispatch_trace (u8 * s, va_list * args) +{ + CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *); + CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *); + crypto_dispatch_trace_t *t = va_arg (*args, crypto_dispatch_trace_t *); + + s = format (s, "%U: %U", format_vnet_crypto_async_op, t->op, + format_vnet_crypto_op_status, t->op_status); + return s; +} + +static void +vnet_crypto_async_add_trace (vlib_main_t * vm, vlib_node_runtime_t * node, + vlib_buffer_t * b, + vnet_crypto_async_op_id_t op_id, + vnet_crypto_op_status_t status) +{ + crypto_dispatch_trace_t *tr = vlib_add_trace (vm, node, b, sizeof (*tr)); + tr->op_status = status; + tr->op = op_id; +} + +static_always_inline u32 +crypto_dequeue_frame (vlib_main_t * vm, vlib_node_runtime_t * node, + vnet_crypto_thread_t * ct, + vnet_crypto_frame_dequeue_t * hdl, + u32 n_cache, u32 * n_total) +{ + vnet_crypto_async_frame_t *cf = (hdl) (vm); + + while (cf) + { + vec_validate (ct->buffer_indice, n_cache + cf->n_elts); + vec_validate (ct->nexts, n_cache + cf->n_elts); + clib_memcpy_fast (ct->buffer_indice + n_cache, cf->buffer_indices, + sizeof (u32) * cf->n_elts); + if (cf->state == VNET_CRYPTO_FRAME_STATE_SUCCESS) + { + clib_memcpy_fast (ct->nexts + n_cache, cf->next_node_index, + sizeof (u16) * cf->n_elts); + } + else + { + u32 i; + for (i = 0; i < cf->n_elts; i++) + { + if (cf->elts[i].status != VNET_CRYPTO_OP_STATUS_COMPLETED) + { + ct->nexts[i + n_cache] = CRYPTO_DISPATCH_NEXT_ERR_DROP; + vlib_node_increment_counter (vm, node->node_index, + cf->elts[i].status, 1); + } + else + ct->nexts[i + n_cache] = cf->next_node_index[i]; + } + } + n_cache += cf->n_elts; + *n_total += cf->n_elts; + if (n_cache >= VLIB_FRAME_SIZE) + { + vlib_buffer_enqueue_to_next (vm, node, ct->buffer_indice, ct->nexts, + n_cache); + n_cache = 0; + } + + if (PREDICT_FALSE (node->flags & VLIB_NODE_FLAG_TRACE)) + { + u32 i; + + for (i = 0; i < cf->n_elts; i++) + { + vlib_buffer_t *b = vlib_get_buffer (vm, cf->buffer_indices[i]); + if (b->flags & VLIB_BUFFER_IS_TRACED) + vnet_crypto_async_add_trace (vm, node, b, cf->op, + cf->elts[i].status); + } + } + vnet_crypto_async_free_frame (vm, cf); + cf = (hdl) (vm); + } + + return n_cache; +} + +VLIB_NODE_FN (crypto_dispatch_node) (vlib_main_t * vm, + vlib_node_runtime_t * node, + vlib_frame_t * frame) +{ + vnet_crypto_main_t *cm = &crypto_main; + vnet_crypto_thread_t *ct = cm->threads + vm->thread_index; + u32 n_dispatched = 0, n_cache = 0; + u32 index; + + /* *INDENT-OFF* */ + clib_bitmap_foreach (index, cm->async_active_ids, ({ + n_cache = crypto_dequeue_frame (vm, node, ct, cm->dequeue_handlers[index], + n_cache, &n_dispatched); + })); + /* *INDENT-ON* */ + if (n_cache) + vlib_buffer_enqueue_to_next (vm, node, ct->buffer_indice, ct->nexts, + n_cache); + + return n_dispatched; +} + +/* *INDENT-OFF* */ +VLIB_REGISTER_NODE (crypto_dispatch_node) = { + .name = "crypto-dispatch", + .type = VLIB_NODE_TYPE_INPUT, + .state = VLIB_NODE_STATE_DISABLED, + .format_trace = format_crypto_dispatch_trace, + + .n_errors = ARRAY_LEN(vnet_crypto_async_error_strings), + .error_strings = vnet_crypto_async_error_strings, + + .n_next_nodes = CRYPTO_DISPATCH_N_NEXT, + .next_nodes = { +#define _(n, s) \ + [CRYPTO_DISPATCH_NEXT_##n] = s, + foreach_crypto_dispatch_next +#undef _ + }, +}; +/* *INDENT-ON* */ + +/* + * fd.io coding-style-patch-verification: ON + * + * Local Variables: + * eval: (c-set-style "gnu") + * End: + */ |