/* * Copyright (c) 2020 Cisco and/or its affiliates. * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at: * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include #include #include typedef enum { #define _(sym,str) VNET_CRYPTO_ASYNC_ERROR_##sym, foreach_crypto_op_status #undef _ VNET_CRYPTO_ASYNC_N_ERROR, } vnet_crypto_async_error_t; static char *vnet_crypto_async_error_strings[] = { #define _(sym,string) string, foreach_crypto_op_status #undef _ }; #define foreach_crypto_dispatch_next \ _(ERR_DROP, "error-drop") typedef enum { #define _(n, s) CRYPTO_DISPATCH_NEXT_##n, foreach_crypto_dispatch_next #undef _ CRYPTO_DISPATCH_N_NEXT, } crypto_dispatch_next_t; typedef struct { vnet_crypto_op_status_t op_status; vnet_crypto_async_op_id_t op; } crypto_dispatch_trace_t; static u8 * format_crypto_dispatch_trace (u8 * s, va_list * args) { CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *); CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *); crypto_dispatch_trace_t *t = va_arg (*args, crypto_dispatch_trace_t *); s = format (s, "%U: %U", format_vnet_crypto_async_op, t->op, format_vnet_crypto_op_status, t->op_status); return s; } static void vnet_crypto_async_add_trace (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_buffer_t * b, vnet_crypto_async_op_id_t op_id, vnet_crypto_op_status_t status) { crypto_dispatch_trace_t *tr = vlib_add_trace (vm, node, b, sizeof (*tr)); tr->op_status = status; tr->op = op_id; } static_always_inline u32 crypto_dequeue_frame (vlib_main_t * vm, vlib_node_runtime_t * node, vnet_crypto_thread_t * ct, vnet_crypto_frame_dequeue_t * hdl, u32 n_cache, u32 * n_total) { vnet_crypto_main_t *cm = &crypto_main; u32 n_elts = 0; u32 enqueue_thread_idx = ~0; vnet_crypto_async_frame_t *cf = (hdl) (vm, &n_elts, &enqueue_thread_idx); *n_total += n_elts; while (cf || n_elts) { if (cf) { vec_validate (ct->buffer_indices, n_cache + cf->n_elts); vec_validate (ct->nexts, n_cache + cf->n_elts); clib_memcpy_fast (ct->buffer_indices + n_cache, cf->buffer_indices, sizeof (u32) * cf->n_elts); if (cf->state == VNET_CRYPTO_FRAME_STATE_SUCCESS) { clib_memcpy_fast (ct->nexts + n_cache, cf->next_node_index, sizeof (u16) * cf->n_elts); } else { u32 i; for (i = 0; i < cf->n_elts; i++) { if (cf->elts[i].status != VNET_CRYPTO_OP_STATUS_COMPLETED) { ct->nexts[i + n_cache] = CRYPTO_DISPATCH_NEXT_ERR_DROP; vlib_node_increment_counter (vm, node->node_index, cf->elts[i].status, 1); } else ct->nexts[i + n_cache] = cf->next_node_index[i]; } } n_cache += cf->n_elts; if (n_cache >= VLIB_FRAME_SIZE) { vlib_buffer_enqueue_to_next_vec (vm, node, &ct->buffer_indices, &ct->nexts, n_cache); n_cache = 0; } if (PREDICT_FALSE (node->flags & VLIB_NODE_FLAG_TRACE)) { u32 i; for (i = 0; i < cf->n_elts; i++) { vlib_buffer_t *b = vlib_get_buffer (vm, cf->buffer_indices[i]); if (b->flags & VLIB_BUFFER_IS_TRACED) vnet_crypto_async_add_trace (vm, node, b, cf->op, cf->elts[i].status); } } vnet_crypto_async_free_frame (vm, cf); } /* signal enqueue-thread to dequeue the processed frame (n_elts>0) */ if (n_elts > 0 && ((node->state == VLIB_NODE_STATE_POLLING && (node->flags & VLIB_NODE_FLAG_SWITCH_FROM_POLLING_TO_INTERRUPT_MODE)) || node->state == VLIB_NODE_STATE_INTERRUPT)) { vlib_node_set_interrupt_pending ( vlib_get_main_by_index (enqueue_thread_idx), cm->crypto_node_index); } n_elts = 0; enqueue_thread_idx = 0; cf = (hdl) (vm, &n_elts, &enqueue_thread_idx); *n_total += n_elts; } return n_cache; } VLIB_NODE_FN (crypto_dispatch_node) (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame) { vnet_crypto_main_t *cm = &crypto_main; vnet_crypto_thread_t *ct = cm->threads + vm->thread_index; u32 n_dispatched = 0, n_cache = 0, index; vec_foreach_index (index, cm->dequeue_handlers) { n_cache = crypto_dequeue_frame ( vm, node, ct, cm->dequeue_handlers[index], n_cache, &n_dispatched); } if (n_cache) vlib_buffer_enqueue_to_next_vec (vm, node, &ct->buffer_indices, &ct->nexts, n_cache); /* if there are still pending tasks and node in interrupt mode, sending current thread signal to dequeue next loop */ if (pool_elts (ct->frame_pool) > 0 && ((node->state == VLIB_NODE_STATE_POLLING && (node->flags & VLIB_NODE_FLAG_SWITCH_FROM_POLLING_TO_INTERRUPT_MODE)) || node->state == VLIB_NODE_STATE_INTERRUPT)) { vlib_node_set_interrupt_pending (vm, node->node_index); } return n_dispatched; } VLIB_REGISTER_NODE (crypto_dispatch_node) = { .name = "crypto-dispatch", .type = VLIB_NODE_TYPE_INPUT, .flags = VLIB_NODE_FLAG_ADAPTIVE_MODE, .state = VLIB_NODE_STATE_INTERRUPT, .format_trace = format_crypto_dispatch_trace, .n_errors = ARRAY_LEN(vnet_crypto_async_error_strings), .error_strings = vnet_crypto_async_error_strings, .n_next_nodes = CRYPTO_DISPATCH_N_NEXT, .next_nodes = { #define _(n, s) \ [CRYPTO_DISPATCH_NEXT_##n] = s, foreach_crypto_dispatch_next #undef _ }, }; /* * fd.io coding-style-patch-verification: ON * * Local Variables: * eval: (c-set-style "gnu") * End: */