summaryrefslogtreecommitdiffstats
path: root/src/vnet/devices/dpdk/ipsec
diff options
context:
space:
mode:
Diffstat (limited to 'src/vnet/devices/dpdk/ipsec')
-rw-r--r--src/vnet/devices/dpdk/ipsec/cli.c154
-rw-r--r--src/vnet/devices/dpdk/ipsec/crypto_node.c215
-rw-r--r--src/vnet/devices/dpdk/ipsec/dir.dox18
-rw-r--r--src/vnet/devices/dpdk/ipsec/dpdk_crypto_ipsec_doc.md86
-rw-r--r--src/vnet/devices/dpdk/ipsec/esp.h249
-rw-r--r--src/vnet/devices/dpdk/ipsec/esp_decrypt.c594
-rw-r--r--src/vnet/devices/dpdk/ipsec/esp_encrypt.c609
-rw-r--r--src/vnet/devices/dpdk/ipsec/ipsec.c430
-rw-r--r--src/vnet/devices/dpdk/ipsec/ipsec.h227
9 files changed, 0 insertions, 2582 deletions
diff --git a/src/vnet/devices/dpdk/ipsec/cli.c b/src/vnet/devices/dpdk/ipsec/cli.c
deleted file mode 100644
index f9d3a5d082e..00000000000
--- a/src/vnet/devices/dpdk/ipsec/cli.c
+++ /dev/null
@@ -1,154 +0,0 @@
-/*
- * Copyright (c) 2016 Intel and/or its affiliates.
- * Licensed under the Apache License, Version 2.0 (the "License");
- * you may not use this file except in compliance with the License.
- * You may obtain a copy of the License at:
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-#include <vnet/vnet.h>
-#include <vnet/devices/dpdk/dpdk.h>
-#include <vnet/devices/dpdk/ipsec/ipsec.h>
-
-static void
-dpdk_ipsec_show_mapping (vlib_main_t * vm, u16 detail_display)
-{
- dpdk_config_main_t *conf = &dpdk_config_main;
- dpdk_crypto_main_t *dcm = &dpdk_crypto_main;
- vlib_thread_main_t *tm = vlib_get_thread_main ();
- u32 i, skip_master;
-
- if (!conf->cryptodev)
- {
- vlib_cli_output (vm, "DPDK Cryptodev support is disabled\n");
- return;
- }
-
- if (detail_display)
- vlib_cli_output (vm, "worker\t%10s\t%15s\tdir\tdev\tqp\n",
- "cipher", "auth");
- else
- vlib_cli_output (vm, "worker\tcrypto device id(type)\n");
-
- skip_master = vlib_num_workers () > 0;
-
- for (i = 0; i < tm->n_vlib_mains; i++)
- {
- uword key, data;
- u32 cpu_index = vlib_mains[i]->cpu_index;
- crypto_worker_main_t *cwm = &dcm->workers_main[cpu_index];
- u8 *s = 0;
-
- if (skip_master)
- {
- skip_master = 0;
- continue;
- }
-
- if (!detail_display)
- {
- i32 last_cdev = -1;
- crypto_qp_data_t *qpd;
-
- s = format (s, "%u\t", cpu_index);
-
- /* *INDENT-OFF* */
- vec_foreach (qpd, cwm->qp_data)
- {
- u32 dev_id = qpd->dev_id;
-
- if ((u16) last_cdev != dev_id)
- {
- struct rte_cryptodev_info cdev_info;
-
- rte_cryptodev_info_get (dev_id, &cdev_info);
-
- s = format(s, "%u(%s)\t", dev_id, cdev_info.feature_flags &
- RTE_CRYPTODEV_FF_HW_ACCELERATED ? "HW" : "SW");
- }
- last_cdev = dev_id;
- }
- /* *INDENT-ON* */
- vlib_cli_output (vm, "%s", s);
- }
- else
- {
- char cipher_str[15], auth_str[15];
- struct rte_cryptodev_capabilities cap;
- crypto_worker_qp_key_t *p_key = (crypto_worker_qp_key_t *) & key;
- /* *INDENT-OFF* */
- hash_foreach (key, data, cwm->algo_qp_map,
- ({
- cap.op = RTE_CRYPTO_OP_TYPE_SYMMETRIC;
- cap.sym.xform_type = RTE_CRYPTO_SYM_XFORM_CIPHER;
- cap.sym.cipher.algo = p_key->cipher_algo;
- check_algo_is_supported (&cap, cipher_str);
- cap.op = RTE_CRYPTO_OP_TYPE_SYMMETRIC;
- cap.sym.xform_type = RTE_CRYPTO_SYM_XFORM_AUTH;
- cap.sym.auth.algo = p_key->auth_algo;
- check_algo_is_supported (&cap, auth_str);
- vlib_cli_output (vm, "%u\t%10s\t%15s\t%3s\t%u\t%u\n",
- vlib_mains[i]->cpu_index, cipher_str, auth_str,
- p_key->is_outbound ? "out" : "in",
- cwm->qp_data[data].dev_id,
- cwm->qp_data[data].qp_id);
- }));
- /* *INDENT-ON* */
- }
- }
-}
-
-static clib_error_t *
-lcore_cryptodev_map_fn (vlib_main_t * vm, unformat_input_t * input,
- vlib_cli_command_t * cmd)
-{
- unformat_input_t _line_input, *line_input = &_line_input;
- u16 detail = 0;
- clib_error_t *error = NULL;
-
- if (!unformat_user (input, unformat_line_input, line_input))
- return 0;
-
- while (unformat_check_input (line_input) != UNFORMAT_END_OF_INPUT)
- {
- if (unformat (line_input, "verbose"))
- detail = 1;
- else
- {
- error = clib_error_return (0, "parse error: '%U'",
- format_unformat_error, line_input);
- goto done;
- }
- }
-
- dpdk_ipsec_show_mapping (vm, detail);
-
-done:
- unformat_free (line_input);
-
- return error;
-}
-
-/* *INDENT-OFF* */
-VLIB_CLI_COMMAND (lcore_cryptodev_map, static) = {
- .path = "show crypto device mapping",
- .short_help =
- "show cryptodev device mapping <verbose>",
- .function = lcore_cryptodev_map_fn,
-};
-/* *INDENT-ON* */
-
-/*
- * fd.io coding-style-patch-verification: ON
- *
- * Local Variables:
- * eval: (c-set-style "gnu")
- * End:
- */
diff --git a/src/vnet/devices/dpdk/ipsec/crypto_node.c b/src/vnet/devices/dpdk/ipsec/crypto_node.c
deleted file mode 100644
index e8fef23553a..00000000000
--- a/src/vnet/devices/dpdk/ipsec/crypto_node.c
+++ /dev/null
@@ -1,215 +0,0 @@
-/*
- *------------------------------------------------------------------
- * crypto_node.c - DPDK Cryptodev input node
- *
- * Copyright (c) 2016 Intel and/or its affiliates.
- * Licensed under the Apache License, Version 2.0 (the "License");
- * you may not use this file except in compliance with the License.
- * You may obtain a copy of the License at:
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- *------------------------------------------------------------------
- */
-
-#include <vlib/vlib.h>
-#include <vnet/ip/ip.h>
-#include <vnet/ethernet/ethernet.h>
-#include <vnet/ipsec/ipsec.h>
-
-#include <vnet/devices/dpdk/dpdk.h>
-#include <vnet/devices/dpdk/dpdk_priv.h>
-#include <vnet/devices/dpdk/ipsec/ipsec.h>
-
-#define foreach_dpdk_crypto_input_next \
- _(DROP, "error-drop") \
- _(ENCRYPT_POST, "dpdk-esp-encrypt-post") \
- _(DECRYPT_POST, "dpdk-esp-decrypt-post")
-
-typedef enum
-{
-#define _(f,s) DPDK_CRYPTO_INPUT_NEXT_##f,
- foreach_dpdk_crypto_input_next
-#undef _
- DPDK_CRYPTO_INPUT_N_NEXT,
-} dpdk_crypto_input_next_t;
-
-#define foreach_dpdk_crypto_input_error \
- _(DQ_COPS, "Crypto ops dequeued") \
- _(COP_FAILED, "Crypto op failed")
-
-typedef enum
-{
-#define _(f,s) DPDK_CRYPTO_INPUT_ERROR_##f,
- foreach_dpdk_crypto_input_error
-#undef _
- DPDK_CRYPTO_INPUT_N_ERROR,
-} dpdk_crypto_input_error_t;
-
-static char *dpdk_crypto_input_error_strings[] = {
-#define _(n, s) s,
- foreach_dpdk_crypto_input_error
-#undef _
-};
-
-vlib_node_registration_t dpdk_crypto_input_node;
-
-typedef struct
-{
- u32 cdev;
- u32 qp;
- u32 status;
- u32 sa_idx;
- u32 next_index;
-} dpdk_crypto_input_trace_t;
-
-static u8 *
-format_dpdk_crypto_input_trace (u8 * s, va_list * args)
-{
- CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
- CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
- dpdk_crypto_input_trace_t *t = va_arg (*args, dpdk_crypto_input_trace_t *);
-
- s = format (s, "dpdk_crypto: cryptodev-id %u queue-pair %u next-index %d",
- t->cdev, t->qp, t->next_index);
-
- s = format (s, " status %u sa-idx %u\n", t->status, t->sa_idx);
-
- return s;
-}
-
-static_always_inline u32
-dpdk_crypto_dequeue (vlib_main_t * vm, vlib_node_runtime_t * node,
- crypto_qp_data_t * qpd)
-{
- u32 n_deq, *to_next = 0, next_index, n_cops, def_next_index;
- struct rte_crypto_op **cops = qpd->cops;
-
- if (qpd->inflights == 0)
- return 0;
-
- if (qpd->is_outbound)
- def_next_index = DPDK_CRYPTO_INPUT_NEXT_ENCRYPT_POST;
- else
- def_next_index = DPDK_CRYPTO_INPUT_NEXT_DECRYPT_POST;
-
- n_cops = rte_cryptodev_dequeue_burst (qpd->dev_id, qpd->qp_id,
- cops, VLIB_FRAME_SIZE);
- n_deq = n_cops;
- next_index = def_next_index;
-
- qpd->inflights -= n_cops;
- ASSERT (qpd->inflights >= 0);
-
- while (n_cops > 0)
- {
- u32 n_left_to_next;
-
- vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
-
- while (n_cops > 0 && n_left_to_next > 0)
- {
- u32 bi0, next0;
- vlib_buffer_t *b0 = 0;
- struct rte_crypto_op *cop;
- struct rte_crypto_sym_op *sym_cop;
-
- cop = cops[0];
- cops += 1;
- n_cops -= 1;
- n_left_to_next -= 1;
-
- next0 = def_next_index;
-
- if (PREDICT_FALSE (cop->status != RTE_CRYPTO_OP_STATUS_SUCCESS))
- {
- next0 = DPDK_CRYPTO_INPUT_NEXT_DROP;
- vlib_node_increment_counter (vm, dpdk_crypto_input_node.index,
- DPDK_CRYPTO_INPUT_ERROR_COP_FAILED,
- 1);
- }
- cop->status = RTE_CRYPTO_OP_STATUS_NOT_PROCESSED;
-
- sym_cop = (struct rte_crypto_sym_op *) (cop + 1);
- b0 = vlib_buffer_from_rte_mbuf (sym_cop->m_src);
- bi0 = vlib_get_buffer_index (vm, b0);
-
- to_next[0] = bi0;
- to_next += 1;
-
- if (PREDICT_FALSE (b0->flags & VLIB_BUFFER_IS_TRACED))
- {
- vlib_trace_next_frame (vm, node, next0);
- dpdk_crypto_input_trace_t *tr =
- vlib_add_trace (vm, node, b0, sizeof (*tr));
- tr->cdev = qpd->dev_id;
- tr->qp = qpd->qp_id;
- tr->status = cop->status;
- tr->next_index = next0;
- tr->sa_idx = vnet_buffer (b0)->ipsec.sad_index;
- }
-
- vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next,
- n_left_to_next, bi0, next0);
- }
- vlib_put_next_frame (vm, node, next_index, n_left_to_next);
- }
-
- crypto_free_cop (qpd, qpd->cops, n_deq);
-
- vlib_node_increment_counter (vm, dpdk_crypto_input_node.index,
- DPDK_CRYPTO_INPUT_ERROR_DQ_COPS, n_deq);
- return n_deq;
-}
-
-static uword
-dpdk_crypto_input_fn (vlib_main_t * vm, vlib_node_runtime_t * node,
- vlib_frame_t * frame)
-{
- u32 cpu_index = os_get_cpu_number ();
- dpdk_crypto_main_t *dcm = &dpdk_crypto_main;
- crypto_worker_main_t *cwm = &dcm->workers_main[cpu_index];
- crypto_qp_data_t *qpd;
- u32 n_deq = 0;
-
- /* *INDENT-OFF* */
- vec_foreach (qpd, cwm->qp_data)
- n_deq += dpdk_crypto_dequeue(vm, node, qpd);
- /* *INDENT-ON* */
-
- return n_deq;
-}
-
-/* *INDENT-OFF* */
-VLIB_REGISTER_NODE (dpdk_crypto_input_node) =
-{
- .function = dpdk_crypto_input_fn,
- .name = "dpdk-crypto-input",
- .format_trace = format_dpdk_crypto_input_trace,
- .type = VLIB_NODE_TYPE_INPUT,
- .state = VLIB_NODE_STATE_DISABLED,
- .n_errors = DPDK_CRYPTO_INPUT_N_ERROR,
- .error_strings = dpdk_crypto_input_error_strings,
- .n_next_nodes = DPDK_CRYPTO_INPUT_N_NEXT,
- .next_nodes =
- {
-#define _(s,n) [DPDK_CRYPTO_INPUT_NEXT_##s] = n,
- foreach_dpdk_crypto_input_next
-#undef _
- },
-};
-/* *INDENT-ON* */
-
-VLIB_NODE_FUNCTION_MULTIARCH (dpdk_crypto_input_node, dpdk_crypto_input_fn)
-/*
- * fd.io coding-style-patch-verification: ON
- *
- * Local Variables:
- * eval: (c-set-style "gnu")
- * End:
- */
diff --git a/src/vnet/devices/dpdk/ipsec/dir.dox b/src/vnet/devices/dpdk/ipsec/dir.dox
deleted file mode 100644
index ffebfc4d62e..00000000000
--- a/src/vnet/devices/dpdk/ipsec/dir.dox
+++ /dev/null
@@ -1,18 +0,0 @@
-/*
- * Copyright (c) 2016 Intel and/or its affiliates.
- * Licensed under the Apache License, Version 2.0 (the "License");
- * you may not use this file except in compliance with the License.
- * You may obtain a copy of the License at:
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-/**
- @dir vnet/vnet/devices/dpdk/ipsec
- @brief IPSec ESP encrypt/decrypt using DPDK Cryptodev API
-*/
diff --git a/src/vnet/devices/dpdk/ipsec/dpdk_crypto_ipsec_doc.md b/src/vnet/devices/dpdk/ipsec/dpdk_crypto_ipsec_doc.md
deleted file mode 100644
index fed2fe0eee8..00000000000
--- a/src/vnet/devices/dpdk/ipsec/dpdk_crypto_ipsec_doc.md
+++ /dev/null
@@ -1,86 +0,0 @@
-# VPP IPSec implementation using DPDK Cryptodev API {#dpdk_crypto_ipsec_doc}
-
-This document is meant to contain all related information about implementation and usability.
-
-
-## VPP IPsec with DPDK Cryptodev
-
-DPDK Cryptodev is an asynchronous crypto API that supports both Hardware and Software implementations (for more details refer to [DPDK Cryptography Device Library documentation](http://dpdk.org/doc/guides/prog_guide/cryptodev_lib.html)).
-
-When DPDK support is enabled and there are enough Cryptodev resources for all workers, the node graph is reconfigured by adding and changing default next nodes.
-
-The following nodes are added:
-* dpdk-crypto-input : polling input node, basically dequeuing from crypto devices.
-* dpdk-esp-encrypt : internal node.
-* dpdk-esp-decrypt : internal node.
-* dpdk-esp-encrypt-post : internal node.
-* dpdk-esp-decrypt-post : internal node.
-
-Set new default next nodes:
-* for esp encryption: esp-encrypt -> dpdk-esp-encrypt
-* for esp decryption: esp-decrypt -> dpdk-esp-decrypt
-
-
-### How to enable VPP IPSec with DPDK Cryptodev support
-
-DPDK Cryptodev is supported in DPDK enabled VPP.
-By default, only HW Cryptodev is supported but needs to be explicetly enabled with the following config option:
-
-```
-dpdk {
- enable-cryptodev
-}
-```
-
-To enable SW Cryptodev support (AESNI-MB-PMD and GCM-PMD), we need the following env option:
-
- vpp_uses_dpdk_cryptodev_sw=yes
-
-A couple of ways to achive this:
-* uncomment/add it in the platforms config (ie. build-data/platforms/vpp.mk)
-* set the option when building vpp (ie. make vpp_uses_dpdk_cryptodev_sw=yes build-release)
-
-When enabling SW Cryptodev support, it means that you need to pre-build the required crypto libraries needed by those SW Cryptodev PMDs.
-
-
-### Crypto Resources allocation
-
-VPP allocates crypto resources based on a best effort approach:
-* first allocate Hardware crypto resources, then Software.
-* if there are not enough crypto resources for all workers, the graph node is not modifed, therefore the default VPP IPsec implementation based in OpenSSL is used. The following message is displayed:
-
- 0: dpdk_ipsec_init: not enough cryptodevs for ipsec
-
-
-### Configuration example
-
-To enable DPDK Cryptodev the user just need to provide the startup.conf option
-as mentioned previously.
-
-Example startup.conf:
-
-```
-dpdk {
- socket-mem 1024,1024
- num-mbufs 131072
- dev 0000:81:00.0
- dev 0000:81:00.1
- enable-cryptodev
- dev 0000:85:01.0
- dev 0000:85:01.1
- vdev cryptodev_aesni_mb_pmd,socket_id=1
- vdev cryptodev_aesni_mb_pmd,socket_id=1
-}
-```
-
-In the above configuration:
-* 0000:85:01.0 and 0000:85:01.1 are crypto BDFs and they require the same driver binding as DPDK Ethernet devices but they do not support any extra configuration options.
-* Two AESNI-MB Software Cryptodev PMDs are created in NUMA node 1.
-
-For further details refer to [DPDK Crypto Device Driver documentation](http://dpdk.org/doc/guides/cryptodevs/index.html)
-
-### Operational data
-
-The following CLI command displays the Cryptodev/Worker mapping:
-
- show crypto device mapping [verbose]
diff --git a/src/vnet/devices/dpdk/ipsec/esp.h b/src/vnet/devices/dpdk/ipsec/esp.h
deleted file mode 100644
index d0b276189c8..00000000000
--- a/src/vnet/devices/dpdk/ipsec/esp.h
+++ /dev/null
@@ -1,249 +0,0 @@
-/*
- * Copyright (c) 2016 Intel and/or its affiliates.
- * Licensed under the Apache License, Version 2.0 (the "License");
- * you may not use this file except in compliance with the License.
- * You may obtain a copy of the License at:
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-#ifndef __DPDK_ESP_H__
-#define __DPDK_ESP_H__
-
-#include <vnet/devices/dpdk/ipsec/ipsec.h>
-#include <vnet/ipsec/ipsec.h>
-#include <vnet/ipsec/esp.h>
-
-typedef struct
-{
- enum rte_crypto_cipher_algorithm algo;
- u8 key_len;
- u8 iv_len;
-} dpdk_esp_crypto_alg_t;
-
-typedef struct
-{
- enum rte_crypto_auth_algorithm algo;
- u8 trunc_size;
-} dpdk_esp_integ_alg_t;
-
-typedef struct
-{
- dpdk_esp_crypto_alg_t *esp_crypto_algs;
- dpdk_esp_integ_alg_t *esp_integ_algs;
-} dpdk_esp_main_t;
-
-dpdk_esp_main_t dpdk_esp_main;
-
-static_always_inline void
-dpdk_esp_init ()
-{
- dpdk_esp_main_t *em = &dpdk_esp_main;
- dpdk_esp_integ_alg_t *i;
- dpdk_esp_crypto_alg_t *c;
-
- vec_validate (em->esp_crypto_algs, IPSEC_CRYPTO_N_ALG - 1);
-
- c = &em->esp_crypto_algs[IPSEC_CRYPTO_ALG_AES_CBC_128];
- c->algo = RTE_CRYPTO_CIPHER_AES_CBC;
- c->key_len = 16;
- c->iv_len = 16;
-
- c = &em->esp_crypto_algs[IPSEC_CRYPTO_ALG_AES_CBC_192];
- c->algo = RTE_CRYPTO_CIPHER_AES_CBC;
- c->key_len = 24;
- c->iv_len = 16;
-
- c = &em->esp_crypto_algs[IPSEC_CRYPTO_ALG_AES_CBC_256];
- c->algo = RTE_CRYPTO_CIPHER_AES_CBC;
- c->key_len = 32;
- c->iv_len = 16;
-
- c = &em->esp_crypto_algs[IPSEC_CRYPTO_ALG_AES_GCM_128];
- c->algo = RTE_CRYPTO_CIPHER_AES_GCM;
- c->key_len = 16;
- c->iv_len = 8;
-
- vec_validate (em->esp_integ_algs, IPSEC_INTEG_N_ALG - 1);
-
- i = &em->esp_integ_algs[IPSEC_INTEG_ALG_SHA1_96];
- i->algo = RTE_CRYPTO_AUTH_SHA1_HMAC;
- i->trunc_size = 12;
-
- i = &em->esp_integ_algs[IPSEC_INTEG_ALG_SHA_256_96];
- i->algo = RTE_CRYPTO_AUTH_SHA256_HMAC;
- i->trunc_size = 12;
-
- i = &em->esp_integ_algs[IPSEC_INTEG_ALG_SHA_256_128];
- i->algo = RTE_CRYPTO_AUTH_SHA256_HMAC;
- i->trunc_size = 16;
-
- i = &em->esp_integ_algs[IPSEC_INTEG_ALG_SHA_384_192];
- i->algo = RTE_CRYPTO_AUTH_SHA384_HMAC;
- i->trunc_size = 24;
-
- i = &em->esp_integ_algs[IPSEC_INTEG_ALG_SHA_512_256];
- i->algo = RTE_CRYPTO_AUTH_SHA512_HMAC;
- i->trunc_size = 32;
-
- i = &em->esp_integ_algs[IPSEC_INTEG_ALG_AES_GCM_128];
- i->algo = RTE_CRYPTO_AUTH_AES_GCM;
- i->trunc_size = 16;
-}
-
-static_always_inline int
-translate_crypto_algo (ipsec_crypto_alg_t crypto_algo,
- struct rte_crypto_sym_xform *cipher_xform)
-{
- switch (crypto_algo)
- {
- case IPSEC_CRYPTO_ALG_NONE:
- cipher_xform->cipher.algo = RTE_CRYPTO_CIPHER_NULL;
- break;
- case IPSEC_CRYPTO_ALG_AES_CBC_128:
- case IPSEC_CRYPTO_ALG_AES_CBC_192:
- case IPSEC_CRYPTO_ALG_AES_CBC_256:
- cipher_xform->cipher.algo = RTE_CRYPTO_CIPHER_AES_CBC;
- break;
- case IPSEC_CRYPTO_ALG_AES_GCM_128:
- cipher_xform->cipher.algo = RTE_CRYPTO_CIPHER_AES_GCM;
- break;
- default:
- return -1;
- }
-
- cipher_xform->type = RTE_CRYPTO_SYM_XFORM_CIPHER;
-
- return 0;
-}
-
-static_always_inline int
-translate_integ_algo (ipsec_integ_alg_t integ_alg,
- struct rte_crypto_sym_xform *auth_xform, int use_esn)
-{
- switch (integ_alg)
- {
- case IPSEC_INTEG_ALG_NONE:
- auth_xform->auth.algo = RTE_CRYPTO_AUTH_NULL;
- auth_xform->auth.digest_length = 0;
- break;
- case IPSEC_INTEG_ALG_SHA1_96:
- auth_xform->auth.algo = RTE_CRYPTO_AUTH_SHA1_HMAC;
- auth_xform->auth.digest_length = 12;
- break;
- case IPSEC_INTEG_ALG_SHA_256_96:
- auth_xform->auth.algo = RTE_CRYPTO_AUTH_SHA256_HMAC;
- auth_xform->auth.digest_length = 12;
- break;
- case IPSEC_INTEG_ALG_SHA_256_128:
- auth_xform->auth.algo = RTE_CRYPTO_AUTH_SHA256_HMAC;
- auth_xform->auth.digest_length = 16;
- break;
- case IPSEC_INTEG_ALG_SHA_384_192:
- auth_xform->auth.algo = RTE_CRYPTO_AUTH_SHA384_HMAC;
- auth_xform->auth.digest_length = 24;
- break;
- case IPSEC_INTEG_ALG_SHA_512_256:
- auth_xform->auth.algo = RTE_CRYPTO_AUTH_SHA512_HMAC;
- auth_xform->auth.digest_length = 32;
- break;
- case IPSEC_INTEG_ALG_AES_GCM_128:
- auth_xform->auth.algo = RTE_CRYPTO_AUTH_AES_GCM;
- auth_xform->auth.digest_length = 16;
- auth_xform->auth.add_auth_data_length = use_esn ? 12 : 8;
- break;
- default:
- return -1;
- }
-
- auth_xform->type = RTE_CRYPTO_SYM_XFORM_AUTH;
-
- return 0;
-}
-
-static_always_inline int
-create_sym_sess (ipsec_sa_t * sa, crypto_sa_session_t * sa_sess,
- u8 is_outbound)
-{
- u32 cpu_index = os_get_cpu_number ();
- dpdk_crypto_main_t *dcm = &dpdk_crypto_main;
- crypto_worker_main_t *cwm = &dcm->workers_main[cpu_index];
- struct rte_crypto_sym_xform cipher_xform = { 0 };
- struct rte_crypto_sym_xform auth_xform = { 0 };
- struct rte_crypto_sym_xform *xfs;
- uword key = 0, *data;
- crypto_worker_qp_key_t *p_key = (crypto_worker_qp_key_t *) & key;
-
- if (sa->crypto_alg == IPSEC_CRYPTO_ALG_AES_GCM_128)
- {
- sa->crypto_key_len -= 4;
- clib_memcpy (&sa->salt, &sa->crypto_key[sa->crypto_key_len], 4);
- }
- else
- {
- u32 seed = (u32) clib_cpu_time_now ();
- sa->salt = random_u32 (&seed);
- }
-
- cipher_xform.type = RTE_CRYPTO_SYM_XFORM_CIPHER;
- cipher_xform.cipher.key.data = sa->crypto_key;
- cipher_xform.cipher.key.length = sa->crypto_key_len;
-
- auth_xform.type = RTE_CRYPTO_SYM_XFORM_AUTH;
- auth_xform.auth.key.data = sa->integ_key;
- auth_xform.auth.key.length = sa->integ_key_len;
-
- if (translate_crypto_algo (sa->crypto_alg, &cipher_xform) < 0)
- return -1;
- p_key->cipher_algo = cipher_xform.cipher.algo;
-
- if (translate_integ_algo (sa->integ_alg, &auth_xform, sa->use_esn) < 0)
- return -1;
- p_key->auth_algo = auth_xform.auth.algo;
-
- if (is_outbound)
- {
- cipher_xform.cipher.op = RTE_CRYPTO_CIPHER_OP_ENCRYPT;
- auth_xform.auth.op = RTE_CRYPTO_AUTH_OP_GENERATE;
- cipher_xform.next = &auth_xform;
- xfs = &cipher_xform;
- }
- else
- {
- cipher_xform.cipher.op = RTE_CRYPTO_CIPHER_OP_DECRYPT;
- auth_xform.auth.op = RTE_CRYPTO_AUTH_OP_VERIFY;
- auth_xform.next = &cipher_xform;
- xfs = &auth_xform;
- }
-
- p_key->is_outbound = is_outbound;
-
- data = hash_get (cwm->algo_qp_map, key);
- if (!data)
- return -1;
-
- sa_sess->sess =
- rte_cryptodev_sym_session_create (cwm->qp_data[*data].dev_id, xfs);
-
- if (!sa_sess->sess)
- return -1;
-
- sa_sess->qp_index = (u8) * data;
-
- return 0;
-}
-
-#endif /* __DPDK_ESP_H__ */
-
-/*
- * fd.io coding-style-patch-verification: ON
- *
- * Local Variables:
- * eval: (c-set-style "gnu")
- * End:
- */
diff --git a/src/vnet/devices/dpdk/ipsec/esp_decrypt.c b/src/vnet/devices/dpdk/ipsec/esp_decrypt.c
deleted file mode 100644
index 760076093cb..00000000000
--- a/src/vnet/devices/dpdk/ipsec/esp_decrypt.c
+++ /dev/null
@@ -1,594 +0,0 @@
-/*
- * esp_decrypt.c : IPSec ESP Decrypt node using DPDK Cryptodev
- *
- * Copyright (c) 2016 Intel and/or its affiliates.
- * Licensed under the Apache License, Version 2.0 (the "License");
- * you may not use this file except in compliance with the License.
- * You may obtain a copy of the License at:
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-#include <vnet/vnet.h>
-#include <vnet/api_errno.h>
-#include <vnet/ip/ip.h>
-
-#include <vnet/ipsec/ipsec.h>
-#include <vnet/devices/dpdk/ipsec/ipsec.h>
-#include <vnet/devices/dpdk/ipsec/esp.h>
-#include <vnet/devices/dpdk/dpdk.h>
-#include <vnet/devices/dpdk/dpdk_priv.h>
-
-#define foreach_esp_decrypt_next \
-_(DROP, "error-drop") \
-_(IP4_INPUT, "ip4-input") \
-_(IP6_INPUT, "ip6-input")
-
-#define _(v, s) ESP_DECRYPT_NEXT_##v,
-typedef enum {
- foreach_esp_decrypt_next
-#undef _
- ESP_DECRYPT_N_NEXT,
-} esp_decrypt_next_t;
-
-#define foreach_esp_decrypt_error \
- _(RX_PKTS, "ESP pkts received") \
- _(DECRYPTION_FAILED, "ESP decryption failed") \
- _(REPLAY, "SA replayed packet") \
- _(NOT_IP, "Not IP packet (dropped)") \
- _(ENQ_FAIL, "Enqueue failed (buffer full)") \
- _(NO_CRYPTODEV, "Cryptodev not configured") \
- _(BAD_LEN, "Invalid ciphertext length") \
- _(UNSUPPORTED, "Cipher/Auth not supported")
-
-
-typedef enum {
-#define _(sym,str) ESP_DECRYPT_ERROR_##sym,
- foreach_esp_decrypt_error
-#undef _
- ESP_DECRYPT_N_ERROR,
-} esp_decrypt_error_t;
-
-static char * esp_decrypt_error_strings[] = {
-#define _(sym,string) string,
- foreach_esp_decrypt_error
-#undef _
-};
-
-vlib_node_registration_t dpdk_esp_decrypt_node;
-
-typedef struct {
- ipsec_crypto_alg_t crypto_alg;
- ipsec_integ_alg_t integ_alg;
-} esp_decrypt_trace_t;
-
-/* packet trace format function */
-static u8 * format_esp_decrypt_trace (u8 * s, va_list * args)
-{
- CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
- CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
- esp_decrypt_trace_t * t = va_arg (*args, esp_decrypt_trace_t *);
-
- s = format (s, "esp: crypto %U integrity %U",
- format_ipsec_crypto_alg, t->crypto_alg,
- format_ipsec_integ_alg, t->integ_alg);
- return s;
-}
-
-static uword
-dpdk_esp_decrypt_node_fn (vlib_main_t * vm,
- vlib_node_runtime_t * node,
- vlib_frame_t * from_frame)
-{
- u32 n_left_from, *from, *to_next, next_index;
- ipsec_main_t *im = &ipsec_main;
- u32 cpu_index = os_get_cpu_number();
- dpdk_crypto_main_t * dcm = &dpdk_crypto_main;
- dpdk_esp_main_t * em = &dpdk_esp_main;
- u32 i;
-
- from = vlib_frame_vector_args (from_frame);
- n_left_from = from_frame->n_vectors;
-
- if (PREDICT_FALSE(!dcm->workers_main))
- {
- vlib_node_increment_counter (vm, dpdk_esp_decrypt_node.index,
- ESP_DECRYPT_ERROR_NO_CRYPTODEV, n_left_from);
- vlib_buffer_free(vm, from, n_left_from);
- return n_left_from;
- }
-
- crypto_worker_main_t *cwm = vec_elt_at_index(dcm->workers_main, cpu_index);
- u32 n_qps = vec_len(cwm->qp_data);
- struct rte_crypto_op ** cops_to_enq[n_qps];
- u32 n_cop_qp[n_qps], * bi_to_enq[n_qps];
-
- for (i = 0; i < n_qps; i++)
- {
- bi_to_enq[i] = cwm->qp_data[i].bi;
- cops_to_enq[i] = cwm->qp_data[i].cops;
- }
-
- memset(n_cop_qp, 0, n_qps * sizeof(u32));
-
- crypto_alloc_cops();
-
- next_index = ESP_DECRYPT_NEXT_DROP;
-
- while (n_left_from > 0)
- {
- u32 n_left_to_next;
-
- vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
-
- while (n_left_from > 0 && n_left_to_next > 0)
- {
- u32 bi0, sa_index0 = ~0, seq, icv_size, iv_size;
- vlib_buffer_t * b0;
- esp_header_t * esp0;
- ipsec_sa_t * sa0;
- struct rte_mbuf * mb0 = 0;
- const int BLOCK_SIZE = 16;
- crypto_sa_session_t * sa_sess;
- void * sess;
- u16 qp_index;
- struct rte_crypto_op * cop = 0;
-
- bi0 = from[0];
- from += 1;
- n_left_from -= 1;
-
- b0 = vlib_get_buffer (vm, bi0);
- esp0 = vlib_buffer_get_current (b0);
-
- sa_index0 = vnet_buffer(b0)->ipsec.sad_index;
- sa0 = pool_elt_at_index (im->sad, sa_index0);
-
- seq = clib_host_to_net_u32(esp0->seq);
-
- /* anti-replay check */
- if (sa0->use_anti_replay)
- {
- int rv = 0;
-
- if (PREDICT_TRUE(sa0->use_esn))
- rv = esp_replay_check_esn(sa0, seq);
- else
- rv = esp_replay_check(sa0, seq);
-
- if (PREDICT_FALSE(rv))
- {
- clib_warning ("anti-replay SPI %u seq %u", sa0->spi, seq);
- vlib_node_increment_counter (vm, dpdk_esp_decrypt_node.index,
- ESP_DECRYPT_ERROR_REPLAY, 1);
- to_next[0] = bi0;
- to_next += 1;
- n_left_to_next -= 1;
- goto trace;
- }
- }
-
- sa0->total_data_size += b0->current_length;
-
- if (PREDICT_FALSE(sa0->integ_alg == IPSEC_INTEG_ALG_NONE) ||
- PREDICT_FALSE(sa0->crypto_alg == IPSEC_CRYPTO_ALG_NONE))
- {
- clib_warning ("SPI %u : only cipher + auth supported", sa0->spi);
- vlib_node_increment_counter (vm, dpdk_esp_decrypt_node.index,
- ESP_DECRYPT_ERROR_UNSUPPORTED, 1);
- to_next[0] = bi0;
- to_next += 1;
- n_left_to_next -= 1;
- goto trace;
- }
-
- sa_sess = pool_elt_at_index(cwm->sa_sess_d[0], sa_index0);
-
- if (PREDICT_FALSE(!sa_sess->sess))
- {
- int ret = create_sym_sess(sa0, sa_sess, 0);
-
- if (PREDICT_FALSE (ret))
- {
- to_next[0] = bi0;
- to_next += 1;
- n_left_to_next -= 1;
- goto trace;
- }
- }
-
- sess = sa_sess->sess;
- qp_index = sa_sess->qp_index;
-
- ASSERT (vec_len (vec_elt (cwm->qp_data, qp_index).free_cops) > 0);
- cop = vec_pop (vec_elt (cwm->qp_data, qp_index).free_cops);
- ASSERT (cop->status == RTE_CRYPTO_OP_STATUS_NOT_PROCESSED);
-
- cops_to_enq[qp_index][0] = cop;
- cops_to_enq[qp_index] += 1;
- n_cop_qp[qp_index] += 1;
- bi_to_enq[qp_index][0] = bi0;
- bi_to_enq[qp_index] += 1;
-
- rte_crypto_op_attach_sym_session(cop, sess);
-
- icv_size = em->esp_integ_algs[sa0->integ_alg].trunc_size;
- iv_size = em->esp_crypto_algs[sa0->crypto_alg].iv_len;
-
- /* Convert vlib buffer to mbuf */
- mb0 = rte_mbuf_from_vlib_buffer(b0);
- mb0->data_len = b0->current_length;
- mb0->pkt_len = b0->current_length;
- mb0->data_off = RTE_PKTMBUF_HEADROOM + b0->current_data;
-
- /* Outer IP header has already been stripped */
- u16 payload_len = rte_pktmbuf_pkt_len(mb0) - sizeof (esp_header_t) -
- iv_size - icv_size;
-
- if ((payload_len & (BLOCK_SIZE - 1)) || (payload_len <= 0))
- {
- clib_warning ("payload %u not multiple of %d\n",
- payload_len, BLOCK_SIZE);
- vlib_node_increment_counter (vm, dpdk_esp_decrypt_node.index,
- ESP_DECRYPT_ERROR_BAD_LEN, 1);
- vec_add (vec_elt (cwm->qp_data, qp_index).free_cops, &cop, 1);
- bi_to_enq[qp_index] -= 1;
- cops_to_enq[qp_index] -= 1;
- n_cop_qp[qp_index] -= 1;
- to_next[0] = bi0;
- to_next += 1;
- n_left_to_next -= 1;
- goto trace;
- }
-
- struct rte_crypto_sym_op *sym_cop = (struct rte_crypto_sym_op *)(cop + 1);
-
- sym_cop->m_src = mb0;
- sym_cop->cipher.data.offset = sizeof (esp_header_t) + iv_size;
- sym_cop->cipher.data.length = payload_len;
-
- u8 *iv = rte_pktmbuf_mtod_offset(mb0, void*, sizeof (esp_header_t));
- dpdk_cop_priv_t * priv = (dpdk_cop_priv_t *)(sym_cop + 1);
-
- if (sa0->crypto_alg == IPSEC_CRYPTO_ALG_AES_GCM_128)
- {
- dpdk_gcm_cnt_blk *icb = &priv->cb;
- icb->salt = sa0->salt;
- clib_memcpy(icb->iv, iv, 8);
- icb->cnt = clib_host_to_net_u32(1);
- sym_cop->cipher.iv.data = (u8 *)icb;
- sym_cop->cipher.iv.phys_addr = cop->phys_addr +
- (uintptr_t)icb - (uintptr_t)cop;
- sym_cop->cipher.iv.length = 16;
-
- u8 *aad = priv->aad;
- clib_memcpy(aad, iv - sizeof(esp_header_t), 8);
- sym_cop->auth.aad.data = aad;
- sym_cop->auth.aad.phys_addr = cop->phys_addr +
- (uintptr_t)aad - (uintptr_t)cop;
- if (sa0->use_esn)
- {
- *((u32*)&aad[8]) = sa0->seq_hi;
- sym_cop->auth.aad.length = 12;
- }
- else
- {
- sym_cop->auth.aad.length = 8;
- }
-
- sym_cop->auth.digest.data = rte_pktmbuf_mtod_offset(mb0, void*,
- rte_pktmbuf_pkt_len(mb0) - icv_size);
- sym_cop->auth.digest.phys_addr = rte_pktmbuf_mtophys_offset(mb0,
- rte_pktmbuf_pkt_len(mb0) - icv_size);
- sym_cop->auth.digest.length = icv_size;
-
- }
- else
- {
- sym_cop->cipher.iv.data = rte_pktmbuf_mtod_offset(mb0, void*,
- sizeof (esp_header_t));
- sym_cop->cipher.iv.phys_addr = rte_pktmbuf_mtophys_offset(mb0,
- sizeof (esp_header_t));
- sym_cop->cipher.iv.length = iv_size;
-
- if (sa0->use_esn)
- {
- dpdk_cop_priv_t* priv = (dpdk_cop_priv_t*) (sym_cop + 1);
- u8* payload_end = rte_pktmbuf_mtod_offset(
- mb0, u8*, sizeof(esp_header_t) + iv_size + payload_len);
-
- clib_memcpy (priv->icv, payload_end, icv_size);
- *((u32*) payload_end) = sa0->seq_hi;
- sym_cop->auth.data.offset = 0;
- sym_cop->auth.data.length = sizeof(esp_header_t) + iv_size
- + payload_len + sizeof(sa0->seq_hi);
- sym_cop->auth.digest.data = priv->icv;
- sym_cop->auth.digest.phys_addr = cop->phys_addr
- + (uintptr_t) priv->icv - (uintptr_t) cop;
- sym_cop->auth.digest.length = icv_size;
- }
- else
- {
- sym_cop->auth.data.offset = 0;
- sym_cop->auth.data.length = sizeof(esp_header_t) +
- iv_size + payload_len;
-
- sym_cop->auth.digest.data = rte_pktmbuf_mtod_offset(mb0, void*,
- rte_pktmbuf_pkt_len(mb0) - icv_size);
- sym_cop->auth.digest.phys_addr = rte_pktmbuf_mtophys_offset(mb0,
- rte_pktmbuf_pkt_len(mb0) - icv_size);
- sym_cop->auth.digest.length = icv_size;
- }
- }
-
-trace:
- if (PREDICT_FALSE(b0->flags & VLIB_BUFFER_IS_TRACED))
- {
- esp_decrypt_trace_t *tr = vlib_add_trace (vm, node, b0, sizeof (*tr));
- tr->crypto_alg = sa0->crypto_alg;
- tr->integ_alg = sa0->integ_alg;
- }
- }
- vlib_put_next_frame (vm, node, next_index, n_left_to_next);
- }
- vlib_node_increment_counter (vm, dpdk_esp_decrypt_node.index,
- ESP_DECRYPT_ERROR_RX_PKTS,
- from_frame->n_vectors);
- crypto_qp_data_t *qpd;
- /* *INDENT-OFF* */
- vec_foreach_index (i, cwm->qp_data)
- {
- u32 enq;
-
- qpd = vec_elt_at_index(cwm->qp_data, i);
- enq = rte_cryptodev_enqueue_burst(qpd->dev_id, qpd->qp_id,
- qpd->cops, n_cop_qp[i]);
- qpd->inflights += enq;
-
- if (PREDICT_FALSE(enq < n_cop_qp[i]))
- {
- crypto_free_cop (qpd, &qpd->cops[enq], n_cop_qp[i] - enq);
- vlib_buffer_free (vm, &qpd->bi[enq], n_cop_qp[i] - enq);
-
- vlib_node_increment_counter (vm, dpdk_esp_decrypt_node.index,
- ESP_DECRYPT_ERROR_ENQ_FAIL,
- n_cop_qp[i] - enq);
- }
- }
- /* *INDENT-ON* */
-
- return from_frame->n_vectors;
-}
-
-VLIB_REGISTER_NODE (dpdk_esp_decrypt_node) = {
- .function = dpdk_esp_decrypt_node_fn,
- .name = "dpdk-esp-decrypt",
- .vector_size = sizeof (u32),
- .format_trace = format_esp_decrypt_trace,
- .type = VLIB_NODE_TYPE_INTERNAL,
-
- .n_errors = ARRAY_LEN(esp_decrypt_error_strings),
- .error_strings = esp_decrypt_error_strings,
-
- .n_next_nodes = ESP_DECRYPT_N_NEXT,
- .next_nodes = {
-#define _(s,n) [ESP_DECRYPT_NEXT_##s] = n,
- foreach_esp_decrypt_next
-#undef _
- },
-};
-
-VLIB_NODE_FUNCTION_MULTIARCH (dpdk_esp_decrypt_node, dpdk_esp_decrypt_node_fn)
-
-/*
- * Decrypt Post Node
- */
-
-#define foreach_esp_decrypt_post_error \
- _(PKTS, "ESP post pkts")
-
-typedef enum {
-#define _(sym,str) ESP_DECRYPT_POST_ERROR_##sym,
- foreach_esp_decrypt_post_error
-#undef _
- ESP_DECRYPT_POST_N_ERROR,
-} esp_decrypt_post_error_t;
-
-static char * esp_decrypt_post_error_strings[] = {
-#define _(sym,string) string,
- foreach_esp_decrypt_post_error
-#undef _
-};
-
-vlib_node_registration_t dpdk_esp_decrypt_post_node;
-
-static u8 * format_esp_decrypt_post_trace (u8 * s, va_list * args)
-{
- return s;
-}
-
-static uword
-dpdk_esp_decrypt_post_node_fn (vlib_main_t * vm,
- vlib_node_runtime_t * node,
- vlib_frame_t * from_frame)
-{
- u32 n_left_from, *from, *to_next = 0, next_index;
- ipsec_sa_t * sa0;
- u32 sa_index0 = ~0;
- ipsec_main_t *im = &ipsec_main;
- dpdk_esp_main_t *em = &dpdk_esp_main;
-
- from = vlib_frame_vector_args (from_frame);
- n_left_from = from_frame->n_vectors;
-
- next_index = node->cached_next_index;
-
- while (n_left_from > 0)
- {
- u32 n_left_to_next;
-
- vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
-
- while (n_left_from > 0 && n_left_to_next > 0)
- {
- esp_footer_t * f0;
- u32 bi0, next0, icv_size, iv_size;
- vlib_buffer_t * b0 = 0;
- ip4_header_t *ih4 = 0, *oh4 = 0;
- ip6_header_t *ih6 = 0, *oh6 = 0;
- u8 tunnel_mode = 1;
- u8 transport_ip6 = 0;
-
- next0 = ESP_DECRYPT_NEXT_DROP;
-
- bi0 = from[0];
- from += 1;
- n_left_from -= 1;
- n_left_to_next -= 1;
-
- b0 = vlib_get_buffer (vm, bi0);
-
- sa_index0 = vnet_buffer(b0)->ipsec.sad_index;
- sa0 = pool_elt_at_index (im->sad, sa_index0);
-
- to_next[0] = bi0;
- to_next += 1;
-
- icv_size = em->esp_integ_algs[sa0->integ_alg].trunc_size;
- iv_size = em->esp_crypto_algs[sa0->crypto_alg].iv_len;
-
- if (sa0->use_anti_replay)
- {
- esp_header_t * esp0 = vlib_buffer_get_current (b0);
- u32 seq;
- seq = clib_host_to_net_u32(esp0->seq);
- if (PREDICT_TRUE(sa0->use_esn))
- esp_replay_advance_esn(sa0, seq);
- else
- esp_replay_advance(sa0, seq);
- }
-
- ih4 = (ip4_header_t *) (b0->data + sizeof(ethernet_header_t));
- vlib_buffer_advance (b0, sizeof (esp_header_t) + iv_size);
-
- b0->current_length -= (icv_size + 2);
- b0->flags = VLIB_BUFFER_TOTAL_LENGTH_VALID;
- f0 = (esp_footer_t *) ((u8 *) vlib_buffer_get_current (b0) +
- b0->current_length);
- b0->current_length -= f0->pad_length;
-
- /* transport mode */
- if (PREDICT_FALSE(!sa0->is_tunnel && !sa0->is_tunnel_ip6))
- {
- tunnel_mode = 0;
-
- if (PREDICT_TRUE((ih4->ip_version_and_header_length & 0xF0) != 0x40))
- {
- if (PREDICT_TRUE((ih4->ip_version_and_header_length & 0xF0) == 0x60))
- transport_ip6 = 1;
- else
- {
- clib_warning("next header: 0x%x", f0->next_header);
- vlib_node_increment_counter (vm, dpdk_esp_decrypt_node.index,
- ESP_DECRYPT_ERROR_NOT_IP, 1);
- goto trace;
- }
- }
- }
-
- if (PREDICT_TRUE (tunnel_mode))
- {
- if (PREDICT_TRUE(f0->next_header == IP_PROTOCOL_IP_IN_IP))
- next0 = ESP_DECRYPT_NEXT_IP4_INPUT;
- else if (f0->next_header == IP_PROTOCOL_IPV6)
- next0 = ESP_DECRYPT_NEXT_IP6_INPUT;
- else
- {
- clib_warning("next header: 0x%x", f0->next_header);
- vlib_node_increment_counter (vm, dpdk_esp_decrypt_node.index,
- ESP_DECRYPT_ERROR_DECRYPTION_FAILED,
- 1);
- goto trace;
- }
- }
- /* transport mode */
- else
- {
- if (PREDICT_FALSE(transport_ip6))
- {
- ih6 = (ip6_header_t *) (b0->data + sizeof(ethernet_header_t));
- vlib_buffer_advance (b0, -sizeof(ip6_header_t));
- oh6 = vlib_buffer_get_current (b0);
- memmove(oh6, ih6, sizeof(ip6_header_t));
-
- next0 = ESP_DECRYPT_NEXT_IP6_INPUT;
- oh6->protocol = f0->next_header;
- oh6->payload_length =
- clib_host_to_net_u16 (
- vlib_buffer_length_in_chain(vm, b0) -
- sizeof (ip6_header_t));
- }
- else
- {
- vlib_buffer_advance (b0, -sizeof(ip4_header_t));
- oh4 = vlib_buffer_get_current (b0);
- memmove(oh4, ih4, sizeof(ip4_header_t));
-
- next0 = ESP_DECRYPT_NEXT_IP4_INPUT;
- oh4->ip_version_and_header_length = 0x45;
- oh4->fragment_id = 0;
- oh4->flags_and_fragment_offset = 0;
- oh4->protocol = f0->next_header;
- oh4->length = clib_host_to_net_u16 (
- vlib_buffer_length_in_chain (vm, b0));
- oh4->checksum = ip4_header_checksum (oh4);
- }
- }
-
- vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32)~0;
-
-trace:
- if (PREDICT_FALSE(b0->flags & VLIB_BUFFER_IS_TRACED))
- {
- esp_decrypt_trace_t *tr = vlib_add_trace (vm, node, b0, sizeof (*tr));
- tr->crypto_alg = sa0->crypto_alg;
- tr->integ_alg = sa0->integ_alg;
- }
-
- vlib_validate_buffer_enqueue_x1 (vm, node, next_index,
- to_next, n_left_to_next, bi0, next0);
- }
- vlib_put_next_frame (vm, node, next_index, n_left_to_next);
- }
- vlib_node_increment_counter (vm, dpdk_esp_decrypt_post_node.index,
- ESP_DECRYPT_POST_ERROR_PKTS,
- from_frame->n_vectors);
-
- return from_frame->n_vectors;
-}
-
-VLIB_REGISTER_NODE (dpdk_esp_decrypt_post_node) = {
- .function = dpdk_esp_decrypt_post_node_fn,
- .name = "dpdk-esp-decrypt-post",
- .vector_size = sizeof (u32),
- .format_trace = format_esp_decrypt_post_trace,
- .type = VLIB_NODE_TYPE_INTERNAL,
-
- .n_errors = ARRAY_LEN(esp_decrypt_post_error_strings),
- .error_strings = esp_decrypt_post_error_strings,
-
- .n_next_nodes = ESP_DECRYPT_N_NEXT,
- .next_nodes = {
-#define _(s,n) [ESP_DECRYPT_NEXT_##s] = n,
- foreach_esp_decrypt_next
-#undef _
- },
-};
-
-VLIB_NODE_FUNCTION_MULTIARCH (dpdk_esp_decrypt_post_node, dpdk_esp_decrypt_post_node_fn)
diff --git a/src/vnet/devices/dpdk/ipsec/esp_encrypt.c b/src/vnet/devices/dpdk/ipsec/esp_encrypt.c
deleted file mode 100644
index 6eb1afc9501..00000000000
--- a/src/vnet/devices/dpdk/ipsec/esp_encrypt.c
+++ /dev/null
@@ -1,609 +0,0 @@
-/*
- * esp_encrypt.c : IPSec ESP encrypt node using DPDK Cryptodev
- *
- * Copyright (c) 2016 Intel and/or its affiliates.
- * Licensed under the Apache License, Version 2.0 (the "License");
- * you may not use this file except in compliance with the License.
- * You may obtain a copy of the License at:
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-#include <vnet/vnet.h>
-#include <vnet/api_errno.h>
-#include <vnet/ip/ip.h>
-
-#include <vnet/ipsec/ipsec.h>
-#include <vnet/devices/dpdk/ipsec/ipsec.h>
-#include <vnet/devices/dpdk/ipsec/esp.h>
-#include <vnet/devices/dpdk/dpdk.h>
-#include <vnet/devices/dpdk/dpdk_priv.h>
-
-#define foreach_esp_encrypt_next \
-_(DROP, "error-drop") \
-_(IP4_LOOKUP, "ip4-lookup") \
-_(IP6_LOOKUP, "ip6-lookup") \
-_(INTERFACE_OUTPUT, "interface-output")
-
-#define _(v, s) ESP_ENCRYPT_NEXT_##v,
-typedef enum
-{
- foreach_esp_encrypt_next
-#undef _
- ESP_ENCRYPT_N_NEXT,
-} esp_encrypt_next_t;
-
-#define foreach_esp_encrypt_error \
- _(RX_PKTS, "ESP pkts received") \
- _(SEQ_CYCLED, "sequence number cycled") \
- _(ENQ_FAIL, "Enqueue failed (buffer full)") \
- _(NO_CRYPTODEV, "Cryptodev not configured") \
- _(UNSUPPORTED, "Cipher/Auth not supported")
-
-
-typedef enum
-{
-#define _(sym,str) ESP_ENCRYPT_ERROR_##sym,
- foreach_esp_encrypt_error
-#undef _
- ESP_ENCRYPT_N_ERROR,
-} esp_encrypt_error_t;
-
-static char *esp_encrypt_error_strings[] = {
-#define _(sym,string) string,
- foreach_esp_encrypt_error
-#undef _
-};
-
-vlib_node_registration_t dpdk_esp_encrypt_node;
-
-typedef struct
-{
- u32 spi;
- u32 seq;
- ipsec_crypto_alg_t crypto_alg;
- ipsec_integ_alg_t integ_alg;
-} esp_encrypt_trace_t;
-
-/* packet trace format function */
-static u8 *
-format_esp_encrypt_trace (u8 * s, va_list * args)
-{
- CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
- CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
- esp_encrypt_trace_t *t = va_arg (*args, esp_encrypt_trace_t *);
-
- s = format (s, "esp: spi %u seq %u crypto %U integrity %U",
- t->spi, t->seq,
- format_ipsec_crypto_alg, t->crypto_alg,
- format_ipsec_integ_alg, t->integ_alg);
- return s;
-}
-
-static uword
-dpdk_esp_encrypt_node_fn (vlib_main_t * vm,
- vlib_node_runtime_t * node,
- vlib_frame_t * from_frame)
-{
- u32 n_left_from, *from, *to_next, next_index;
- ipsec_main_t *im = &ipsec_main;
- u32 cpu_index = os_get_cpu_number ();
- dpdk_crypto_main_t *dcm = &dpdk_crypto_main;
- dpdk_esp_main_t *em = &dpdk_esp_main;
- u32 i;
-
- from = vlib_frame_vector_args (from_frame);
- n_left_from = from_frame->n_vectors;
-
- if (PREDICT_FALSE (!dcm->workers_main))
- {
- /* Likely there are not enough cryptodevs, so drop frame */
- vlib_node_increment_counter (vm, dpdk_esp_encrypt_node.index,
- ESP_ENCRYPT_ERROR_NO_CRYPTODEV,
- n_left_from);
- vlib_buffer_free (vm, from, n_left_from);
- return n_left_from;
- }
-
- crypto_worker_main_t *cwm = vec_elt_at_index (dcm->workers_main, cpu_index);
- u32 n_qps = vec_len (cwm->qp_data);
- struct rte_crypto_op **cops_to_enq[n_qps];
- u32 n_cop_qp[n_qps], *bi_to_enq[n_qps];
-
- for (i = 0; i < n_qps; i++)
- {
- bi_to_enq[i] = cwm->qp_data[i].bi;
- cops_to_enq[i] = cwm->qp_data[i].cops;
- }
-
- memset (n_cop_qp, 0, n_qps * sizeof (u32));
-
- crypto_alloc_cops ();
-
- next_index = ESP_ENCRYPT_NEXT_DROP;
-
- while (n_left_from > 0)
- {
- u32 n_left_to_next;
-
- vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
-
- while (n_left_from > 0 && n_left_to_next > 0)
- {
- u32 bi0, next0;
- vlib_buffer_t *b0 = 0;
- u32 sa_index0;
- ipsec_sa_t *sa0;
- ip4_and_esp_header_t *ih0, *oh0 = 0;
- ip6_and_esp_header_t *ih6_0, *oh6_0 = 0;
- struct rte_mbuf *mb0 = 0;
- esp_footer_t *f0;
- u8 is_ipv6;
- u8 ip_hdr_size;
- u8 next_hdr_type;
- u8 transport_mode = 0;
- const int BLOCK_SIZE = 16;
- u32 iv_size;
- u16 orig_sz;
- crypto_sa_session_t *sa_sess;
- void *sess;
- struct rte_crypto_op *cop = 0;
- u16 qp_index;
-
- bi0 = from[0];
- from += 1;
- n_left_from -= 1;
-
- b0 = vlib_get_buffer (vm, bi0);
- sa_index0 = vnet_buffer (b0)->ipsec.sad_index;
- sa0 = pool_elt_at_index (im->sad, sa_index0);
-
- if (PREDICT_FALSE (esp_seq_advance (sa0)))
- {
- clib_warning ("sequence number counter has cycled SPI %u",
- sa0->spi);
- vlib_node_increment_counter (vm, dpdk_esp_encrypt_node.index,
- ESP_ENCRYPT_ERROR_SEQ_CYCLED, 1);
- //TODO: rekey SA
- to_next[0] = bi0;
- to_next += 1;
- n_left_to_next -= 1;
- goto trace;
- }
-
- sa0->total_data_size += b0->current_length;
-
- sa_sess = pool_elt_at_index (cwm->sa_sess_d[1], sa_index0);
- if (PREDICT_FALSE (!sa_sess->sess))
- {
- int ret = create_sym_sess (sa0, sa_sess, 1);
-
- if (PREDICT_FALSE (ret))
- {
- to_next[0] = bi0;
- to_next += 1;
- n_left_to_next -= 1;
- goto trace;
- }
- }
-
- qp_index = sa_sess->qp_index;
- sess = sa_sess->sess;
-
- ASSERT (vec_len (vec_elt (cwm->qp_data, qp_index).free_cops) > 0);
- cop = vec_pop (vec_elt (cwm->qp_data, qp_index).free_cops);
- ASSERT (cop->status == RTE_CRYPTO_OP_STATUS_NOT_PROCESSED);
-
- cops_to_enq[qp_index][0] = cop;
- cops_to_enq[qp_index] += 1;
- n_cop_qp[qp_index] += 1;
- bi_to_enq[qp_index][0] = bi0;
- bi_to_enq[qp_index] += 1;
-
- ssize_t adv;
- iv_size = em->esp_crypto_algs[sa0->crypto_alg].iv_len;
- ih0 = vlib_buffer_get_current (b0);
- orig_sz = b0->current_length;
- is_ipv6 = (ih0->ip4.ip_version_and_header_length & 0xF0) == 0x60;
- /* is ipv6 */
- if (PREDICT_TRUE (sa0->is_tunnel))
- {
- if (PREDICT_TRUE (!is_ipv6))
- adv = -sizeof (ip4_and_esp_header_t);
- else
- adv = -sizeof (ip6_and_esp_header_t);
- }
- else
- {
- adv = -sizeof (esp_header_t);
- if (PREDICT_TRUE (!is_ipv6))
- orig_sz -= sizeof (ip4_header_t);
- else
- orig_sz -= sizeof (ip6_header_t);
- }
-
- /*transport mode save the eth header before it is overwritten */
- if (PREDICT_FALSE (!sa0->is_tunnel))
- {
- ethernet_header_t *ieh0 = (ethernet_header_t *)
- ((u8 *) vlib_buffer_get_current (b0) -
- sizeof (ethernet_header_t));
- ethernet_header_t *oeh0 =
- (ethernet_header_t *) ((u8 *) ieh0 + (adv - iv_size));
- clib_memcpy (oeh0, ieh0, sizeof (ethernet_header_t));
- }
-
- vlib_buffer_advance (b0, adv - iv_size);
-
- /* XXX IP6/ip4 and IP4/IP6 not supported, only IP4/IP4 and IP6/IP6 */
-
- /* is ipv6 */
- if (PREDICT_FALSE (is_ipv6))
- {
- ih6_0 = (ip6_and_esp_header_t *) ih0;
- ip_hdr_size = sizeof (ip6_header_t);
- oh6_0 = vlib_buffer_get_current (b0);
-
- if (PREDICT_TRUE (sa0->is_tunnel))
- {
- next_hdr_type = IP_PROTOCOL_IPV6;
- oh6_0->ip6.ip_version_traffic_class_and_flow_label =
- ih6_0->ip6.ip_version_traffic_class_and_flow_label;
- }
- else
- {
- next_hdr_type = ih6_0->ip6.protocol;
- memmove (oh6_0, ih6_0, sizeof (ip6_header_t));
- }
-
- oh6_0->ip6.protocol = IP_PROTOCOL_IPSEC_ESP;
- oh6_0->ip6.hop_limit = 254;
- oh6_0->esp.spi = clib_net_to_host_u32 (sa0->spi);
- oh6_0->esp.seq = clib_net_to_host_u32 (sa0->seq);
- }
- else
- {
- ip_hdr_size = sizeof (ip4_header_t);
- oh0 = vlib_buffer_get_current (b0);
-
- if (PREDICT_TRUE (sa0->is_tunnel))
- {
- next_hdr_type = IP_PROTOCOL_IP_IN_IP;
- oh0->ip4.tos = ih0->ip4.tos;
- }
- else
- {
- next_hdr_type = ih0->ip4.protocol;
- memmove (oh0, ih0, sizeof (ip4_header_t));
- }
-
- oh0->ip4.ip_version_and_header_length = 0x45;
- oh0->ip4.fragment_id = 0;
- oh0->ip4.flags_and_fragment_offset = 0;
- oh0->ip4.ttl = 254;
- oh0->ip4.protocol = IP_PROTOCOL_IPSEC_ESP;
- oh0->esp.spi = clib_net_to_host_u32 (sa0->spi);
- oh0->esp.seq = clib_net_to_host_u32 (sa0->seq);
- }
-
- if (PREDICT_TRUE (sa0->is_tunnel && !sa0->is_tunnel_ip6))
- {
- oh0->ip4.src_address.as_u32 = sa0->tunnel_src_addr.ip4.as_u32;
- oh0->ip4.dst_address.as_u32 = sa0->tunnel_dst_addr.ip4.as_u32;
-
- /* in tunnel mode send it back to FIB */
- next0 = ESP_ENCRYPT_NEXT_IP4_LOOKUP;
- vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0;
- }
- else if (sa0->is_tunnel && sa0->is_tunnel_ip6)
- {
- oh6_0->ip6.src_address.as_u64[0] =
- sa0->tunnel_src_addr.ip6.as_u64[0];
- oh6_0->ip6.src_address.as_u64[1] =
- sa0->tunnel_src_addr.ip6.as_u64[1];
- oh6_0->ip6.dst_address.as_u64[0] =
- sa0->tunnel_dst_addr.ip6.as_u64[0];
- oh6_0->ip6.dst_address.as_u64[1] =
- sa0->tunnel_dst_addr.ip6.as_u64[1];
-
- /* in tunnel mode send it back to FIB */
- next0 = ESP_ENCRYPT_NEXT_IP6_LOOKUP;
- vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0;
- }
- else
- {
- next0 = ESP_ENCRYPT_NEXT_INTERFACE_OUTPUT;
- transport_mode = 1;
- }
-
- ASSERT (sa0->crypto_alg < IPSEC_CRYPTO_N_ALG);
- ASSERT (sa0->crypto_alg != IPSEC_CRYPTO_ALG_NONE);
-
- int blocks = 1 + (orig_sz + 1) / BLOCK_SIZE;
-
- /* pad packet in input buffer */
- u8 pad_bytes = BLOCK_SIZE * blocks - 2 - orig_sz;
- u8 i;
- u8 *padding = vlib_buffer_get_current (b0) + b0->current_length;
-
- for (i = 0; i < pad_bytes; ++i)
- padding[i] = i + 1;
-
- f0 = vlib_buffer_get_current (b0) + b0->current_length + pad_bytes;
- f0->pad_length = pad_bytes;
- f0->next_header = next_hdr_type;
- b0->current_length += pad_bytes + 2 +
- em->esp_integ_algs[sa0->integ_alg].trunc_size;
-
- vnet_buffer (b0)->sw_if_index[VLIB_RX] =
- vnet_buffer (b0)->sw_if_index[VLIB_RX];
- b0->flags |= VLIB_BUFFER_TOTAL_LENGTH_VALID;
-
- struct rte_crypto_sym_op *sym_cop;
- sym_cop = (struct rte_crypto_sym_op *) (cop + 1);
-
- dpdk_cop_priv_t *priv = (dpdk_cop_priv_t *) (sym_cop + 1);
-
- vnet_buffer (b0)->unused[0] = next0;
-
- mb0 = rte_mbuf_from_vlib_buffer (b0);
- mb0->data_len = b0->current_length;
- mb0->pkt_len = b0->current_length;
- mb0->data_off = RTE_PKTMBUF_HEADROOM + b0->current_data;
-
- rte_crypto_op_attach_sym_session (cop, sess);
-
- sym_cop->m_src = mb0;
-
- dpdk_gcm_cnt_blk *icb = &priv->cb;
- icb->salt = sa0->salt;
- icb->iv[0] = sa0->seq;
- icb->iv[1] = sa0->seq_hi;
-
- if (sa0->crypto_alg == IPSEC_CRYPTO_ALG_AES_GCM_128)
- {
- icb->cnt = clib_host_to_net_u32 (1);
- clib_memcpy (vlib_buffer_get_current (b0) + ip_hdr_size +
- sizeof (esp_header_t), icb->iv, 8);
- sym_cop->cipher.data.offset =
- ip_hdr_size + sizeof (esp_header_t) + iv_size;
- sym_cop->cipher.data.length = BLOCK_SIZE * blocks;
- sym_cop->cipher.iv.length = 16;
- }
- else
- {
- sym_cop->cipher.data.offset =
- ip_hdr_size + sizeof (esp_header_t);
- sym_cop->cipher.data.length = BLOCK_SIZE * blocks + iv_size;
- sym_cop->cipher.iv.length = iv_size;
- }
-
- sym_cop->cipher.iv.data = (u8 *) icb;
- sym_cop->cipher.iv.phys_addr = cop->phys_addr + (uintptr_t) icb
- - (uintptr_t) cop;
-
-
- ASSERT (sa0->integ_alg < IPSEC_INTEG_N_ALG);
- ASSERT (sa0->integ_alg != IPSEC_INTEG_ALG_NONE);
-
- if (PREDICT_FALSE (sa0->integ_alg == IPSEC_INTEG_ALG_AES_GCM_128))
- {
- u8 *aad = priv->aad;
- clib_memcpy (aad, vlib_buffer_get_current (b0) + ip_hdr_size,
- 8);
- sym_cop->auth.aad.data = aad;
- sym_cop->auth.aad.phys_addr = cop->phys_addr +
- (uintptr_t) aad - (uintptr_t) cop;
-
- if (PREDICT_FALSE (sa0->use_esn))
- {
- *((u32 *) & aad[8]) = sa0->seq_hi;
- sym_cop->auth.aad.length = 12;
- }
- else
- {
- sym_cop->auth.aad.length = 8;
- }
- }
- else
- {
- sym_cop->auth.data.offset = ip_hdr_size;
- sym_cop->auth.data.length = b0->current_length - ip_hdr_size
- - em->esp_integ_algs[sa0->integ_alg].trunc_size;
-
- if (PREDICT_FALSE (sa0->use_esn))
- {
- u8 *payload_end =
- vlib_buffer_get_current (b0) + b0->current_length;
- *((u32 *) payload_end) = sa0->seq_hi;
- sym_cop->auth.data.length += sizeof (sa0->seq_hi);
- }
- }
- sym_cop->auth.digest.data = vlib_buffer_get_current (b0) +
- b0->current_length -
- em->esp_integ_algs[sa0->integ_alg].trunc_size;
- sym_cop->auth.digest.phys_addr = rte_pktmbuf_mtophys_offset (mb0,
- b0->current_length
- -
- em->esp_integ_algs
- [sa0->integ_alg].trunc_size);
- sym_cop->auth.digest.length =
- em->esp_integ_algs[sa0->integ_alg].trunc_size;
-
-
- if (PREDICT_FALSE (is_ipv6))
- {
- oh6_0->ip6.payload_length =
- clib_host_to_net_u16 (vlib_buffer_length_in_chain (vm, b0) -
- sizeof (ip6_header_t));
- }
- else
- {
- oh0->ip4.length =
- clib_host_to_net_u16 (vlib_buffer_length_in_chain (vm, b0));
- oh0->ip4.checksum = ip4_header_checksum (&oh0->ip4);
- }
-
- if (transport_mode)
- vlib_buffer_advance (b0, -sizeof (ethernet_header_t));
-
- trace:
- if (PREDICT_FALSE (b0->flags & VLIB_BUFFER_IS_TRACED))
- {
- esp_encrypt_trace_t *tr =
- vlib_add_trace (vm, node, b0, sizeof (*tr));
- tr->spi = sa0->spi;
- tr->seq = sa0->seq - 1;
- tr->crypto_alg = sa0->crypto_alg;
- tr->integ_alg = sa0->integ_alg;
- }
- }
- vlib_put_next_frame (vm, node, next_index, n_left_to_next);
- }
- vlib_node_increment_counter (vm, dpdk_esp_encrypt_node.index,
- ESP_ENCRYPT_ERROR_RX_PKTS,
- from_frame->n_vectors);
- crypto_qp_data_t *qpd;
- /* *INDENT-OFF* */
- vec_foreach_index (i, cwm->qp_data)
- {
- u32 enq;
-
- qpd = vec_elt_at_index(cwm->qp_data, i);
- enq = rte_cryptodev_enqueue_burst(qpd->dev_id, qpd->qp_id,
- qpd->cops, n_cop_qp[i]);
- qpd->inflights += enq;
-
- if (PREDICT_FALSE(enq < n_cop_qp[i]))
- {
- crypto_free_cop (qpd, &qpd->cops[enq], n_cop_qp[i] - enq);
- vlib_buffer_free (vm, &qpd->bi[enq], n_cop_qp[i] - enq);
-
- vlib_node_increment_counter (vm, dpdk_esp_encrypt_node.index,
- ESP_ENCRYPT_ERROR_ENQ_FAIL,
- n_cop_qp[i] - enq);
- }
- }
- /* *INDENT-ON* */
-
- return from_frame->n_vectors;
-}
-
-VLIB_REGISTER_NODE (dpdk_esp_encrypt_node) =
-{
- .function = dpdk_esp_encrypt_node_fn,.name = "dpdk-esp-encrypt",.flags =
- VLIB_NODE_FLAG_IS_OUTPUT,.vector_size = sizeof (u32),.format_trace =
- format_esp_encrypt_trace,.n_errors =
- ARRAY_LEN (esp_encrypt_error_strings),.error_strings =
- esp_encrypt_error_strings,.n_next_nodes = 1,.next_nodes =
- {
- [ESP_ENCRYPT_NEXT_DROP] = "error-drop",}
-};
-
-VLIB_NODE_FUNCTION_MULTIARCH (dpdk_esp_encrypt_node, dpdk_esp_encrypt_node_fn)
-/*
- * ESP Encrypt Post Node
- */
-#define foreach_esp_encrypt_post_error \
- _(PKTS, "ESP post pkts")
- typedef enum
- {
-#define _(sym,str) ESP_ENCRYPT_POST_ERROR_##sym,
- foreach_esp_encrypt_post_error
-#undef _
- ESP_ENCRYPT_POST_N_ERROR,
- } esp_encrypt_post_error_t;
-
- static char *esp_encrypt_post_error_strings[] = {
-#define _(sym,string) string,
- foreach_esp_encrypt_post_error
-#undef _
- };
-
-vlib_node_registration_t dpdk_esp_encrypt_post_node;
-
-static u8 *
-format_esp_encrypt_post_trace (u8 * s, va_list * args)
-{
- return s;
-}
-
-static uword
-dpdk_esp_encrypt_post_node_fn (vlib_main_t * vm,
- vlib_node_runtime_t * node,
- vlib_frame_t * from_frame)
-{
- u32 n_left_from, *from, *to_next = 0, next_index;
-
- from = vlib_frame_vector_args (from_frame);
- n_left_from = from_frame->n_vectors;
-
- next_index = node->cached_next_index;
-
- while (n_left_from > 0)
- {
- u32 n_left_to_next;
-
- vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
-
- while (n_left_from > 0 && n_left_to_next > 0)
- {
- u32 bi0, next0;
- vlib_buffer_t *b0 = 0;
-
- bi0 = from[0];
- from += 1;
- n_left_from -= 1;
- n_left_to_next -= 1;
-
- b0 = vlib_get_buffer (vm, bi0);
-
- to_next[0] = bi0;
- to_next += 1;
-
- next0 = vnet_buffer (b0)->unused[0];
-
- vlib_validate_buffer_enqueue_x1 (vm, node, next_index,
- to_next, n_left_to_next, bi0,
- next0);
- }
- vlib_put_next_frame (vm, node, next_index, n_left_to_next);
- }
-
- vlib_node_increment_counter (vm, dpdk_esp_encrypt_post_node.index,
- ESP_ENCRYPT_POST_ERROR_PKTS,
- from_frame->n_vectors);
-
- return from_frame->n_vectors;
-}
-
-VLIB_REGISTER_NODE (dpdk_esp_encrypt_post_node) =
-{
- .function = dpdk_esp_encrypt_post_node_fn,.name =
- "dpdk-esp-encrypt-post",.vector_size = sizeof (u32),.format_trace =
- format_esp_encrypt_post_trace,.type = VLIB_NODE_TYPE_INTERNAL,.n_errors =
- ARRAY_LEN (esp_encrypt_post_error_strings),.error_strings =
- esp_encrypt_post_error_strings,.n_next_nodes =
- ESP_ENCRYPT_N_NEXT,.next_nodes =
- {
-#define _(s,n) [ESP_ENCRYPT_NEXT_##s] = n,
- foreach_esp_encrypt_next
-#undef _
- }
-};
-
-VLIB_NODE_FUNCTION_MULTIARCH (dpdk_esp_encrypt_post_node,
- dpdk_esp_encrypt_post_node_fn)
-/*
- * fd.io coding-style-patch-verification: ON
- *
- * Local Variables:
- * eval: (c-set-style "gnu")
- * End:
- */
diff --git a/src/vnet/devices/dpdk/ipsec/ipsec.c b/src/vnet/devices/dpdk/ipsec/ipsec.c
deleted file mode 100644
index 05c17c995d8..00000000000
--- a/src/vnet/devices/dpdk/ipsec/ipsec.c
+++ /dev/null
@@ -1,430 +0,0 @@
-/*
- * Copyright (c) 2016 Intel and/or its affiliates.
- * Licensed under the Apache License, Version 2.0 (the "License");
- * you may not use this file except in compliance with the License.
- * You may obtain a copy of the License at:
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-#include <vnet/vnet.h>
-#include <vnet/ip/ip.h>
-#include <vnet/api_errno.h>
-#include <vnet/ipsec/ipsec.h>
-#include <vlib/node_funcs.h>
-
-#include <vnet/devices/dpdk/dpdk.h>
-#include <vnet/devices/dpdk/ipsec/ipsec.h>
-#include <vnet/devices/dpdk/ipsec/esp.h>
-
-#define DPDK_CRYPTO_NB_SESS_OBJS 20000
-#define DPDK_CRYPTO_CACHE_SIZE 512
-#define DPDK_CRYPTO_PRIV_SIZE 128
-#define DPDK_CRYPTO_N_QUEUE_DESC 1024
-#define DPDK_CRYPTO_NB_COPS (1024 * 4)
-
-static int
-add_del_sa_sess (u32 sa_index, u8 is_add)
-{
- dpdk_crypto_main_t *dcm = &dpdk_crypto_main;
- crypto_worker_main_t *cwm;
- u8 skip_master = vlib_num_workers () > 0;
-
- /* *INDENT-OFF* */
- vec_foreach (cwm, dcm->workers_main)
- {
- crypto_sa_session_t *sa_sess;
- u8 is_outbound;
-
- if (skip_master)
- {
- skip_master = 0;
- continue;
- }
-
- for (is_outbound = 0; is_outbound < 2; is_outbound++)
- {
- if (is_add)
- {
- pool_get (cwm->sa_sess_d[is_outbound], sa_sess);
- }
- else
- {
- u8 dev_id;
-
- sa_sess = pool_elt_at_index (cwm->sa_sess_d[is_outbound], sa_index);
- dev_id = cwm->qp_data[sa_sess->qp_index].dev_id;
-
- if (!sa_sess->sess)
- continue;
-
- if (rte_cryptodev_sym_session_free(dev_id, sa_sess->sess))
- {
- clib_warning("failed to free session");
- return -1;
- }
- memset(sa_sess, 0, sizeof(sa_sess[0]));
- }
- }
- }
- /* *INDENT-OFF* */
-
- return 0;
-}
-
-static void
-update_qp_data (crypto_worker_main_t * cwm,
- u8 cdev_id, u16 qp_id, u8 is_outbound, u16 * idx)
-{
- crypto_qp_data_t *qpd;
-
- /* *INDENT-OFF* */
- vec_foreach_index (*idx, cwm->qp_data)
- {
- qpd = vec_elt_at_index(cwm->qp_data, *idx);
-
- if (qpd->dev_id == cdev_id && qpd->qp_id == qp_id &&
- qpd->is_outbound == is_outbound)
- return;
- }
- /* *INDENT-ON* */
-
- vec_add2 (cwm->qp_data, qpd, 1);
-
- qpd->dev_id = cdev_id;
- qpd->qp_id = qp_id;
- qpd->is_outbound = is_outbound;
-}
-
-/*
- * return:
- * 0: already exist
- * 1: mapped
- */
-static int
-add_mapping (crypto_worker_main_t * cwm,
- u8 cdev_id, u16 qp, u8 is_outbound,
- const struct rte_cryptodev_capabilities *cipher_cap,
- const struct rte_cryptodev_capabilities *auth_cap)
-{
- u16 qp_index;
- uword key = 0, data, *ret;
- crypto_worker_qp_key_t *p_key = (crypto_worker_qp_key_t *) & key;
-
- p_key->cipher_algo = (u8) cipher_cap->sym.cipher.algo;
- p_key->auth_algo = (u8) auth_cap->sym.auth.algo;
- p_key->is_outbound = is_outbound;
-
- ret = hash_get (cwm->algo_qp_map, key);
- if (ret)
- return 0;
-
- update_qp_data (cwm, cdev_id, qp, is_outbound, &qp_index);
-
- data = (uword) qp_index;
- hash_set (cwm->algo_qp_map, key, data);
-
- return 1;
-}
-
-/*
- * return:
- * 0: already exist
- * 1: mapped
- */
-static int
-add_cdev_mapping (crypto_worker_main_t * cwm,
- struct rte_cryptodev_info *dev_info, u8 cdev_id,
- u16 qp, u8 is_outbound)
-{
- const struct rte_cryptodev_capabilities *i, *j;
- u32 mapped = 0;
-
- for (i = dev_info->capabilities; i->op != RTE_CRYPTO_OP_TYPE_UNDEFINED; i++)
- {
- if (i->sym.xform_type != RTE_CRYPTO_SYM_XFORM_CIPHER)
- continue;
-
- if (check_algo_is_supported (i, NULL) != 0)
- continue;
-
- for (j = dev_info->capabilities; j->op != RTE_CRYPTO_OP_TYPE_UNDEFINED;
- j++)
- {
- if (j->sym.xform_type != RTE_CRYPTO_SYM_XFORM_AUTH)
- continue;
-
- if (check_algo_is_supported (j, NULL) != 0)
- continue;
-
- mapped |= add_mapping (cwm, cdev_id, qp, is_outbound, i, j);
- }
- }
-
- return mapped;
-}
-
-static int
-check_cryptodev_queues ()
-{
- u32 n_qs = 0;
- u8 cdev_id;
- u32 n_req_qs = 2;
-
- if (vlib_num_workers () > 0)
- n_req_qs = vlib_num_workers () * 2;
-
- for (cdev_id = 0; cdev_id < rte_cryptodev_count (); cdev_id++)
- {
- struct rte_cryptodev_info cdev_info;
-
- rte_cryptodev_info_get (cdev_id, &cdev_info);
-
- if (!
- (cdev_info.feature_flags & RTE_CRYPTODEV_FF_SYM_OPERATION_CHAINING))
- continue;
-
- n_qs += cdev_info.max_nb_queue_pairs;
- }
-
- if (n_qs >= n_req_qs)
- return 0;
- else
- return -1;
-}
-
-static clib_error_t *
-dpdk_ipsec_check_support (ipsec_sa_t * sa)
-{
- if (sa->crypto_alg == IPSEC_CRYPTO_ALG_AES_GCM_128)
- {
- if (sa->integ_alg != IPSEC_INTEG_ALG_NONE)
- return clib_error_return (0, "unsupported integ-alg %U with "
- "crypto-algo aes-gcm-128",
- format_ipsec_integ_alg, sa->integ_alg);
- sa->integ_alg = IPSEC_INTEG_ALG_AES_GCM_128;
- }
- else
- {
- if (sa->integ_alg == IPSEC_INTEG_ALG_NONE ||
- sa->integ_alg == IPSEC_INTEG_ALG_AES_GCM_128)
- return clib_error_return (0, "unsupported integ-alg %U",
- format_ipsec_integ_alg, sa->integ_alg);
- }
-
- return 0;
-}
-
-static uword
-dpdk_ipsec_process (vlib_main_t * vm, vlib_node_runtime_t * rt,
- vlib_frame_t * f)
-{
- dpdk_config_main_t *conf = &dpdk_config_main;
- ipsec_main_t *im = &ipsec_main;
- dpdk_crypto_main_t *dcm = &dpdk_crypto_main;
- vlib_thread_main_t *tm = vlib_get_thread_main ();
- struct rte_cryptodev_config dev_conf;
- struct rte_cryptodev_qp_conf qp_conf;
- struct rte_cryptodev_info cdev_info;
- struct rte_mempool *rmp;
- i32 dev_id, ret;
- u32 i, skip_master;
-
- if (!conf->cryptodev)
- {
- clib_warning ("DPDK Cryptodev support is disabled, "
- "default to OpenSSL IPsec");
- return 0;
- }
-
- if (check_cryptodev_queues () < 0)
- {
- conf->cryptodev = 0;
- clib_warning ("not enough Cryptodevs, default to OpenSSL IPsec");
- return 0;
- }
-
- vec_alloc (dcm->workers_main, tm->n_vlib_mains);
- _vec_len (dcm->workers_main) = tm->n_vlib_mains;
-
- fprintf (stdout, "DPDK Cryptodevs info:\n");
- fprintf (stdout, "dev_id\tn_qp\tnb_obj\tcache_size\n");
- /* HW cryptodevs have higher dev_id, use HW first */
- for (dev_id = rte_cryptodev_count () - 1; dev_id >= 0; dev_id--)
- {
- u16 max_nb_qp, qp = 0;
- skip_master = vlib_num_workers () > 0;
-
- rte_cryptodev_info_get (dev_id, &cdev_info);
-
- if (!
- (cdev_info.feature_flags & RTE_CRYPTODEV_FF_SYM_OPERATION_CHAINING))
- continue;
-
- max_nb_qp = cdev_info.max_nb_queue_pairs;
-
- for (i = 0; i < tm->n_vlib_mains; i++)
- {
- u8 is_outbound;
- crypto_worker_main_t *cwm;
- uword *map;
-
- if (skip_master)
- {
- skip_master = 0;
- continue;
- }
-
- cwm = vec_elt_at_index (dcm->workers_main, i);
- map = cwm->algo_qp_map;
-
- if (!map)
- {
- map = hash_create (0, sizeof (crypto_worker_qp_key_t));
- if (!map)
- {
- clib_warning ("unable to create hash table for worker %u",
- vlib_mains[i]->cpu_index);
- goto error;
- }
- cwm->algo_qp_map = map;
- }
-
- for (is_outbound = 0; is_outbound < 2 && qp < max_nb_qp;
- is_outbound++)
- qp += add_cdev_mapping (cwm, &cdev_info, dev_id, qp, is_outbound);
- }
-
- if (qp == 0)
- continue;
-
- dev_conf.socket_id = rte_cryptodev_socket_id (dev_id);
- dev_conf.nb_queue_pairs = cdev_info.max_nb_queue_pairs;
- dev_conf.session_mp.nb_objs = DPDK_CRYPTO_NB_SESS_OBJS;
- dev_conf.session_mp.cache_size = DPDK_CRYPTO_CACHE_SIZE;
-
- ret = rte_cryptodev_configure (dev_id, &dev_conf);
- if (ret < 0)
- {
- clib_warning ("cryptodev %u config error", dev_id);
- goto error;
- }
-
- qp_conf.nb_descriptors = DPDK_CRYPTO_N_QUEUE_DESC;
- for (qp = 0; qp < dev_conf.nb_queue_pairs; qp++)
- {
- ret = rte_cryptodev_queue_pair_setup (dev_id, qp, &qp_conf,
- dev_conf.socket_id);
- if (ret < 0)
- {
- clib_warning ("cryptodev %u qp %u setup error", dev_id, qp);
- goto error;
- }
- }
- vec_validate_aligned (dcm->cop_pools, dev_conf.socket_id,
- CLIB_CACHE_LINE_BYTES);
-
- if (!vec_elt (dcm->cop_pools, dev_conf.socket_id))
- {
- u8 *pool_name = format (0, "crypto_op_pool_socket%u%c",
- dev_conf.socket_id, 0);
-
- rmp = rte_crypto_op_pool_create ((char *) pool_name,
- RTE_CRYPTO_OP_TYPE_SYMMETRIC,
- DPDK_CRYPTO_NB_COPS *
- (1 + vlib_num_workers ()),
- DPDK_CRYPTO_CACHE_SIZE,
- DPDK_CRYPTO_PRIV_SIZE,
- dev_conf.socket_id);
- vec_free (pool_name);
-
- if (!rmp)
- {
- clib_warning ("failed to allocate mempool on socket %u",
- dev_conf.socket_id);
- goto error;
- }
- vec_elt (dcm->cop_pools, dev_conf.socket_id) = rmp;
- }
-
- fprintf (stdout, "%u\t%u\t%u\t%u\n", dev_id, dev_conf.nb_queue_pairs,
- DPDK_CRYPTO_NB_SESS_OBJS, DPDK_CRYPTO_CACHE_SIZE);
- }
-
- dpdk_esp_init ();
-
- /* Add new next node and set as default */
- vlib_node_t *node, *next_node;
-
- next_node = vlib_get_node_by_name (vm, (u8 *) "dpdk-esp-encrypt");
- ASSERT (next_node);
- node = vlib_get_node_by_name (vm, (u8 *) "ipsec-output-ip4");
- ASSERT (node);
- im->esp_encrypt_node_index = next_node->index;
- im->esp_encrypt_next_index =
- vlib_node_add_next (vm, node->index, next_node->index);
-
- next_node = vlib_get_node_by_name (vm, (u8 *) "dpdk-esp-decrypt");
- ASSERT (next_node);
- node = vlib_get_node_by_name (vm, (u8 *) "ipsec-input-ip4");
- ASSERT (node);
- im->esp_decrypt_node_index = next_node->index;
- im->esp_decrypt_next_index =
- vlib_node_add_next (vm, node->index, next_node->index);
-
- im->cb.check_support_cb = dpdk_ipsec_check_support;
- im->cb.add_del_sa_sess_cb = add_del_sa_sess;
-
- if (vec_len (vlib_mains) == 0)
- vlib_node_set_state (&vlib_global_main, dpdk_crypto_input_node.index,
- VLIB_NODE_STATE_POLLING);
- else
- for (i = 1; i < tm->n_vlib_mains; i++)
- vlib_node_set_state (vlib_mains[i], dpdk_crypto_input_node.index,
- VLIB_NODE_STATE_POLLING);
-
- /* TODO cryptodev counters */
-
- return 0;
-
-error:
- ;
- crypto_worker_main_t *cwm;
- struct rte_mempool **mp;
- /* *INDENT-OFF* */
- vec_foreach (cwm, dcm->workers_main)
- hash_free (cwm->algo_qp_map);
-
- vec_foreach (mp, dcm->cop_pools)
- {
- if (mp)
- rte_mempool_free (mp[0]);
- }
- /* *INDENT-ON* */
- vec_free (dcm->workers_main);
- vec_free (dcm->cop_pools);
-
- return 0;
-}
-
-/* *INDENT-OFF* */
-VLIB_REGISTER_NODE (dpdk_ipsec_process_node,static) = {
- .function = dpdk_ipsec_process,
- .type = VLIB_NODE_TYPE_PROCESS,
- .name = "dpdk-ipsec-process",
- .process_log2_n_stack_bytes = 17,
-};
-/* *INDENT-ON* */
-
-/*
- * fd.io coding-style-patch-verification: ON
- *
- * Local Variables:
- * eval: (c-set-style "gnu")
- * End:
- */
diff --git a/src/vnet/devices/dpdk/ipsec/ipsec.h b/src/vnet/devices/dpdk/ipsec/ipsec.h
deleted file mode 100644
index 3465b3618ad..00000000000
--- a/src/vnet/devices/dpdk/ipsec/ipsec.h
+++ /dev/null
@@ -1,227 +0,0 @@
-/*
- * Copyright (c) 2016 Intel and/or its affiliates.
- * Licensed under the Apache License, Version 2.0 (the "License");
- * you may not use this file except in compliance with the License.
- * You may obtain a copy of the License at:
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-#ifndef __DPDK_IPSEC_H__
-#define __DPDK_IPSEC_H__
-
-#include <vnet/vnet.h>
-
-#undef always_inline
-#include <rte_crypto.h>
-#include <rte_cryptodev.h>
-
-#if CLIB_DEBUG > 0
-#define always_inline static inline
-#else
-#define always_inline static inline __attribute__ ((__always_inline__))
-#endif
-
-
-#define MAX_QP_PER_LCORE 16
-
-typedef struct
-{
- u32 salt;
- u32 iv[2];
- u32 cnt;
-} dpdk_gcm_cnt_blk;
-
-typedef struct
-{
- dpdk_gcm_cnt_blk cb;
- union
- {
- u8 aad[12];
- u8 icv[64];
- };
-} dpdk_cop_priv_t;
-
-typedef struct
-{
- u8 cipher_algo;
- u8 auth_algo;
- u8 is_outbound;
-} crypto_worker_qp_key_t;
-
-typedef struct
-{
- u16 dev_id;
- u16 qp_id;
- u16 is_outbound;
- i16 inflights;
- u32 bi[VLIB_FRAME_SIZE];
- struct rte_crypto_op *cops[VLIB_FRAME_SIZE];
- struct rte_crypto_op **free_cops;
-} crypto_qp_data_t;
-
-typedef struct
-{
- u8 qp_index;
- void *sess;
-} crypto_sa_session_t;
-
-typedef struct
-{
- crypto_sa_session_t *sa_sess_d[2];
- crypto_qp_data_t *qp_data;
- uword *algo_qp_map;
-} crypto_worker_main_t;
-
-typedef struct
-{
- struct rte_mempool **cop_pools;
- crypto_worker_main_t *workers_main;
-} dpdk_crypto_main_t;
-
-dpdk_crypto_main_t dpdk_crypto_main;
-
-extern vlib_node_registration_t dpdk_crypto_input_node;
-
-#define CRYPTO_N_FREE_COPS (VLIB_FRAME_SIZE * 3)
-
-static_always_inline void
-crypto_alloc_cops ()
-{
- dpdk_crypto_main_t *dcm = &dpdk_crypto_main;
- u32 cpu_index = os_get_cpu_number ();
- crypto_worker_main_t *cwm = &dcm->workers_main[cpu_index];
- unsigned socket_id = rte_socket_id ();
- crypto_qp_data_t *qpd;
-
- /* *INDENT-OFF* */
- vec_foreach (qpd, cwm->qp_data)
- {
- u32 l = vec_len (qpd->free_cops);
-
- if (PREDICT_FALSE (l < VLIB_FRAME_SIZE))
- {
- u32 n_alloc;
-
- if (PREDICT_FALSE (!qpd->free_cops))
- vec_alloc (qpd->free_cops, CRYPTO_N_FREE_COPS);
-
- n_alloc = rte_crypto_op_bulk_alloc (dcm->cop_pools[socket_id],
- RTE_CRYPTO_OP_TYPE_SYMMETRIC,
- &qpd->free_cops[l],
- CRYPTO_N_FREE_COPS - l - 1);
-
- _vec_len (qpd->free_cops) = l + n_alloc;
- }
- }
- /* *INDENT-ON* */
-}
-
-static_always_inline void
-crypto_free_cop (crypto_qp_data_t * qpd, struct rte_crypto_op **cops, u32 n)
-{
- u32 l = vec_len (qpd->free_cops);
-
- if (l + n >= CRYPTO_N_FREE_COPS)
- {
- l -= VLIB_FRAME_SIZE;
- rte_mempool_put_bulk (cops[0]->mempool,
- (void **) &qpd->free_cops[l], VLIB_FRAME_SIZE);
- }
- clib_memcpy (&qpd->free_cops[l], cops, sizeof (*cops) * n);
-
- _vec_len (qpd->free_cops) = l + n;
-}
-
-static_always_inline int
-check_algo_is_supported (const struct rte_cryptodev_capabilities *cap,
- char *name)
-{
- struct
- {
- uint8_t cipher_algo;
- enum rte_crypto_sym_xform_type type;
- union
- {
- enum rte_crypto_auth_algorithm auth;
- enum rte_crypto_cipher_algorithm cipher;
- };
- char *name;
- } supported_algo[] =
- {
- {
- .type = RTE_CRYPTO_SYM_XFORM_CIPHER,.cipher =
- RTE_CRYPTO_CIPHER_NULL,.name = "NULL"},
- {
- .type = RTE_CRYPTO_SYM_XFORM_CIPHER,.cipher =
- RTE_CRYPTO_CIPHER_AES_CBC,.name = "AES_CBC"},
- {
- .type = RTE_CRYPTO_SYM_XFORM_CIPHER,.cipher =
- RTE_CRYPTO_CIPHER_AES_CTR,.name = "AES_CTR"},
- {
- .type = RTE_CRYPTO_SYM_XFORM_CIPHER,.cipher =
- RTE_CRYPTO_CIPHER_3DES_CBC,.name = "3DES-CBC"},
- {
- .type = RTE_CRYPTO_SYM_XFORM_CIPHER,.cipher =
- RTE_CRYPTO_CIPHER_AES_GCM,.name = "AES-GCM"},
- {
- .type = RTE_CRYPTO_SYM_XFORM_AUTH,.auth =
- RTE_CRYPTO_AUTH_SHA1_HMAC,.name = "HMAC-SHA1"},
- {
- .type = RTE_CRYPTO_SYM_XFORM_AUTH,.auth =
- RTE_CRYPTO_AUTH_SHA256_HMAC,.name = "HMAC-SHA256"},
- {
- .type = RTE_CRYPTO_SYM_XFORM_AUTH,.auth =
- RTE_CRYPTO_AUTH_SHA384_HMAC,.name = "HMAC-SHA384"},
- {
- .type = RTE_CRYPTO_SYM_XFORM_AUTH,.auth =
- RTE_CRYPTO_AUTH_SHA512_HMAC,.name = "HMAC-SHA512"},
- {
- .type = RTE_CRYPTO_SYM_XFORM_AUTH,.auth =
- RTE_CRYPTO_AUTH_AES_XCBC_MAC,.name = "AES-XCBC-MAC"},
- {
- .type = RTE_CRYPTO_SYM_XFORM_AUTH,.auth =
- RTE_CRYPTO_AUTH_AES_GCM,.name = "AES-GCM"},
- {
- /* tail */
- .type = RTE_CRYPTO_SYM_XFORM_NOT_SPECIFIED},};
- uint32_t i = 0;
-
- if (cap->op != RTE_CRYPTO_OP_TYPE_SYMMETRIC)
- return -1;
-
- while (supported_algo[i].type != RTE_CRYPTO_SYM_XFORM_NOT_SPECIFIED)
- {
- if (cap->sym.xform_type == supported_algo[i].type)
- {
- if ((cap->sym.xform_type == RTE_CRYPTO_SYM_XFORM_CIPHER &&
- cap->sym.cipher.algo == supported_algo[i].cipher) ||
- (cap->sym.xform_type == RTE_CRYPTO_SYM_XFORM_AUTH &&
- cap->sym.auth.algo == supported_algo[i].auth))
- {
- if (name)
- strcpy (name, supported_algo[i].name);
- return 0;
- }
- }
-
- i++;
- }
-
- return -1;
-}
-
-#endif /* __DPDK_IPSEC_H__ */
-
-/*
- * fd.io coding-style-patch-verification: ON
- *
- * Local Variables:
- * eval: (c-set-style "gnu")
- * End:
- */