diff options
Diffstat (limited to 'src/plugins/dpdk/ipsec/ipsec.c')
-rw-r--r-- | src/plugins/dpdk/ipsec/ipsec.c | 1269 |
1 files changed, 940 insertions, 329 deletions
diff --git a/src/plugins/dpdk/ipsec/ipsec.c b/src/plugins/dpdk/ipsec/ipsec.c index c9fce3d8392..2fd331c1ccd 100644 --- a/src/plugins/dpdk/ipsec/ipsec.c +++ b/src/plugins/dpdk/ipsec/ipsec.c @@ -1,5 +1,5 @@ /* - * Copyright (c) 2016 Intel and/or its affiliates. + * Copyright (c) 2017 Intel and/or its affiliates. * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at: @@ -20,426 +20,1058 @@ #include <dpdk/device/dpdk.h> #include <dpdk/ipsec/ipsec.h> -#include <dpdk/ipsec/esp.h> -#define DPDK_CRYPTO_NB_SESS_OBJS 20000 -#define DPDK_CRYPTO_CACHE_SIZE 512 -#define DPDK_CRYPTO_PRIV_SIZE 128 -#define DPDK_CRYPTO_N_QUEUE_DESC 1024 -#define DPDK_CRYPTO_NB_COPS (1024 * 4) +#define EMPTY_STRUCT {0} -static int -add_del_sa_sess (u32 sa_index, u8 is_add) +static void +algos_init (u32 n_mains) { dpdk_crypto_main_t *dcm = &dpdk_crypto_main; - crypto_worker_main_t *cwm; - u8 skip_master = vlib_num_workers () > 0; + crypto_alg_t *a; + + vec_validate_aligned (dcm->cipher_algs, IPSEC_CRYPTO_N_ALG - 1, 8); + + { +#define _(v,f,str) \ + dcm->cipher_algs[IPSEC_CRYPTO_ALG_##f].name = str; \ + dcm->cipher_algs[IPSEC_CRYPTO_ALG_##f].disabled = n_mains; + foreach_ipsec_crypto_alg +#undef _ + } + + /* Minimum boundary for ciphers is 4B, required by ESP */ + a = &dcm->cipher_algs[IPSEC_CRYPTO_ALG_NONE]; + a->type = RTE_CRYPTO_SYM_XFORM_CIPHER; + a->alg = RTE_CRYPTO_CIPHER_NULL; + a->boundary = 4; /* 1 */ + a->key_len = 0; + a->iv_len = 0; + + a = &dcm->cipher_algs[IPSEC_CRYPTO_ALG_AES_CBC_128]; + a->type = RTE_CRYPTO_SYM_XFORM_CIPHER; + a->alg = RTE_CRYPTO_CIPHER_AES_CBC; + a->boundary = 16; + a->key_len = 16; + a->iv_len = 16; + + a = &dcm->cipher_algs[IPSEC_CRYPTO_ALG_AES_CBC_192]; + a->type = RTE_CRYPTO_SYM_XFORM_CIPHER; + a->alg = RTE_CRYPTO_CIPHER_AES_CBC; + a->boundary = 16; + a->key_len = 24; + a->iv_len = 16; + + a = &dcm->cipher_algs[IPSEC_CRYPTO_ALG_AES_CBC_256]; + a->type = RTE_CRYPTO_SYM_XFORM_CIPHER; + a->alg = RTE_CRYPTO_CIPHER_AES_CBC; + a->boundary = 16; + a->key_len = 32; + a->iv_len = 16; + + a = &dcm->cipher_algs[IPSEC_CRYPTO_ALG_AES_CTR_128]; + a->type = RTE_CRYPTO_SYM_XFORM_CIPHER; + a->alg = RTE_CRYPTO_CIPHER_AES_CTR; + a->boundary = 4; /* 1 */ + a->key_len = 16; + a->iv_len = 8; + + a = &dcm->cipher_algs[IPSEC_CRYPTO_ALG_AES_CTR_192]; + a->type = RTE_CRYPTO_SYM_XFORM_CIPHER; + a->alg = RTE_CRYPTO_CIPHER_AES_CTR; + a->boundary = 4; /* 1 */ + a->key_len = 24; + a->iv_len = 8; + + a = &dcm->cipher_algs[IPSEC_CRYPTO_ALG_AES_CTR_256]; + a->type = RTE_CRYPTO_SYM_XFORM_CIPHER; + a->alg = RTE_CRYPTO_CIPHER_AES_CTR; + a->boundary = 4; /* 1 */ + a->key_len = 32; + a->iv_len = 8; - /* *INDENT-OFF* */ - vec_foreach (cwm, dcm->workers_main) - { - crypto_sa_session_t *sa_sess; - u8 is_outbound; +#if DPDK_NO_AEAD +#define AES_GCM_TYPE RTE_CRYPTO_SYM_XFORM_CIPHER +#define AES_GCM_ALG RTE_CRYPTO_CIPHER_AES_GCM +#else +#define AES_GCM_TYPE RTE_CRYPTO_SYM_XFORM_AEAD +#define AES_GCM_ALG RTE_CRYPTO_AEAD_AES_GCM +#endif - if (skip_master) - { - skip_master = 0; - continue; - } + a = &dcm->cipher_algs[IPSEC_CRYPTO_ALG_AES_GCM_128]; + a->type = AES_GCM_TYPE; + a->alg = AES_GCM_ALG; + a->boundary = 4; /* 1 */ + a->key_len = 16; + a->iv_len = 8; + a->trunc_size = 16; + + a = &dcm->cipher_algs[IPSEC_CRYPTO_ALG_AES_GCM_192]; + a->type = AES_GCM_TYPE; + a->alg = AES_GCM_ALG; + a->boundary = 4; /* 1 */ + a->key_len = 24; + a->iv_len = 8; + a->trunc_size = 16; + + a = &dcm->cipher_algs[IPSEC_CRYPTO_ALG_AES_GCM_256]; + a->type = AES_GCM_TYPE; + a->alg = AES_GCM_ALG; + a->boundary = 4; /* 1 */ + a->key_len = 32; + a->iv_len = 8; + a->trunc_size = 16; + + vec_validate (dcm->auth_algs, IPSEC_INTEG_N_ALG - 1); + + { +#define _(v,f,str) \ + dcm->auth_algs[IPSEC_INTEG_ALG_##f].name = str; \ + dcm->auth_algs[IPSEC_INTEG_ALG_##f].disabled = n_mains; + foreach_ipsec_integ_alg +#undef _ + } + + a = &dcm->auth_algs[IPSEC_INTEG_ALG_NONE]; + a->type = RTE_CRYPTO_SYM_XFORM_AUTH; + a->alg = RTE_CRYPTO_AUTH_NULL; + a->key_len = 0; + a->trunc_size = 0; + + a = &dcm->auth_algs[IPSEC_INTEG_ALG_MD5_96]; + a->type = RTE_CRYPTO_SYM_XFORM_AUTH; + a->alg = RTE_CRYPTO_AUTH_MD5_HMAC; + a->key_len = 16; + a->trunc_size = 12; + + a = &dcm->auth_algs[IPSEC_INTEG_ALG_SHA1_96]; + a->type = RTE_CRYPTO_SYM_XFORM_AUTH; + a->alg = RTE_CRYPTO_AUTH_SHA1_HMAC; + a->key_len = 20; + a->trunc_size = 12; + + a = &dcm->auth_algs[IPSEC_INTEG_ALG_SHA_256_96]; + a->type = RTE_CRYPTO_SYM_XFORM_AUTH; + a->alg = RTE_CRYPTO_AUTH_SHA256_HMAC; + a->key_len = 32; + a->trunc_size = 12; + + a = &dcm->auth_algs[IPSEC_INTEG_ALG_SHA_256_128]; + a->type = RTE_CRYPTO_SYM_XFORM_AUTH; + a->alg = RTE_CRYPTO_AUTH_SHA256_HMAC; + a->key_len = 32; + a->trunc_size = 16; + + a = &dcm->auth_algs[IPSEC_INTEG_ALG_SHA_384_192]; + a->type = RTE_CRYPTO_SYM_XFORM_AUTH; + a->alg = RTE_CRYPTO_AUTH_SHA384_HMAC; + a->key_len = 48; + a->trunc_size = 24; + + a = &dcm->auth_algs[IPSEC_INTEG_ALG_SHA_512_256]; + a->type = RTE_CRYPTO_SYM_XFORM_AUTH; + a->alg = RTE_CRYPTO_AUTH_SHA512_HMAC; + a->key_len = 64; + a->trunc_size = 32; +} - for (is_outbound = 0; is_outbound < 2; is_outbound++) - { - if (is_add) - { - pool_get (cwm->sa_sess_d[is_outbound], sa_sess); - } - else - { - u8 dev_id; - i32 ret; +static u8 +cipher_alg_index (const crypto_alg_t * alg) +{ + dpdk_crypto_main_t *dcm = &dpdk_crypto_main; - sa_sess = pool_elt_at_index (cwm->sa_sess_d[is_outbound], sa_index); - dev_id = cwm->qp_data[sa_sess->qp_index].dev_id; + return (alg - dcm->cipher_algs); +} - if (!sa_sess->sess) - continue; -#if DPDK_NO_AEAD - ret = (rte_cryptodev_sym_session_free(dev_id, sa_sess->sess) == NULL); - ASSERT (ret); -#else - ret = rte_cryptodev_sym_session_clear(dev_id, sa_sess->sess); - ASSERT (!ret); +static u8 +auth_alg_index (const crypto_alg_t * alg) +{ + dpdk_crypto_main_t *dcm = &dpdk_crypto_main; + + return (alg - dcm->auth_algs); +} + +static crypto_alg_t * +cipher_cap_to_alg (const struct rte_cryptodev_capabilities *cap, u8 key_len) +{ + dpdk_crypto_main_t *dcm = &dpdk_crypto_main; + crypto_alg_t *alg; + + if (cap->op != RTE_CRYPTO_OP_TYPE_SYMMETRIC) + return NULL; - ret = rte_cryptodev_sym_session_free(sa_sess->sess); - ASSERT (!ret); + /* *INDENT-OFF* */ + vec_foreach (alg, dcm->cipher_algs) + { + if ((cap->sym.xform_type == RTE_CRYPTO_SYM_XFORM_CIPHER) && + (alg->type == RTE_CRYPTO_SYM_XFORM_CIPHER) && + (cap->sym.cipher.algo == alg->alg) && + (alg->key_len == key_len)) + return alg; +#if ! DPDK_NO_AEAD + if ((cap->sym.xform_type == RTE_CRYPTO_SYM_XFORM_AEAD) && + (alg->type == RTE_CRYPTO_SYM_XFORM_AEAD) && + (cap->sym.aead.algo == alg->alg) && + (alg->key_len == key_len)) + return alg; #endif - memset(sa_sess, 0, sizeof(sa_sess[0])); - } - } } - /* *INDENT-OFF* */ + /* *INDENT-ON* */ - return 0; + return NULL; } -static void -update_qp_data (crypto_worker_main_t * cwm, - u8 cdev_id, u16 qp_id, u8 is_outbound, u16 * idx) +static crypto_alg_t * +auth_cap_to_alg (const struct rte_cryptodev_capabilities *cap, u8 trunc_size) { - crypto_qp_data_t *qpd; + dpdk_crypto_main_t *dcm = &dpdk_crypto_main; + crypto_alg_t *alg; + + if ((cap->op != RTE_CRYPTO_OP_TYPE_SYMMETRIC) || + (cap->sym.xform_type != RTE_CRYPTO_SYM_XFORM_AUTH)) + return NULL; /* *INDENT-OFF* */ - vec_foreach_index (*idx, cwm->qp_data) + vec_foreach (alg, dcm->auth_algs) { - qpd = vec_elt_at_index(cwm->qp_data, *idx); - - if (qpd->dev_id == cdev_id && qpd->qp_id == qp_id && - qpd->is_outbound == is_outbound) - return; + if ((cap->sym.auth.algo == alg->alg) && + (alg->trunc_size == trunc_size)) + return alg; } /* *INDENT-ON* */ - vec_add2_aligned (cwm->qp_data, qpd, 1, CLIB_CACHE_LINE_BYTES); + return NULL; +} + +#if ! DPDK_NO_AEAD +static void +crypto_set_aead_xform (struct rte_crypto_sym_xform *xform, + ipsec_sa_t * sa, u8 is_outbound) +{ + dpdk_crypto_main_t *dcm = &dpdk_crypto_main; + crypto_alg_t *c; + + c = vec_elt_at_index (dcm->cipher_algs, sa->crypto_alg); + + ASSERT (c->type == RTE_CRYPTO_SYM_XFORM_AEAD); + + xform->type = RTE_CRYPTO_SYM_XFORM_AEAD; + xform->aead.algo = c->alg; + xform->aead.key.data = sa->crypto_key; + xform->aead.key.length = c->key_len; + xform->aead.iv.offset = + crypto_op_get_priv_offset () + offsetof (dpdk_op_priv_t, cb); + xform->aead.iv.length = 12; + xform->aead.digest_length = c->trunc_size; + xform->aead.aad_length = sa->use_esn ? 12 : 8; + xform->next = NULL; - qpd->dev_id = cdev_id; - qpd->qp_id = qp_id; - qpd->is_outbound = is_outbound; + if (is_outbound) + xform->aead.op = RTE_CRYPTO_AEAD_OP_ENCRYPT; + else + xform->aead.op = RTE_CRYPTO_AEAD_OP_DECRYPT; } +#endif -/* - * return: - * 0: already exist - * 1: mapped - */ -static int -add_mapping (crypto_worker_main_t * cwm, - u8 cdev_id, u16 qp, u8 is_outbound, - const struct rte_cryptodev_capabilities *cipher_cap, - const struct rte_cryptodev_capabilities *auth_cap) +static void +crypto_set_cipher_xform (struct rte_crypto_sym_xform *xform, + ipsec_sa_t * sa, u8 is_outbound) { - u16 qp_index; - uword key = 0, data, *ret; - crypto_worker_qp_key_t *p_key = (crypto_worker_qp_key_t *) & key; + dpdk_crypto_main_t *dcm = &dpdk_crypto_main; + crypto_alg_t *c; - p_key->cipher_algo = (u8) cipher_cap->sym.cipher.algo; - p_key->auth_algo = (u8) auth_cap->sym.auth.algo; - p_key->is_outbound = is_outbound; + c = vec_elt_at_index (dcm->cipher_algs, sa->crypto_alg); + + ASSERT (c->type == RTE_CRYPTO_SYM_XFORM_CIPHER); + + xform->type = RTE_CRYPTO_SYM_XFORM_CIPHER; + xform->cipher.algo = c->alg; + xform->cipher.key.data = sa->crypto_key; + xform->cipher.key.length = c->key_len; #if ! DPDK_NO_AEAD - p_key->is_aead = cipher_cap->sym.xform_type == RTE_CRYPTO_SYM_XFORM_AEAD; + xform->cipher.iv.offset = + crypto_op_get_priv_offset () + offsetof (dpdk_op_priv_t, cb); + xform->cipher.iv.length = c->iv_len; #endif + xform->next = NULL; - ret = hash_get (cwm->algo_qp_map, key); - if (ret) - return 0; + if (is_outbound) + xform->cipher.op = RTE_CRYPTO_CIPHER_OP_ENCRYPT; + else + xform->cipher.op = RTE_CRYPTO_CIPHER_OP_DECRYPT; +} - update_qp_data (cwm, cdev_id, qp, is_outbound, &qp_index); +static void +crypto_set_auth_xform (struct rte_crypto_sym_xform *xform, + ipsec_sa_t * sa, u8 is_outbound) +{ + dpdk_crypto_main_t *dcm = &dpdk_crypto_main; + crypto_alg_t *a; + + a = vec_elt_at_index (dcm->auth_algs, sa->integ_alg); - data = (uword) qp_index; - hash_set (cwm->algo_qp_map, key, data); + ASSERT (a->type == RTE_CRYPTO_SYM_XFORM_AUTH); + + xform->type = RTE_CRYPTO_SYM_XFORM_AUTH; + xform->auth.algo = a->alg; + xform->auth.key.data = sa->integ_key; + xform->auth.key.length = a->key_len; + xform->auth.digest_length = a->trunc_size; +#if DPDK_NO_AEAD + if (sa->crypto_alg == IPSEC_CRYPTO_ALG_AES_GCM_128 | + sa->crypto_alg == IPSEC_CRYPTO_ALG_AES_GCM_192 | + sa->crypto_alg == IPSEC_CRYPTO_ALG_AES_GCM_256) + xform->auth.algo = RTE_CRYPTO_AUTH_AES_GCM; + xform->auth.add_auth_data_length = sa->use_esn ? 12 : 8; +#else +#if 0 + xform->auth.iv.offset = + sizeof (struct rte_crypto_op) + sizeof (struct rte_crypto_sym_op) + + offsetof (dpdk_op_priv_t, cb); + xform->auth.iv.length = a->iv_len; +#endif +#endif + xform->next = NULL; - return 1; + if (is_outbound) + xform->auth.op = RTE_CRYPTO_AUTH_OP_GENERATE; + else + xform->auth.op = RTE_CRYPTO_AUTH_OP_VERIFY; } -/* - * return: - * 0: already exist - * 1: mapped - */ -static int -add_cdev_mapping (crypto_worker_main_t * cwm, - struct rte_cryptodev_info *dev_info, u8 cdev_id, - u16 qp, u8 is_outbound) +clib_error_t * +create_sym_session (struct rte_cryptodev_sym_session **session, + u32 sa_idx, + crypto_resource_t * res, + crypto_worker_main_t * cwm, u8 is_outbound) { - const struct rte_cryptodev_capabilities *i, *j; - u32 mapped = 0; + dpdk_crypto_main_t *dcm = &dpdk_crypto_main; + ipsec_main_t *im = &ipsec_main; + crypto_data_t *data; + ipsec_sa_t *sa; + struct rte_crypto_sym_xform cipher_xform = { 0 }; + struct rte_crypto_sym_xform auth_xform = { 0 }; + struct rte_crypto_sym_xform *xfs; + crypto_session_key_t key = { 0 }; + + key.drv_id = res->drv_id; + key.sa_idx = sa_idx; + + sa = pool_elt_at_index (im->sad, sa_idx); - for (i = dev_info->capabilities; i->op != RTE_CRYPTO_OP_TYPE_UNDEFINED; i++) - { #if ! DPDK_NO_AEAD - if (i->sym.xform_type == RTE_CRYPTO_SYM_XFORM_AEAD) + if ((sa->crypto_alg == IPSEC_CRYPTO_ALG_AES_GCM_128) | + (sa->crypto_alg == IPSEC_CRYPTO_ALG_AES_GCM_192) | + (sa->crypto_alg == IPSEC_CRYPTO_ALG_AES_GCM_256)) + { + crypto_set_aead_xform (&cipher_xform, sa, is_outbound); + xfs = &cipher_xform; + } + else +#endif /* ! DPDK_NO_AEAD */ + { + crypto_set_cipher_xform (&cipher_xform, sa, is_outbound); + crypto_set_auth_xform (&auth_xform, sa, is_outbound); + + if (is_outbound) + { + cipher_xform.next = &auth_xform; + xfs = &cipher_xform; + } + else { - struct rte_cryptodev_capabilities none = { 0 }; + auth_xform.next = &cipher_xform; + xfs = &auth_xform; + } + } - if (check_algo_is_supported (i, NULL) != 0) - continue; + data = vec_elt_at_index (dcm->data, res->numa); + +#if DPDK_NO_AEAD + /* + * DPDK_VER <= 1705: + * Each worker/thread has its own session per device driver + */ + session[0] = rte_cryptodev_sym_session_create (res->dev_id, xfs); + if (!session[0]) + { + data->session_drv_failed[res->drv_id] += 1; + return clib_error_return (0, "failed to create session for dev %u", + res->dev_id); + } +#else + /* + * DPDK_VER >= 1708: + * Multiple worker/threads share the session for an SA + * Single session per SA, initialized for each device driver + */ + session[0] = (void *) hash_get (data->session_by_sa_index, sa_idx); + + if (!session[0]) + { + session[0] = rte_cryptodev_sym_session_create (data->session_h); + if (!session[0]) + { + data->session_h_failed += 1; + return clib_error_return (0, "failed to create session header"); + } + hash_set (data->session_by_sa_index, sa_idx, session[0]); + } - none.sym.auth.algo = RTE_CRYPTO_AUTH_NULL; + struct rte_mempool **mp; + mp = vec_elt_at_index (data->session_drv, res->drv_id); + ASSERT (mp[0] != NULL); - mapped |= add_mapping (cwm, cdev_id, qp, is_outbound, i, &none); - continue; + i32 ret = + rte_cryptodev_sym_session_init (res->dev_id, session[0], xfs, mp[0]); + if (ret) + { + data->session_drv_failed[res->drv_id] += 1; + return clib_error_return (0, "failed to init session for drv %u", + res->drv_id); + } +#endif /* DPDK_NO_AEAD */ + + hash_set (cwm->session_by_drv_id_and_sa_index, key.val, session[0]); + + return 0; +} + +static void __attribute__ ((unused)) clear_and_free_obj (void *obj) +{ + struct rte_mempool *mp = rte_mempool_from_obj (obj); + + memset (obj, 0, mp->elt_size); + + rte_mempool_put (mp, obj); +} + +#if ! DPDK_NO_AEAD +/* This is from rte_cryptodev_pmd.h */ +static inline void * +get_session_private_data (const struct rte_cryptodev_sym_session *sess, + uint8_t driver_id) +{ + return sess->sess_private_data[driver_id]; +} + +/* This is from rte_cryptodev_pmd.h */ +static inline void +set_session_private_data (struct rte_cryptodev_sym_session *sess, + uint8_t driver_id, void *private_data) +{ + sess->sess_private_data[driver_id] = private_data; +} +#endif + +static clib_error_t * +add_del_sa_session (u32 sa_index, u8 is_add) +{ + ipsec_main_t *im = &ipsec_main; + dpdk_crypto_main_t *dcm = &dpdk_crypto_main; + crypto_worker_main_t *cwm; + struct rte_cryptodev_sym_session *s; + crypto_session_key_t key = { 0 }; + uword *val; + u32 drv_id; + i32 ret; + + key.sa_idx = sa_index; + + if (is_add) + { +#if 1 + ipsec_sa_t *sa = pool_elt_at_index (im->sad, sa_index); + u32 seed; + switch (sa->crypto_alg) + { + case IPSEC_CRYPTO_ALG_AES_GCM_128: + case IPSEC_CRYPTO_ALG_AES_GCM_192: + case IPSEC_CRYPTO_ALG_AES_GCM_256: + clib_memcpy (&sa->salt, &sa->crypto_key[sa->crypto_key_len - 4], 4); + break; + default: + seed = (u32) clib_cpu_time_now (); + sa->salt = random_u32 (&seed); } #endif - if (i->sym.xform_type != RTE_CRYPTO_SYM_XFORM_CIPHER) - continue; + return 0; + } - if (check_algo_is_supported (i, NULL) != 0) - continue; + /* XXX Wait N cycles to be sure session is not in use OR + * keep refcnt at SA level per worker/thread ? */ + unix_sleep (0.2); - for (j = dev_info->capabilities; j->op != RTE_CRYPTO_OP_TYPE_UNDEFINED; - j++) + /* *INDENT-OFF* */ + vec_foreach (cwm, dcm->workers_main) + { + for (drv_id = 0; drv_id < dcm->max_drv_id; drv_id++) { - if (j->sym.xform_type != RTE_CRYPTO_SYM_XFORM_AUTH) - continue; + key.drv_id = drv_id; + val = hash_get (cwm->session_by_drv_id_and_sa_index, key.val); + s = (struct rte_cryptodev_sym_session *) val; - if (check_algo_is_supported (j, NULL) != 0) + if (!s) continue; - mapped |= add_mapping (cwm, cdev_id, qp, is_outbound, i, j); +#if DPDK_NO_AEAD + ret = (rte_cryptodev_sym_session_free (s->dev_id, s) == NULL); + ASSERT (ret); +#endif + hash_unset (cwm->session_by_drv_id_and_sa_index, key.val); } } + /* *INDENT-ON* */ - return mapped; -} +#if ! DPDK_NO_AEAD + crypto_data_t *data; + /* *INDENT-OFF* */ + vec_foreach (data, dcm->data) + { + val = hash_get (data->session_by_sa_index, sa_index); + s = (struct rte_cryptodev_sym_session *) val; -static int -check_cryptodev_queues () -{ - u32 n_qs = 0; - u8 cdev_id; - u32 n_req_qs = 2; + if (!s) + continue; - if (vlib_num_workers () > 0) - n_req_qs = vlib_num_workers () * 2; + hash_unset (data->session_by_sa_index, sa_index); - for (cdev_id = 0; cdev_id < rte_cryptodev_count (); cdev_id++) - { - struct rte_cryptodev_info cdev_info; + void *drv_session; + vec_foreach_index (drv_id, dcm->drv) + { + drv_session = get_session_private_data (s, drv_id); + if (!drv_session) + continue; - rte_cryptodev_info_get (cdev_id, &cdev_info); + /* + * Custom clear to avoid finding a dev_id for drv_id: + * ret = rte_cryptodev_sym_session_clear (dev_id, drv_session); + * ASSERT (!ret); + */ + clear_and_free_obj (drv_session); - if (! - (cdev_info.feature_flags & RTE_CRYPTODEV_FF_SYM_OPERATION_CHAINING)) - continue; + set_session_private_data (s, drv_id, NULL); + } - n_qs += cdev_info.max_nb_queue_pairs; + ret = rte_cryptodev_sym_session_free(s); + ASSERT (!ret); } + /* *INDENT-ON* */ +#endif - if (n_qs >= n_req_qs) - return 0; - else - return -1; + return 0; } static clib_error_t * dpdk_ipsec_check_support (ipsec_sa_t * sa) { - if (sa->crypto_alg == IPSEC_CRYPTO_ALG_AES_GCM_128) + dpdk_crypto_main_t *dcm = &dpdk_crypto_main; + + if (sa->integ_alg == IPSEC_INTEG_ALG_NONE) + switch (sa->crypto_alg) + { + case IPSEC_CRYPTO_ALG_AES_GCM_128: + case IPSEC_CRYPTO_ALG_AES_GCM_192: + case IPSEC_CRYPTO_ALG_AES_GCM_256: + break; + default: + return clib_error_return (0, "unsupported integ-alg %U crypto-alg %U", + format_ipsec_integ_alg, sa->integ_alg, + format_ipsec_crypto_alg, sa->crypto_alg); + } + + /* XXX do we need the NONE check? */ + if (sa->crypto_alg != IPSEC_CRYPTO_ALG_NONE && + dcm->cipher_algs[sa->crypto_alg].disabled) + return clib_error_return (0, "disabled crypto-alg %U", + format_ipsec_crypto_alg, sa->crypto_alg); + + /* XXX do we need the NONE check? */ + if (sa->integ_alg != IPSEC_INTEG_ALG_NONE && + dcm->auth_algs[sa->integ_alg].disabled) + return clib_error_return (0, "disabled integ-alg %U", + format_ipsec_integ_alg, sa->integ_alg); + return NULL; +} + +static void +crypto_parse_capabilities (crypto_dev_t * dev, + const struct rte_cryptodev_capabilities *cap, + u32 n_mains) +{ + dpdk_crypto_main_t *dcm = &dpdk_crypto_main; + crypto_alg_t *alg; + u8 len, inc; + + for (; cap->op != RTE_CRYPTO_OP_TYPE_UNDEFINED; cap++) { - if (sa->integ_alg != IPSEC_INTEG_ALG_NONE) - return clib_error_return (0, "unsupported integ-alg %U with " - "crypto-alg aes-gcm-128", - format_ipsec_integ_alg, sa->integ_alg); -#if DPDK_NO_AEAD - sa->integ_alg = IPSEC_INTEG_ALG_AES_GCM_128; + /* A single capability maps to multiple cipher/auth algorithms */ + switch (cap->sym.xform_type) + { +#if ! DPDK_NO_AEAD + case RTE_CRYPTO_SYM_XFORM_AEAD: #endif + case RTE_CRYPTO_SYM_XFORM_CIPHER: + inc = cap->sym.cipher.key_size.increment; + inc = inc ? inc : 1; + for (len = cap->sym.cipher.key_size.min; + len <= cap->sym.cipher.key_size.max; len += inc) + { + alg = cipher_cap_to_alg (cap, len); + if (!alg) + continue; + dev->cipher_support[cipher_alg_index (alg)] = 1; + alg->resources += vec_len (dev->free_resources); + /* At least enough resources to support one algo */ + dcm->enabled |= (alg->resources >= n_mains); + } + break; + case RTE_CRYPTO_SYM_XFORM_AUTH: + inc = cap->sym.auth.digest_size.increment; + inc = inc ? inc : 1; + for (len = cap->sym.auth.digest_size.min; + len <= cap->sym.auth.digest_size.max; len += inc) + { + alg = auth_cap_to_alg (cap, len); + if (!alg) + continue; + dev->auth_support[auth_alg_index (alg)] = 1; + alg->resources += vec_len (dev->free_resources); + /* At least enough resources to support one algo */ + dcm->enabled |= (alg->resources >= n_mains); + } + break; + default: + ; + } } +} + +#define DPDK_CRYPTO_N_QUEUE_DESC 2048 +#define DPDK_CRYPTO_NB_SESS_OBJS 20000 + +static clib_error_t * +crypto_dev_conf (u8 dev, u16 n_qp, u8 numa) +{ + struct rte_cryptodev_config dev_conf; + struct rte_cryptodev_qp_conf qp_conf; + i32 ret; + u16 qp; + i8 *error_str; + + dev_conf.socket_id = numa; + dev_conf.nb_queue_pairs = n_qp; +#if DPDK_NO_AEAD + dev_conf.session_mp.nb_objs = DPDK_CRYPTO_NB_SESS_OBJS; + dev_conf.session_mp.cache_size = 512; +#endif + + error_str = "failed to configure crypto device %u"; + ret = rte_cryptodev_configure (dev, &dev_conf); + if (ret < 0) + return clib_error_return (0, error_str, dev); + + error_str = "failed to setup crypto device %u queue pair %u"; + qp_conf.nb_descriptors = DPDK_CRYPTO_N_QUEUE_DESC; + for (qp = 0; qp < n_qp; qp++) + { #if DPDK_NO_AEAD - else if (sa->crypto_alg == IPSEC_CRYPTO_ALG_NONE || - sa->integ_alg == IPSEC_INTEG_ALG_NONE || - sa->integ_alg == IPSEC_INTEG_ALG_AES_GCM_128) + ret = rte_cryptodev_queue_pair_setup (dev, qp, &qp_conf, numa); #else - else if (sa->integ_alg == IPSEC_INTEG_ALG_NONE) + ret = rte_cryptodev_queue_pair_setup (dev, qp, &qp_conf, numa, NULL); #endif - return clib_error_return (0, - "unsupported integ-alg %U with crypto-alg %U", - format_ipsec_integ_alg, sa->integ_alg, - format_ipsec_crypto_alg, sa->crypto_alg); + if (ret < 0) + return clib_error_return (0, error_str, dev, qp); + } return 0; } -static uword -dpdk_ipsec_process (vlib_main_t * vm, vlib_node_runtime_t * rt, - vlib_frame_t * f) +static void +crypto_scan_devs (u32 n_mains) { - ipsec_main_t *im = &ipsec_main; dpdk_crypto_main_t *dcm = &dpdk_crypto_main; - vlib_thread_main_t *tm = vlib_get_thread_main (); - struct rte_cryptodev_config dev_conf; - struct rte_cryptodev_qp_conf qp_conf; - struct rte_cryptodev_info cdev_info; - struct rte_mempool *rmp; - i32 dev_id, ret; - u32 i, skip_master; -#if ! DPDK_NO_AEAD - u32 max_sess_size = 0, sess_size; - i8 socket_id; + struct rte_cryptodev *cryptodev; + struct rte_cryptodev_info info; + crypto_dev_t *dev; + crypto_resource_t *res; + clib_error_t *error; + u32 i; + u16 max_res_idx, res_idx, j; + u8 drv_id; + + vec_validate_init_empty (dcm->dev, rte_cryptodev_count () - 1, + (crypto_dev_t) EMPTY_STRUCT); + + for (i = 0; i < rte_cryptodev_count (); i++) + { + dev = vec_elt_at_index (dcm->dev, i); + + cryptodev = &rte_cryptodevs[i]; + rte_cryptodev_info_get (i, &info); + + dev->id = i; + dev->name = cryptodev->data->name; + dev->numa = rte_cryptodev_socket_id (i); + dev->features = info.feature_flags; + dev->max_qp = info.max_nb_queue_pairs; +#if DPDK_NO_AEAD + drv_id = cryptodev->dev_type; +#else + drv_id = info.driver_id; #endif + if (drv_id >= vec_len (dcm->drv)) + vec_validate_init_empty (dcm->drv, drv_id, + (crypto_drv_t) EMPTY_STRUCT); + vec_elt_at_index (dcm->drv, drv_id)->name = info.driver_name; + dev->drv_id = drv_id; + vec_add1 (vec_elt_at_index (dcm->drv, drv_id)->devs, i); + + if (!(info.feature_flags & RTE_CRYPTODEV_FF_SYM_OPERATION_CHAINING)) + continue; - if (check_cryptodev_queues () < 0) - { - clib_warning ("not enough Cryptodevs, default to OpenSSL IPsec"); - return 0; - } - dcm->enabled = 1; + if ((error = crypto_dev_conf (i, dev->max_qp, dev->numa))) + { + clib_error_report (error); + continue; + } - vec_alloc (dcm->workers_main, tm->n_vlib_mains); - _vec_len (dcm->workers_main) = tm->n_vlib_mains; + max_res_idx = (dev->max_qp / 2) - 1; - skip_master = vlib_num_workers () > 0; + vec_validate (dev->free_resources, max_res_idx); - fprintf (stdout, "DPDK Cryptodevs info:\n"); - fprintf (stdout, "dev_id\tn_qp\tnb_obj\tcache_size\n"); - /* HW cryptodevs have higher dev_id, use HW first */ - for (dev_id = rte_cryptodev_count () - 1; dev_id >= 0; dev_id--) - { - u16 max_nb_qp, qp = 0; + res_idx = vec_len (dcm->resource); + vec_validate_init_empty_aligned (dcm->resource, res_idx + max_res_idx, + (crypto_resource_t) EMPTY_STRUCT, + CLIB_CACHE_LINE_BYTES); - rte_cryptodev_info_get (dev_id, &cdev_info); + for (j = 0; j <= max_res_idx; j++, res_idx++) + { + vec_elt (dev->free_resources, max_res_idx - j) = res_idx; + res = &dcm->resource[res_idx]; + res->dev_id = i; + res->drv_id = drv_id; + res->qp_id = j * 2; + res->numa = dev->numa; + res->thread_idx = (u16) ~ 0; + } - if (! - (cdev_info.feature_flags & RTE_CRYPTODEV_FF_SYM_OPERATION_CHAINING)) - continue; + crypto_parse_capabilities (dev, info.capabilities, n_mains); + } +} - max_nb_qp = cdev_info.max_nb_queue_pairs; +void +crypto_auto_placement (void) +{ + dpdk_crypto_main_t *dcm = &dpdk_crypto_main; + crypto_resource_t *res; + crypto_worker_main_t *cwm; + crypto_dev_t *dev; + u32 thread_idx, skip_master; + u16 res_idx, *idx; + u8 used; + u16 i; - for (i = 0; i < tm->n_vlib_mains; i++) + skip_master = vlib_num_workers () > 0; + + /* *INDENT-OFF* */ + vec_foreach (dev, dcm->dev) + { + vec_foreach_index (thread_idx, dcm->workers_main) { - u8 is_outbound; - crypto_worker_main_t *cwm; - uword *map; + if (vec_len (dev->free_resources) == 0) + break; + + if (thread_idx < skip_master) + continue; + + /* Check thread is not already using the device */ + vec_foreach (idx, dev->used_resources) + if (dcm->resource[idx[0]].thread_idx == thread_idx) + continue; - if (skip_master) + cwm = vec_elt_at_index (dcm->workers_main, thread_idx); + + used = 0; + res_idx = vec_pop (dev->free_resources); + + /* Set device only for supported algos */ + for (i = 0; i < IPSEC_CRYPTO_N_ALG; i++) + if (dev->cipher_support[i] && + cwm->cipher_resource_idx[i] == (u16) ~0) + { + dcm->cipher_algs[i].disabled--; + cwm->cipher_resource_idx[i] = res_idx; + used = 1; + } + + for (i = 0; i < IPSEC_INTEG_N_ALG; i++) + if (dev->auth_support[i] && + cwm->auth_resource_idx[i] == (u16) ~0) + { + dcm->auth_algs[i].disabled--; + cwm->auth_resource_idx[i] = res_idx; + used = 1; + } + + if (!used) { - skip_master = 0; + vec_add1 (dev->free_resources, res_idx); continue; } - cwm = vec_elt_at_index (dcm->workers_main, i); - map = cwm->algo_qp_map; + vec_add1 (dev->used_resources, res_idx); - if (!map) - { - map = hash_create (0, sizeof (crypto_worker_qp_key_t)); - if (!map) - { - clib_warning ("unable to create hash table for worker %u", - vlib_mains[i]->thread_index); - goto error; - } - cwm->algo_qp_map = map; - } + res = vec_elt_at_index (dcm->resource, res_idx); + + ASSERT (res->thread_idx == (u16) ~0); + res->thread_idx = thread_idx; - for (is_outbound = 0; is_outbound < 2 && qp < max_nb_qp; - is_outbound++) - qp += add_cdev_mapping (cwm, &cdev_info, dev_id, qp, is_outbound); + /* Add device to vector of polling resources */ + vec_add1 (cwm->resource_idx, res_idx); } + } + /* *INDENT-ON* */ +} - if (qp == 0) - continue; +static void +crypto_op_init (struct rte_mempool *mempool, + void *_arg __attribute__ ((unused)), + void *_obj, unsigned i __attribute__ ((unused))) +{ + struct rte_crypto_op *op = _obj; - dev_conf.socket_id = rte_cryptodev_socket_id (dev_id); - dev_conf.nb_queue_pairs = cdev_info.max_nb_queue_pairs; #if DPDK_NO_AEAD - dev_conf.session_mp.nb_objs = DPDK_CRYPTO_NB_SESS_OBJS; - dev_conf.session_mp.cache_size = DPDK_CRYPTO_CACHE_SIZE; + op->sym = (struct rte_crypto_sym_op *) (op + 1); + op->sym->sess_type = RTE_CRYPTO_SYM_OP_WITH_SESSION; +#else + op->sess_type = RTE_CRYPTO_OP_WITH_SESSION; #endif - ret = rte_cryptodev_configure (dev_id, &dev_conf); - if (ret < 0) - { - clib_warning ("cryptodev %u config error", dev_id); - goto error; - } + op->type = RTE_CRYPTO_OP_TYPE_SYMMETRIC; + op->status = RTE_CRYPTO_OP_STATUS_NOT_PROCESSED; + op->phys_addr = rte_mem_virt2phy (_obj); + op->mempool = mempool; +} - qp_conf.nb_descriptors = DPDK_CRYPTO_N_QUEUE_DESC; - for (qp = 0; qp < dev_conf.nb_queue_pairs; qp++) - { +static clib_error_t * +crypto_create_crypto_op_pool (u8 numa) +{ + dpdk_crypto_main_t *dcm = &dpdk_crypto_main; + dpdk_config_main_t *conf = &dpdk_config_main; + crypto_data_t *data; + u8 *pool_name; + u32 pool_priv_size = sizeof (struct rte_crypto_op_pool_private); + struct rte_crypto_op_pool_private *priv; + clib_error_t *error = NULL; + + data = vec_elt_at_index (dcm->data, numa); + + if (data->crypto_op) + return NULL; + + pool_name = format (0, "crypto_pool_numa%u%c", numa, 0); + + data->crypto_op = + rte_mempool_create ((i8 *) pool_name, conf->num_mbufs, crypto_op_len (), + 512, pool_priv_size, NULL, NULL, crypto_op_init, NULL, + numa, 0); + + if (!data->crypto_op) + { + error = clib_error_return (0, "failed to allocate %s", pool_name); + goto done; + } + + priv = rte_mempool_get_priv (data->crypto_op); + + priv->priv_size = pool_priv_size; + priv->type = RTE_CRYPTO_OP_TYPE_SYMMETRIC; + +done: + vec_free (pool_name); + + return error; +} + +static clib_error_t * +crypto_create_session_h_pool (u8 numa) +{ #if DPDK_NO_AEAD - ret = rte_cryptodev_queue_pair_setup (dev_id, qp, &qp_conf, - dev_conf.socket_id); + return NULL; #else - ret = rte_cryptodev_queue_pair_setup (dev_id, qp, &qp_conf, - dev_conf.socket_id, NULL); + dpdk_crypto_main_t *dcm = &dpdk_crypto_main; + crypto_data_t *data; + u8 *pool_name; + u32 elt_size; + clib_error_t *error = NULL; + + data = vec_elt_at_index (dcm->data, numa); + + if (data->session_h) + return NULL; + + pool_name = format (0, "session_h_pool_numa%u%c", numa, 0); + elt_size = rte_cryptodev_get_header_session_size (); + + data->session_h = + rte_mempool_create ((i8 *) pool_name, DPDK_CRYPTO_NB_SESS_OBJS, elt_size, + 512, 0, NULL, NULL, NULL, NULL, numa, 0); + + if (!data->session_h) + error = clib_error_return (0, "failed to allocate %s", pool_name); + + vec_free (pool_name); + + return error; #endif - if (ret < 0) - { - clib_warning ("cryptodev %u qp %u setup error", dev_id, qp); - goto error; - } - } - vec_validate (dcm->cop_pools, dev_conf.socket_id); +} -#if ! DPDK_NO_AEAD - sess_size = rte_cryptodev_get_private_session_size (dev_id); - if (sess_size > max_sess_size) - max_sess_size = sess_size; +static clib_error_t * +crypto_create_session_drv_pool (crypto_dev_t * dev) +{ +#if DPDK_NO_AEAD + return NULL; +#else + dpdk_crypto_main_t *dcm = &dpdk_crypto_main; + crypto_data_t *data; + u8 *pool_name; + u32 elt_size; + clib_error_t *error = NULL; + u8 numa = dev->numa; + + data = vec_elt_at_index (dcm->data, numa); + + vec_validate (data->session_drv, dev->drv_id); + vec_validate (data->session_drv_failed, dev->drv_id); + + if (data->session_drv[dev->drv_id]) + return NULL; + + pool_name = format (0, "session_drv%u_pool_numa%u%c", dev->drv_id, numa, 0); + elt_size = rte_cryptodev_get_private_session_size (dev->id); + + data->session_drv[dev->drv_id] = + rte_mempool_create ((i8 *) pool_name, DPDK_CRYPTO_NB_SESS_OBJS, elt_size, + 512, 0, NULL, NULL, NULL, NULL, numa, 0); + + if (!data->session_drv[dev->drv_id]) + error = clib_error_return (0, "failed to allocate %s", pool_name); + + vec_free (pool_name); + + return error; #endif +} - if (!vec_elt (dcm->cop_pools, dev_conf.socket_id)) - { - u8 *pool_name = format (0, "crypto_op_pool_socket%u%c", - dev_conf.socket_id, 0); - - rmp = rte_crypto_op_pool_create ((char *) pool_name, - RTE_CRYPTO_OP_TYPE_SYMMETRIC, - DPDK_CRYPTO_NB_COPS * - (1 + vlib_num_workers ()), - DPDK_CRYPTO_CACHE_SIZE, - DPDK_CRYPTO_PRIV_SIZE, - dev_conf.socket_id); - - if (!rmp) - { - clib_warning ("failed to allocate %s", pool_name); - vec_free (pool_name); - goto error; - } - vec_free (pool_name); - vec_elt (dcm->cop_pools, dev_conf.socket_id) = rmp; - } +static clib_error_t * +crypto_create_pools (void) +{ + dpdk_crypto_main_t *dcm = &dpdk_crypto_main; + clib_error_t *error = NULL; + crypto_dev_t *dev; + + /* *INDENT-OFF* */ + vec_foreach (dev, dcm->dev) + { + vec_validate (dcm->data, dev->numa); + + error = crypto_create_crypto_op_pool (dev->numa); + if (error) + return error; + + error = crypto_create_session_h_pool (dev->numa); + if (error) + return error; - fprintf (stdout, "%u\t%u\t%u\t%u\n", dev_id, dev_conf.nb_queue_pairs, - DPDK_CRYPTO_NB_SESS_OBJS, DPDK_CRYPTO_CACHE_SIZE); + error = crypto_create_session_drv_pool (dev); + if (error) + return error; } + /* *INDENT-ON* */ + + return NULL; +} + +static void +crypto_disable (void) +{ + dpdk_crypto_main_t *dcm = &dpdk_crypto_main; + crypto_data_t *data; + u8 i; + + dcm->enabled = 0; -#if ! DPDK_NO_AEAD /* *INDENT-OFF* */ - vec_foreach_index (socket_id, dcm->cop_pools) + vec_foreach (data, dcm->data) { - u8 *pool_name; + rte_mempool_free (data->crypto_op); + rte_mempool_free (data->session_h); - if (!vec_elt (dcm->cop_pools, socket_id)) - continue; + vec_foreach_index (i, data->session_drv) + rte_mempool_free (data->session_drv[i]); - vec_validate (dcm->sess_h_pools, socket_id); - pool_name = format (0, "crypto_sess_h_socket%u%c", - socket_id, 0); - rmp = - rte_mempool_create((i8 *)pool_name, DPDK_CRYPTO_NB_SESS_OBJS, - rte_cryptodev_get_header_session_size (), - 512, 0, NULL, NULL, NULL, NULL, - socket_id, 0); - if (!rmp) - { - clib_warning ("failed to allocate %s", pool_name); - vec_free (pool_name); - goto error; - } - vec_free (pool_name); - vec_elt (dcm->sess_h_pools, socket_id) = rmp; - - vec_validate (dcm->sess_pools, socket_id); - pool_name = format (0, "crypto_sess_socket%u%c", - socket_id, 0); - rmp = - rte_mempool_create((i8 *)pool_name, DPDK_CRYPTO_NB_SESS_OBJS, - max_sess_size, 512, 0, NULL, NULL, NULL, NULL, - socket_id, 0); - if (!rmp) - { - clib_warning ("failed to allocate %s", pool_name); - vec_free (pool_name); - goto error; - } - vec_free (pool_name); - vec_elt (dcm->sess_pools, socket_id) = rmp; + vec_free (data->session_drv); } /* *INDENT-ON* */ -#endif - dpdk_esp_init (); + vec_free (dcm->data); - /* Add new next node and set as default */ + vec_free (dcm->workers_main); + vec_free (dcm->sa_session); + vec_free (dcm->dev); + vec_free (dcm->resource); + vec_free (dcm->cipher_algs); + vec_free (dcm->auth_algs); +} + +static uword +dpdk_ipsec_process (vlib_main_t * vm, vlib_node_runtime_t * rt, + vlib_frame_t * f) +{ + ipsec_main_t *im = &ipsec_main; + dpdk_crypto_main_t *dcm = &dpdk_crypto_main; + vlib_thread_main_t *tm = vlib_get_thread_main (); + crypto_worker_main_t *cwm; + clib_error_t *error = NULL; + u32 i, skip_master, n_mains; + + n_mains = tm->n_vlib_mains; + skip_master = vlib_num_workers () > 0; + + algos_init (n_mains - skip_master); + + crypto_scan_devs (n_mains - skip_master); + + if (!(dcm->enabled)) + { + clib_warning ("not enough DPDK crypto resources, default to OpenSSL"); + crypto_disable (); + return 0; + } + + vec_validate_init_empty (dcm->workers_main, n_mains - 1, + (crypto_worker_main_t) EMPTY_STRUCT); + + /* *INDENT-OFF* */ + vec_foreach (cwm, dcm->workers_main) + { + memset (cwm->cipher_resource_idx, ~0, + IPSEC_CRYPTO_N_ALG * sizeof(*cwm->cipher_resource_idx)); + memset (cwm->auth_resource_idx, ~0, + IPSEC_INTEG_N_ALG * sizeof(*cwm->auth_resource_idx)); + } + /* *INDENT-ON* */ + + crypto_auto_placement (); + + error = crypto_create_pools (); + if (error) + { + clib_error_report (error); + crypto_disable (); + return 0; + } + + /* Add new next node and set it as default */ vlib_node_t *node, *next_node; next_node = vlib_get_node_by_name (vm, (u8 *) "dpdk-esp-encrypt"); @@ -459,33 +1091,12 @@ dpdk_ipsec_process (vlib_main_t * vm, vlib_node_runtime_t * rt, vlib_node_add_next (vm, node->index, next_node->index); im->cb.check_support_cb = dpdk_ipsec_check_support; - im->cb.add_del_sa_sess_cb = add_del_sa_sess; - - for (i = skip_master; i < tm->n_vlib_mains; i++) - vlib_node_set_state (vlib_mains[i], dpdk_crypto_input_node.index, - VLIB_NODE_STATE_POLLING); - - /* TODO cryptodev counters */ - - return 0; - -error: - ; - crypto_worker_main_t *cwm; - struct rte_mempool **mp; - /* *INDENT-OFF* */ - vec_foreach (cwm, dcm->workers_main) - hash_free (cwm->algo_qp_map); - - vec_foreach (mp, dcm->cop_pools) - { - if (mp) - rte_mempool_free (mp[0]); - } - /* *INDENT-ON* */ - vec_free (dcm->workers_main); - vec_free (dcm->cop_pools); + im->cb.add_del_sa_sess_cb = add_del_sa_session; + node = vlib_get_node_by_name (vm, (u8 *) "dpdk-crypto-input"); + ASSERT (node); + for (i = skip_master; i < n_mains; i++) + vlib_node_set_state (vlib_mains[i], node->index, VLIB_NODE_STATE_POLLING); return 0; } |