aboutsummaryrefslogtreecommitdiffstats
path: root/src/plugins/dpdk/ipsec/ipsec.h
blob: dc11dfc2c14c0e493b7de42c435ae5c41f9ce804 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
16
/* SPDX-License-Identifier: Apache-2.0
 * Copyright(c) 2022 Cisco Systems, Inc.
 */

#include <vnet/fib/ip6_fib.h>
#include <vnet/dpo/load_balance.h>
#include <vnet/l2/feat_bitmap.h>
#include <vnet/fib/fib_table.h>
#include <vnet/srv6/sr.h>
#include <vnet/srv6/sr_pt.h>

/**
 * @brief PT node trace
 */
typedef struct
{
  u32 iface;
  u16 id;
  u8 load;
  timestamp_64_t t64;
  u8 tts_template;
  u8 tts;
  u8 behavior;
} pt_trace_t;

static u8 *
format_pt_trace (u8 *s, va_list *args)
{
  CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
  CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
  pt_trace_t *t = va_arg (*args, pt_trace_t *);
  switch (t->behavior)
    {
    case PT_BEHAVIOR_MID:
      s = format (
	s,
	"Behavior Midpoint, outgoing interface %U, outgoing interface id %u, "
	"outgoing interface load %u, t64_sec %u, t64_nsec %u, tts_template "
	"%u, tts %u",
	format_vnet_sw_if_index_name, vnet_get_main (), t->iface, t->id,
	t->load, clib_host_to_net_u32 (t->t64.sec),
	clib_host_to_net_u32 (t->t64.nsec), t->tts_template, t->tts);
      break;
    default:
      break;
    }
  return s;
}

static_always_inline void
pt_midpoint_processing (vlib_main_t *vm, vlib_node_runtime_t *node,
			vlib_buffer_t *b0, ip6_header_t *ip0,
			sr_pt_iface_t *ls, timestamp_64_t t64)
{
  ip6_hop_by_hop_header_t *hbh;
  ip6_hop_by_hop_option_t *hbh_opt;
  ip6_hop_by_hop_option_pt_t *hbh_opt_pt;

  if (ip0->protocol == IP_PROTOCOL_IP6_HOP_BY_HOP_OPTIONS)
    {
      hbh = (void *) (ip0 + 1);
      hbh_opt = (void *) (hbh + 1);
      if (hbh_opt
/*
 * Copyright (c) 2017 Intel and/or its affiliates.
 * Licensed under the Apache License, Version 2.0 (the "License");
 * you may not use this file except in compliance with the License.
 * You may obtain a copy of the License at:
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
#ifndef __DPDK_IPSEC_H__
#define __DPDK_IPSEC_H__

#include <vnet/vnet.h>
#include <vppinfra/cache.h>
#include <vnet/ipsec/ipsec.h>

#undef always_inline
#include <rte_config.h>
#include <rte_crypto.h>
#include <rte_cryptodev.h>

#if CLIB_DEBUG > 0
#define always_inline static inline
#else
#define always_inline static inline __attribute__ ((__always_inline__))
#endif

#define foreach_dpdk_crypto_input_next		\
  _(DROP, "error-drop")				\
  _(IP4_LOOKUP, "ip4-lookup")                   \
  _(IP6_LOOKUP, "ip6-lookup")                   \
  _(INTERFACE_OUTPUT, "interface-output")	\
  _(DECRYPT_POST, "dpdk-esp-decrypt-post")

typedef enum
{
#define _(f,s) DPDK_CRYPTO_INPUT_NEXT_##f,
  foreach_dpdk_crypto_input_next
#undef _
    DPDK_CRYPTO_INPUT_N_NEXT,
} dpdk_crypto_input_next_t;

#define MAX_QP_PER_LCORE 16

typedef struct
{
  u32 salt;
  u32 iv[2];
  u32 cnt;
} dpdk_gcm_cnt_blk;

typedef struct
{
  u32 next;
  dpdk_gcm_cnt_blk cb __attribute__ ((aligned (16)));
  u8 aad[16];
  u8 icv[32];
} dpdk_op_priv_t;

typedef struct
{
  u16 *resource_idx;
  struct rte_crypto_op **ops;
  u16 cipher_resource_idx[IPSEC_CRYPTO_N_ALG];
  u16 auth_resource_idx[IPSEC_INTEG_N_ALG];
    CLIB_CACHE_LINE_ALIGN_MARK (pad);
} crypto_worker_main_t __attribute__ ((aligned (CLIB_CACHE_LINE_BYTES)));

typedef struct
{
  char *name;
  enum rte_crypto_sym_xform_type type;
  u32 alg;
  u8 key_len;
  u8 iv_len;
  u8 trunc_size;
  u8 boundary;
  u8 disabled;
  u8 resources;
} crypto_alg_t __attribute__ ((aligned (8)));

typedef struct
{
  u16 *free_resources;
  u16 *used_resources;
  u8 cipher_support[IPSEC_CRYPTO_N_ALG];
  u8 auth_support[IPSEC_INTEG_N_ALG];
  u8 drv_id;
  u8 numa;
  u16 id;
  const char *name;
  u32 max_qp;
  u64 features;
} crypto_dev_t;

typedef struct
{
  const char *name;
  u16 *devs;
} crypto_drv_t;

typedef struct
{
  u16 thread_idx;
  u8 remove;
  u8 drv_id;
  u8 dev_id;
  u8 numa;
  u16 qp_id;
  u16 inflights[2];
  u16 n_ops;
  u16 __unused;
  struct rte_crypto_op *ops[VLIB_FRAME_SIZE];
  u32 bi[VLIB_FRAME_SIZE];
} crypto_resource_t __attribute__ ((aligned (CLIB_CACHE_LINE_BYTES)));

typedef struct
{
  u64 ts;
  struct rte_cryptodev_sym_session *session;
} crypto_session_disposal_t;

typedef struct
{
  struct rte_cryptodev_sym_session *session;
  u64 dev_mask;
} crypto_session_by_drv_t;

typedef struct
{
  /* Required for vec_validate_aligned */
  CLIB_CACHE_LINE_ALIGN_MARK (cacheline0);
  struct rte_mempool *crypto_op;
  struct rte_mempool *session_h;
  struct rte_mempool **session_drv;
  crypto_session_disposal_t *session_disposal;
  uword *session_by_sa_index;
  u64 crypto_op_get_failed;
  u64 session_h_failed;
  u64 *session_drv_failed;
  crypto_session_by_drv_t *session_by_drv_id_and_sa_index;
  clib_spinlock_t lockp;
} crypto_data_t;

typedef struct
{
  crypto_worker_main_t *workers_main;
  crypto_dev_t *dev;
  crypto_resource_t *resource;
  crypto_alg_t *cipher_algs;
  crypto_alg_t *auth_algs;
  crypto_data_t *data;
  crypto_drv_t *drv;
  u64 session_timeout;		/* nsec */
  u8 enabled;
} dpdk_crypto_main_t;

extern dpdk_crypto_main_t dpdk_crypto_main;

static const u8 pad_data[] =
  { 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 0 };

void crypto_auto_placement (void);

clib_error_t *create_sym_session (struct rte_cryptodev_sym_session **session,
				  u32 sa_idx, crypto_resource_t * res,
				  crypto_worker_main_t * cwm, u8 is_outbound);

static_always_inline u32
crypto_op_len (void)
{
  const u32 align = 4;
  u32 op_size =
    sizeof (struct rte_crypto_op) + sizeof (struct rte_crypto_sym_op);

  return ((op_size + align - 1) & ~(align - 1)) + sizeof (dpdk_op_priv_t);
}

static_always_inline u32
crypto_op_get_priv_offset (
="o">- 1) & ~(align - 1); return offset; } static_always_inline dpdk_op_priv_t * crypto_op_get_priv (struct rte_crypto_op * op) { return (dpdk_op_priv_t *) (((u8 *) op) + crypto_op_get_priv_offset ()); } static_always_inline void add_session_by_drv_and_sa_idx (struct rte_cryptodev_sym_session *session, crypto_data_t * data, u32 drv_id, u32 sa_idx) { crypto_session_by_drv_t *sbd; vec_validate_aligned (data->session_by_drv_id_and_sa_index, sa_idx, CLIB_CACHE_LINE_BYTES); sbd = vec_elt_at_index (data->session_by_drv_id_and_sa_index, sa_idx); sbd->dev_mask |= 1L << drv_id; sbd->session = session; } static_always_inline struct rte_cryptodev_sym_session * get_session_by_drv_and_sa_idx (crypto_data_t * data, u32 drv_id, u32 sa_idx) { crypto_session_by_drv_t *sess_by_sa; if (_vec_len (data->session_by_drv_id_and_sa_index) <= sa_idx) return NULL; sess_by_sa = vec_elt_at_index (data->session_by_drv_id_and_sa_index, sa_idx); return (sess_by_sa->dev_mask & (1L << drv_id)) ? sess_by_sa->session : NULL; } static_always_inline clib_error_t * crypto_get_session (struct rte_cryptodev_sym_session ** session, u32 sa_idx, crypto_resource_t * res, crypto_worker_main_t * cwm, u8 is_outbound) { dpdk_crypto_main_t *dcm = &dpdk_crypto_main; crypto_data_t *data; struct rte_cryptodev_sym_session *sess; data = vec_elt_at_index (dcm->data, res->numa); sess = get_session_by_drv_and_sa_idx (data, res->drv_id, sa_idx); if (PREDICT_FALSE (!sess)) return create_sym_session (session, sa_idx, res, cwm, is_outbound); session[0] = sess; return NULL; } static_always_inline u16 get_resource (crypto_worker_main_t * cwm, ipsec_sa_t * sa) { u16 cipher_res = cwm->cipher_resource_idx[sa->crypto_alg]; u16 auth_res = cwm->auth_resource_idx[sa->integ_alg]; u8 is_aead; /* Not allowed to setup SA with no-aead-cipher/NULL or NULL/NULL */ is_aead = ((sa->crypto_alg == IPSEC_CRYPTO_ALG_AES_GCM_128) || (sa->crypto_alg == IPSEC_CRYPTO_ALG_AES_GCM_192) || (sa->crypto_alg == IPSEC_CRYPTO_ALG_AES_GCM_256)); if (sa->crypto_alg == IPSEC_CRYPTO_ALG_NONE) return auth_res; if (cipher_res == auth_res) return cipher_res; if (is_aead) return cipher_res; return (u16) ~ 0; } static_always_inline i32 crypto_alloc_ops (u8 numa, struct rte_crypto_op ** ops, u32 n) { dpdk_crypto_main_t *dcm = &dpdk_crypto_main; crypto_data_t *data = vec_elt_at_index (dcm->data, numa); i32 ret; ret = rte_mempool_get_bulk (data->crypto_op, (void **) ops, n); /* *INDENT-OFF* */ data->crypto_op_get_failed += ! !ret; /* *INDENT-ON* */ return ret; } static_always_inline void crypto_free_ops (u8 numa, struct rte_crypto_op **ops, u32 n) { dpdk_crypto_main_t *dcm = &dpdk_crypto_main; crypto_data_t *data = vec_elt_at_index (dcm->data, numa); if (!n) return; rte_mempool_put_bulk (data->crypto_op, (void **) ops, n); } static_always_inline void crypto_enqueue_ops (vlib_main_t * vm, crypto_worker_main_t * cwm, u8 outbound, u32 node_index, u32 error, u8 numa) { dpdk_crypto_main_t *dcm = &dpdk_crypto_main; crypto_resource_t *res; u16 *res_idx; /* *INDENT-OFF* */ vec_foreach (res_idx, cwm->resource_idx) { u16 enq; res = vec_elt_at_index (dcm->resource, res_idx[0]); if (!res->n_ops) continue; enq = rte_cryptodev_enqueue_burst (res->dev_id, res->qp_id + outbound, res->ops, res->n_ops); res->inflights[outbound] += enq; if (PREDICT_FALSE (enq < res->n_ops)) { crypto_free_ops (numa, &res->ops[enq], res->n_ops - enq); vlib_buffer_free (vm, &res->bi[enq], res->n_ops - enq); vlib_node_increment_counter (vm, node_index, error, res->n_ops - enq); } res->n_ops = 0; } /* *INDENT-ON* */ } static_always_inline void crypto_set_icb (dpdk_gcm_cnt_blk * icb, u32 salt, u32 seq, u32 seq_hi) { icb->salt = salt; icb->iv[0] = seq; icb->iv[1] = seq_hi; } static_always_inline void crypto_op_setup (u8 is_aead, struct rte_mbuf *mb0, struct rte_crypto_op *op, void *session, u32 cipher_off, u32 cipher_len, u32 auth_off, u32 auth_len, u8 * aad, u8 * digest, u64 digest_paddr) { struct rte_crypto_sym_op *sym_op; sym_op = (struct rte_crypto_sym_op *) (op + 1); sym_op->m_src = mb0; sym_op->session = session; if (is_aead) { sym_op->aead.data.offset = cipher_off; sym_op->aead.data.length = cipher_len; sym_op->aead.aad.data = aad; sym_op->aead.aad.phys_addr = op->phys_addr + (uintptr_t) aad - (uintptr_t) op; sym_op->aead.digest.data = digest; sym_op->aead.digest.phys_addr = digest_paddr; } else { sym_op->cipher.data.offset = cipher_off; sym_op->cipher.data.length = cipher_len; sym_op->auth.data.offset = auth_off; sym_op->auth.data.length = auth_len; sym_op->auth.digest.data = digest; sym_op->auth.digest.phys_addr = digest_paddr; } } #endif /* __DPDK_IPSEC_H__ */ /* * fd.io coding-style-patch-verification: ON * * Local Variables: * eval: (c-set-style "gnu") * End: */