diff options
author | Nathan Skrzypczak <nathan.skrzypczak@gmail.com> | 2019-05-03 14:20:27 +0200 |
---|---|---|
committer | Dave Wallace <dwallacelf@gmail.com> | 2019-05-06 13:08:16 +0000 |
commit | 3a34b1d9366551bbff7eabff3ab48ec8189249e2 (patch) | |
tree | 3452571d43f045df0ebcb2ed713df75c26d39c93 /src | |
parent | 7107d7ef036b48cf910cb90a03978b9371b5c3ab (diff) |
Add QUIC multistream support
Change-Id: I9dc746b8c62c3e7ee2f65e34a1a1dca243ed4bd9
Signed-off-by: Nathan Skrzypczak <nathan.skrzypczak@gmail.com>
Diffstat (limited to 'src')
-rw-r--r-- | src/plugins/quic/quic.c | 950 | ||||
-rw-r--r-- | src/plugins/quic/quic.h | 36 | ||||
-rw-r--r-- | src/tests/vnet/session/quic_echo.c | 242 |
3 files changed, 890 insertions, 338 deletions
diff --git a/src/plugins/quic/quic.c b/src/plugins/quic/quic.c index d00c0e762ad..b77a1d37d8e 100644 --- a/src/plugins/quic/quic.c +++ b/src/plugins/quic/quic.c @@ -33,14 +33,32 @@ static quic_main_t quic_main; static void quic_update_timer (quic_ctx_t * ctx); -static int64_t quic_get_time (quicly_now_cb * self); static void quic_connection_closed (u32 conn_index); static void quic_disconnect (u32 ctx_index, u32 thread_index); static int quic_connect_new_stream (session_endpoint_cfg_t * sep); static int quic_connect_new_connection (session_endpoint_cfg_t * sep); +static int64_t quic_get_time (quicly_now_cb * self); +static quicly_now_cb quicly_vpp_now_cb = { quic_get_time }; + +static void quic_transfer_connection (u32 ctx_index, u32 dest_thread); + +#define QUIC_TIMER_HANDLE_INVALID ((u32) ~0) +#define QUIC_SESSION_INVALID ((u32) ~0 - 1) +#define QUIC_MAX_PACKET_SIZE 1280 + #define QUIC_INT_MAX 0x3FFFFFFFFFFFFFFF +/* Taken from quicly.c */ +#define QUICLY_QUIC_BIT 0x40 + +#define QUICLY_PACKET_TYPE_INITIAL (QUICLY_LONG_HEADER_BIT | QUICLY_QUIC_BIT | 0) +#define QUICLY_PACKET_TYPE_0RTT (QUICLY_LONG_HEADER_BIT | QUICLY_QUIC_BIT | 0x10) +#define QUICLY_PACKET_TYPE_HANDSHAKE (QUICLY_LONG_HEADER_BIT | QUICLY_QUIC_BIT | 0x20) +#define QUICLY_PACKET_TYPE_RETRY (QUICLY_LONG_HEADER_BIT | QUICLY_QUIC_BIT | 0x30) +#define QUICLY_PACKET_TYPE_BITMASK 0xf0 + + static u32 quic_ctx_alloc () { @@ -73,7 +91,7 @@ quic_ctx_get (u32 ctx_index) } static quic_ctx_t * -quic_ctx_get_w_thread (u32 ctx_index, u8 thread_index) +quic_ctx_get_w_thread (u32 ctx_index, u32 thread_index) { return pool_elt_at_index (quic_main.ctx_pool[thread_index], ctx_index); } @@ -94,11 +112,10 @@ quic_disconnect_transport (quic_ctx_t * ctx) static int quic_send_datagram (session_t * udp_session, quicly_datagram_t * packet) { - // QUIC_DBG (2, "Called quic_send_datagram at %ld", quic_get_time (NULL)); + /* QUIC_DBG (2, "Called quic_send_datagram at %ld", quic_get_time (NULL)); */ u32 max_enqueue; session_dgram_hdr_t hdr; - int rv; - u32 len; + u32 len, ret; svm_fifo_t *f; transport_connection_t *tc; @@ -108,21 +125,28 @@ quic_send_datagram (session_t * udp_session, quicly_datagram_t * packet) max_enqueue = svm_fifo_max_enqueue (f); if (max_enqueue <= sizeof (session_dgram_hdr_t)) - return 1; + { + QUIC_DBG (1, "Not enough space to enqueue header"); + return 1; + } max_enqueue -= sizeof (session_dgram_hdr_t); if (max_enqueue < len) - return 1; + { + QUIC_DBG (1, "Too much data to send, max_enqueue %u, len %u", + max_enqueue, len); + return 1; + } - // Build packet header for fifo + /* Build packet header for fifo */ hdr.data_length = len; hdr.data_offset = 0; hdr.is_ip4 = tc->is_ip4; clib_memcpy (&hdr.lcl_ip, &tc->lcl_ip, sizeof (ip46_address_t)); hdr.lcl_port = tc->lcl_port; - // Read dest address from quicly-provided sockaddr + /* Read dest address from quicly-provided sockaddr */ if (hdr.is_ip4) { ASSERT (packet->sa.sa_family == AF_INET); @@ -138,28 +162,43 @@ quic_send_datagram (session_t * udp_session, quicly_datagram_t * packet) clib_memcpy (&hdr.rmt_ip.ip6, &sa6->sin6_addr, 16); } - rv = svm_fifo_enqueue (f, sizeof (hdr), (u8 *) & hdr); - ASSERT (rv == sizeof (hdr)); - if (svm_fifo_enqueue (f, len, packet->data.base) != len) - return 1; + ret = svm_fifo_enqueue (f, sizeof (hdr), (u8 *) & hdr); + if (ret != sizeof (hdr)) + { + QUIC_DBG (2, "Not enough space to enqueue header"); + return 1; + } + ret = svm_fifo_enqueue (f, len, packet->data.base); + if (ret != len) + { + QUIC_DBG (2, "Not enough space to enqueue payload"); + return 1; + } return 0; } +#define QUIC_SEND_PACKET_VEC_SIZE 16 + +static int +quic_sendable_packet_count (session_t * udp_session) +{ + u32 max_enqueue; + max_enqueue = svm_fifo_max_enqueue (udp_session->tx_fifo); + return clib_min (max_enqueue / + (QUIC_MAX_PACKET_SIZE + sizeof (session_dgram_hdr_t)), + QUIC_SEND_PACKET_VEC_SIZE); +} + static int quic_send_packets (quic_ctx_t * ctx) { - //QUIC_DBG (2, "Called quic_send_packets"); - quicly_datagram_t *packets[16]; + quicly_datagram_t *packets[QUIC_SEND_PACKET_VEC_SIZE]; session_t *udp_session; quicly_conn_t *conn; - size_t num_packets, i; - int ret; - + size_t num_packets, i, max_packets; + /* We have sctx, get qctx */ if (ctx->c_quic_ctx_id.is_stream) - { - // We have sctx, get qctx - ctx = quic_ctx_get (ctx->c_quic_ctx_id.quic_connection_ctx_id); - } + ctx = quic_ctx_get (ctx->c_quic_ctx_id.quic_connection_ctx_id); ASSERT (!ctx->c_quic_ctx_id.is_stream); @@ -170,37 +209,42 @@ quic_send_packets (quic_ctx_t * ctx) if (!conn) return 0; + /* TODO : quicly can assert it can send min_packets up to 2 */ + if (quic_sendable_packet_count (udp_session) < 2) + goto stop_sending; + do { - num_packets = sizeof (packets) / sizeof (packets[0]); - if ((ret = quicly_send (conn, packets, &num_packets)) == 0) + max_packets = quic_sendable_packet_count (udp_session); + if (max_packets < 2) + break; + num_packets = max_packets; + if (quicly_send (conn, packets, &num_packets)) + goto quicly_error; + + for (i = 0; i != num_packets; ++i) { - for (i = 0; i != num_packets; ++i) - { - if (quic_send_datagram (udp_session, packets[i])) - { - QUIC_DBG (2, "quic_send_datagram failed"); - goto stop_sending; - } - ret = 0; - quicly_default_free_packet_cb.cb - (&quicly_default_free_packet_cb, packets[i]); - } - } - else - { - QUIC_DBG (2, "quicly_send returned %d, closing connection\n", ret); - return ret; + if (quic_send_datagram (udp_session, packets[i])) + goto quicly_error; + + quicly_default_free_packet_cb.cb + (&quicly_default_free_packet_cb, packets[i]); } + } - while (ret == 0 && num_packets == sizeof (packets) / sizeof (packets[0])); + while (num_packets > 0 && num_packets == max_packets); -stop_sending: if (svm_fifo_set_event (udp_session->tx_fifo)) - session_send_io_evt_to_thread (udp_session->tx_fifo, FIFO_EVENT_APP_TX); + session_send_io_evt_to_thread (udp_session->tx_fifo, SESSION_IO_EVT_TX); +stop_sending: quic_update_timer (ctx); return 0; + +quicly_error: + QUIC_DBG (1, "Error sending packets closing connection"); + quic_connection_closed (ctx->c_c_index); + return 1; } /***************************************************************************** @@ -208,6 +252,18 @@ stop_sending: * Called from QUIC lib *****************************************************************************/ +static void +quic_on_stream_destroy (quicly_stream_t * stream) +{ + quic_stream_data_t *stream_data = stream->data; + u32 sctx_id = stream_data->ctx_id; + quic_ctx_t *sctx = quic_ctx_get (sctx_id); + QUIC_DBG (2, "Stream %ld (ctx %u) destroyed", stream->stream_id, sctx_id); + session_transport_delete_notify (&sctx->connection); + quic_ctx_free (sctx); + quicly_streambuf_destroy (stream); +} + static int quic_on_stop_sending (quicly_stream_t * stream, int error_code) { @@ -226,36 +282,93 @@ static int quic_on_receive (quicly_stream_t * stream, size_t off, const void *src, size_t len) { - QUIC_DBG (2, "received data: %lu bytes", len); - u32 to_enqueue, ctx_id; +/* QUIC_DBG (2, "received data: %lu bytes", len); */ + u32 max_enq, ctx_id, ret; quic_ctx_t *sctx; session_t *stream_session; svm_fifo_t *rx_fifo; app_worker_t *app_wrk; + int rv; + ptls_iovec_t input; + + if ((rv = quicly_streambuf_ingress_receive (stream, off, src, len)) != 0) + { + clib_warning ("quicly_streambuf_ingress_receive Failed"); + return rv; + } + + if ((input = quicly_streambuf_ingress_get (stream)).len == 0) + { + QUIC_DBG (2, "No data, ignoring"); + return QUICLY_ERROR_PACKET_IGNORED; + } + + QUIC_DBG (1, "Got %lu bytes in RX", input.len); ctx_id = ((quic_stream_data_t *) stream->data)->ctx_id; sctx = quic_ctx_get (ctx_id); stream_session = session_get (sctx->c_s_index, vlib_get_thread_index ()); rx_fifo = stream_session->rx_fifo; - to_enqueue = svm_fifo_max_enqueue (rx_fifo); - if (to_enqueue > len) - to_enqueue = len; - // TODO what happens to the excess bytes? + max_enq = svm_fifo_max_enqueue (rx_fifo); - svm_fifo_enqueue (rx_fifo, to_enqueue, src); + if (!max_enq) + { + QUIC_DBG (2, "RX FIFO full on stream receive"); + rv = QUICLY_ERROR_PACKET_IGNORED; + goto notify_app; + } + len = clib_min (max_enq, input.len); + ret = svm_fifo_enqueue (rx_fifo, len, input.base); + if (ret != len) + { + QUIC_DBG (2, "Not enough data to dequeue"); + return 1; + } + quicly_streambuf_ingress_shift (stream, len); + rv = 0; - // Notify app +notify_app: app_wrk = app_worker_get_if_valid (stream_session->app_wrk_index); if (PREDICT_TRUE (app_wrk != 0)) app_worker_lock_and_send_event (app_wrk, stream_session, SESSION_IO_EVT_RX); + return rv; +} + +void +quic_streambuf_egress_shift (quicly_stream_t * stream, size_t delta) +{ + quicly_streambuf_egress_shift (stream, delta); +} + +int +quic_streambuf_egress_emit (quicly_stream_t * stream, size_t off, void *dst, + size_t * len, int *wrote_all) +{ + quicly_streambuf_t *sbuf = stream->data; + + assert (off < sbuf->egress.buf.off); + + if (off + *len < sbuf->egress.buf.off) + { + *wrote_all = 0; + } + else + { + *len = sbuf->egress.buf.off - off; + *wrote_all = 1; + } + + QUIC_DBG (1, "EGRESS %lu bytes in TX", *len); + memcpy (dst, sbuf->egress.buf.base + off, *len); + return 0; } static const quicly_stream_callbacks_t quic_stream_callbacks = { - .on_destroy = quicly_streambuf_destroy, - .on_send_shift = quicly_streambuf_egress_shift, - .on_send_emit = quicly_streambuf_egress_emit, + .on_destroy = quic_on_stream_destroy, + .on_send_shift = quic_streambuf_egress_shift, + .on_send_emit = quic_streambuf_egress_emit, .on_send_stop = quic_on_stop_sending, .on_receive = quic_on_receive, .on_receive_reset = quic_on_receive_reset @@ -265,11 +378,12 @@ static void quic_accept_stream (void *s) { quicly_stream_t *stream = (quicly_stream_t *) s; - session_t *stream_session, *quic_session; + session_t *stream_session; quic_stream_data_t *stream_data; app_worker_t *app_wrk; quic_ctx_t *qctx, *sctx; u32 qctx_id, sctx_id; + quic_main_t *qm = &quic_main; int rv; sctx_id = quic_ctx_alloc (); @@ -292,20 +406,18 @@ quic_accept_stream (void *s) sctx->c_quic_ctx_id.stream = stream; sctx->c_quic_ctx_id.stream_session_handle = session_handle (stream_session); - quic_session = - session_get_from_handle (qctx->c_quic_ctx_id.quic_session_handle); stream_data = (quic_stream_data_t *) stream->data; stream_data->ctx_id = sctx_id; sctx->c_s_index = stream_session->session_index; stream_session->session_state = SESSION_STATE_CREATED; + stream_session->flags |= SESSION_F_QUIC_STREAM; stream_session->app_wrk_index = sctx->c_quic_ctx_id.parent_app_wrk_id; stream_session->connection_index = sctx->c_c_index; stream_session->session_type = session_type_from_proto_and_ip (TRANSPORT_PROTO_QUIC, qctx->c_quic_ctx_id.udp_is_ip4); - stream_session->opaque = QUIC_SESSION_TYPE_STREAM; - stream_session->listener_index = quic_session->session_index; + stream_session->listener_index = qm->fake_app_listener_index; stream_session->app_index = sctx->c_quic_ctx_id.parent_app_id; app_wrk = app_worker_get (stream_session->app_wrk_index); @@ -340,8 +452,8 @@ quic_on_stream_open (quicly_stream_open_cb * self, quicly_stream_t * stream) return ret; } stream->callbacks = &quic_stream_callbacks; - // Notify accept on parent qsession, but only if this is not a locally - // initiated stream + /* Notify accept on parent qsession, but only if this is not a locally + * initiated stream */ if (!quicly_stream_is_self_initiated (stream)) { quic_accept_stream (stream); @@ -356,7 +468,7 @@ quic_on_conn_close (quicly_closed_by_peer_cb * self, quicly_conn_t * conn, int code, uint64_t frame_type, const char *reason, size_t reason_len) { - QUIC_DBG (2, "connection closed, reason: %s", reason); + QUIC_DBG (2, "connection closed, reason: %.*s", reason, reason_len); u32 ctx_index = (u64) * quicly_get_data (conn); quic_ctx_t *ctx = quic_ctx_get (ctx_index); session_transport_closing_notify (&ctx->connection); @@ -642,37 +754,50 @@ load_bio_private_key (ptls_context_t * ctx, const char *pk_data) ctx->sign_certificate = &sc.super; } +static inline void +quic_make_connection_key (clib_bihash_kv_16_8_t * kv, + const quicly_cid_plaintext_t * id) +{ + kv->key[0] = ((u64) id->master_id) << 32 | (u64) id->thread_id; + kv->key[1] = id->node_id; +} + static void quic_connection_closed (u32 ctx_index) { + /* TODO : free fifos */ QUIC_DBG (2, "QUIC connection closed"); + tw_timer_wheel_1t_3w_1024sl_ov_t *tw; + clib_bihash_kv_16_8_t kv; + quicly_conn_t *conn; quic_ctx_t *ctx; ctx = quic_ctx_get (ctx_index); ASSERT (!ctx->c_quic_ctx_id.is_stream); - // TODO if connection is not established, just delete the session? + /* TODO if connection is not established, just delete the session? */ - // TODO: close all streams? or is the streams closed cb called by quicly? + /* TODO: close all streams? or is the streams closed cb called by quicly? */ + + /* Stop the timer */ + if (ctx->timer_handle != QUIC_TIMER_HANDLE_INVALID) + { + tw = &quic_main.wrk_ctx[vlib_get_thread_index ()].timer_wheel; + tw_timer_stop_1t_3w_1024sl_ov (tw, ctx->timer_handle); + } + + /* Delete the connection from the connection map */ + conn = ctx->c_quic_ctx_id.conn; + quic_make_connection_key (&kv, quicly_get_master_id (conn)); + clib_bihash_add_del_16_8 (&quic_main.connection_hash, &kv, 0 /* is_add */ ); session_transport_delete_notify (&ctx->connection); - // Do not try to send anything anymore + /* Do not try to send anything anymore */ quicly_free (ctx->c_quic_ctx_id.conn); ctx->c_quic_ctx_id.conn = NULL; quic_ctx_free (ctx); } -static int64_t -quic_get_time (quicly_now_cb * self) -{ - // TODO read value set by set_time_now? - // (needs to change it not to call this function) - vlib_main_t *vlib_main = vlib_get_main (); - f64 time = vlib_time_now (vlib_main); - return (int64_t) (time * 1000.f); -} -quicly_now_cb quicly_vpp_now_cb = { quic_get_time }; - static void allocate_quicly_ctx (application_t * app, u8 is_client) { @@ -690,6 +815,7 @@ allocate_quicly_ctx (application_t * app, u8 is_client) app->quicly_ctx = (u64 *) quicly_ctx; memcpy (quicly_ctx, &quicly_default_context, sizeof (quicly_context_t)); + quicly_ctx->max_packet_size = QUIC_MAX_PACKET_SIZE; quicly_ctx->tls = &quic_tlsctx; quicly_ctx->stream_open = &on_stream_open; quicly_ctx->closed_by_peer = &on_closed_by_peer; @@ -697,7 +823,7 @@ allocate_quicly_ctx (application_t * app, u8 is_client) quicly_amend_ptls_context (quicly_ctx->tls); - quicly_ctx->event_log.mask = 0; + quicly_ctx->event_log.mask = 0; /* logs */ quicly_ctx->event_log.cb = quicly_new_default_event_log_cb (stderr); quicly_ctx->transport_params.max_data = QUIC_INT_MAX; @@ -734,47 +860,92 @@ allocate_quicly_ctx (application_t * app, u8 is_client) * BEGIN TIMERS HANDLING *****************************************************************************/ +static int64_t +quic_get_thread_time (u8 thread_index) +{ + return quic_main.wrk_ctx[thread_index].time_now; +} + +static int64_t +quic_get_time (quicly_now_cb * self) +{ + u8 thread_index = vlib_get_thread_index (); + return quic_get_thread_time (thread_index); +} + static u32 quic_set_time_now (u32 thread_index) { - quic_main.wrk_ctx[thread_index].time_now = quic_get_time (NULL); + vlib_main_t *vlib_main = vlib_get_main (); + f64 time = vlib_time_now (vlib_main); + quic_main.wrk_ctx[thread_index].time_now = (int64_t) (time * 1000.f); return quic_main.wrk_ctx[thread_index].time_now; } +/* Transport proto callback */ +static void +quic_update_time (f64 now, u8 thread_index) +{ + tw_timer_wheel_1t_3w_1024sl_ov_t *tw; + + tw = &quic_main.wrk_ctx[thread_index].timer_wheel; + quic_set_time_now (thread_index); + tw_timer_expire_timers_1t_3w_1024sl_ov (tw, now); +} + static void quic_timer_expired (u32 conn_index) { quic_ctx_t *ctx; - QUIC_DBG (2, "Timer expired for conn %u at %ld", conn_index, + QUIC_DBG (5, "Timer expired for conn %u at %ld", conn_index, quic_get_time (NULL)); ctx = quic_ctx_get (conn_index); ctx->timer_handle = QUIC_TIMER_HANDLE_INVALID; - if (quic_send_packets (ctx)) - { - quic_connection_closed (conn_index); - } + quic_send_packets (ctx); } static void quic_update_timer (quic_ctx_t * ctx) { tw_timer_wheel_1t_3w_1024sl_ov_t *tw; - int64_t next_timeout; + int64_t next_timeout, next_interval; + session_t *quic_session; - // This timeout is in ms which is the unit of our timer + /* This timeout is in ms which is the unit of our timer */ next_timeout = quicly_get_first_timeout (ctx->c_quic_ctx_id.conn); + next_interval = next_timeout - quic_get_time (NULL); + + if (next_timeout == 0 || next_interval <= 0) + { + if (ctx->c_s_index == QUIC_SESSION_INVALID) + { + next_interval = 1; + } + else + { + quic_session = session_get (ctx->c_s_index, ctx->c_thread_index); + if (svm_fifo_set_event (quic_session->tx_fifo)) + session_send_io_evt_to_thread_custom (quic_session, + quic_session->thread_index, + SESSION_IO_EVT_BUILTIN_TX); + return; + } + } + tw = &quic_main.wrk_ctx[vlib_get_thread_index ()].timer_wheel; - f64 next_timeout_f = ((f64) next_timeout) / 1000.f; - // clib_warning ("Timer set to %ld (%lf)", next_timeout, next_timeout_f); + QUIC_DBG (5, "Timer set to %ld (int %ld) for ctx %u", next_timeout, + next_interval, ctx->c_c_index); if (ctx->timer_handle == QUIC_TIMER_HANDLE_INVALID) { if (next_timeout == INT64_MAX) - return; + { + QUIC_DBG (5, "timer for ctx %u already stopped", ctx->c_c_index); + return; + } ctx->timer_handle = - tw_timer_start_1t_3w_1024sl_ov (tw, ctx->c_c_index, 0, - next_timeout_f); + tw_timer_start_1t_3w_1024sl_ov (tw, ctx->c_c_index, 0, next_interval); } else { @@ -782,11 +953,13 @@ quic_update_timer (quic_ctx_t * ctx) { tw_timer_stop_1t_3w_1024sl_ov (tw, ctx->timer_handle); ctx->timer_handle = QUIC_TIMER_HANDLE_INVALID; + QUIC_DBG (5, "Stopping timer for ctx %u", ctx->c_c_index); } else tw_timer_update_1t_3w_1024sl_ov (tw, ctx->timer_handle, - next_timeout_f); + next_interval); } + return; } static void @@ -818,7 +991,7 @@ quic_connect (transport_endpoint_cfg_t * tep) if (sep->port == 0) { - // TODO: better logic to detect if this is a stream or a connection request + /* TODO: better logic to detect if this is a stream or a connection request */ connect_stream = 1; } @@ -843,9 +1016,10 @@ quic_connect_new_stream (session_endpoint_cfg_t * sep) app_worker_t *app_wrk; quic_ctx_t *qctx, *sctx; u32 sctx_index; + quic_main_t *qm = &quic_main; int rv; - // Find base session to which the user want to attach a stream + /* Find base session to which the user want to attach a stream */ quic_session_handle = sep->transport_opts; QUIC_DBG (2, "Opening new stream (qsession %u)", sep->transport_opts); quic_session = session_get_from_handle (quic_session_handle); @@ -857,7 +1031,14 @@ quic_connect_new_stream (session_endpoint_cfg_t * sep) return -1; } - sctx_index = quic_ctx_alloc (); // Allocate before we get pointers + app_wrk = app_worker_get_if_valid (quic_session->app_wrk_index); + if (!app_wrk) + { + QUIC_DBG (1, "Invalid app worker :("); + return -1; + } + + sctx_index = quic_ctx_alloc (); /* Allocate before we get pointers */ sctx = quic_ctx_get (sctx_index); qctx = quic_ctx_get (quic_session->connection_index); if (qctx->c_quic_ctx_id.is_stream) @@ -888,17 +1069,17 @@ quic_connect_new_stream (session_endpoint_cfg_t * sep) QUIC_DBG (2, "Opened stream %d, creating session", stream->stream_id); - app_wrk = app_worker_get_if_valid (quic_session->app_wrk_index); - stream_session = session_alloc (qctx->c_thread_index); QUIC_DBG (1, "Created stream_session, id %u ctx %u", stream_session->session_index, sctx_index); - stream_session->app_wrk_index = quic_session->app_wrk_index; + stream_session->flags |= SESSION_F_QUIC_STREAM; + stream_session->app_wrk_index = app_wrk->wrk_index; stream_session->connection_index = sctx_index; + stream_session->listener_index = qm->fake_app_listener_index; stream_session->session_type = session_type_from_proto_and_ip (TRANSPORT_PROTO_QUIC, qctx->c_quic_ctx_id.udp_is_ip4); - stream_session->opaque = QUIC_SESSION_TYPE_STREAM; + sctx->c_s_index = stream_session->session_index; sctx->c_quic_ctx_id.stream_session_handle = session_handle (stream_session); @@ -941,7 +1122,8 @@ quic_connect_new_connection (session_endpoint_cfg_t * sep) ctx_index = quic_ctx_alloc (); ctx = quic_ctx_get (ctx_index); ctx->c_quic_ctx_id.parent_app_wrk_id = sep->app_wrk_index; - ctx->c_s_index = 0xFAFAFAFA; + ctx->c_s_index = QUIC_SESSION_INVALID; + ctx->c_c_index = ctx_index; ctx->c_quic_ctx_id.udp_is_ip4 = sep->is_ip4; ctx->timer_handle = QUIC_TIMER_HANDLE_INVALID; ctx->conn_state = QUIC_CONN_STATE_HANDSHAKE; @@ -953,13 +1135,13 @@ quic_connect_new_connection (session_endpoint_cfg_t * sep) } else { - // needed by quic for crypto + determining client / server + /* needed by quic for crypto + determining client / server */ ctx->srv_hostname = format (0, "%U", format_ip46_address, &sep->ip, sep->is_ip4); } clib_memcpy (&cargs->sep, sep, sizeof (session_endpoint_cfg_t)); - cargs->sep.transport_proto = TRANSPORT_PROTO_UDP; + cargs->sep.transport_proto = TRANSPORT_PROTO_UDPC; cargs->app_index = qm->app_index; cargs->api_context = ctx_index; @@ -973,7 +1155,6 @@ quic_connect_new_connection (session_endpoint_cfg_t * sep) if ((error = vnet_connect (cargs))) return error; - QUIC_DBG (1, "New connect request %u", ctx_index); return 0; } @@ -981,26 +1162,28 @@ static void quic_disconnect (u32 ctx_index, u32 thread_index) { QUIC_DBG (2, "Called quic_disconnect"); - //tw_timer_wheel_1t_3w_1024sl_ov_t *tw; quic_ctx_t *ctx; - QUIC_DBG (1, "Closing connection %x", ctx_index); - ctx = quic_ctx_get (ctx_index); if (ctx->c_quic_ctx_id.is_stream) { + QUIC_DBG (1, "Closing stream %x", ctx_index); quicly_stream_t *stream = ctx->c_quic_ctx_id.stream; quicly_reset_stream (stream, 0x30000); - session_transport_delete_notify (&ctx->connection); - quic_ctx_free (ctx); } else { + QUIC_DBG (1, "Closing connection %x", ctx_index); quicly_conn_t *conn = ctx->c_quic_ctx_id.conn; - // Start connection closing. Keep sending packets until quicly_send - // returns QUICLY_ERROR_FREE_CONNECTION + /* Start connection closing. Keep sending packets until quicly_send + returns QUICLY_ERROR_FREE_CONNECTION */ quicly_close (conn, 0, ""); - quic_send_packets (ctx); + /* This also causes all streams to be closed (and the cb called) */ + if (quic_send_packets (ctx)) + { + QUIC_DBG (2, "closing connection in disconnect"); + quic_connection_closed (ctx->c_c_index); + } } } @@ -1020,15 +1203,15 @@ quic_start_listen (u32 quic_listen_session_index, transport_endpoint_t * tep) sep = (session_endpoint_cfg_t *) tep; app_wrk = app_worker_get (sep->app_wrk_index); - // We need to call this because we call app_worker_init_connected in - // quic_accept_stream, which assumes the connect segment manager exists + /* We need to call this because we call app_worker_init_connected in + * quic_accept_stream, which assumes the connect segment manager exists */ app_worker_alloc_connects_segment_manager (app_wrk); app = application_get (app_wrk->app_index); QUIC_DBG (2, "Called quic_start_listen for app %d", app_wrk->app_index); allocate_quicly_ctx (app, 0 /* is_client */ ); - sep->transport_proto = TRANSPORT_PROTO_UDP; + sep->transport_proto = TRANSPORT_PROTO_UDPC; memset (args, 0, sizeof (*args)); args->app_index = qm->app_index; args->sep_ext = *sep; @@ -1036,16 +1219,15 @@ quic_start_listen (u32 quic_listen_session_index, transport_endpoint_t * tep) if (vnet_listen (args)) return -1; - lctx_index = quic_ctx_alloc (); // listener + lctx_index = quic_ctx_alloc (); /* listener */ udp_handle = args->handle; app_listener = app_listener_get_w_handle (udp_handle); udp_listen_session = app_listener_get_session (app_listener); udp_listen_session->opaque = lctx_index; quic_listen_session = listen_session_get (quic_listen_session_index); - quic_listen_session->opaque = QUIC_SESSION_TYPE_LISTEN; - lctx = quic_ctx_get (lctx_index); // listener + lctx = quic_ctx_get (lctx_index); /* listener */ lctx->is_listener = 1; lctx->c_quic_ctx_id.parent_app_wrk_id = sep->app_wrk_index; lctx->c_quic_ctx_id.parent_app_id = app_wrk->app_index; @@ -1064,7 +1246,7 @@ quic_stop_listen (u32 lctx_index) QUIC_DBG (2, "Called quic_stop_listen"); quic_ctx_t *lctx; - lctx = quic_ctx_get (lctx_index); // listener + lctx = quic_ctx_get (lctx_index); /* listener */ vnet_unlisten_args_t a = { .handle = lctx->c_quic_ctx_id.udp_session_handle, .app_index = quic_main.app_index, @@ -1073,9 +1255,9 @@ quic_stop_listen (u32 lctx_index) if (vnet_unlisten (&a)) clib_warning ("unlisten errored"); - // TODO: crypto state cleanup + /* TODO: crypto state cleanup */ - quic_ctx_free (lctx); // listener + quic_ctx_free (lctx); /* listener */ return 0; } @@ -1097,20 +1279,10 @@ quic_listener_get (u32 listener_index) return &ctx->connection; } -static void -quic_update_time (f64 now, u8 thread_index) -{ - tw_timer_wheel_1t_3w_1024sl_ov_t *tw; - - tw = &quic_main.wrk_ctx[thread_index].timer_wheel; - quic_set_time_now (thread_index); - tw_timer_expire_timers_1t_3w_1024sl_ov (tw, now); -} - static u8 * format_quic_connection (u8 * s, va_list * args) { - s = format (s, "[QUIC] connection"); //TODO + s = format (s, "[QUIC] connection"); /* TODO */ return s; } @@ -1123,11 +1295,11 @@ format_quic_half_open (u8 * s, va_list * args) return s; } -// TODO improve +/* TODO improve */ static u8 * format_quic_listener (u8 * s, va_list * args) { - s = format (s, "[QUIC] listener"); // TODO + s = format (s, "[QUIC] listener"); /* TODO */ return s; } @@ -1167,6 +1339,7 @@ quic_notify_app_connected (quic_ctx_t * ctx) session_t *quic_session; app_worker_t *app_wrk; u32 ctx_id = ctx->c_c_index; + quic_main_t *qm = &quic_main; app_wrk = app_worker_get_if_valid (ctx->c_quic_ctx_id.parent_app_wrk_id); if (!app_wrk) @@ -1181,7 +1354,7 @@ quic_notify_app_connected (quic_ctx_t * ctx) ctx->c_s_index = quic_session->session_index; quic_session->app_wrk_index = ctx->c_quic_ctx_id.parent_app_wrk_id; quic_session->connection_index = ctx->c_c_index; - quic_session->opaque = QUIC_SESSION_TYPE_QUIC; + quic_session->listener_index = qm->fake_app_listener_index; quic_session->session_type = session_type_from_proto_and_ip (TRANSPORT_PROTO_QUIC, ctx->c_quic_ctx_id.udp_is_ip4); @@ -1201,7 +1374,7 @@ quic_notify_app_connected (quic_ctx_t * ctx) return -1; } - // If the app opens a stream in its callback it may invalidate ctx + /* If the app opens a stream in its callback it may invalidate ctx */ ctx = quic_ctx_get (ctx_id); ctx->c_quic_ctx_id.quic_session_handle = session_handle (quic_session); quic_session->session_state = SESSION_STATE_LISTENING; @@ -1217,16 +1390,18 @@ quic_session_connected_callback (u32 quic_app_index, u32 ctx_index, { QUIC_DBG (2, "QSession is now connected (id %u)", udp_session->session_index); - // This should always be called before quic_connect returns since UDP always - // connects instantly. + /* This should always be called before quic_connect returns since UDP always + * connects instantly. */ + clib_bihash_kv_16_8_t kv; struct sockaddr_in6 sa6; struct sockaddr *sa = (struct sockaddr *) &sa6; socklen_t salen; transport_connection_t *tc; + app_worker_t *app_wrk; + quicly_conn_t *conn; + application_t *app; quic_ctx_t *ctx; int ret; - application_t *app; - app_worker_t *app_wrk; ctx = quic_ctx_get (ctx_index); if (is_fail) @@ -1259,11 +1434,11 @@ quic_session_connected_callback (u32 quic_app_index, u32 ctx_index, is_fail, vlib_get_thread_index (), (ctx) ? ctx_index : ~0); ctx->c_quic_ctx_id.udp_session_handle = session_handle (udp_session); - udp_session->opaque = ctx_index; + udp_session->opaque = ctx->c_quic_ctx_id.parent_app_id; udp_session->session_state = SESSION_STATE_READY; - // Init QUIC lib connection - // Generate required sockaddr & salen + /* Init QUIC lib connection + * Generate required sockaddr & salen */ tc = session_get_transport (udp_session); quic_build_sockaddr (sa, &salen, &tc->rmt_ip, tc->rmt_port, tc->is_ip4); @@ -1273,17 +1448,116 @@ quic_session_connected_callback (u32 quic_app_index, u32 ctx_index, (char *) ctx->srv_hostname, sa, salen, &quic_main.next_cid, &quic_main.hs_properties, NULL); ++quic_main.next_cid.master_id; - // Save context handle in quicly connection + /* Save context handle in quicly connection */ *quicly_get_data (ctx->c_quic_ctx_id.conn) = (void *) (u64) ctx_index; assert (ret == 0); - int rv = quic_send_packets (ctx); - if (rv) + /* Register connection in connections map */ + conn = ctx->c_quic_ctx_id.conn; + quic_make_connection_key (&kv, quicly_get_master_id (conn)); + kv.value = ((u64) vlib_get_thread_index ()) << 32 | (u64) ctx_index; + QUIC_DBG (1, "Registering conn with id %lu %lu", kv.key[0], kv.key[1]); + clib_bihash_add_del_16_8 (&quic_main.connection_hash, &kv, 1 /* is_add */ ); + + quic_send_packets (ctx); + + /* UDP stack quirk? preemptively transfer connection if that happens */ + if (udp_session->thread_index != vlib_get_thread_index ()) + quic_transfer_connection (ctx_index, udp_session->thread_index); + + return ret; +} + +static void +quic_receive_connection (void *arg) +{ + u32 new_ctx_id, thread_index = vlib_get_thread_index (); + quic_ctx_t *temp_ctx, *new_ctx; + clib_bihash_kv_16_8_t kv; + quicly_conn_t *conn; + + temp_ctx = arg; + new_ctx_id = quic_ctx_alloc (); + new_ctx = quic_ctx_get (new_ctx_id); + + QUIC_DBG (3, "Received conn %u (now %u)", temp_ctx->c_thread_index, + new_ctx_id); + + + memcpy (new_ctx, temp_ctx, sizeof (quic_ctx_t)); + free (temp_ctx); + + new_ctx->c_thread_index = thread_index; + new_ctx->c_c_index = new_ctx_id; + + conn = new_ctx->c_quic_ctx_id.conn; + quic_make_connection_key (&kv, quicly_get_master_id (conn)); + kv.value = ((u64) thread_index) << 32 | (u64) new_ctx_id; + QUIC_DBG (1, "Registering conn with id %lu %lu", kv.key[0], kv.key[1]); + clib_bihash_add_del_16_8 (&quic_main.connection_hash, &kv, 1 /* is_add */ ); + new_ctx->timer_handle = QUIC_TIMER_HANDLE_INVALID; + quic_update_timer (new_ctx); + + /* Trigger read on this connection ? */ +} + +static void +quic_transfer_connection (u32 ctx_index, u32 dest_thread) +{ + tw_timer_wheel_1t_3w_1024sl_ov_t *tw; + quic_ctx_t *ctx, *temp_ctx; + clib_bihash_kv_16_8_t kv; + quicly_conn_t *conn; + + QUIC_DBG (3, "Transferring conn %u to thread %u", ctx_index, dest_thread); + + temp_ctx = malloc (sizeof (quic_ctx_t)); + ASSERT (temp_ctx); + ctx = quic_ctx_get (ctx_index); + + memcpy (temp_ctx, ctx, sizeof (quic_ctx_t)); + + /* Remove from lookup hash, timer wheel and thread-local pool */ + conn = ctx->c_quic_ctx_id.conn; + quic_make_connection_key (&kv, quicly_get_master_id (conn)); + clib_bihash_add_del_16_8 (&quic_main.connection_hash, &kv, 0 /* is_add */ ); + if (ctx->timer_handle != QUIC_TIMER_HANDLE_INVALID) { - QUIC_DBG (1, "Error sending packets %d, closing connection", rv); - quic_connection_closed (ctx_index); + tw = &quic_main.wrk_ctx[vlib_get_thread_index ()].timer_wheel; + tw_timer_stop_1t_3w_1024sl_ov (tw, ctx->timer_handle); } - return ret; + quic_ctx_free (ctx); + + /* Send connection to destination thread */ + session_send_rpc_evt_to_thread (dest_thread, quic_receive_connection, + (void *) temp_ctx); +} + +static void +quic_transfer_connection_rpc (void *arg) +{ + u64 arg_int = (u64) arg; + u32 ctx_index, dest_thread; + + ctx_index = (u32) (arg_int >> 32); + dest_thread = (u32) (arg_int & UINT32_MAX); + quic_transfer_connection (ctx_index, dest_thread); +} + +/* + * This assumes that the connection is not yet associated to a session + * So currently it only works on the client side when receiving the first packet + * from the server + */ +static void +quic_move_connection_to_thread (u32 ctx_index, u32 owner_thread, + u32 to_thread) +{ + QUIC_DBG (3, "Requesting transfer of conn %u from thread %u", ctx_index, + owner_thread); + u64 arg = ((u64) ctx_index) << 32 | to_thread; + session_send_rpc_evt_to_thread (owner_thread, quic_transfer_connection_rpc, + (void *) arg); } static void @@ -1299,9 +1573,41 @@ quic_session_reset_callback (session_t * s) } int -quic_session_accepted_callback (session_t * s) +quic_session_accepted_callback (session_t * udp_session) { - /* never called */ + /* New UDP connection, try to accept it */ + QUIC_DBG (1, "UDP session accepted"); + u32 ctx_index; + u32 *pool_index; + quic_ctx_t *ctx, *lctx; + session_t *udp_listen_session; + + udp_listen_session = listen_session_get (udp_session->listener_index); + + ctx_index = quic_ctx_alloc (); + ctx = quic_ctx_get (ctx_index); + ctx->c_thread_index = udp_session->thread_index; + ctx->c_c_index = ctx_index; + ctx->c_s_index = QUIC_SESSION_INVALID; + ctx->c_quic_ctx_id.udp_session_handle = session_handle (udp_session); + ctx->c_quic_ctx_id.listener_ctx_id = udp_listen_session->opaque; + lctx = quic_ctx_get_w_thread (udp_listen_session->opaque, + udp_listen_session->thread_index); + ctx->c_quic_ctx_id.udp_is_ip4 = lctx->c_quic_ctx_id.udp_is_ip4; + ctx->c_quic_ctx_id.parent_app_id = lctx->c_quic_ctx_id.parent_app_id; + ctx->c_quic_ctx_id.parent_app_wrk_id = + lctx->c_quic_ctx_id.parent_app_wrk_id; + ctx->timer_handle = QUIC_TIMER_HANDLE_INVALID; + ctx->conn_state = QUIC_CONN_STATE_OPENED; + + udp_session->opaque = ctx->c_quic_ctx_id.parent_app_id; + + /* Put this ctx in the "opening" pool */ + pool_get (quic_main.wrk_ctx[ctx->c_thread_index].opening_ctx_pool, + pool_index); + *pool_index = ctx_index; + + /* TODO timeout to delete these if they never connect */ return 0; } @@ -1326,12 +1632,11 @@ quic_del_segment_callback (u32 client_index, u64 seg_handle) static int quic_custom_tx_callback (void *s) { - QUIC_DBG (2, "Called quic_custom_tx_callback"); session_t *stream_session = (session_t *) s; quic_ctx_t *ctx; svm_fifo_t *f; quicly_stream_t *stream; - u32 deq_max; + u32 deq_max, rv; u8 *data; if (PREDICT_FALSE @@ -1340,69 +1645,109 @@ quic_custom_tx_callback (void *s) ctx = quic_ctx_get (stream_session->connection_index); if (PREDICT_FALSE (!ctx->c_quic_ctx_id.is_stream)) { - QUIC_DBG (1, "Error: trying to send on quic session not stream"); - return -1; + goto tx_end; /* Most probably a reschedule */ } - stream = ctx->c_quic_ctx_id.stream; + if (!ctx->c_quic_ctx_id.stream->sendstate.is_open) + { + QUIC_DBG (3, "Warning: tried to send on closed stream"); + return -1; + } f = stream_session->tx_fifo; deq_max = svm_fifo_max_dequeue (f); if (!deq_max) - return 0; + goto tx_end; - data = svm_fifo_head (f); - if (quicly_streambuf_egress_write (stream, data, deq_max)) + data = malloc (deq_max); + rv = svm_fifo_peek (f, 0, deq_max, data); + if (rv != deq_max) { - assert (0); - return 0; + QUIC_DBG (2, "Not enough data dequeued in TX"); + return 1; } - QUIC_DBG (2, "Sent %u bytes", deq_max); - svm_fifo_dequeue_drop (f, deq_max); - int rv = quic_send_packets (ctx); + stream = ctx->c_quic_ctx_id.stream; + + rv = quicly_streambuf_egress_write (stream, data, deq_max); + free (data); if (rv) + return rv; + rv = svm_fifo_dequeue_drop (f, deq_max); + if (rv != deq_max) { - QUIC_DBG (1, "TX error sending packets %d, closing connection", rv); - quic_connection_closed (ctx->c_quic_ctx_id.quic_connection_ctx_id); + QUIC_DBG (2, "Not enough data dropped in TX"); + return 1; } + QUIC_DBG (2, "Sent %u bytes", deq_max); + +tx_end: + quic_send_packets (ctx); return 0; } + +/* + * Returns 0 if a matching connection is found and is on the right thread. + * If a connection is found, even on the wrong thread, ctx_thread and ctx_index + * will be set. + */ static inline int -quic_find_packet_ctx (quic_ctx_t ** ctx, quicly_conn_t ** conn, +quic_find_packet_ctx (u32 * ctx_thread, u32 * ctx_index, struct sockaddr *sa, socklen_t salen, - quicly_decoded_packet_t packet) + quicly_decoded_packet_t * packet, + u32 caller_thread_index) { quic_ctx_t *ctx_; quicly_conn_t *conn_; - /* *INDENT-OFF* */ - pool_foreach (ctx_, quic_main.ctx_pool[vlib_get_thread_index()], - ({ - conn_ = ctx_->c_quic_ctx_id.conn; - if (!ctx_->c_quic_ctx_id.is_stream && conn_ && !ctx_->is_listener) - { - if (quicly_is_destination(conn_, sa, salen, &packet)) - { - *conn = conn_; - *ctx = ctx_; - // QUIC_DBG (2, "connection_found"); - return 0; - } - } - })); - /* *INDENT-ON* */ - return 0; + clib_bihash_kv_16_8_t kv; + clib_bihash_16_8_t *h; + + h = &quic_main.connection_hash; + quic_make_connection_key (&kv, &packet->cid.dest.plaintext); + + if (clib_bihash_search_16_8 (h, &kv, &kv) == 0) + { + u32 index = kv.value & UINT32_MAX; + u32 thread_id = kv.value >> 32; + /* Check if this connection belongs to this thread, otherwise + * ask for it to be moved */ + if (thread_id != caller_thread_index) + { + QUIC_DBG (3, "Connection is on wrong thread"); + /* Cannot make full check with quicly_is_destination... */ + *ctx_index = index; + *ctx_thread = thread_id; + return -1; + } + ctx_ = quic_ctx_get (index); + conn_ = ctx_->c_quic_ctx_id.conn; + if (conn_ && quicly_is_destination (conn_, sa, salen, packet)) + { + QUIC_DBG (4, "Connection found"); + *ctx_index = index; + *ctx_thread = thread_id; + return 0; + } + } + QUIC_DBG (3, "connection not found"); + return -1; } static int quic_receive (quic_ctx_t * ctx, quicly_conn_t * conn, quicly_decoded_packet_t packet) { + int rv; u32 ctx_id = ctx->c_c_index; - quicly_receive (conn, &packet); - // ctx pointer may change if a new stream is opened + rv = quicly_receive (conn, &packet); + if (rv) /* TOOD : QUICLY_ERROR_PACKET_IGNORED sould be handled */ + { + QUIC_DBG (1, "Quicly receive ignored packet code : %u", rv); + return 0; + } + /* ctx pointer may change if a new stream is opened */ ctx = quic_ctx_get (ctx_id); - // Conn may be set to null if the connection is terminated + /* Conn may be set to null if the connection is terminated */ if (ctx->c_quic_ctx_id.conn && ctx->conn_state == QUIC_CONN_STATE_HANDSHAKE) { if (quicly_connection_is_ready (conn)) @@ -1415,11 +1760,7 @@ quic_receive (quic_ctx_t * ctx, quicly_conn_t * conn, } } } - if (quic_send_packets (ctx)) - { - quic_connection_closed (ctx->c_c_index); - } - return 0; + return quic_send_packets (ctx); } static int @@ -1447,10 +1788,9 @@ quic_create_quic_session (quic_ctx_t * ctx) ctx->c_quic_ctx_id.udp_is_ip4); quic_session->listener_index = quic_listen_session->session_index; quic_session->app_index = quic_main.app_index; - quic_session->opaque = QUIC_SESSION_TYPE_QUIC; - // TODO: don't alloc fifos when we don't transfer data on this session - // but we still need fifos for the events? + /* TODO: don't alloc fifos when we don't transfer data on this session + * but we still need fifos for the events? */ if ((rv = app_worker_init_accepted (quic_session))) { QUIC_DBG (1, "failed to allocate fifos"); @@ -1458,9 +1798,6 @@ quic_create_quic_session (quic_ctx_t * ctx) return rv; } ctx->c_quic_ctx_id.quic_session_handle = session_handle (quic_session); - ctx->c_quic_ctx_id.parent_app_id = lctx->c_quic_ctx_id.parent_app_id; - ctx->c_quic_ctx_id.udp_is_ip4 = lctx->c_quic_ctx_id.udp_is_ip4; - ctx->c_quic_ctx_id.parent_app_wrk_id = quic_session->app_wrk_index; session_lookup_add_connection (&ctx->connection, session_handle (quic_session)); app_wrk = app_worker_get (quic_session->app_wrk_index); @@ -1475,49 +1812,45 @@ quic_create_quic_session (quic_ctx_t * ctx) static int quic_create_connection (quicly_context_t * quicly_ctx, - u64 udp_session_handle, u32 lctx_index, + u64 udp_session_handle, u32 ctx_index, struct sockaddr *sa, socklen_t salen, quicly_decoded_packet_t packet) { + clib_bihash_kv_16_8_t kv; quic_ctx_t *ctx; - u32 ctx_index; quicly_conn_t *conn; int rv; - /* new connection, accept and create context if packet is valid */ - // TODO: check if socket is actually listening? + /* new connection, accept and create context if packet is valid + * TODO: check if socket is actually listening? */ if ((rv = quicly_accept (&conn, quicly_ctx, sa, salen, &packet, ptls_iovec_init (NULL, 0), &quic_main.next_cid, NULL))) { - // Invalid packet, pass + /* Invalid packet, pass */ assert (conn == NULL); QUIC_DBG (2, "Accept failed with %d", rv); + /* TODO: cleanup created quic ctx and UDP session */ return 0; } assert (conn != NULL); ++quic_main.next_cid.master_id; - // Create context - ctx_index = quic_ctx_alloc (); ctx = quic_ctx_get (ctx_index); - // Save ctx handle in quicly connection + /* Save ctx handle in quicly connection */ *quicly_get_data (conn) = (void *) (u64) ctx_index; - - ctx->c_thread_index = vlib_get_thread_index (); - ctx->c_c_index = ctx_index; - ctx->c_quic_ctx_id.udp_session_handle = udp_session_handle; - ctx->c_quic_ctx_id.listener_ctx_id = lctx_index; - ctx->timer_handle = QUIC_TIMER_HANDLE_INVALID; ctx->c_quic_ctx_id.conn = conn; + ctx->conn_state = QUIC_CONN_STATE_HANDSHAKE; quic_create_quic_session (ctx); - if (quic_send_packets (ctx)) - { - quic_connection_closed (ctx_index); - } - return 0; + /* Register connection in connections map */ + quic_make_connection_key (&kv, quicly_get_master_id (conn)); + kv.value = ((u64) vlib_get_thread_index ()) << 32 | (u64) ctx_index; + clib_bihash_add_del_16_8 (&quic_main.connection_hash, &kv, 1 /* is_add */ ); + QUIC_DBG (1, "Registering conn with id %lu %lu", kv.key[0], kv.key[1]); + + return quic_send_packets (ctx); } static int @@ -1539,8 +1872,7 @@ quic_reset_connection (quicly_context_t * quicly_ctx, u64 udp_session_handle, dgram = quicly_send_stateless_reset (quicly_ctx, sa, salen, &packet.cid.dest.plaintext); udp_session = session_get_from_handle (udp_session_handle); - if (quic_send_datagram (udp_session, dgram)) // TODO : set event on fifo - QUIC_DBG (2, "Send reset failed"); + return quic_send_datagram (udp_session, dgram); /* TODO : set event on fifo */ } return 0; } @@ -1548,76 +1880,138 @@ quic_reset_connection (quicly_context_t * quicly_ctx, u64 udp_session_handle, static int quic_app_rx_callback (session_t * udp_session) { - // Read data from UDP rx_fifo and pass it to the quicly conn. + /* Read data from UDP rx_fifo and pass it to the quicly conn. */ quicly_decoded_packet_t packet; session_dgram_hdr_t ph; application_t *app; - quicly_conn_t *conn = NULL; - quic_ctx_t *lctx, *ctx = NULL; + quic_ctx_t *ctx = NULL; svm_fifo_t *f; size_t plen; struct sockaddr_in6 sa6; struct sockaddr *sa = (struct sockaddr *) &sa6; socklen_t salen; - u32 max_deq, len; + u32 max_deq, len, full_len, ctx_index, ctx_thread = UINT32_MAX, ret; u8 *data; - u32 lctx_index = udp_session->opaque; + int err; + u32 *opening_ctx_pool, *ctx_index_ptr; + u32 app_index = udp_session->opaque; u64 udp_session_handle = session_handle (udp_session); + int rv = 0; + u32 thread_index = vlib_get_thread_index (); - // DEBUG - // lctx = quic_ctx_get (lctx_index); - // QUIC_DBG (2, "Got RX data on session %d", - // lctx->c_quic_ctx_id.udp_session_handle); - - f = udp_session->rx_fifo; + app = application_get_if_valid (app_index); + if (!app) + { + QUIC_DBG (1, "Got RX on detached app"); + /* TODO: close this session, cleanup state? */ + return 1; + } do { - conn = NULL; + udp_session = session_get_from_handle (udp_session_handle); /* session alloc might have happened */ + f = udp_session->rx_fifo; + svm_fifo_unset_event (f); max_deq = svm_fifo_max_dequeue (f); if (max_deq < sizeof (session_dgram_hdr_t)) + return 0; + + ret = svm_fifo_peek (f, 0, SESSION_CONN_HDR_LEN, (u8 *) & ph); + if (ret != SESSION_CONN_HDR_LEN) { - svm_fifo_unset_event (f); - return 0; + QUIC_DBG (2, "Not enough data for header in RX"); + return 1; + } + if (ph.data_length < ph.data_offset) + { + QUIC_DBG (2, "Not enough data vs offset in RX"); + return 1; } - // QUIC_DBG (2, "Processing one packet at %ld", quic_get_time (NULL)); - - svm_fifo_unset_event (f); - svm_fifo_peek (f, 0, sizeof (ph), (u8 *) & ph); - ASSERT (ph.data_length >= ph.data_offset); len = ph.data_length - ph.data_offset; + full_len = ph.data_length + ph.data_offset + SESSION_CONN_HDR_LEN; + if (full_len > max_deq) + { + QUIC_DBG (2, "Not enough data in fifo RX"); + return 1; + } - quic_build_sockaddr (sa, &salen, &ph.rmt_ip, ph.rmt_port, ph.is_ip4); - - // Quicly can read len bytes from the fifo at offset: - // ph.data_offset + SESSION_CONN_HDR_LEN - data = svm_fifo_head (f) + ph.data_offset + SESSION_CONN_HDR_LEN; + /* Quicly can read len bytes from the fifo at offset: + * ph.data_offset + SESSION_CONN_HDR_LEN */ + data = malloc (ph.data_length); + ret = + svm_fifo_peek (f, ph.data_offset + SESSION_CONN_HDR_LEN, + ph.data_length, data); + if (ret != ph.data_length) + { + QUIC_DBG (2, "Not enough data peeked in RX"); + return 1; + free (data); + } - lctx = quic_ctx_get (lctx_index); - app = application_get (lctx->c_quic_ctx_id.parent_app_id); + plen = + quicly_decode_packet ((quicly_context_t *) app->quicly_ctx, &packet, + data, len); + rv = 0; + quic_build_sockaddr (sa, &salen, &ph.rmt_ip, ph.rmt_port, ph.is_ip4); plen = quicly_decode_packet ((quicly_context_t *) app->quicly_ctx, &packet, data, len); + if (plen != SIZE_MAX) { - quic_find_packet_ctx (&ctx, &conn, sa, salen, packet); - if (conn != NULL) - quic_receive (ctx, conn, packet); - else if (QUICLY_PACKET_IS_LONG_HEADER (packet.octets.base[0])) - quic_create_connection ((quicly_context_t *) app->quicly_ctx, - udp_session_handle, lctx_index, - sa, salen, packet); - else if (((quicly_context_t *) app->quicly_ctx)->encrypt_cid) - quic_reset_connection ((quicly_context_t *) app->quicly_ctx, - udp_session_handle, sa, salen, packet); + + err = quic_find_packet_ctx (&ctx_thread, &ctx_index, sa, salen, + &packet, thread_index); + if (err == 0) + { + ctx = quic_ctx_get_w_thread (ctx_index, thread_index); + quic_receive (ctx, ctx->c_quic_ctx_id.conn, packet); + } + else if (ctx_thread != UINT32_MAX) + { + /* Connection found but on wrong thread, ask move */ + quic_move_connection_to_thread (ctx_index, ctx_thread, + thread_index); + } + else if ((packet.octets.base[0] & QUICLY_PACKET_TYPE_BITMASK) == + QUICLY_PACKET_TYPE_INITIAL) + { + /* Try to find matching "opening" ctx */ + opening_ctx_pool = + quic_main.wrk_ctx[thread_index].opening_ctx_pool; + + /* *INDENT-OFF* */ + pool_foreach (ctx_index_ptr, opening_ctx_pool, + ({ + ctx = quic_ctx_get_w_thread (*ctx_index_ptr, thread_index); + if (ctx->c_quic_ctx_id.udp_session_handle == udp_session_handle) + { + /* Right ctx found, create conn & remove from pool */ + quic_create_connection ((quicly_context_t *) app->quicly_ctx, + udp_session_handle, *ctx_index_ptr, + sa, salen, packet); + pool_put (opening_ctx_pool, ctx_index_ptr); + goto ctx_search_done; + } + })); + /* *INDENT-ON* */ + + } + else + { + quic_reset_connection ((quicly_context_t *) app->quicly_ctx, + udp_session_handle, sa, salen, packet); + } } + ctx_search_done: svm_fifo_dequeue_drop (f, ph.data_length + ph.data_offset + SESSION_CONN_HDR_LEN); + free (data); } while (1); - return 0; + return rv; } always_inline void @@ -1627,7 +2021,9 @@ quic_common_get_transport_endpoint (quic_ctx_t * ctx, session_t *udp_session; QUIC_DBG (2, "Called quic_get_transport_endpoint"); if (ctx->c_quic_ctx_id.is_stream) - tep->is_ip4 = 255; /* well this is ugly */ + { + tep->is_ip4 = 255; /* well this is ugly */ + } else { udp_session = @@ -1669,7 +2065,7 @@ static session_cb_vft_t quic_app_cb_vft = { .builtin_app_rx_callback = quic_app_rx_callback, }; -const static transport_proto_vft_t quic_proto = { +static const transport_proto_vft_t quic_proto = { .connect = quic_connect, .close = quic_disconnect, .start_listen = quic_start_listen, @@ -1694,11 +2090,13 @@ quic_init (vlib_main_t * vm) QUIC_DBG (2, "Called quic_init"); u32 add_segment_size = (4096ULL << 20) - 1, segment_size = 512 << 20; vlib_thread_main_t *vtm = vlib_get_thread_main (); + tw_timer_wheel_1t_3w_1024sl_ov_t *tw; vnet_app_attach_args_t _a, *a = &_a; u64 options[APP_OPTIONS_N_OPTIONS]; quic_main_t *qm = &quic_main; u32 fifo_size = 64 << 10; - u32 num_threads; + u32 num_threads, i; + application_t *app; num_threads = 1 /* main thread */ + vtm->n_threads; @@ -1725,21 +2123,35 @@ quic_init (vlib_main_t * vm) vec_validate (qm->ctx_pool, num_threads - 1); vec_validate (qm->wrk_ctx, num_threads - 1); - // Timers, one per thread. - tw_timer_wheel_1t_3w_1024sl_ov_t *tw; - /* *INDENT-OFF* */ - foreach_vlib_main (({ - tw = &qm->wrk_ctx[ii].timer_wheel; - tw_timer_wheel_init_1t_3w_1024sl_ov (tw, quic_expired_timers_dispatch, - 10e-3 /* timer period 1ms */ , ~0); - tw->last_run_time = vlib_time_now (this_vlib_main); - })); - /* *INDENT-ON* */ + /* Timer wheels, one per thread. */ + for (i = 0; i < num_threads; i++) + { + tw = &qm->wrk_ctx[i].timer_wheel; + tw_timer_wheel_init_1t_3w_1024sl_ov (tw, quic_expired_timers_dispatch, + 1e-3 /* timer period 1ms */ , ~0); + tw->last_run_time = vlib_time_now (vlib_get_main ()); + } + + clib_bihash_init_16_8 (&qm->connection_hash, "quic connections", 1024, + 4 << 20); if (!qm->ca_cert_path) qm->ca_cert_path = QUIC_DEFAULT_CA_CERT_PATH; qm->app_index = a->app_index; + + /* Fake app listener hack, to remove */ + app = application_get (a->app_index); + app_listener_t *fake_app_listener; + pool_get (app->listeners, fake_app_listener); + clib_memset (fake_app_listener, 0, sizeof (*fake_app_listener)); + fake_app_listener->al_index = fake_app_listener - app->listeners; + fake_app_listener->app_index = app->app_index; + fake_app_listener->session_index = SESSION_INVALID_INDEX; + fake_app_listener->local_index = SESSION_INVALID_INDEX; + qm->fake_app_listener_index = fake_app_listener->al_index; + /* End fake listener hack */ + qm->tstamp_ticks_per_clock = vm->clib_time.seconds_per_clock / QUIC_TSTAMP_RESOLUTION; @@ -1758,7 +2170,7 @@ VLIB_INIT_FUNCTION (quic_init); VLIB_PLUGIN_REGISTER () = { .version = VPP_BUILD_VER, - .description = "Quic Transport Protocol", + .description = "Quic transport protocol", }; /* *INDENT-ON* */ diff --git a/src/plugins/quic/quic.h b/src/plugins/quic/quic.h index a4615946d94..512ffd95eeb 100644 --- a/src/plugins/quic/quic.h +++ b/src/plugins/quic/quic.h @@ -20,19 +20,17 @@ #include <vppinfra/lock.h> #include <vppinfra/tw_timer_1t_3w_1024sl_ov.h> +#include <vppinfra/bihash_16_8.h> #include <quicly.h> #include <quicly/streambuf.h> - -#define QUIC_DEBUG 0 -#define QUIC_DEBUG_LEVEL_CLIENT 0 -#define QUIC_DEBUG_LEVEL_SERVER 0 +#define QUIC_DEBUG 0 +#define QUIC_DEBUG_LEVEL_CLIENT 0 +#define QUIC_DEBUG_LEVEL_SERVER 0 #define QUIC_DEFAULT_CA_CERT_PATH "/etc/ssl/certs/ca-certificates.crt" -#define QUIC_TIMER_HANDLE_INVALID ((u32) ~0) - #define QUIC_TSTAMP_RESOLUTION 0.001 /* QUIC tick resolution (1ms) */ @@ -44,16 +42,9 @@ #define QUIC_DBG(_lvl, _fmt, _args...) #endif -#define QUIC_CONN_STATE_HANDSHAKE 0 -#define QUIC_CONN_STATE_READY 1 - -enum quic_session_type_t -{ - QUIC_SESSION_TYPE_QUIC = 0, - QUIC_SESSION_TYPE_STREAM = 1, - QUIC_SESSION_TYPE_LISTEN = INT32_MAX, -}; - +#define QUIC_CONN_STATE_OPENED 0 +#define QUIC_CONN_STATE_HANDSHAKE 1 +#define QUIC_CONN_STATE_READY 2 /* *INDENT-OFF* */ typedef CLIB_PACKED (struct quic_ctx_id_ @@ -62,14 +53,14 @@ typedef CLIB_PACKED (struct quic_ctx_id_ u32 parent_app_id; union { CLIB_PACKED (struct { - session_handle_t quic_session_handle; // TODO: remove + session_handle_t quic_session_handle; /* TODO: remove */ session_handle_t udp_session_handle; quicly_conn_t *conn; u32 listener_ctx_id; u8 udp_is_ip4; }); CLIB_PACKED (struct { - session_handle_t stream_session_handle; // TODO: remove + session_handle_t stream_session_handle; /* TODO: remove */ quicly_stream_t *stream; u32 quic_connection_ctx_id; }); @@ -80,8 +71,8 @@ typedef CLIB_PACKED (struct quic_ctx_id_ STATIC_ASSERT (sizeof (quic_ctx_id_t) <= 42, "ctx id must be less than 42"); -// This structure is used to implement the concept of VPP connection for QUIC. -// We create one per connection and one per stream. +/* This structure is used to implement the concept of VPP connection for QUIC. + * We create one per connection and one per stream. */ typedef struct quic_ctx_ { union @@ -105,8 +96,9 @@ typedef struct quic_stream_data_ typedef struct quic_worker_ctx_ { CLIB_CACHE_LINE_ALIGN_MARK (cacheline0); - u32 time_now; /**< worker time */ + int64_t time_now; /**< worker time */ tw_timer_wheel_1t_3w_1024sl_ov_t timer_wheel; /**< worker timer wheel */ + u32 *opening_ctx_pool; } quic_worker_ctx_t; typedef struct quic_main_ @@ -114,7 +106,9 @@ typedef struct quic_main_ u32 app_index; quic_ctx_t **ctx_pool; quic_worker_ctx_t *wrk_ctx; + clib_bihash_16_8_t connection_hash; /* quicly connection id -> conn handle */ f64 tstamp_ticks_per_clock; + u32 fake_app_listener_index; /* ugly hack for accept cb */ /* * Config diff --git a/src/tests/vnet/session/quic_echo.c b/src/tests/vnet/session/quic_echo.c index 7d0cedd1d6b..29b33f5aeb7 100644 --- a/src/tests/vnet/session/quic_echo.c +++ b/src/tests/vnet/session/quic_echo.c @@ -38,10 +38,10 @@ #include <vpp/api/vpe_all_api_h.h> #undef vl_printfun -#define TCP_ECHO_DBG 0 -#define DBG(_fmt,_args...) \ - if (TCP_ECHO_DBG) \ - clib_warning (_fmt, _args) +#define QUIC_ECHO_DBG 0 +#define DBG(_fmt, _args...) \ + if (QUIC_ECHO_DBG) \ + clib_warning (_fmt, ##_args) typedef struct { @@ -68,6 +68,13 @@ typedef enum STATE_DETACHED } connection_state_t; +enum quic_session_type_t +{ + QUIC_SESSION_TYPE_QUIC = 0, + QUIC_SESSION_TYPE_STREAM = 1, + QUIC_SESSION_TYPE_LISTEN = INT32_MAX, +}; + typedef struct { /* vpe input queue */ @@ -85,6 +92,10 @@ typedef struct /* Hash table for disconnect processing */ uword *session_index_by_vpp_handles; + /* Hash table for shared segment_names */ + uword *shared_segment_names; + clib_spinlock_t segment_names_lock; + /* intermediate rx buffer */ u8 *rx_buf; @@ -140,6 +151,7 @@ typedef struct * vpp. If sock api is used, shm binary api is subsequently bootstrapped * and all other messages are exchanged using shm IPC. */ u8 use_sock_api; + int max_test_msg; fifo_segment_main_t segment_main; } echo_main_t; @@ -182,19 +194,54 @@ init_error_string_table (echo_main_t * em) static void handle_mq_event (session_event_t * e); +#if CLIB_DEBUG > 0 +#define TIMEOUT 10.0 +#else +#define TIMEOUT 10.0 +#endif + +static int +wait_for_segment_allocation (u64 segment_handle) +{ + echo_main_t *em = &echo_main; + f64 timeout; + timeout = clib_time_now (&em->clib_time) + TIMEOUT; + uword *segment_present; + DBG ("ASKING for %lu", segment_handle); + while (clib_time_now (&em->clib_time) < timeout) + { + clib_spinlock_lock (&em->segment_names_lock); + segment_present = hash_get (em->shared_segment_names, segment_handle); + clib_spinlock_unlock (&em->segment_names_lock); + if (segment_present != 0) + return 0; + if (em->time_to_stop == 1) + return 0; + } + DBG ("timeout waiting for segment_allocation %lu", segment_handle); + return -1; +} + +static int +wait_for_disconnected_sessions (echo_main_t * em) +{ + f64 timeout; + timeout = clib_time_now (&em->clib_time) + TIMEOUT; + while (clib_time_now (&em->clib_time) < timeout) + { + if (hash_elts (em->session_index_by_vpp_handles) == 0) + return 0; + } + DBG ("timeout waiting for disconnected_sessions"); + return -1; +} + static int wait_for_state_change (echo_main_t * em, connection_state_t state) { svm_msg_q_msg_t msg; session_event_t *e; f64 timeout; - -#if CLIB_DEBUG > 0 -#define TIMEOUT 600.0 -#else -#define TIMEOUT 600.0 -#endif - timeout = clib_time_now (&em->clib_time) + TIMEOUT; while (clib_time_now (&em->clib_time) < timeout) @@ -306,7 +353,6 @@ ssvm_segment_attach (char *name, ssvm_segment_type_t type, int fd) clib_warning ("svm_fifo_segment_attach ('%s') failed", name); return rv; } - vec_reset_length (a->new_segment_indices); return 0; } @@ -318,6 +364,9 @@ vl_api_application_attach_reply_t_handler (vl_api_application_attach_reply_t * echo_main_t *em = &echo_main; int *fds = 0; u32 n_fds = 0; + u64 segment_handle; + segment_handle = clib_net_to_host_u64 (mp->segment_handle); + DBG ("Attached returned app %u", htons (mp->app_index)); if (mp->retval) { @@ -361,6 +410,10 @@ vl_api_application_attach_reply_t_handler (vl_api_application_attach_reply_t * -1)) goto failed; } + DBG ("SETTING for %lu", segment_handle); + clib_spinlock_lock (&em->segment_names_lock); + hash_set (em->shared_segment_names, segment_handle, 1); + clib_spinlock_unlock (&em->segment_names_lock); em->state = STATE_ATTACHED; return; @@ -456,7 +509,26 @@ vl_api_map_another_segment_t_handler (vl_api_map_another_segment_t * mp) { fifo_segment_main_t *sm = &echo_main.segment_main; fifo_segment_create_args_t _a, *a = &_a; + echo_main_t *em = &echo_main; int rv; + int *fds = 0; + + if (mp->fd_flags & SESSION_FD_F_MEMFD_SEGMENT) + { + vec_validate (fds, 1); + vl_socket_client_recv_fd_msg (fds, 1, 5); + if (ssvm_segment_attach + ((char *) mp->segment_name, SSVM_SEGMENT_MEMFD, fds[0])) + clib_warning + ("svm_fifo_segment_attach ('%s') failed on SSVM_SEGMENT_MEMFD", + mp->segment_name); + DBG ("SETTING for %lu", mp->segment_name); + clib_spinlock_lock (&em->segment_names_lock); + hash_set (em->shared_segment_names, mp->segment_name, 1); + clib_spinlock_unlock (&em->segment_names_lock); + vec_free (fds); + return; + } clib_memset (a, 0, sizeof (*a)); a->segment_name = (char *) mp->segment_name; @@ -471,6 +543,9 @@ vl_api_map_another_segment_t_handler (vl_api_map_another_segment_t * mp) } clib_warning ("Mapped new segment '%s' size %d", mp->segment_name, mp->segment_size); + clib_spinlock_lock (&em->segment_names_lock); + hash_set (em->shared_segment_names, mp->segment_name, 1); + clib_spinlock_unlock (&em->segment_names_lock); } static void @@ -486,16 +561,21 @@ session_print_stats (echo_main_t * em, echo_session_t * session) } static void -test_recv_bytes (echo_session_t * s, u8 * rx_buf, u32 n_read) +test_recv_bytes (echo_main_t * em, echo_session_t * s, u8 * rx_buf, + u32 n_read) { int i; for (i = 0; i < n_read; i++) { - if (rx_buf[i] != ((s->bytes_received + i) & 0xff)) + if (rx_buf[i] != ((s->bytes_received + i) & 0xff) + && em->max_test_msg > 0) { clib_warning ("error at byte %lld, 0x%x not 0x%x", s->bytes_received + i, rx_buf[i], ((s->bytes_received + i) & 0xff)); + em->max_test_msg--; + if (em->max_test_msg == 0) + clib_warning ("Too many errors, hiding next ones"); } } } @@ -517,12 +597,13 @@ recv_data_chunk (echo_main_t * em, echo_session_t * s, u8 * rx_buf) if (n_read > 0) { if (em->test_return_packets) - test_recv_bytes (s, rx_buf, n_read); + test_recv_bytes (em, s, rx_buf, n_read); n_to_read -= n_read; s->bytes_received += n_read; s->bytes_to_receive -= n_read; + ASSERT (s->bytes_to_receive >= 0); } else break; @@ -577,8 +658,9 @@ client_thread_fn (void *arg) break; } - clib_warning ("GOT OUT"); - DBG ("session %d done", session_index); + DBG ("session %d done send %lu to do, %lu done || recv %lu to do, %lu done", + session_index, s->bytes_to_send, s->bytes_sent, s->bytes_to_receive, + s->bytes_received); em->tx_total += s->bytes_sent; em->rx_total += s->bytes_received; em->n_active_clients--; @@ -587,7 +669,7 @@ client_thread_fn (void *arg) } void -client_send_connect (echo_main_t * em) +client_send_connect (echo_main_t * em, u8 * uri, u32 opaque) { vl_api_connect_uri_t *cmp; cmp = vl_msg_api_alloc (sizeof (*cmp)); @@ -595,8 +677,8 @@ client_send_connect (echo_main_t * em) cmp->_vl_msg_id = ntohs (VL_API_CONNECT_URI); cmp->client_index = em->my_client_index; - cmp->context = ntohl (0xfeedface); - memcpy (cmp->uri, em->connect_uri, vec_len (em->connect_uri)); + cmp->context = ntohl (opaque); + memcpy (cmp->uri, uri, vec_len (uri)); vl_msg_api_send_shmem (em->vl_input_queue, (u8 *) & cmp); } @@ -641,6 +723,13 @@ session_bound_handler (session_bound_msg_t * mp) } static void +quic_qsession_accepted_handler (session_accepted_msg_t * mp) +{ + DBG ("Accept on QSession index %u", mp->handle); +} + + +static void session_accepted_handler (session_accepted_msg_t * mp) { app_session_evt_t _app_evt, *app_evt = &_app_evt; @@ -650,8 +739,11 @@ session_accepted_handler (session_accepted_msg_t * mp) echo_session_t *session; static f64 start_time; u32 session_index; + u64 segment_handle; u8 *ip_str; + segment_handle = mp->segment_handle; + if (start_time == 0.0) start_time = clib_time_now (&em->clib_time); @@ -662,7 +754,14 @@ session_accepted_handler (session_accepted_msg_t * mp) /* Allocate local session and set it up */ pool_get (em->sessions, session); session_index = session - em->sessions; + DBG ("Setting session_index %lu", session_index); + if (wait_for_segment_allocation (segment_handle)) + { + clib_warning ("timeout waiting for segment allocation %lu", + segment_handle); + return; + } rx_fifo = uword_to_pointer (mp->server_rx_fifo, svm_fifo_t *); rx_fifo->client_session_index = session_index; tx_fifo = uword_to_pointer (mp->server_tx_fifo, svm_fifo_t *); @@ -670,12 +769,28 @@ session_accepted_handler (session_accepted_msg_t * mp) session->rx_fifo = rx_fifo; session->tx_fifo = tx_fifo; + session->vpp_session_handle = mp->handle; session->vpp_evt_q = uword_to_pointer (mp->vpp_event_queue_address, svm_msg_q_t *); /* Add it to lookup table */ hash_set (em->session_index_by_vpp_handles, mp->handle, session_index); + /* + * Send accept reply to vpp + */ + app_alloc_ctrl_evt_to_vpp (session->vpp_evt_q, app_evt, + SESSION_CTRL_EVT_ACCEPTED_REPLY); + rmp = (session_accepted_reply_msg_t *) app_evt->evt->data; + rmp->handle = mp->handle; + rmp->context = mp->context; + app_send_ctrl_evt_to_vpp (session->vpp_evt_q, app_evt); + + /* TODO : this is very ugly */ + if (mp->rmt.is_ip4 != 255) + return quic_qsession_accepted_handler (mp); + DBG ("SSession handle is %lu", mp->handle); + em->state = STATE_READY; /* Stats printing */ @@ -687,21 +802,20 @@ session_accepted_handler (session_accepted_msg_t * mp) (f64) pool_elts (em->sessions) / (now - start_time)); } - /* - * Send accept reply to vpp - */ - app_alloc_ctrl_evt_to_vpp (session->vpp_evt_q, app_evt, - SESSION_CTRL_EVT_ACCEPTED_REPLY); - rmp = (session_accepted_reply_msg_t *) app_evt->evt->data; - rmp->handle = mp->handle; - rmp->context = mp->context; - app_send_ctrl_evt_to_vpp (session->vpp_evt_q, app_evt); - session->bytes_received = 0; session->start = clib_time_now (&em->clib_time); } static void +quic_session_connected_handler (session_connected_msg_t * mp) +{ + echo_main_t *em = &echo_main; + u8 *uri = format (0, "QUIC://session/%lu", mp->handle); + DBG ("QSession Connect : %s", uri); + client_send_connect (em, uri, QUIC_SESSION_TYPE_STREAM); +} + +static void session_connected_handler (session_connected_msg_t * mp) { echo_main_t *em = &echo_main; @@ -709,6 +823,8 @@ session_connected_handler (session_connected_msg_t * mp) u32 session_index; svm_fifo_t *rx_fifo, *tx_fifo; int rv; + u64 segment_handle; + segment_handle = mp->segment_handle; if (mp->retval) { @@ -725,7 +841,14 @@ session_connected_handler (session_connected_msg_t * mp) pool_get (em->sessions, session); clib_memset (session, 0, sizeof (*session)); session_index = session - em->sessions; + DBG ("Setting session_index %lu", session_index); + if (wait_for_segment_allocation (segment_handle)) + { + clib_warning ("timeout waiting for segment allocation %lu", + segment_handle); + return; + } rx_fifo = uword_to_pointer (mp->server_rx_fifo, svm_fifo_t *); rx_fifo->client_session_index = session_index; tx_fifo = uword_to_pointer (mp->server_tx_fifo, svm_fifo_t *); @@ -740,6 +863,11 @@ session_connected_handler (session_connected_msg_t * mp) hash_set (em->session_index_by_vpp_handles, mp->handle, session_index); + if (mp->context == QUIC_SESSION_TYPE_QUIC) + return quic_session_connected_handler (mp); + + DBG ("SSession Connected"); + /* * Start RX thread */ @@ -768,6 +896,7 @@ session_disconnected_handler (session_disconnected_msg_t * mp) echo_session_t *session = 0; uword *p; int rv = 0; + DBG ("Got a SESSION_CTRL_EVT_DISCONNECTED for session %lu", mp->handle); p = hash_get (em->session_index_by_vpp_handles, mp->handle); if (!p) @@ -778,6 +907,7 @@ session_disconnected_handler (session_disconnected_msg_t * mp) session = pool_elt_at_index (em->sessions, p[0]); hash_unset (em->session_index_by_vpp_handles, mp->handle); + pool_put (em->sessions, session); app_alloc_ctrl_evt_to_vpp (session->vpp_evt_q, app_evt, @@ -830,18 +960,23 @@ handle_mq_event (session_event_t * e) switch (e->event_type) { case SESSION_CTRL_EVT_BOUND: + DBG ("SESSION_CTRL_EVT_BOUND"); session_bound_handler ((session_bound_msg_t *) e->data); break; case SESSION_CTRL_EVT_ACCEPTED: + DBG ("SESSION_CTRL_EVT_ACCEPTED"); session_accepted_handler ((session_accepted_msg_t *) e->data); break; case SESSION_CTRL_EVT_CONNECTED: + DBG ("SESSION_CTRL_EVT_CONNECTED"); session_connected_handler ((session_connected_msg_t *) e->data); break; case SESSION_CTRL_EVT_DISCONNECTED: + DBG ("SESSION_CTRL_EVT_DISCONNECTED"); session_disconnected_handler ((session_disconnected_msg_t *) e->data); break; case SESSION_CTRL_EVT_RESET: + DBG ("SESSION_CTRL_EVT_RESET"); session_reset_handler ((session_reset_msg_t *) e->data); break; default: @@ -856,6 +991,7 @@ clients_run (echo_main_t * em) svm_msg_q_msg_t msg; session_event_t *e; echo_session_t *s; + hash_pair_t *p; int i; /* Init test data */ @@ -870,7 +1006,7 @@ clients_run (echo_main_t * em) return; for (i = 0; i < em->n_clients; i++) - client_send_connect (em); + client_send_connect (em, em->connect_uri, QUIC_SESSION_TYPE_QUIC); start_time = clib_time_now (&em->clib_time); while (em->n_clients_connected < em->n_clients @@ -897,18 +1033,23 @@ clients_run (echo_main_t * em) /* * Initialize connections */ - for (i = 0; i < em->n_clients; i++) - { - s = pool_elt_at_index (em->sessions, i); + DBG ("Initialize connections on %u clients", em->n_clients); + + /* *INDENT-OFF* */ + hash_foreach_pair (p, em->session_index_by_vpp_handles, + ({ + s = pool_elt_at_index (em->sessions, p->value[0]); s->bytes_to_send = em->bytes_to_send; if (!em->no_return) s->bytes_to_receive = em->bytes_to_send; - } + })); + /* *INDENT-ON* */ em->n_active_clients = em->n_clients_connected; /* * Wait for client threads to send the data */ + DBG ("Waiting for data on %u clients", em->n_active_clients); start_time = clib_time_now (&em->clib_time); em->state = STATE_READY; while (em->n_active_clients) @@ -925,11 +1066,14 @@ clients_run (echo_main_t * em) svm_msg_q_free_msg (em->our_event_queue, &msg); } - for (i = 0; i < em->n_clients; i++) - { - s = pool_elt_at_index (em->sessions, i); + /* *INDENT-OFF* */ + hash_foreach_pair (p, em->session_index_by_vpp_handles, + ({ + s = pool_elt_at_index (em->sessions, p->value[0]); + DBG ("Sending disconnect on session %lu", p->key); client_disconnect (em, s); - } + })); + /* *INDENT-ON* */ /* * Stats and detach @@ -940,6 +1084,7 @@ clients_run (echo_main_t * em) em->tx_total / (1ULL << 30), deltat); fformat (stdout, "%.4f Gbit/second\n", (em->tx_total * 8.0) / deltat / 1e9); + wait_for_disconnected_sessions (em); application_detach (em); } @@ -1059,7 +1204,6 @@ server_handle_rx (echo_main_t * em, session_event_t * e) int n_read, max_dequeue, n_sent; u32 offset, to_dequeue; echo_session_t *s; - s = pool_elt_at_index (em->sessions, e->session_index); /* Clear event only once. Otherwise, if we do it in the loop by calling @@ -1070,7 +1214,6 @@ server_handle_rx (echo_main_t * em, session_event_t * e) max_dequeue = svm_fifo_max_dequeue (s->rx_fifo); if (PREDICT_FALSE (!max_dequeue)) return; - do { /* The options here are to limit ourselves to max_dequeue or read @@ -1079,8 +1222,12 @@ server_handle_rx (echo_main_t * em, session_event_t * e) to_dequeue = clib_min (max_dequeue, vec_len (em->rx_buf)); n_read = app_recv_stream_raw (s->rx_fifo, em->rx_buf, to_dequeue, 0 /* clear evt */ , 0 /* peek */ ); + if (n_read > 0) { + if (em->test_return_packets) + test_recv_bytes (em, s, em->rx_buf, n_read); + max_dequeue -= n_read; s->bytes_received += n_read; } @@ -1127,10 +1274,10 @@ server_handle_mq (echo_main_t * em) if (rc == ETIMEDOUT) continue; e = svm_msg_q_msg_data (em->our_event_queue, &msg); - clib_warning ("Event %d", e->event_type); switch (e->event_type) { - case FIFO_EVENT_APP_RX: + case SESSION_IO_EVT_RX: + DBG ("SESSION_IO_EVT_RX"); server_handle_rx (em, e); break; default: @@ -1220,6 +1367,7 @@ vl_api_disconnect_session_reply_t_handler (vl_api_disconnect_session_reply_t * { echo_main_t *em = &echo_main; uword *p; + DBG ("Got disonnected reply for session %lu", mp->handle); if (mp->retval) { @@ -1300,12 +1448,14 @@ main (int argc, char **argv) clib_memset (em, 0, sizeof (*em)); em->session_index_by_vpp_handles = hash_create (0, sizeof (uword)); + em->shared_segment_names = hash_create (0, sizeof (uword)); + clib_spinlock_init (&em->segment_names_lock); em->my_pid = getpid (); - em->configured_segment_size = 1 << 20; em->socket_name = 0; em->use_sock_api = 1; em->fifo_size = 64 << 10; em->n_clients = 1; + em->max_test_msg = 50; clib_time_init (&em->clib_time); init_error_string_table (em); @@ -1320,10 +1470,6 @@ main (int argc, char **argv) } else if (unformat (a, "uri %s", &uri)) ; - else if (unformat (a, "segment-size %dM", &tmp)) - em->configured_segment_size = tmp << 20; - else if (unformat (a, "segment-size %dG", &tmp)) - em->configured_segment_size = tmp << 30; else if (unformat (a, "server")) i_am_server = 1; else if (unformat (a, "client")) |