diff options
Diffstat (limited to 'src/plugins/hs_apps/sapi/vpp_echo.c')
-rw-r--r-- | src/plugins/hs_apps/sapi/vpp_echo.c | 1049 |
1 files changed, 1049 insertions, 0 deletions
diff --git a/src/plugins/hs_apps/sapi/vpp_echo.c b/src/plugins/hs_apps/sapi/vpp_echo.c new file mode 100644 index 00000000000..5c9690f299f --- /dev/null +++ b/src/plugins/hs_apps/sapi/vpp_echo.c @@ -0,0 +1,1049 @@ +/* + * Copyright (c) 2019 Cisco and/or its affiliates. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at: + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +#include <stdio.h> +#include <signal.h> + +#include <vlibmemory/api.h> +#include <svm/fifo_segment.h> + +#include <hs_apps/sapi/vpp_echo_common.h> + +echo_main_t echo_main; + +static void +echo_session_prealloc (echo_main_t * em) +{ + /* We need to prealloc to avoid vec resize in threads */ + echo_session_t *session; + int i; + for (i = 0; i < em->n_sessions; i++) + { + pool_get (em->sessions, session); + clib_memset (session, 0, sizeof (*session)); + session->session_index = session - em->sessions; + session->listener_index = SESSION_INVALID_INDEX; + session->session_state = ECHO_SESSION_STATE_INITIAL; + } +} + +static void +echo_assert_test_suceeded (echo_main_t * em) +{ + CHECK (em->n_clients * em->bytes_to_receive, + em->stats.rx_total, "Not enough data received"); + CHECK (em->n_clients * em->bytes_to_send, + em->stats.tx_total, "Not enough data sent"); + clib_spinlock_lock (&em->sid_vpp_handles_lock); + CHECK (0, hash_elts (em->session_index_by_vpp_handles), + "Some sessions are still open"); + clib_spinlock_unlock (&em->sid_vpp_handles_lock); +} + +always_inline void +echo_session_dequeue_notify (echo_session_t * s) +{ + int rv; + if (!svm_fifo_set_event (s->rx_fifo)) + return; + if ((rv = + app_send_io_evt_to_vpp (s->vpp_evt_q, s->rx_fifo->master_session_index, + SESSION_IO_EVT_RX, SVM_Q_WAIT))) + ECHO_FAIL ("app_send_io_evt_to_vpp errored %d", rv); + svm_fifo_clear_deq_ntf (s->rx_fifo); +} + +static void +stop_signal (int signum) +{ + echo_main_t *em = &echo_main; + em->time_to_stop = 1; +} + +int +connect_to_vpp (char *name) +{ + echo_main_t *em = &echo_main; + api_main_t *am = &api_main; + + if (em->use_sock_api) + { + if (vl_socket_client_connect ((char *) em->socket_name, name, + 0 /* default rx, tx buffer */ )) + { + ECHO_FAIL ("socket connect failed"); + return -1; + } + + if (vl_socket_client_init_shm (0, 1 /* want_pthread */ )) + { + ECHO_FAIL ("init shm api failed"); + return -1; + } + } + else + { + if (vl_client_connect_to_vlib ("/vpe-api", name, 32) < 0) + { + ECHO_FAIL ("shmem connect failed"); + return -1; + } + } + em->vl_input_queue = am->shmem_hdr->vl_input_queue; + em->my_client_index = am->my_client_index; + return 0; +} + +static void +echo_event_didnt_happen (u8 e) +{ + echo_main_t *em = &echo_main; + u8 *s = format (0, "%U", echo_format_timing_event, e); + ECHO_LOG (0, "Expected event %s to happend, which did not", s); + em->has_failed = 1; +} + +static void +print_global_json_stats (echo_main_t * em) +{ + if (!(em->timing.events_sent & em->timing.start_event)) + return echo_event_didnt_happen (em->timing.start_event); + if (!(em->timing.events_sent & em->timing.end_event)) + return echo_event_didnt_happen (em->timing.end_event); + f64 deltat = em->timing.end_time - em->timing.start_time; + u8 *start_evt = + format (0, "%U", echo_format_timing_event, em->timing.start_event); + u8 *end_evt = + format (0, "%U", echo_format_timing_event, em->timing.end_event); + fformat (stdout, "{\n"); + fformat (stdout, "\"time\": \"%.9f\",\n", deltat); + fformat (stdout, "\"start_evt\": \"%s\",\n", start_evt); + fformat (stdout, "\"end_evt\": \"%s\",\n", end_evt); + fformat (stdout, "\"rx_data\": %lld,\n", em->stats.rx_total); + fformat (stdout, "\"tx_rx\": %lld,\n", em->stats.tx_total); + fformat (stdout, "\"closing\": {\n"); + fformat (stdout, " \"reset\": { \"q\": %d, \"s\": %d },\n", + em->stats.reset_count.q, em->stats.reset_count.s); + fformat (stdout, " \"close\": { \"q\": %d, \"s\": %d },\n", + em->stats.close_count.q, em->stats.close_count.s); + fformat (stdout, " \"active\": { \"q\": %d, \"s\": %d },\n", + em->stats.active_count.q, em->stats.active_count.s); + fformat (stdout, " \"clean\": { \"q\": %d, \"s\": %d }\n", + em->stats.clean_count.q, em->stats.clean_count.s); + fformat (stdout, "}\n"); + fformat (stdout, "}\n"); +} + +static void +print_global_stats (echo_main_t * em) +{ + u8 *s; + if (!(em->timing.events_sent & em->timing.start_event)) + return echo_event_didnt_happen (em->timing.start_event); + if (!(em->timing.events_sent & em->timing.end_event)) + return echo_event_didnt_happen (em->timing.end_event); + f64 deltat = em->timing.end_time - em->timing.start_time; + s = format (0, "%U:%U", + echo_format_timing_event, em->timing.start_event, + echo_format_timing_event, em->timing.end_event); + fformat (stdout, "Timing %s\n", s); + fformat (stdout, "-------- TX --------\n"); + fformat (stdout, "%lld bytes (%lld mbytes, %lld gbytes) in %.6f seconds\n", + em->stats.tx_total, em->stats.tx_total / (1ULL << 20), + em->stats.tx_total / (1ULL << 30), deltat); + fformat (stdout, "%.4f Gbit/second\n", + (em->stats.tx_total * 8.0) / deltat / 1e9); + fformat (stdout, "-------- RX --------\n"); + fformat (stdout, "%lld bytes (%lld mbytes, %lld gbytes) in %.6f seconds\n", + em->stats.rx_total, em->stats.rx_total / (1ULL << 20), + em->stats.rx_total / (1ULL << 30), deltat); + fformat (stdout, "%.4f Gbit/second\n", + (em->stats.rx_total * 8.0) / deltat / 1e9); + fformat (stdout, "--------------------\n"); + fformat (stdout, "Received close on %d streams (and %d Quic conn)\n", + em->stats.close_count.s, em->stats.close_count.q); + fformat (stdout, "Received reset on %d streams (and %d Quic conn)\n", + em->stats.reset_count.s, em->stats.reset_count.q); + fformat (stdout, "Sent close on %d streams (and %d Quic conn)\n", + em->stats.active_count.s, em->stats.active_count.q); + fformat (stdout, "Discarded %d streams (and %d Quic conn)\n", + em->stats.clean_count.s, em->stats.clean_count.q); +} + +void +echo_update_count_on_session_close (echo_main_t * em, echo_session_t * s) +{ + + ECHO_LOG (1, "[%lu/%lu] -> S(%x) -> [%lu/%lu]", + s->bytes_received, s->bytes_received + s->bytes_to_receive, + s->session_index, s->bytes_sent, + s->bytes_sent + s->bytes_to_send); + clib_atomic_fetch_add (&em->stats.tx_total, s->bytes_sent); + clib_atomic_fetch_add (&em->stats.rx_total, s->bytes_received); + + if (PREDICT_FALSE (em->stats.rx_total == + em->n_clients * em->bytes_to_receive)) + echo_notify_event (em, ECHO_EVT_LAST_BYTE); +} + +static void +echo_free_sessions (echo_main_t * em) +{ + /* Free marked sessions */ + echo_session_t *s; + u32 *session_indexes = 0, *session_index; + + /* *INDENT-OFF* */ + pool_foreach (s, em->sessions, + ({ + if (s->session_state == ECHO_SESSION_STATE_CLOSED) + vec_add1 (session_indexes, s->session_index);} + )); + /* *INDENT-ON* */ + vec_foreach (session_index, session_indexes) + { + /* Free session */ + s = pool_elt_at_index (em->sessions, *session_index); + echo_session_handle_add_del (em, s->vpp_session_handle, + SESSION_INVALID_INDEX); + pool_put (em->sessions, s); + clib_memset (s, 0xfe, sizeof (*s)); + } +} + +static void +test_recv_bytes (echo_main_t * em, echo_session_t * s, u8 * rx_buf, + u32 n_read) +{ + u32 i; + u8 expected; + for (i = 0; i < n_read; i++) + { + expected = (s->bytes_received + i) & 0xff; + if (rx_buf[i] == expected || em->max_test_msg > 0) + continue; + ECHO_LOG (0, "Session 0x%lx byte %lld was 0x%x expected 0x%x", + s->vpp_session_handle, s->bytes_received + i, rx_buf[i], + expected); + em->max_test_msg--; + if (em->max_test_msg == 0) + ECHO_LOG (0, "Too many errors, hiding next ones"); + if (em->test_return_packets == RETURN_PACKETS_ASSERT) + ECHO_FAIL ("test-bytes errored"); + } +} + +static int +recv_data_chunk (echo_main_t * em, echo_session_t * s, u8 * rx_buf) +{ + int n_read; + n_read = app_recv ((app_session_t *) s, rx_buf, vec_len (rx_buf)); + if (n_read <= 0) + return 0; + if (svm_fifo_needs_deq_ntf (s->rx_fifo, n_read)) + echo_session_dequeue_notify (s); + + if (em->test_return_packets) + test_recv_bytes (em, s, rx_buf, n_read); + + s->bytes_received += n_read; + s->bytes_to_receive -= n_read; + return n_read; +} + +static int +send_data_chunk (echo_session_t * s, u8 * tx_buf, int offset, int len) +{ + int n_sent; + int bytes_this_chunk = clib_min (s->bytes_to_send, len - offset); + if (!bytes_this_chunk) + return 0; + n_sent = app_send ((app_session_t *) s, tx_buf + offset, + bytes_this_chunk, SVM_Q_WAIT); + if (n_sent < 0) + return 0; + s->bytes_to_send -= n_sent; + s->bytes_sent += n_sent; + return n_sent; +} + +static int +mirror_data_chunk (echo_main_t * em, echo_session_t * s, u8 * tx_buf, u64 len) +{ + u64 n_sent = 0; + while (n_sent < len && !em->time_to_stop) + n_sent += send_data_chunk (s, tx_buf, n_sent, len); + return n_sent; +} + +static inline void +echo_check_closed_listener (echo_main_t * em, echo_session_t * s) +{ + echo_session_t *ls; + /* if parent has died, terminate gracefully */ + if (s->listener_index == SESSION_INVALID_INDEX) + return; + ls = pool_elt_at_index (em->sessions, s->listener_index); + if (ls->session_state < ECHO_SESSION_STATE_CLOSING) + return; + ECHO_LOG (2, "Session 0%lx died, close child 0x%lx", ls->vpp_session_handle, + s->vpp_session_handle); + echo_update_count_on_session_close (em, s); + em->proto_cb_vft->cleanup_cb (s, 1 /* parent_died */ ); +} + +/* + * Rx/Tx polling thread per connection + */ +static void +echo_handle_data (echo_main_t * em, echo_session_t * s, u8 * rx_buf) +{ + int n_read, n_sent = 0; + + n_read = recv_data_chunk (em, s, rx_buf); + if (em->data_source == ECHO_TEST_DATA_SOURCE) + n_sent = send_data_chunk (s, em->connect_test_data, + s->bytes_sent % em->tx_buf_size, + em->tx_buf_size); + else if (em->data_source == ECHO_RX_DATA_SOURCE) + n_sent = mirror_data_chunk (em, s, rx_buf, n_read); + if (!s->bytes_to_send && !s->bytes_to_receive) + { + /* Session is done, need to close */ + if (s->session_state == ECHO_SESSION_STATE_AWAIT_DATA) + s->session_state = ECHO_SESSION_STATE_CLOSING; + else + { + s->session_state = ECHO_SESSION_STATE_AWAIT_CLOSING; + if (em->send_stream_disconnects == ECHO_CLOSE_F_ACTIVE) + { + echo_send_rpc (em, echo_send_disconnect_session, + (void *) s->vpp_session_handle, 0); + clib_atomic_fetch_add (&em->stats.active_count.s, 1); + } + else if (em->send_stream_disconnects == ECHO_CLOSE_F_NONE) + { + s->session_state = ECHO_SESSION_STATE_CLOSING; + clib_atomic_fetch_add (&em->stats.clean_count.s, 1); + } + } + return; + } + + /* Check for idle clients */ + if (em->log_lvl > 1) + { + if (n_sent || n_read) + s->idle_cycles = 0; + else if (s->idle_cycles++ == 1e7) + { + s->idle_cycles = 0; + ECHO_LOG (1, "Idle client TX:%dB RX:%dB", s->bytes_to_send, + s->bytes_to_receive); + ECHO_LOG (1, "Idle FIFOs TX:%dB RX:%dB", + svm_fifo_max_dequeue (s->tx_fifo), + svm_fifo_max_dequeue (s->rx_fifo)); + ECHO_LOG (1, "Session 0x%lx state %u", s->vpp_session_handle, + s->session_state); + } + } +} + +static void * +echo_data_thread_fn (void *arg) +{ + clib_mem_set_thread_index (); + echo_main_t *em = &echo_main; + u32 N = em->n_clients; + u32 n = (N + em->n_rx_threads - 1) / em->n_rx_threads; + u32 idx = (u64) arg; + if (n * idx >= N) + { + ECHO_LOG (1, "Thread %u exiting, no sessions to care for", idx); + pthread_exit (0); + } + u32 thread_n_sessions = clib_min (n, N - n * idx); + + u32 i = 0; + u32 n_closed_sessions = 0; + u32 session_index; + u8 *rx_buf = 0; + echo_session_t *s; + vec_validate (rx_buf, em->rx_buf_size); + + for (i = 0; !em->time_to_stop; i = (i + 1) % thread_n_sessions) + { + n_closed_sessions = i == 0 ? 0 : n_closed_sessions; + session_index = em->data_thread_args[n * idx + i]; + if (session_index == SESSION_INVALID_INDEX) + continue; + s = pool_elt_at_index (em->sessions, session_index); + switch (s->session_state) + { + case ECHO_SESSION_STATE_READY: + case ECHO_SESSION_STATE_AWAIT_DATA: + echo_handle_data (em, s, rx_buf); + echo_check_closed_listener (em, s); + break; + case ECHO_SESSION_STATE_AWAIT_CLOSING: + echo_check_closed_listener (em, s); + break; + case ECHO_SESSION_STATE_CLOSING: + echo_update_count_on_session_close (em, s); + em->proto_cb_vft->cleanup_cb (s, 0 /* parent_died */ ); + break; + case ECHO_SESSION_STATE_CLOSED: + n_closed_sessions++; + break; + } + if (n_closed_sessions == thread_n_sessions) + break; + } + pthread_exit (0); +} + +static void +session_bound_handler (session_bound_msg_t * mp) +{ + echo_main_t *em = &echo_main; + echo_session_t *listen_session; + if (mp->retval) + { + ECHO_FAIL ("bind failed: %U", format_api_error, + clib_net_to_host_u32 (mp->retval)); + return; + } + ECHO_LOG (0, "listening on %U:%u", format_ip46_address, mp->lcl_ip, + mp->lcl_is_ip4 ? IP46_TYPE_IP4 : IP46_TYPE_IP6, + clib_net_to_host_u16 (mp->lcl_port)); + + /* Allocate local session and set it up */ + listen_session = echo_session_new (em); + listen_session->session_type = ECHO_SESSION_TYPE_LISTEN; + echo_session_handle_add_del (em, mp->handle, listen_session->session_index); + em->state = STATE_LISTEN; + em->listen_session_index = listen_session->session_index; + if (em->proto_cb_vft->bound_uri_cb) + em->proto_cb_vft->bound_uri_cb (mp, listen_session); +} + +static void +session_accepted_handler (session_accepted_msg_t * mp) +{ + app_session_evt_t _app_evt, *app_evt = &_app_evt; + session_accepted_reply_msg_t *rmp; + svm_fifo_t *rx_fifo, *tx_fifo; + echo_main_t *em = &echo_main; + echo_session_t *session, *ls; + /* Allocate local session and set it up */ + session = echo_session_new (em); + + if (wait_for_segment_allocation (mp->segment_handle)) + { + ECHO_FAIL ("wait_for_segment_allocation errored"); + return; + } + + rx_fifo = uword_to_pointer (mp->server_rx_fifo, svm_fifo_t *); + rx_fifo->client_session_index = session->session_index; + tx_fifo = uword_to_pointer (mp->server_tx_fifo, svm_fifo_t *); + tx_fifo->client_session_index = session->session_index; + + session->rx_fifo = rx_fifo; + session->tx_fifo = tx_fifo; + + /* session->transport needed by app_send_dgram */ + clib_memcpy_fast (&session->transport.rmt_ip, &mp->rmt.ip, + sizeof (ip46_address_t)); + session->transport.is_ip4 = mp->rmt.is_ip4; + session->transport.rmt_port = mp->rmt.port; + clib_memcpy_fast (&session->transport.lcl_ip, &em->uri_elts.ip, + sizeof (ip46_address_t)); + session->transport.lcl_port = em->uri_elts.port; + + session->vpp_session_handle = mp->handle; + session->start = clib_time_now (&em->clib_time); + session->vpp_evt_q = uword_to_pointer (mp->vpp_event_queue_address, + svm_msg_q_t *); + if (!(ls = echo_get_session_from_handle (em, mp->listener_handle))) + return; + session->listener_index = ls->session_index; + + /* Add it to lookup table */ + ECHO_LOG (1, "Accepted session 0x%lx -> 0x%lx", mp->handle, + mp->listener_handle); + echo_session_handle_add_del (em, mp->handle, session->session_index); + + app_alloc_ctrl_evt_to_vpp (session->vpp_evt_q, app_evt, + SESSION_CTRL_EVT_ACCEPTED_REPLY); + rmp = (session_accepted_reply_msg_t *) app_evt->evt->data; + rmp->handle = mp->handle; + rmp->context = mp->context; + app_send_ctrl_evt_to_vpp (session->vpp_evt_q, app_evt); + em->proto_cb_vft->accepted_cb (mp, session); +} + +static void +session_connected_handler (session_connected_msg_t * mp) +{ + echo_main_t *em = &echo_main; + echo_session_t *session; + u32 listener_index = htonl (mp->context); + svm_fifo_t *rx_fifo, *tx_fifo; + + if (mp->retval) + { + ECHO_FAIL ("connection failed with code: %U", format_api_error, + clib_net_to_host_u32 (mp->retval)); + return; + } + + session = echo_session_new (em); + if (wait_for_segment_allocation (mp->segment_handle)) + { + ECHO_FAIL ("wait_for_segment_allocation errored"); + return; + } + + rx_fifo = uword_to_pointer (mp->server_rx_fifo, svm_fifo_t *); + rx_fifo->client_session_index = session->session_index; + tx_fifo = uword_to_pointer (mp->server_tx_fifo, svm_fifo_t *); + tx_fifo->client_session_index = session->session_index; + + session->rx_fifo = rx_fifo; + session->tx_fifo = tx_fifo; + session->vpp_session_handle = mp->handle; + session->start = clib_time_now (&em->clib_time); + session->vpp_evt_q = uword_to_pointer (mp->vpp_event_queue_address, + svm_msg_q_t *); + session->listener_index = listener_index; + /* session->transport needed by app_send_dgram */ + clib_memcpy_fast (&session->transport.lcl_ip, &mp->lcl.ip, + sizeof (ip46_address_t)); + session->transport.is_ip4 = mp->lcl.is_ip4; + session->transport.lcl_port = mp->lcl.port; + clib_memcpy_fast (&session->transport.rmt_ip, &em->uri_elts.ip, + sizeof (ip46_address_t)); + session->transport.rmt_port = em->uri_elts.port; + + echo_session_handle_add_del (em, mp->handle, session->session_index); + em->proto_cb_vft->connected_cb ((session_connected_bundled_msg_t *) mp, + session->session_index, 0 /* is_failed */ ); +} + +/* + * + * End of ECHO callback definitions + * + */ + +static void +session_disconnected_handler (session_disconnected_msg_t * mp) +{ + app_session_evt_t _app_evt, *app_evt = &_app_evt; + session_disconnected_reply_msg_t *rmp; + echo_main_t *em = &echo_main; + echo_session_t *s; + ECHO_LOG (1, "passive close session 0x%lx", mp->handle); + if (!(s = echo_get_session_from_handle (em, mp->handle))) + return; + em->proto_cb_vft->disconnected_cb (mp, s); + + app_alloc_ctrl_evt_to_vpp (s->vpp_evt_q, app_evt, + SESSION_CTRL_EVT_DISCONNECTED_REPLY); + rmp = (session_disconnected_reply_msg_t *) app_evt->evt->data; + rmp->retval = 0; + rmp->handle = mp->handle; + rmp->context = mp->context; + app_send_ctrl_evt_to_vpp (s->vpp_evt_q, app_evt); +} + +static void +session_reset_handler (session_reset_msg_t * mp) +{ + app_session_evt_t _app_evt, *app_evt = &_app_evt; + echo_main_t *em = &echo_main; + session_reset_reply_msg_t *rmp; + echo_session_t *s = 0; + ECHO_LOG (1, "Reset session 0x%lx", mp->handle); + if (!(s = echo_get_session_from_handle (em, mp->handle))) + return; + em->proto_cb_vft->reset_cb (mp, s); + + app_alloc_ctrl_evt_to_vpp (s->vpp_evt_q, app_evt, + SESSION_CTRL_EVT_RESET_REPLY); + rmp = (session_reset_reply_msg_t *) app_evt->evt->data; + rmp->retval = 0; + rmp->handle = mp->handle; + app_send_ctrl_evt_to_vpp (s->vpp_evt_q, app_evt); +} + +static void +handle_mq_event (session_event_t * e) +{ + switch (e->event_type) + { + case SESSION_CTRL_EVT_BOUND: + session_bound_handler ((session_bound_msg_t *) e->data); + break; + case SESSION_CTRL_EVT_ACCEPTED: + session_accepted_handler ((session_accepted_msg_t *) e->data); + break; + case SESSION_CTRL_EVT_CONNECTED: + session_connected_handler ((session_connected_msg_t *) e->data); + break; + case SESSION_CTRL_EVT_DISCONNECTED: + session_disconnected_handler ((session_disconnected_msg_t *) e->data); + break; + case SESSION_CTRL_EVT_RESET: + session_reset_handler ((session_reset_msg_t *) e->data); + break; + case SESSION_IO_EVT_RX: + break; + default: + ECHO_LOG (0, "unhandled event %u", e->event_type); + } +} + +static void +echo_process_rpcs (echo_main_t * em) +{ + echo_rpc_msg_t *rpc; + svm_msg_q_msg_t msg; + while (em->state < STATE_DATA_DONE && !em->time_to_stop) + { + if (svm_msg_q_sub (em->rpc_msq_queue, &msg, SVM_Q_TIMEDWAIT, 1)) + continue; + rpc = svm_msg_q_msg_data (em->rpc_msq_queue, &msg); + ((echo_rpc_t) rpc->fp) (rpc->arg, rpc->opaque); + svm_msg_q_free_msg (em->rpc_msq_queue, &msg); + } +} + +static void * +echo_mq_thread_fn (void *arg) +{ + clib_mem_set_thread_index (); + echo_main_t *em = &echo_main; + session_event_t *e; + svm_msg_q_msg_t msg; + int rv; + wait_for_state_change (em, STATE_ATTACHED, 0); + if (em->state < STATE_ATTACHED || !em->our_event_queue) + { + ECHO_FAIL ("Application failed to attach"); + pthread_exit (0); + } + + while (1) + { + if (!(rv = svm_msg_q_sub (em->our_event_queue, + &msg, SVM_Q_TIMEDWAIT, 1))) + { + e = svm_msg_q_msg_data (em->our_event_queue, &msg); + handle_mq_event (e); + svm_msg_q_free_msg (em->our_event_queue, &msg); + } + if (rv == ETIMEDOUT + && (em->time_to_stop || em->state == STATE_DETACHED)) + break; + } + pthread_exit (0); +} + +static void +clients_run (echo_main_t * em) +{ + u64 i; + echo_notify_event (em, ECHO_EVT_FIRST_QCONNECT); + for (i = 0; i < em->n_connects; i++) + echo_send_connect (em->uri, SESSION_INVALID_INDEX); + wait_for_state_change (em, STATE_READY, 0); + ECHO_LOG (1, "App is ready"); + echo_process_rpcs (em); +} + +static void +server_run (echo_main_t * em) +{ + echo_send_listen (em); + wait_for_state_change (em, STATE_READY, 0); + ECHO_LOG (1, "App is ready"); + echo_process_rpcs (em); + /* Cleanup */ + echo_send_unbind (em); + if (wait_for_state_change (em, STATE_DISCONNECTED, TIMEOUT)) + { + ECHO_FAIL ("Timeout waiting for state disconnected"); + return; + } +} + +static void +print_usage_and_exit (void) +{ + echo_main_t *em = &echo_main; + int i; + fprintf (stderr, + "Usage: vpp_echo [socket-name SOCKET] [client|server] [uri URI] [OPTIONS]\n" + "Generates traffic and assert correct teardown of the QUIC hoststack\n" + "\n" + " socket-name PATH Specify the binary socket path to connect to VPP\n" + " use-svm-api Use SVM API to connect to VPP\n" + " test-bytes[:assert] Check data correctness when receiving (assert fails on first error)\n" + " fifo-size N Use N Kb fifos\n" + " rx-buf N Use N Kb RX buffer\n" + " tx-buf N Use N Kb TX test buffer\n" + " appns NAMESPACE Use the namespace NAMESPACE\n" + " all-scope all-scope option\n" + " local-scope local-scope option\n" + " global-scope global-scope option\n" + " secret SECRET set namespace secret\n" + " chroot prefix PATH Use PATH as memory root path\n" + " sclose=[Y|N|W] When a stream is done, pass[N] send[Y] or wait[W] for close\n" + "\n" + " time START:END Time between evts START & END, events being :\n" + " start - Start of the app\n" + " qconnect - first Connection connect sent\n" + " qconnected - last Connection connected\n" + " sconnect - first Stream connect sent\n" + " sconnected - last Stream got connected\n" + " lastbyte - Last expected byte received\n" + " exit - Exiting of the app\n" + " json Output global stats in json\n" + " log=N Set the log level to [0: no output, 1:errors, 2:log]\n" + "\n" + " nclients N Open N clients sending data\n" + " nthreads N Use N busy loop threads for data [in addition to main & msg queue]\n" + " TX=1337[Kb|Mb|GB] Send 1337 [K|M|G]bytes, use TX=RX to reflect the data\n" + " RX=1337[Kb|Mb|GB] Expect 1337 [K|M|G]bytes\n" "\n"); + for (i = 0; i < TRANSPORT_N_PROTO; i++) + { + echo_proto_cb_vft_t *vft = em->available_proto_cb_vft[i]; + if (vft && vft->print_usage_cb) + vft->print_usage_cb (); + } + fprintf (stderr, "\nDefault configuration is :\n" + " server nclients 1/1 RX=64Kb TX=RX\n" + " client nclients 1/1 RX=64Kb TX=64Kb\n"); + exit (1); +} + +static int +echo_process_each_proto_opts (unformat_input_t * a) +{ + echo_main_t *em = &echo_main; + int i, rv; + for (i = 0; i < TRANSPORT_N_PROTO; i++) + { + echo_proto_cb_vft_t *vft = em->available_proto_cb_vft[i]; + if (vft && vft->process_opts_cb) + if ((rv = vft->process_opts_cb (a))) + return rv; + } + return 0; +} + +static void +echo_set_each_proto_defaults_before_opts (echo_main_t * em) +{ + int i; + for (i = 0; i < TRANSPORT_N_PROTO; i++) + { + echo_proto_cb_vft_t *vft = em->available_proto_cb_vft[i]; + if (vft && vft->set_defaults_before_opts_cb) + vft->set_defaults_before_opts_cb (); + } +} + +void +echo_process_opts (int argc, char **argv) +{ + echo_main_t *em = &echo_main; + unformat_input_t _argv, *a = &_argv; + u32 tmp; + u8 *chroot_prefix; + u8 *uri = 0; + u8 default_f_active; + + unformat_init_command_line (a, argv); + while (unformat_check_input (a) != UNFORMAT_END_OF_INPUT) + { + if (echo_process_each_proto_opts (a)) + ; + else if (unformat (a, "chroot prefix %s", &chroot_prefix)) + vl_set_memory_root_path ((char *) chroot_prefix); + else if (unformat (a, "uri %s", &uri)) + em->uri = format (0, "%s%c", uri, 0); + else if (unformat (a, "server")) + em->i_am_master = 1; + else if (unformat (a, "client")) + em->i_am_master = 0; + else if (unformat (a, "test-bytes:assert")) + em->test_return_packets = RETURN_PACKETS_ASSERT; + else if (unformat (a, "test-bytes")) + em->test_return_packets = RETURN_PACKETS_LOG_WRONG; + else if (unformat (a, "socket-name %s", &em->socket_name)) + ; + else if (unformat (a, "use-svm-api")) + em->use_sock_api = 0; + else if (unformat (a, "fifo-size %d", &tmp)) + em->fifo_size = tmp << 10; + else if (unformat (a, "rx-buf %d", &tmp)) + em->rx_buf_size = tmp << 10; + else if (unformat (a, "tx-buf %d", &tmp)) + em->rx_buf_size = tmp << 10; + else if (unformat (a, "nclients %d", &em->n_clients)) + { + em->n_sessions = em->n_clients + 1; + em->n_connects = em->n_clients; + } + else if (unformat (a, "nthreads %d", &em->n_rx_threads)) + ; + else if (unformat (a, "appns %_%v%_", &em->appns_id)) + ; + else if (unformat (a, "all-scope")) + em->appns_flags |= (APP_OPTIONS_FLAGS_USE_GLOBAL_SCOPE + | APP_OPTIONS_FLAGS_USE_LOCAL_SCOPE); + else if (unformat (a, "local-scope")) + em->appns_flags = APP_OPTIONS_FLAGS_USE_LOCAL_SCOPE; + else if (unformat (a, "global-scope")) + em->appns_flags = APP_OPTIONS_FLAGS_USE_GLOBAL_SCOPE; + else if (unformat (a, "secret %lu", &em->appns_secret)) + ; + else if (unformat (a, "TX=RX")) + em->data_source = ECHO_RX_DATA_SOURCE; + else if (unformat (a, "TX=%U", unformat_data, &em->bytes_to_send)) + ; + else if (unformat (a, "RX=%U", unformat_data, &em->bytes_to_receive)) + ; + else if (unformat (a, "json")) + em->output_json = 1; + else if (unformat (a, "log=%d", &em->log_lvl)) + ; + else if (unformat (a, "sclose=%U", + echo_unformat_close, &em->send_stream_disconnects)) + ; + else if (unformat (a, "time %U:%U", + echo_unformat_timing_event, &em->timing.start_event, + echo_unformat_timing_event, &em->timing.end_event)) + ; + else + print_usage_and_exit (); + } + + /* setting default for unset values + * + * bytes_to_send / bytes_to_receive & data_source */ + if (em->bytes_to_receive == (u64) ~ 0) + em->bytes_to_receive = 64 << 10; /* default */ + if (em->bytes_to_send == (u64) ~ 0) + em->bytes_to_send = 64 << 10; /* default */ + else if (em->bytes_to_send == 0) + em->data_source = ECHO_NO_DATA_SOURCE; + else + em->data_source = ECHO_TEST_DATA_SOURCE; + + if (em->data_source == ECHO_INVALID_DATA_SOURCE) + em->data_source = + em->i_am_master ? ECHO_RX_DATA_SOURCE : ECHO_TEST_DATA_SOURCE; + if (em->data_source == ECHO_RX_DATA_SOURCE) + em->bytes_to_send = em->bytes_to_receive; + + /* disconnect flags */ + if (em->i_am_master) + default_f_active = + em->bytes_to_send == 0 ? ECHO_CLOSE_F_ACTIVE : ECHO_CLOSE_F_PASSIVE; + else + default_f_active = + em->bytes_to_receive == 0 ? ECHO_CLOSE_F_PASSIVE : ECHO_CLOSE_F_ACTIVE; + if (em->send_stream_disconnects == ECHO_CLOSE_F_INVALID) + em->send_stream_disconnects = default_f_active; +} + +void +echo_process_uri (echo_main_t * em) +{ + unformat_input_t _input, *input = &_input; + u32 port; + unformat_init_string (input, (char *) em->uri, strlen ((char *) em->uri)); + if (unformat + (input, "%U://%U/%d", unformat_transport_proto, + &em->uri_elts.transport_proto, unformat_ip4_address, + &em->uri_elts.ip.ip4, &port)) + em->uri_elts.is_ip4 = 1; + else + if (unformat + (input, "%U://%U/%d", unformat_transport_proto, + &em->uri_elts.transport_proto, unformat_ip6_address, + &em->uri_elts.ip.ip6, &port)) + em->uri_elts.is_ip4 = 0; + else + ECHO_FAIL ("Unable to process uri"); + em->uri_elts.port = clib_host_to_net_u16 (port); + unformat_free (input); +} + +static void __clib_constructor +vpp_echo_init () +{ + /* init memory before proto register themselves */ + echo_main_t *em = &echo_main; + clib_mem_init_thread_safe (0, 256 << 20); + clib_memset (em, 0, sizeof (*em)); +} + +int +main (int argc, char **argv) +{ + echo_main_t *em = &echo_main; + fifo_segment_main_t *sm = &em->segment_main; + char *app_name; + u64 i; + svm_msg_q_cfg_t _cfg, *cfg = &_cfg; + u32 rpc_queue_size = 64 << 10; + + em->session_index_by_vpp_handles = hash_create (0, sizeof (uword)); + clib_spinlock_init (&em->sid_vpp_handles_lock); + em->shared_segment_handles = hash_create (0, sizeof (uword)); + clib_spinlock_init (&em->segment_handles_lock); + em->socket_name = format (0, "%s%c", API_SOCKET_FILE, 0); + em->use_sock_api = 1; + em->fifo_size = 64 << 10; + em->n_clients = 1; + em->n_connects = 1; + em->n_sessions = 2; + em->max_test_msg = 50; + em->time_to_stop = 0; + em->i_am_master = 1; + em->n_rx_threads = 4; + em->test_return_packets = RETURN_PACKETS_NOTEST; + em->timing.start_event = ECHO_EVT_FIRST_QCONNECT; + em->timing.end_event = ECHO_EVT_LAST_BYTE; + em->bytes_to_receive = ~0; /* defaulted when we know if server/client */ + em->bytes_to_send = ~0; /* defaulted when we know if server/client */ + em->rx_buf_size = 1 << 20; + em->tx_buf_size = 1 << 20; + em->data_source = ECHO_INVALID_DATA_SOURCE; + em->uri = format (0, "%s%c", "tcp://0.0.0.0/1234", 0); + echo_set_each_proto_defaults_before_opts (em); + echo_process_opts (argc, argv); + echo_process_uri (em); + em->proto_cb_vft = em->available_proto_cb_vft[em->uri_elts.transport_proto]; + if (!em->proto_cb_vft) + { + ECHO_FAIL ("Protocol %U is not supported", + format_transport_proto, em->uri_elts.transport_proto); + exit (1); + } + if (em->proto_cb_vft->set_defaults_after_opts_cb) + em->proto_cb_vft->set_defaults_after_opts_cb (); + + vec_validate (em->data_thread_handles, em->n_rx_threads); + vec_validate (em->data_thread_args, em->n_clients); + for (i = 0; i < em->n_clients; i++) + em->data_thread_args[i] = SESSION_INVALID_INDEX; + clib_time_init (&em->clib_time); + init_error_string_table (); + fifo_segment_main_init (sm, HIGH_SEGMENT_BASEVA, 20); + vec_validate (em->connect_test_data, em->tx_buf_size); + for (i = 0; i < em->tx_buf_size; i++) + em->connect_test_data[i] = i & 0xff; + + /* *INDENT-OFF* */ + svm_msg_q_ring_cfg_t rc[1] = { + {rpc_queue_size, sizeof (echo_rpc_msg_t), 0}, + }; + /* *INDENT-ON* */ + cfg->consumer_pid = getpid (); + cfg->n_rings = 1; + cfg->q_nitems = rpc_queue_size; + cfg->ring_cfgs = rc; + em->rpc_msq_queue = svm_msg_q_alloc (cfg); + + signal (SIGINT, stop_signal); + signal (SIGQUIT, stop_signal); + signal (SIGTERM, stop_signal); + echo_api_hookup (em); + + app_name = em->i_am_master ? "echo_server" : "echo_client"; + if (connect_to_vpp (app_name)) + { + svm_region_exit (); + ECHO_FAIL ("Couldn't connect to vpe, exiting...\n"); + exit (1); + } + + echo_session_prealloc (em); + echo_notify_event (em, ECHO_EVT_START); + + echo_send_attach (em); + if (wait_for_state_change (em, STATE_ATTACHED, TIMEOUT)) + { + ECHO_FAIL ("Couldn't attach to vpp, did you run <session enable> ?\n"); + exit (1); + } + if (pthread_create (&em->mq_thread_handle, + NULL /*attr */ , echo_mq_thread_fn, 0)) + { + ECHO_FAIL ("pthread create errored\n"); + exit (1); + } + for (i = 0; i < em->n_rx_threads; i++) + if (pthread_create (&em->data_thread_handles[i], + NULL /*attr */ , echo_data_thread_fn, (void *) i)) + { + ECHO_FAIL ("pthread create errored\n"); + exit (1); + } + if (em->i_am_master) + server_run (em); + else + clients_run (em); + echo_notify_event (em, ECHO_EVT_EXIT); + if (em->output_json) + print_global_json_stats (em); + else + print_global_stats (em); + echo_free_sessions (em); + echo_assert_test_suceeded (em); + echo_send_detach (em); + if (wait_for_state_change (em, STATE_DETACHED, TIMEOUT)) + { + ECHO_FAIL ("ECHO-ERROR: Couldn't detach from vpp, exiting...\n"); + exit (1); + } + int *rv; + pthread_join (em->mq_thread_handle, (void **) &rv); + if (rv) + { + ECHO_FAIL ("mq pthread errored %d", rv); + exit (1); + } + if (em->use_sock_api) + vl_socket_client_disconnect (); + else + vl_client_disconnect_from_vlib (); + ECHO_LOG (0, "Test complete !\n"); + exit (em->has_failed); +} + +/* + * fd.io coding-style-patch-verification: ON + * + * Local Variables: + * eval: (c-set-style "gnu") + * End: + */ |