diff options
author | Florin Coras <fcoras@cisco.com> | 2019-01-09 20:04:38 -0800 |
---|---|---|
committer | Dave Barach <openvpp@barachs.net> | 2019-01-10 21:01:05 +0000 |
commit | 1bcad5cdb78fb04d41f97603312161b2bc2e7bae (patch) | |
tree | 87fa6d1a3be5510dd6bc9ee75300e54623af8883 /src | |
parent | 45db885a6a050aa809ae333bf6c14349db644822 (diff) |
session/vcl: improve fifo tx notifications
Change-Id: Ied476c417c8ba5b4d7c615125c36eecd9ca64fbc
Signed-off-by: Florin Coras <fcoras@cisco.com>
Diffstat (limited to 'src')
-rw-r--r-- | src/svm/svm_fifo.h | 110 | ||||
-rw-r--r-- | src/vcl/vppcom.c | 18 | ||||
-rw-r--r-- | src/vnet/session/session.c | 6 | ||||
-rw-r--r-- | src/vnet/session/session.h | 29 | ||||
-rw-r--r-- | src/vnet/session/session_node.c | 64 |
5 files changed, 144 insertions, 83 deletions
diff --git a/src/svm/svm_fifo.h b/src/svm/svm_fifo.h index 9bba85fddf4..613e0cad5eb 100644 --- a/src/svm/svm_fifo.h +++ b/src/svm/svm_fifo.h @@ -41,6 +41,13 @@ format_function_t format_ooo_list; #define SVM_FIFO_INVALID_SESSION_INDEX ((u32)~0) #define SVM_FIFO_INVALID_INDEX ((u32)~0) +enum +{ + SVM_FIFO_NO_TX_NOTIF = 0, + SVM_FIFO_WANT_TX_NOTIF = 1, + SVM_FIFO_WANT_TX_NOTIF_IF_FULL = 2, +}; + typedef struct { u32 offset; @@ -66,7 +73,8 @@ typedef struct _svm_fifo u32 ct_session_index; /**< Local session index for vpp */ CLIB_CACHE_LINE_ALIGN_MARK (end_shared); u32 head; - volatile u32 want_tx_evt; /**< producer wants nudge */ + volatile u32 want_tx_ntf; /**< producer wants nudge */ + volatile u32 has_tx_ntf; CLIB_CACHE_LINE_ALIGN_MARK (end_consumer); /* producer */ @@ -174,18 +182,6 @@ svm_fifo_unset_event (svm_fifo_t * f) clib_atomic_release (&f->has_event); } -static inline void -svm_fifo_set_want_tx_evt (svm_fifo_t * f, u8 want_evt) -{ - f->want_tx_evt = want_evt; -} - -static inline u8 -svm_fifo_want_tx_evt (svm_fifo_t * f) -{ - return f->want_tx_evt; -} - svm_fifo_t *svm_fifo_create (u32 data_size_in_bytes); void svm_fifo_free (svm_fifo_t * f); @@ -200,27 +196,10 @@ int svm_fifo_dequeue_drop (svm_fifo_t * f, u32 max_bytes); void svm_fifo_dequeue_drop_all (svm_fifo_t * f); int svm_fifo_segments (svm_fifo_t * f, svm_fifo_segment_t * fs); void svm_fifo_segments_free (svm_fifo_t * f, svm_fifo_segment_t * fs); -u32 svm_fifo_number_ooo_segments (svm_fifo_t * f); -ooo_segment_t *svm_fifo_first_ooo_segment (svm_fifo_t * f); void svm_fifo_init_pointers (svm_fifo_t * f, u32 pointer); void svm_fifo_overwrite_head (svm_fifo_t * f, u8 * data, u32 len); - format_function_t format_svm_fifo; -always_inline ooo_segment_t * -svm_fifo_newest_ooo_segment (svm_fifo_t * f) -{ - if (f->ooos_newest == OOO_SEGMENT_INVALID_INDEX) - return 0; - return pool_elt_at_index (f->ooo_segments, f->ooos_newest); -} - -always_inline void -svm_fifo_newest_ooo_segment_reset (svm_fifo_t * f) -{ - f->ooos_newest = OOO_SEGMENT_INVALID_INDEX; -} - /** * Max contiguous chunk of data that can be read */ @@ -265,6 +244,77 @@ svm_fifo_tail (svm_fifo_t * f) } always_inline u32 +svm_fifo_nitems (svm_fifo_t * f) +{ + return f->nitems; +} + +static inline void +svm_fifo_add_want_tx_ntf (svm_fifo_t * f, u8 ntf_type) +{ + f->want_tx_ntf |= ntf_type; +} + +static inline void +svm_fifo_del_want_tx_ntf (svm_fifo_t * f, u8 ntf_type) +{ + f->want_tx_ntf &= ~ntf_type; +} + +static inline void +svm_fifo_clear_tx_ntf (svm_fifo_t * f) +{ + /* Set the flag if want_tx_notif_if_full was the only ntf requested */ + f->has_tx_ntf = f->want_tx_ntf == SVM_FIFO_WANT_TX_NOTIF_IF_FULL; + svm_fifo_del_want_tx_ntf (f, SVM_FIFO_WANT_TX_NOTIF); +} + +static inline void +svm_fifo_reset_tx_ntf (svm_fifo_t * f) +{ + f->has_tx_ntf = 0; +} + +static inline u8 +svm_fifo_needs_tx_ntf (svm_fifo_t * f, u32 n_last_deq) +{ + u8 want_ntf = f->want_tx_ntf; + + if (PREDICT_TRUE (want_ntf == SVM_FIFO_NO_TX_NOTIF)) + return 0; + else if (want_ntf & SVM_FIFO_WANT_TX_NOTIF) + return 1; + else if (want_ntf & SVM_FIFO_WANT_TX_NOTIF_IF_FULL) + { + u32 max_deq = svm_fifo_max_dequeue (f); + u32 nitems = svm_fifo_nitems (f); + if (!f->has_tx_ntf && max_deq < nitems + && max_deq + n_last_deq >= nitems) + return 1; + + return 0; + } + return 0; +} + +u32 svm_fifo_number_ooo_segments (svm_fifo_t * f); +ooo_segment_t *svm_fifo_first_ooo_segment (svm_fifo_t * f); + +always_inline ooo_segment_t * +svm_fifo_newest_ooo_segment (svm_fifo_t * f) +{ + if (f->ooos_newest == OOO_SEGMENT_INVALID_INDEX) + return 0; + return pool_elt_at_index (f->ooo_segments, f->ooos_newest); +} + +always_inline void +svm_fifo_newest_ooo_segment_reset (svm_fifo_t * f) +{ + f->ooos_newest = OOO_SEGMENT_INVALID_INDEX; +} + +always_inline u32 ooo_segment_distance_from_tail (svm_fifo_t * f, u32 pos) { /* Ambiguous. Assumption is that ooo segments don't touch tail */ diff --git a/src/vcl/vppcom.c b/src/vcl/vppcom.c index 54c5f10284d..1797d93c683 100644 --- a/src/vcl/vppcom.c +++ b/src/vcl/vppcom.c @@ -1744,9 +1744,9 @@ vppcom_session_read_internal (uint32_t session_handle, void *buf, int n, if (svm_fifo_is_empty (rx_fifo)) svm_fifo_unset_event (rx_fifo); - if (is_ct && svm_fifo_want_tx_evt (rx_fifo)) + if (is_ct && svm_fifo_needs_tx_ntf (rx_fifo, n_read)) { - svm_fifo_set_want_tx_evt (s->rx_fifo, 0); + svm_fifo_clear_tx_ntf (s->rx_fifo); app_send_io_evt_to_vpp (s->vpp_evt_q, s->rx_fifo, SESSION_IO_EVT_CT_RX, SVM_Q_WAIT); } @@ -1959,7 +1959,7 @@ vppcom_session_write_inline (uint32_t session_handle, void *buf, size_t n, } while (svm_fifo_is_full (tx_fifo)) { - svm_fifo_set_want_tx_evt (tx_fifo, 1); + svm_fifo_add_want_tx_ntf (tx_fifo, SVM_FIFO_WANT_TX_NOTIF); svm_msg_q_lock (mq); if (svm_msg_q_is_empty (mq)) svm_msg_q_wait (mq); @@ -2351,7 +2351,7 @@ vppcom_select (int n_bits, vcl_si_set * read_map, vcl_si_set * write_map, bits_set++; } else - svm_fifo_set_want_tx_evt (session->tx_fifo, 1); + svm_fifo_add_want_tx_ntf (session->tx_fifo, SVM_FIFO_WANT_TX_NOTIF); })); check_rd: @@ -2570,6 +2570,10 @@ vppcom_epoll_ctl (uint32_t vep_handle, int op, uint32_t session_handle, session->is_vep_session = 1; vep_session->vep.next_sh = session_handle; + if (session->tx_fifo) + svm_fifo_add_want_tx_ntf (session->tx_fifo, + SVM_FIFO_WANT_TX_NOTIF_IF_FULL); + VDBG (1, "EPOLL_CTL_ADD: vep_sh %u, sh %u, events 0x%x, data 0x%llx!", vep_handle, session_handle, event->events, event->data.u64); vcl_evt (VCL_EVT_EPOLL_CTLADD, session, event->events, event->data.u64); @@ -2655,6 +2659,10 @@ vppcom_epoll_ctl (uint32_t vep_handle, int op, uint32_t session_handle, session->vep.prev_sh = ~0; session->vep.vep_sh = ~0; session->is_vep_session = 0; + + if (session->tx_fifo) + svm_fifo_del_want_tx_ntf (session->tx_fifo, SVM_FIFO_NO_TX_NOTIF); + VDBG (1, "EPOLL_CTL_DEL: vep_idx %u, sid %u!", vep_handle, session_handle); vcl_evt (VCL_EVT_EPOLL_CTLDEL, session, vep_sh); @@ -2708,6 +2716,7 @@ vcl_epoll_wait_handle_mq_event (vcl_worker_t * wrk, session_event_t * e, add_event = 1; events[*num_ev].events |= EPOLLOUT; session_evt_data = session->vep.ev.data.u64; + svm_fifo_reset_tx_ntf (session->tx_fifo); break; case SESSION_IO_EVT_CT_TX: vcl_fifo_rx_evt_valid_or_break (e->fifo); @@ -2734,6 +2743,7 @@ vcl_epoll_wait_handle_mq_event (vcl_worker_t * wrk, session_event_t * e, add_event = 1; events[*num_ev].events |= EPOLLOUT; session_evt_data = session->vep.ev.data.u64; + svm_fifo_reset_tx_ntf (session->tx_fifo); break; case SESSION_CTRL_EVT_ACCEPTED: session = vcl_session_accepted (wrk, diff --git a/src/vnet/session/session.c b/src/vnet/session/session.c index 4cf0f9e7e49..c56712bbf87 100644 --- a/src/vnet/session/session.c +++ b/src/vnet/session/session.c @@ -557,7 +557,11 @@ session_dequeue_notify (stream_session_t * s) if (PREDICT_FALSE (!app)) return -1; - return app_worker_lock_and_send_event (app, s, FIFO_EVENT_APP_TX); + if (app_worker_lock_and_send_event (app, s, FIFO_EVENT_APP_TX)) + return -1; + + svm_fifo_clear_tx_ntf (s->server_tx_fifo); + return 0; } /** diff --git a/src/vnet/session/session.h b/src/vnet/session/session.h index cf1b3e99f4d..2bbc380282c 100644 --- a/src/vnet/session/session.h +++ b/src/vnet/session/session.h @@ -176,20 +176,6 @@ typedef struct session_tx_context_ session_dgram_hdr_t hdr; } session_tx_context_t; -/* Forward definition */ -typedef struct _session_manager_main session_manager_main_t; - -typedef int - (session_fifo_rx_fn) (vlib_main_t * vm, vlib_node_runtime_t * node, - session_event_t * e0, stream_session_t * s0, - int *n_tx_pkts); - -extern session_fifo_rx_fn session_tx_fifo_peek_and_snd; -extern session_fifo_rx_fn session_tx_fifo_dequeue_and_snd; -extern session_fifo_rx_fn session_tx_fifo_dequeue_internal; - -u8 session_node_lookup_fifo_event (svm_fifo_t * f, session_event_t * e); - typedef struct session_manager_worker_ { CLIB_CACHE_LINE_ALIGN_MARK (cacheline0); @@ -237,7 +223,18 @@ typedef struct session_manager_worker_ } session_manager_worker_t; -struct _session_manager_main +typedef int (session_fifo_rx_fn) (vlib_main_t * vm, + vlib_node_runtime_t * node, + session_manager_worker_t * wrk, + session_event_t * e, int *n_tx_pkts); + +extern session_fifo_rx_fn session_tx_fifo_peek_and_snd; +extern session_fifo_rx_fn session_tx_fifo_dequeue_and_snd; +extern session_fifo_rx_fn session_tx_fifo_dequeue_internal; + +u8 session_node_lookup_fifo_event (svm_fifo_t * f, session_event_t * e); + +typedef struct session_manager_main_ { /** Worker contexts */ session_manager_worker_t *wrk; @@ -297,7 +294,7 @@ struct _session_manager_main f64 *last_event_poll_by_thread; #endif -}; +} session_manager_main_t; extern session_manager_main_t session_manager_main; extern vlib_node_registration_t session_queue_node; diff --git a/src/vnet/session/session_node.c b/src/vnet/session/session_node.c index 880f16388b8..f5e5efeaf0f 100644 --- a/src/vnet/session/session_node.c +++ b/src/vnet/session/session_node.c @@ -632,32 +632,30 @@ session_tx_set_dequeue_params (vlib_main_t * vm, session_tx_context_t * ctx, always_inline int session_tx_fifo_read_and_snd_i (vlib_main_t * vm, vlib_node_runtime_t * node, - session_event_t * e, - stream_session_t * s, int *n_tx_packets, + session_manager_worker_t * wrk, + session_event_t * e, int *n_tx_packets, u8 peek_data) { u32 next_index, next0, next1, *to_next, n_left_to_next; u32 n_trace = vlib_get_trace_count (vm, node), n_bufs_needed = 0; - u32 thread_index = s->thread_index, n_left, pbi; + u32 thread_index = vm->thread_index, n_left, pbi; session_manager_main_t *smm = &session_manager_main; - session_manager_worker_t *wrk = &smm->wrk[thread_index]; session_tx_context_t *ctx = &wrk->ctx; transport_proto_t tp; vlib_buffer_t *pb; u16 n_bufs, rv; - if (PREDICT_FALSE ((rv = session_tx_not_ready (s, peek_data)))) + if (PREDICT_FALSE ((rv = session_tx_not_ready (ctx->s, peek_data)))) { if (rv < 2) vec_add1 (wrk->pending_event_vector, *e); return SESSION_TX_NO_DATA; } - next_index = smm->session_type_to_next[s->session_type]; + next_index = smm->session_type_to_next[ctx->s->session_type]; next0 = next1 = next_index; - tp = session_get_transport_proto (s); - ctx->s = s; + tp = session_get_transport_proto (ctx->s); ctx->transport_vft = transport_protocol_get_vft (tp); ctx->tc = session_tx_get_transport (ctx, peek_data); ctx->snd_mss = ctx->transport_vft->send_mss (ctx->tc); @@ -679,7 +677,7 @@ session_tx_fifo_read_and_snd_i (vlib_main_t * vm, vlib_node_runtime_t * node, } /* Allow enqueuing of a new event */ - svm_fifo_unset_event (s->server_tx_fifo); + svm_fifo_unset_event (ctx->s->server_tx_fifo); /* Check how much we can pull. */ session_tx_set_dequeue_params (vm, ctx, VLIB_FRAME_SIZE - *n_tx_packets, @@ -784,7 +782,7 @@ session_tx_fifo_read_and_snd_i (vlib_main_t * vm, vlib_node_runtime_t * node, if (PREDICT_FALSE (n_trace > 0)) session_tx_trace_frame (vm, node, next_index, to_next, - ctx->n_segs_per_evt, s, n_trace); + ctx->n_segs_per_evt, ctx->s, n_trace); _vec_len (wrk->tx_buffers) = n_bufs; *n_tx_packets += ctx->n_segs_per_evt; @@ -794,18 +792,18 @@ session_tx_fifo_read_and_snd_i (vlib_main_t * vm, vlib_node_runtime_t * node, /* If we couldn't dequeue all bytes mark as partially read */ ASSERT (ctx->left_to_snd == 0); if (ctx->max_len_to_snd < ctx->max_dequeue) - if (svm_fifo_set_event (s->server_tx_fifo)) + if (svm_fifo_set_event (ctx->s->server_tx_fifo)) vec_add1 (wrk->pending_event_vector, *e); if (!peek_data && ctx->transport_vft->tx_type == TRANSPORT_TX_DGRAM) { /* Fix dgram pre header */ if (ctx->max_len_to_snd < ctx->max_dequeue) - svm_fifo_overwrite_head (s->server_tx_fifo, (u8 *) & ctx->hdr, + svm_fifo_overwrite_head (ctx->s->server_tx_fifo, (u8 *) & ctx->hdr, sizeof (session_dgram_pre_hdr_t)); /* More data needs to be read */ - else if (svm_fifo_max_dequeue (s->server_tx_fifo) > 0) - if (svm_fifo_set_event (s->server_tx_fifo)) + else if (svm_fifo_max_dequeue (ctx->s->server_tx_fifo) > 0) + if (svm_fifo_set_event (ctx->s->server_tx_fifo)) vec_add1 (wrk->pending_event_vector, *e); } return SESSION_TX_OK; @@ -813,27 +811,29 @@ session_tx_fifo_read_and_snd_i (vlib_main_t * vm, vlib_node_runtime_t * node, int session_tx_fifo_peek_and_snd (vlib_main_t * vm, vlib_node_runtime_t * node, - session_event_t * e, - stream_session_t * s, int *n_tx_pkts) + session_manager_worker_t * wrk, + session_event_t * e, int *n_tx_pkts) { - return session_tx_fifo_read_and_snd_i (vm, node, e, s, n_tx_pkts, 1); + return session_tx_fifo_read_and_snd_i (vm, node, wrk, e, n_tx_pkts, 1); } int session_tx_fifo_dequeue_and_snd (vlib_main_t * vm, vlib_node_runtime_t * node, - session_event_t * e, - stream_session_t * s, int *n_tx_pkts) + session_manager_worker_t * wrk, + session_event_t * e, int *n_tx_pkts) { - return session_tx_fifo_read_and_snd_i (vm, node, e, s, n_tx_pkts, 0); + return session_tx_fifo_read_and_snd_i (vm, node, wrk, e, n_tx_pkts, 0); } int session_tx_fifo_dequeue_internal (vlib_main_t * vm, vlib_node_runtime_t * node, - session_event_t * e, - stream_session_t * s, int *n_tx_pkts) + session_manager_worker_t * wrk, + session_event_t * e, int *n_tx_pkts) { + stream_session_t *s = wrk->ctx.s; application_t *app; + if (PREDICT_FALSE (s->session_state == SESSION_STATE_CLOSED)) return 0; app = application_get (s->t_app_index); @@ -923,7 +923,7 @@ session_queue_node_fn (vlib_main_t * vm, vlib_node_runtime_t * node, { stream_session_t *s; /* $$$ prefetch 1 ahead maybe */ session_event_t *e; - u8 want_tx_evt; + u8 need_tx_ntf; e = &fifo_events[i]; switch (e->event_type) @@ -943,19 +943,17 @@ session_queue_node_fn (vlib_main_t * vm, vlib_node_runtime_t * node, clib_warning ("session was freed!"); continue; } - - want_tx_evt = svm_fifo_want_tx_evt (s->server_tx_fifo); + wrk->ctx.s = s; /* Spray packets in per session type frames, since they go to * different nodes */ - rv = (smm->session_tx_fns[s->session_type]) (vm, node, e, s, + rv = (smm->session_tx_fns[s->session_type]) (vm, node, wrk, e, &n_tx_packets); if (PREDICT_TRUE (rv == SESSION_TX_OK)) { - if (PREDICT_FALSE (want_tx_evt)) - { - svm_fifo_set_want_tx_evt (s->server_tx_fifo, 0); - session_dequeue_notify (s); - } + need_tx_ntf = svm_fifo_needs_tx_ntf (s->server_tx_fifo, + wrk->ctx.max_len_to_snd); + if (PREDICT_FALSE (need_tx_ntf)) + session_dequeue_notify (s); } else if (PREDICT_FALSE (rv == SESSION_TX_NO_BUFFERS)) { @@ -995,8 +993,10 @@ session_queue_node_fn (vlib_main_t * vm, vlib_node_runtime_t * node, break; case FIFO_EVENT_BUILTIN_TX: s = session_get_from_handle_if_valid (e->session_handle); + wrk->ctx.s = s; if (PREDICT_TRUE (s != 0)) - session_tx_fifo_dequeue_internal (vm, node, e, s, &n_tx_packets); + session_tx_fifo_dequeue_internal (vm, node, wrk, e, + &n_tx_packets); break; case FIFO_EVENT_RPC: fp = e->rpc_args.fp; |