/* * Copyright (c) 2017 Cisco and/or its affiliates. * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at: * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include <vnet/session/segment_manager.h> #include <vnet/session/session.h> #include <vnet/session/application.h> segment_manager_main_t segment_manager_main; /** * Counter used to build segment names */ static u32 segment_name_counter = 0; /** * Default fifo and segment size. TODO config. */ static u32 default_fifo_size = 1 << 12; static u32 default_segment_size = 1 << 20; static u32 default_app_evt_queue_size = 128; segment_manager_properties_t * segment_manager_properties_get (segment_manager_t * sm) { return application_get_segment_manager_properties (sm->app_index); } segment_manager_properties_t * segment_manager_properties_init (segment_manager_properties_t * props) { props->add_segment_size = default_segment_size; props->rx_fifo_size = default_fifo_size; props->tx_fifo_size = default_fifo_size; props->evt_q_size = default_app_evt_queue_size; return props; } static u8 segment_manager_app_detached (segment_manager_t * sm) { return (sm->app_index == SEGMENT_MANAGER_INVALID_APP_INDEX); } void segment_manager_app_detach (segment_manager_t * sm) { sm->app_index = SEGMENT_MANAGER_INVALID_APP_INDEX; } always_inline u32 segment_manager_segment_index (segment_manager_t * sm, svm_fifo_segment_private_t * seg) { return (seg - sm->segments); } /** * Remove segment without lock */ void segment_manager_del_segment (segment_manager_t * sm, svm_fifo_segment_private_t * fs) { segment_manager_main_t *smm = &segment_manager_main; if (ssvm_type (&fs->ssvm) != SSVM_SEGMENT_PRIVATE) clib_valloc_free (&smm->va_allocator, fs->ssvm.requested_va); ssvm_delete (&fs->ssvm); if (CLIB_DEBUG) memset (fs, 0xfb, sizeof (*fs)); pool_put (sm->segments, fs); } /** * Removes segment after acquiring writer lock */ always_inline void segment_manager_lock_and_del_segment (segment_manager_t * sm, u32 fs_index) { svm_fifo_segment_private_t *fs; u8 is_prealloc; clib_rwlock_writer_lock (&sm->segments_rwlock); fs = segment_manager_get_segment (sm, fs_index); is_prealloc = svm_fifo_segment_flags (fs) & FIFO_SEGMENT_F_IS_PREALLOCATED; if (is_prealloc && !segment_manager_app_detached (sm)) { clib_rwlock_writer_unlock (&sm->segments_rwlock); return; } segment_manager_del_segment (sm, fs); clib_rwlock_writer_unlock (&sm->segments_rwlock); } /** * Reads a segment from the segment manager's pool without lock */ svm_fifo_segment_private_t * segment_manager_get_segment (segment_manager_t * sm, u32 segment_index) { return pool_elt_at_index (sm->segments, segment_index); } /** * Reads a segment from the segment manager's pool and acquires reader lock * * Caller must drop the reader's lock by calling * @ref segment_manager_segment_reader_unlock once it finishes working with * the segment. */ svm_fifo_segment_private_t * segment_manager_get_segment_w_lock (segment_manager_t * sm, u32 segment_index) { clib_rwlock_reader_lock (&sm->segments_rwlock); return pool_elt_at_index (sm->segments, segment_index); } void segment_manager_segment_reader_unlock (segment_manager_t * sm) { ASSERT (sm->segments_rwlock->n_readers > 0); clib_rwlock_reader_unlock (&sm->segments_rwlock); } void segment_manager_segment_writer_unlock (segment_manager_t * sm) { clib_rwlock_writer_unlock (&sm->segments_rwlock); } /** * Adds segment to segment manager's pool * * If needed a writer's lock is acquired before allocating a new segment * to avoid affecting any of the segments pool readers. */ int segment_manager_add_segment (segment_manager_t * sm, u32 segment_size) { segment_manager_main_t *smm = &segment_manager_main; u32 rnd_margin = 128 << 10, seg_index; segment_manager_properties_t *props; uword baseva = (u64) ~ 0, alloc_size; svm_fifo_segment_private_t *seg; u8 *seg_name; int rv; props = segment_manager_properties_get (sm); /* Not configured for addition of new segments and not first */ if (!props->add_segment && !segment_size) { clib_warning ("cannot allocate new segment"); return VNET_API_ERROR_INVALID_VALUE; } /* * Allocate fifo segment and lock if needed */ if (vlib_num_workers ()) { clib_rwlock_writer_lock (&sm->segments_rwlock); pool_get (sm->segments, seg); } else { pool_get (sm->segments, seg); } memset (seg, 0, sizeof (*seg)); /* * Initialize ssvm segment and svm fifo private header */ segment_size = segment_size ? segment_size : props->add_segment_size; if (props->segment_type != SSVM_SEGMENT_PRIVATE) { seg_name = format (0, "%d-%d%c", getpid (), segment_name_counter++, 0); alloc_size = segment_size + rnd_margin; baseva = clib_valloc_alloc (&smm->va_allocator, alloc_size, 0); if (!baseva) { clib_warning ("out of space for segments"); return -1; } } else seg_name = format (0, "%s%c", "process-private-segment", 0); seg->ssvm.ssvm_size = segment_size; seg->ssvm.name = seg_name; seg->ssvm.requested_va = baseva; if ((rv = ssvm_master_init (&seg->ssvm, props->segment_type))) { clib_warning ("svm_master_init ('%v', %u) failed", seg_name, segment_size); if (props->segment_type != SSVM_SEGMENT_PRIVATE) clib_valloc_free (&smm->va_allocator, baseva); pool_put (sm->segments, seg); return (rv); } svm_fifo_segment_init (seg); /* * Save segment index before dropping lock, if any held */ seg_index = seg - sm->segments; if (vlib_num_workers ()) clib_rwlock_writer_unlock (&sm->segments_rwlock); return seg_index; } segment_manager_t * segment_manager_new () { segment_manager_main_t *smm = &segment_manager_main; segment_manager_t *sm; pool_get (smm->segment_managers, sm); memset (sm, 0, sizeof (*sm)); clib_rwlock_init (&sm->segments_rwlock); return sm; } /** * Initializes segment manager based on options provided. * Returns error if ssvm segment(s) allocation fails. */ int segment_manager_init (segment_manager_t * sm, u32 first_seg_size, u32 prealloc_fifo_pairs) { u32 rx_fifo_size, tx_fifo_size, pair_size; u32 rx_rounded_data_size, tx_rounded_data_size; u64 approx_total_size, max_seg_size = ((u64) 1 << 32) - clib_mem_get_page_size (); segment_manager_properties_t *props; svm_fifo_segment_private_t *segment; u32 approx_segment_count; int seg_index, i; props = segment_manager_properties_get (sm); first_seg_size = clib_max (first_seg_size, default_segment_size); if (prealloc_fifo_pairs) { /* Figure out how many segments should be preallocated */ rx_rounded_data_size = (1 << (max_log2 (props->rx_fifo_size))); tx_rounded_data_size = (1 << (max_log2 (props->tx_fifo_size))); rx_fifo_size = sizeof (svm_fifo_t) + rx_rounded_data_size; tx_fifo_size = sizeof (svm_fifo_t) + tx_rounded_data_size; pair_size = rx_fifo_size + tx_fifo_size; approx_total_size = (u64) prealloc_fifo_pairs *pair_size; if (first_seg_size > approx_total_size) max_seg_size = first_seg_size; approx_segment_count = (approx_total_size + (max_seg_size - 1)) / max_seg_size; /* Allocate the segments */ for (i = 0; i < approx_segment_count + 1; i++) { seg_index = segment_manager_add_segment (sm, max_seg_size); if (seg_index < 0) { clib_warning ("Failed to preallocate segment %d", i); return seg_index; } segment = segment_manager_get_segment (sm, seg_index); if (i == 0) sm->event_queue = segment_manager_alloc_queue (segment, props->evt_q_size); svm_fifo_segment_preallocate_fifo_pairs (segment, props->rx_fifo_size, props->tx_fifo_size, &prealloc_fifo_pairs); svm_fifo_segment_flags (segment) = FIFO_SEGMENT_F_IS_PREALLOCATED; if (prealloc_fifo_pairs == 0) break; } } else { seg_index = segment_manager_add_segment (sm, first_seg_size); if (seg_index) { clib_warning ("Failed to allocate segment"); return seg_index; } segment = segment_manager_get_segment (sm, seg_index); sm->event_queue = segment_manager_alloc_queue (segment, props->evt_q_size); } return 0; } u8 segment_manager_has_fifos (segment_manager_t * sm) { svm_fifo_segment_private_t *seg; u8 first = 1; /* *INDENT-OFF* */ segment_manager_foreach_segment_w_lock (seg, sm, ({ if (CLIB_DEBUG && !first && !svm_fifo_segment_has_fifos (seg) && !(svm_fifo_segment_flags (seg) & FIFO_SEGMENT_F_IS_PREALLOCATED)) { clib_warning ("segment %d has no fifos!", segment_manager_segment_index (sm, seg)); first = 0; } if (svm_fifo_segment_has_fifos (seg)) { segment_manager_segment_reader_unlock (sm); return 1; } })); /* *INDENT-ON* */ return 0; } /** * Initiate disconnects for all sessions 'owned' by a segment manager */ void segment_manager_del_sessions (segment_manager_t * sm) { svm_fifo_segment_private_t *fifo_segment; stream_session_t *session; svm_fifo_t *fifo; ASSERT (pool_elts (sm->segments) != 0); /* Across all fifo segments used by the server */ /* *INDENT-OFF* */ segment_manager_foreach_segment_w_lock (fifo_segment, sm, ({ fifo = svm_fifo_segment_get_fifo_list (fifo_segment); /* * Remove any residual sessions from the session lookup table * Don't bother deleting the individual fifos, we're going to * throw away the fifo segment in a minute. */ while (fifo) { session = session_get (fifo->master_session_index, fifo->master_thread_index); stream_session_disconnect (session); fifo = fifo->next; } /* Instead of removing the segment, test when cleaning up disconnected * sessions if the segment can be removed. */ })); /* *INDENT-ON* */ } /** * Removes segment manager. * * Since the fifos allocated in the segment keep backpointers to the sessions * prior to removing the segment, we call session disconnect. This * subsequently propagates into transport. */ void segment_manager_del (segment_manager_t * sm) { segment_manager_main_t *smm = &segment_manager_main; svm_fifo_segment_private_t *fifo_segment; ASSERT (!segment_manager_has_fifos (sm) && segment_manager_app_detached (sm)); /* If we have empty preallocated segments that haven't been removed, remove * them now. Apart from that, the first segment in the first segment manager * is not removed when all fifos are removed. It can only be removed when * the manager is explicitly deleted/detached by the app. */ clib_rwlock_writer_lock (&sm->segments_rwlock); /* *INDENT-OFF* */ pool_foreach (fifo_segment, sm->segments, ({ segment_manager_del_segment (sm, fifo_segment); })); /* *INDENT-ON* */ clib_rwlock_writer_unlock (&sm->segments_rwlock); clib_rwlock_free (&sm->segments_rwlock); if (CLIB_DEBUG) memset (sm, 0xfe, sizeof (*sm)); pool_put (smm->segment_managers, sm); } void segment_manager_init_del (segment_manager_t * sm) { segment_manager_app_detach (sm); if (segment_manager_has_fifos (sm)) segment_manager_del_sessions (sm); else { ASSERT (!sm->first_is_protected || segment_manager_app_detached (sm)); segment_manager_del (sm); } } int segment_manager_try_alloc_fifos (svm_fifo_segment_private_t * fifo_segment, u32 rx_fifo_size, u32 tx_fifo_size, svm_fifo_t ** rx_fifo, svm_fifo_t ** tx_fifo) { rx_fifo_size = clib_max (rx_fifo_size, default_fifo_size); *rx_fifo = svm_fifo_segment_alloc_fifo (fifo_segment, rx_fifo_size, FIFO_SEGMENT_RX_FREELIST); tx_fifo_size = clib_max (tx_fifo_size, default_fifo_size); *tx_fifo = svm_fifo_segment_alloc_fifo (fifo_segment, tx_fifo_size, FIFO_SEGMENT_TX_FREELIST); if (*rx_fifo == 0) { /* This would be very odd, but handle it... */ if (*tx_fifo != 0) { svm_fifo_segment_free_fifo (fifo_segment, *tx_fifo, FIFO_SEGMENT_TX_FREELIST); *tx_fifo = 0; } return -1; } if (*tx_fifo == 0) { if (*rx_fifo != 0) { svm_fifo_segment_free_fifo (fifo_segment, *rx_fifo, FIFO_SEGMENT_RX_FREELIST); *rx_fifo = 0; } return -1; } return 0; } int segment_manager_alloc_session_fifos (segment_manager_t * sm, svm_fifo_t ** rx_fifo, svm_fifo_t ** tx_fifo, u32 * fifo_segment_index) { svm_fifo_segment_private_t *fifo_segment = 0; int alloc_fail = 1, rv = 0, new_fs_index; segment_manager_properties_t *props; u8 added_a_segment = 0; u32 sm_index; props = segment_manager_properties_get (sm); /* * Find the first free segment to allocate the fifos in */ /* *INDENT-OFF* */ segment_manager_foreach_segment_w_lock (fifo_segment, sm, ({ alloc_fail = segment_manager_try_alloc_fifos (fifo_segment, props->rx_fifo_size, props->tx_fifo_size, rx_fifo, tx_fifo); /* Exit with lock held, drop it after notifying app */ if (!alloc_fail) goto alloc_success; })); /* *INDENT-ON* */ alloc_check: if (!alloc_fail) { alloc_success: ASSERT (rx_fifo && tx_fifo); sm_index = segment_manager_index (sm); (*tx_fifo)->segment_manager = sm_index; (*rx_fifo)->segment_manager = sm_index; *fifo_segment_index = segment_manager_segment_index (sm, fifo_segment); if (added_a_segment) rv = application_add_segment_notify (sm->app_index, &fifo_segment->ssvm); /* Drop the lock after app is notified */ segment_manager_segment_reader_unlock (sm); return rv; } /* * Allocation failed, see if we can add a new segment */ if (props->add_segment) { if (added_a_segment) { clib_warning ("Added a segment, still can't allocate a fifo"); segment_manager_segment_reader_unlock (sm); return SESSION_ERROR_NEW_SEG_NO_SPACE; } if ((new_fs_index = segment_manager_add_segment (sm, 0)) < 0) { clib_warning ("Failed to add new segment"); return SESSION_ERROR_SEG_CREATE; } fifo_segment = segment_manager_get_segment_w_lock (sm, new_fs_index); alloc_fail = segment_manager_try_alloc_fifos (fifo_segment, props->rx_fifo_size, props->tx_fifo_size, rx_fifo, tx_fifo); added_a_segment = 1; goto alloc_check; } else { clib_warning ("Can't add new seg and no space to allocate fifos!"); return SESSION_ERROR_NO_SPACE; } } void segment_manager_dealloc_fifos (u32 segment_index, svm_fifo_t * rx_fifo, svm_fifo_t * tx_fifo) { svm_fifo_segment_private_t *fifo_segment; segment_manager_t *sm; /* It's possible to have no segment manager if the session was removed * as result of a detach. */ if (!(sm = segment_manager_get_if_valid (rx_fifo->segment_manager))) return; fifo_segment = segment_manager_get_segment_w_lock (sm, segment_index); svm_fifo_segment_free_fifo (fifo_segment, rx_fifo, FIFO_SEGMENT_RX_FREELIST); svm_fifo_segment_free_fifo (fifo_segment, tx_fifo, FIFO_SEGMENT_TX_FREELIST); /* * Try to remove svm segment if it has no fifos. This can be done only if * the segment is not the first in the segment manager or if it is first * and it is not protected. Moreover, if the segment is first and the app * has detached from the segment manager, remove the segment manager. */ if (!svm_fifo_segment_has_fifos (fifo_segment)) { segment_manager_segment_reader_unlock (sm); /* Remove segment if it holds no fifos or first but not protected */ if (segment_index != 0 || !sm->first_is_protected) segment_manager_lock_and_del_segment (sm, segment_index); /* Remove segment manager if no sessions and detached from app */ if (segment_manager_app_detached (sm) && !segment_manager_has_fifos (sm)) segment_manager_del (sm); } else segment_manager_segment_reader_unlock (sm); } /** * Allocates shm queue in the first segment * * Must be called with lock held */ svm_queue_t * segment_manager_alloc_queue (svm_fifo_segment_private_t * segment, u32 queue_size) { ssvm_shared_header_t *sh; svm_queue_t *q; void *oldheap; sh = segment->ssvm.sh; oldheap = ssvm_push_heap (sh); q = svm_queue_init (queue_size, sizeof (session_fifo_event_t), 0 /* consumer pid */ , 0 /* signal when queue non-empty */ ); ssvm_pop_heap (oldheap); return q; } /** * Frees shm queue allocated in the first segment */ void segment_manager_dealloc_queue (segment_manager_t * sm, svm_queue_t * q) { svm_fifo_segment_private_t *segment; ssvm_shared_header_t *sh; void *oldheap; ASSERT (!pool_is_free_index (sm->segments, 0)); segment = segment_manager_get_segment_w_lock (sm, 0); sh = segment->ssvm.sh; oldheap = ssvm_push_heap (sh); svm_queue_free (q); ssvm_pop_heap (oldheap); segment_manager_segment_reader_unlock (sm); } /* * Init segment vm address allocator */ void segment_manager_main_init (segment_manager_main_init_args_t * a) { segment_manager_main_t *sm = &segment_manager_main; clib_valloc_chunk_t _ip, *ip = &_ip; ip->baseva = a->baseva; ip->size = a->size; clib_valloc_init (&sm->va_allocator, ip, 1 /* lock */ ); } static clib_error_t * segment_manager_show_fn (vlib_main_t * vm, unformat_input_t * input, vlib_cli_command_t * cmd) { segment_manager_main_t *smm = &segment_manager_main; svm_fifo_segment_private_t *seg; segment_manager_t *sm; u8 show_segments = 0, verbose = 0; uword address; u64 size; u32 active_fifos; u32 free_fifos; while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { if (unformat (input, "segments")) show_segments = 1; else if (unformat (input, "verbose")) verbose = 1; else return clib_error_return (0, "unknown input `%U'", format_unformat_error, input); } vlib_cli_output (vm, "%d segment managers allocated", pool_elts (smm->segment_managers)); if (verbose && pool_elts (smm->segment_managers)) { vlib_cli_output (vm, "%-10s%=15s%=12s", "Index", "App Index", "Segments"); /* *INDENT-OFF* */ pool_foreach (sm, smm->segment_managers, ({ vlib_cli_output (vm, "%-10d%=15d%=12d", segment_manager_index(sm), sm->app_index, pool_elts (sm->segments)); })); /* *INDENT-ON* */ } if (show_segments) { vlib_cli_output (vm, "%-15s%15s%15s%15s%15s%15s", "Name", "Type", "HeapSize (M)", "ActiveFifos", "FreeFifos", "Address"); /* *INDENT-OFF* */ pool_foreach (sm, smm->segment_managers, ({ segment_manager_foreach_segment_w_lock (seg, sm, ({ svm_fifo_segment_info (seg, &address, &size); active_fifos = svm_fifo_segment_num_fifos (seg); free_fifos = svm_fifo_segment_num_free_fifos (seg, ~0 /* size */); vlib_cli_output (vm, "%-15v%15U%15llu%15u%15u%15llx", ssvm_name (&seg->ssvm), format_svm_fifo_segment_type, seg, size >> 20ULL, active_fifos, free_fifos, address); if (verbose) vlib_cli_output (vm, "%U", format_svm_fifo_segment, seg, verbose); })); })); /* *INDENT-ON* */ } return 0; } /* *INDENT-OFF* */ VLIB_CLI_COMMAND (segment_manager_show_command, static) = { .path = "show segment-manager", .short_help = "show segment-manager [segments][verbose]", .function = segment_manager_show_fn, }; /* *INDENT-ON* */ /* * fd.io coding-style-patch-verification: ON * * Local Variables: * eval: (c-set-style "gnu") * End: */