summaryrefslogtreecommitdiffstats
path: root/test/Makefile
blob: b570e50a8d5b843cb906c07384999ea5ed929435 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
.PHONY: verify-python-path

VPP_TEST_FAILED_DIR=/tmp/vpp-failed-unittests/

verify-python-path:
ifndef VPP_PYTHON_PREFIX
	$(error VPP_PYTHON_PREFIX is not set)
endif

.PHONY: verify-no-running-vpp

ifdef VPP_ZOMBIE_NOCHECK
VPP_PIDS=
else
VPP_PIDS=$(shell pgrep -d, -x vpp_main)
endif

ifeq ($(DEBUG),gdb)
FORCE_FOREGROUND=1
else ifeq ($(DEBUG),gdbserver)
FORCE_FOREGROUND=1
else ifeq ($(DEBUG),core)
FORCE_FOREGROUND=1
else
FORCE_FOREGROUND=0
endif

verify-no-running-vpp:
	@if [ "$(VPP_PIDS)" != "" ]; then \
		echo; \
		echo "*** Existing vpp processes detected (PID(s): $(VPP_PIDS)). Running tests under these conditions is not supported. ***"; \
		echo; \
		ps -fp $(VPP_PIDS);\
		echo; \
		false; \
	fi

UNITTEST_EXTRA_OPTS=
UNITTEST_FAILFAST_OPTS=

ifeq ($(FAILFAST),1)
UNITTEST_EXTRA_OPTS=-f
endif

ifneq ($(EXTERN_TESTS),)
UNITTEST_EXTRA_OPTS=$(UNITTEST_FAILFAST_OPTS) -d $(EXTERN_TESTS)
endif

ifeq ($(TEST_DEBUG),1)
VPP_PYTHON_PREFIX:=$(VPP_PYTHON_PREFIX)/debug
PYTHON_EXTRA_DEPENDS=objgraph pympler
else
PYTHON_EXTRA_DEPENDS=
endif

PYTHON_VENV_PATH=$(VPP_PYTHON_PREFIX)/virtualenv
PYTHON_DEPENDS=$(PYTHON_EXTRA_DEPENDS) faulthandler six scapy==2.3.3 pexpect pycrypto subprocess32 cffi git+https://github.com/klement/py-lispnetworking@setup
SCAPY_SOURCE=$(shell find $(PYTHON_VENV_PATH) -name site-packages)
BUILD_COV_DIR=$(BR)/test-cov

GET_PIP_SCRIPT=$(VPP_PYTHON_PREFIX)/get-pip.py
PIP_INSTALL_DONE=$(VPP_PYTHON_PREFIX)/pip-install.done
PIP_PATCH_DONE=$(VPP_PYTHON_PREFIX)/pip-patch.done
PAPI_INSTALL_DONE=$(VPP_PYTHON_PREFIX)/papi-install.done

PAPI_INSTALL_FLAGS=$(PIP_INSTALL_DONE) $(PIP_PATCH_DONE) $(PAPI_INSTALL_DONE)

ifeq ($(PYTHON),)
PYTHON_INTERP=python2.7
else
PYTHON_INTERP=$(PYTHON)
endif

$(GET_PIP_SCRIPT):
	@mkdir -p $(VPP_PYTHON_PREFIX)
	@bash -c "cd $(VPP_PYTHON_PREFIX) && curl -O https://bootstrap.pypa.io/get-pip.py"

$(PIP_INSTALL_DONE): $(GET_PIP_SCRIPT)
	@virtualenv $(PYTHON_VENV_PATH) -p $(PYTHON_INTERP)
	@bash -c "source $(PYTHON_VENV_PATH)/bin/activate && python $(GET_PIP_SCRIPT)"
	@bash -c "source $(PYTHON_VENV_PATH)/bin/activate && pip install $(PYTHON_DEPENDS)"
	@touch $@

$(PIP_PATCH_DONE): $(PIP_INSTALL_DONE)
	@echo --- patching ---
	@sleep 1 # Ensure python recompiles patched *.py files -> *.pyc
	for f in $(CURDIR)/patches/scapy-2.3.3/*.patch ; do \
		echo Applying patch: $$(basename $$f) ; \
		patch -p1 -d $(SCAPY_SOURCE) < $$f ; \
	done
	@touch $@

$(PAPI_INSTALL_DONE): $(PIP_PATCH_DONE)
	@bash -c "source $(PYTHON_VENV_PATH)/bin/activate && cd $(WS_ROOT)/src/vpp-api/python && python setup.py install"
	@touch $@

define retest-func
@env VPP_TEST_FAILED_DIR=$(VPP_TEST_FAILED_DIR) scripts/setsid_wrapper.sh $(FORCE_FOREGROUND) $(PYTHON_VENV_PATH)/bin/activate python run_tests.py -d $(TEST_DIR) $(UNITTEST_EXTRA_OPTS) || env VPP_TEST_FAILED_DIR=$(VPP_TEST_FAILED_DIR) COMPRESS_FAILED_TEST_LOGS=$(COMPRESS_FAILED_TEST_LOGS) scripts/compress_failed.sh
endef

.PHONY: sanity

ifeq ($(SANITY),no)
SANITY_IMPORT_VPP_PAPI_CMD=true
SANITY_RUN_VPP_CMD=true
else
SANITY_IMPORT_VPP_PAPI_CMD=source $(PYTHON_VENV_PATH)/bin/activate && python sanity_import_vpp_papi.py
SANITY_RUN_VPP_CMD=source $(PYTHON_VENV_PATH)/bin/activate && python sanity_run_vpp.py
endif

sanity: verify-no-running-vpp
	@sys_req/dev_shm_size.sh
	@bash -c "$(SANITY_IMPORT_VPP_PAPI_CMD) ||\
		(echo \"*******************************************************************\" &&\
		 echo \"* Sanity check failed, cannot import vpp_papi\" &&\
		 echo \"* to debug: \" &&\
		 echo \"* 1. enter test shell:   make test-shell\" &&\
		 echo \"* 2. execute debugger:   gdb python -ex 'run sanity_import_vpp_papi.py'\" &&\
	         echo \"*******************************************************************\" &&\
		 false)"
	@bash -c "$(SANITY_RUN_VPP_CMD) ||\
		(echo \"*******************************************************************\" &&\
		 echo \"* Sanity check failed, cannot run vpp\" &&\
	         echo \"*******************************************************************\" &&\
		 false)"

.PHONY: ext
ext:
	make -C ext

test: verify-python-path $(PAPI_INSTALL_DONE) ext sanity reset
	$(call retest-func)

retest: verify-python-path sanity reset
	$(call retest-func)

shell: verify-python-path $(PAPI_INSTALL_DONE)
	@echo "source $(PYTHON_VENV_PATH)/bin/activate;\
		echo '***';\
		echo VPP_TEST_BUILD_DIR=$(VPP_TEST_BUILD_DIR);\
		echo VPP_TEST_BIN=$(VPP_TEST_BIN);\
		echo VPP_TEST_PLUGIN_PATH=$(VPP_TEST_PLUGIN_PATH);\
		echo VPP_TEST_INSTALL_PATH=$(VPP_TEST_INSTALL_PATH);\
		echo EXTERN_TESTS=$(EXTERN_TESTS);\
		echo EXTERN_PLUGINS=$(EXTERN_PLUGINS);\
                echo EXTERN_COV_DIR=$(EXTERN_COV_DIR);\
		echo LD_LIBRARY_PATH=$(LD_LIBRARY_PATH);\
		echo '***';\
		exec </dev/tty" | bash -i

.PHONY: wipe doc

reset:
	@rm -f /dev/shm/vpp-unittest-*
	@rm -rf /tmp/vpp-unittest-*
	@rm -rf $(VPP_TEST_FAILED_DIR)
	@mkdir $(VPP_TEST_FAILED_DIR)

wipe: reset
	@make -C ext clean
	@rm -rf $(PYTHON_VENV_PATH)
	@rm -f $(PAPI_INSTALL_FLAGS)

doc: verify-python-path $(PIP_PATCH_DONE)
	@virtualenv $(PYTHON_VENV_PATH) -p $(PYTHON_INTERP)
	@bash -c "source $(PYTHON_VENV_PATH)/bin/activate && pip install sphinx sphinx-rtd-theme"
	@bash -c "source $(PYTHON_VENV_PATH)/bin/activate && make -C doc WS_ROOT=$(WS_ROOT) BR=$(BR) NO_VPP_PAPI=1 html"

.PHONY: wipe-doc

wipe-doc:
	@make -C doc wipe BR=$(BR)

cov: wipe-cov reset ext verify-python-path $(PAPI_INSTALL_DONE)
	@lcov --zerocounters --directory $(VPP_TEST_BUILD_DIR)
	@test -z "$(EXTERN_COV_DIR)" || lcov --zerocounters --directory $(EXTERN_COV_DIR)
	$(call retest-func)
	@mkdir $(BUILD_COV_DIR)
	@lcov --capture --directory $(VPP_TEST_BUILD_DIR) --output-file $(BUILD_COV_DIR)/coverage.info
	@test -z "$(EXTERN_COV_DIR)" || lcov --capture --directory $(EXTERN_COV_DIR) --output-file $(BUILD_COV_DIR)/extern-coverage.info
	@genhtml $(BUILD_COV_DIR)/coverage.info --output-directory $(BUILD_COV_DIR)/html
	@test -z "$(EXTERN_COV_DIR)" || genhtml $(BUILD_COV_DIR)/extern-coverage.info --output-directory $(BUILD_COV_DIR)/extern-html
	@echo
	@echo "Build finished. Code coverage report is in $(BUILD_COV_DIR)/html/index.html"
	@test -z "$(EXTERN_COV_DIR)" || echo "Code coverage report for out-of-tree objects is in $(BUILD_COV_DIR)/extern-html/index.html"

.PHONY: wipe-cov

wipe-cov: wipe
	@rm -rf $(BUILD_COV_DIR)

.PHONY: checkstyle
checkstyle: verify-python-path
	@virtualenv $(PYTHON_VENV_PATH) -p $(PYTHON_INTERP)
	@bash -c "source $(PYTHON_VENV_PATH)/bin/activate && pip install pycodestyle"
	@bash -c "source $(PYTHON_VENV_PATH)/bin/activate &&\
		pycodestyle --show-source --ignore=E126,E241,E226,E305,E704,E741,E722 -v $(WS_ROOT)/test/*.py ||\
		(echo \"*******************************************************************\" &&\
		 echo \"* Test framework PEP8 compliance check FAILED \" &&\
	         echo \"*******************************************************************\" &&\
		 false)"
	@echo "*******************************************************************"
	@echo "* Test framework PEP8 compliance check passed"
	@echo "*******************************************************************"

help:
	@echo "Running tests:"
	@echo ""
	@echo " test                - build and run (basic) functional tests"
	@echo " test-debug          - build and run (basic) functional tests (debug build)"
	@echo " test-all            - build and run (all) functional tests"
	@echo " test-all-debug      - build and run (all) functional tests (debug build)"
	@echo " retest              - run functional tests"
	@echo " retest-debug        - run functional tests (debug build)"
	@echo " test-wipe           - wipe (temporary) files generated by unit tests"
	@echo " test-shell          - enter shell with test environment"
	@echo " test-shell-debug    - enter shell with test environment (debug build)"
	@echo ""
	@echo "Arguments controlling test runs:"
	@echo " V=[0|1|2]            - set test verbosity level"
	@echo " CACHE_OUTPUT=[0|1]   - cache VPP stdout/stderr and log as one block after test finishes (default: 1)"
	@echo " FAILFAST=[0|1]       - fail fast if 1, complete all tests if 0"
	@echo " TIMEOUT=<timeout>    - fail test suite if any single test takes longer than <timeout> (in seconds) to finish"
	@echo " RETRIES=<n>          - retry failed tests <n> times"
	@echo " DEBUG=<type>         - set VPP debugging kind"
	@echo "    DEBUG=core        - detect coredump and load it in gdb on crash"
	@echo "    DEBUG=gdb         - allow easy debugging by printing VPP PID "
	@echo "                        and waiting for user input before running "
	@echo "                        and tearing down a testcase"
	@echo "    DEBUG=gdbserver   - run gdb inside a gdb server, otherwise "
	@echo "                        same as above"
	@echo " STEP=[yes|no]        - ease debugging by stepping through a testcase "
	@echo " SANITY=[yes|no]      - perform sanity import of vpp-api/sanity vpp run before running tests (default: yes)"
	@echo " EXTENDED_TESTS=[1|y] - run extended tests"
	@echo " TEST=<filter>        - filter the set of tests:"
	@echo "    by file-name      - only run tests from specified file, e.g. TEST=test_bfd selects all tests from test_bfd.py"
	@echo "    by file-suffix    - same as file-name, but 'test_' is omitted e.g. TEST=bfd selects all tests from test_bfd.py"
	@echo "    by wildcard       - wildcard filter is <file>.<class>.<test function>, each can be replaced by '*'"
	@echo "                        e.g. TEST='test_bfd.*.*' is equivalent to above example of filter by file-name"
	@echo "                             TEST='bfd.*.*' is equivalent to above example of filter by file-suffix"
	@echo "                             TEST='bfd.BFDAPITestCase.*' selects all tests from test_bfd.py which are part of BFDAPITestCase class"
	@echo "                             TEST='bfd.BFDAPITestCase.test_add_bfd' selects a single test named test_add_bfd from test_bfd.py/BFDAPITestCase"
	@echo "                             TEST='*.*.test_add_bfd' selects all test functions named test_add_bfd from all files/classes"
	@echo ""
	@echo " VPP_ZOMBIE_NOCHECK=1 - skip checking for vpp (zombie) processes (CAUTION)"
	@echo " COREDUMP_SIZE=<size> - pass <size> as unix { coredump-size <size> } argument to vpp"
	@echo "                        e.g. COREDUMP_SIZE=4g"
	@echo "                             COREDUMP_SIZE=unlimited"
	@echo " EXTERN_TESTS=<path>  - path to out-of-tree test_<name>.py files containing test cases"
	@echo " EXTERN_PLUGINS=<path>- path to out-of-tree plugins to be loaded by vpp under test"
	@echo " EXTERN_COV_DIR=<path>- path to out-of-tree prefix, where source, object and .gcda files can be found for coverage report"
	@echo ""
	@echo " TEST_DEBUG=1         - turn on debugging of the test framework itself (expert)"
	@echo ""
	@echo "Creating test documentation"
	@echo " test-doc            - generate documentation for test framework"
	@echo " test-wipe-doc       - wipe documentation for test framework"
	@echo ""
	@echo "Creating test code coverage report"
	@echo " test-cov            - generate code coverage report for test framework"
	@echo " test-wipe-cov       - wipe code coverage report for test framework"
	@echo ""
	@echo "Verifying code-style"
	@echo " test-checkstyle     - check PEP8 compliance"
	@echo ""
ass="p">{ return pool_elt_at_index (sm->segments, segment_index); } /** * Reads a segment from the segment manager's pool and acquires reader lock * * Caller must drop the reader's lock by calling * @ref segment_manager_segment_reader_unlock once it finishes working with * the segment. */ svm_fifo_segment_private_t * segment_manager_get_segment_w_lock (segment_manager_t * sm, u32 segment_index) { clib_rwlock_reader_lock (&sm->segments_rwlock); return pool_elt_at_index (sm->segments, segment_index); } void segment_manager_segment_reader_unlock (segment_manager_t * sm) { ASSERT (sm->segments_rwlock->n_readers > 0); clib_rwlock_reader_unlock (&sm->segments_rwlock); } void segment_manager_segment_writer_unlock (segment_manager_t * sm) { clib_rwlock_writer_unlock (&sm->segments_rwlock); } /** * Adds segment to segment manager's pool * * If needed a writer's lock is acquired before allocating a new segment * to avoid affecting any of the segments pool readers. */ int segment_manager_add_segment (segment_manager_t * sm, u32 segment_size) { segment_manager_main_t *smm = &segment_manager_main; u32 rnd_margin = 128 << 10, seg_index; segment_manager_properties_t *props; uword baseva = (u64) ~ 0, alloc_size; svm_fifo_segment_private_t *seg; u8 *seg_name; int rv; props = segment_manager_properties_get (sm); /* Not configured for addition of new segments and not first */ if (!props->add_segment && !segment_size) { clib_warning ("cannot allocate new segment"); return VNET_API_ERROR_INVALID_VALUE; } /* * Allocate fifo segment and lock if needed */ if (vlib_num_workers ()) { clib_rwlock_writer_lock (&sm->segments_rwlock); pool_get (sm->segments, seg); } else { pool_get (sm->segments, seg); } memset (seg, 0, sizeof (*seg)); /* * Initialize ssvm segment and svm fifo private header */ segment_size = segment_size ? segment_size : props->add_segment_size; if (props->segment_type != SSVM_SEGMENT_PRIVATE) { seg_name = format (0, "%d-%d%c", getpid (), segment_name_counter++, 0); alloc_size = segment_size + rnd_margin; baseva = clib_valloc_alloc (&smm->va_allocator, alloc_size, 0); if (!baseva) { clib_warning ("out of space for segments"); return -1; } } else seg_name = format (0, "%s%c", "process-private-segment", 0); seg->ssvm.ssvm_size = segment_size; seg->ssvm.name = seg_name; seg->ssvm.requested_va = baseva; if ((rv = ssvm_master_init (&seg->ssvm, props->segment_type))) { clib_warning ("svm_master_init ('%v', %u) failed", seg_name, segment_size); if (props->segment_type != SSVM_SEGMENT_PRIVATE) clib_valloc_free (&smm->va_allocator, baseva); pool_put (sm->segments, seg); return (rv); } svm_fifo_segment_init (seg); /* * Save segment index before dropping lock, if any held */ seg_index = seg - sm->segments; if (vlib_num_workers ()) clib_rwlock_writer_unlock (&sm->segments_rwlock); return seg_index; } segment_manager_t * segment_manager_new () { segment_manager_main_t *smm = &segment_manager_main; segment_manager_t *sm; pool_get (smm->segment_managers, sm); memset (sm, 0, sizeof (*sm)); clib_rwlock_init (&sm->segments_rwlock); return sm; } /** * Initializes segment manager based on options provided. * Returns error if ssvm segment(s) allocation fails. */ int segment_manager_init (segment_manager_t * sm, u32 first_seg_size, u32 prealloc_fifo_pairs) { u32 rx_fifo_size, tx_fifo_size, pair_size; u32 rx_rounded_data_size, tx_rounded_data_size; u64 approx_total_size, max_seg_size = ((u64) 1 << 32) - clib_mem_get_page_size (); segment_manager_properties_t *props; svm_fifo_segment_private_t *segment; u32 approx_segment_count; int seg_index, i; props = segment_manager_properties_get (sm); first_seg_size = clib_max (first_seg_size, default_segment_size); if (prealloc_fifo_pairs) { /* Figure out how many segments should be preallocated */ rx_rounded_data_size = (1 << (max_log2 (props->rx_fifo_size))); tx_rounded_data_size = (1 << (max_log2 (props->tx_fifo_size))); rx_fifo_size = sizeof (svm_fifo_t) + rx_rounded_data_size; tx_fifo_size = sizeof (svm_fifo_t) + tx_rounded_data_size; pair_size = rx_fifo_size + tx_fifo_size; approx_total_size = (u64) prealloc_fifo_pairs *pair_size; if (first_seg_size > approx_total_size) max_seg_size = first_seg_size; approx_segment_count = (approx_total_size + (max_seg_size - 1)) / max_seg_size; /* Allocate the segments */ for (i = 0; i < approx_segment_count + 1; i++) { seg_index = segment_manager_add_segment (sm, max_seg_size); if (seg_index < 0) { clib_warning ("Failed to preallocate segment %d", i); return seg_index; } segment = segment_manager_get_segment (sm, seg_index); if (i == 0) sm->event_queue = segment_manager_alloc_queue (segment, props->evt_q_size); svm_fifo_segment_preallocate_fifo_pairs (segment, props->rx_fifo_size, props->tx_fifo_size, &prealloc_fifo_pairs); svm_fifo_segment_flags (segment) = FIFO_SEGMENT_F_IS_PREALLOCATED; if (prealloc_fifo_pairs == 0) break; } } else { seg_index = segment_manager_add_segment (sm, first_seg_size); if (seg_index) { clib_warning ("Failed to allocate segment"); return seg_index; } segment = segment_manager_get_segment (sm, seg_index); sm->event_queue = segment_manager_alloc_queue (segment, props->evt_q_size); } return 0; } u8 segment_manager_has_fifos (segment_manager_t * sm) { svm_fifo_segment_private_t *seg; u8 first = 1; /* *INDENT-OFF* */ segment_manager_foreach_segment_w_lock (seg, sm, ({ if (CLIB_DEBUG && !first && !svm_fifo_segment_has_fifos (seg) && !(svm_fifo_segment_flags (seg) & FIFO_SEGMENT_F_IS_PREALLOCATED)) { clib_warning ("segment %d has no fifos!", segment_manager_segment_index (sm, seg)); first = 0; } if (svm_fifo_segment_has_fifos (seg)) { segment_manager_segment_reader_unlock (sm); return 1; } })); /* *INDENT-ON* */ return 0; } /** * Initiate disconnects for all sessions 'owned' by a segment manager */ void segment_manager_del_sessions (segment_manager_t * sm) { svm_fifo_segment_private_t *fifo_segment; stream_session_t *session; svm_fifo_t *fifo; ASSERT (pool_elts (sm->segments) != 0); /* Across all fifo segments used by the server */ /* *INDENT-OFF* */ segment_manager_foreach_segment_w_lock (fifo_segment, sm, ({ fifo = svm_fifo_segment_get_fifo_list (fifo_segment); /* * Remove any residual sessions from the session lookup table * Don't bother deleting the individual fifos, we're going to * throw away the fifo segment in a minute. */ while (fifo) { session = session_get (fifo->master_session_index, fifo->master_thread_index); stream_session_disconnect (session); fifo = fifo->next; } /* Instead of removing the segment, test when cleaning up disconnected * sessions if the segment can be removed. */ })); /* *INDENT-ON* */ } /** * Removes segment manager. * * Since the fifos allocated in the segment keep backpointers to the sessions * prior to removing the segment, we call session disconnect. This * subsequently propagates into transport. */ void segment_manager_del (segment_manager_t * sm) { segment_manager_main_t *smm = &segment_manager_main; svm_fifo_segment_private_t *fifo_segment; ASSERT (!segment_manager_has_fifos (sm) && segment_manager_app_detached (sm)); /* If we have empty preallocated segments that haven't been removed, remove * them now. Apart from that, the first segment in the first segment manager * is not removed when all fifos are removed. It can only be removed when * the manager is explicitly deleted/detached by the app. */ clib_rwlock_writer_lock (&sm->segments_rwlock); /* *INDENT-OFF* */ pool_foreach (fifo_segment, sm->segments, ({ segment_manager_del_segment (sm, fifo_segment); })); /* *INDENT-ON* */ clib_rwlock_writer_unlock (&sm->segments_rwlock); clib_rwlock_free (&sm->segments_rwlock); if (CLIB_DEBUG) memset (sm, 0xfe, sizeof (*sm)); pool_put (smm->segment_managers, sm); } void segment_manager_init_del (segment_manager_t * sm) { segment_manager_app_detach (sm); if (segment_manager_has_fifos (sm)) segment_manager_del_sessions (sm); else { ASSERT (!sm->first_is_protected || segment_manager_app_detached (sm)); segment_manager_del (sm); } } int segment_manager_try_alloc_fifos (svm_fifo_segment_private_t * fifo_segment, u32 rx_fifo_size, u32 tx_fifo_size, svm_fifo_t ** rx_fifo, svm_fifo_t ** tx_fifo) { rx_fifo_size = clib_max (rx_fifo_size, default_fifo_size); *rx_fifo = svm_fifo_segment_alloc_fifo (fifo_segment, rx_fifo_size, FIFO_SEGMENT_RX_FREELIST); tx_fifo_size = clib_max (tx_fifo_size, default_fifo_size); *tx_fifo = svm_fifo_segment_alloc_fifo (fifo_segment, tx_fifo_size, FIFO_SEGMENT_TX_FREELIST); if (*rx_fifo == 0) { /* This would be very odd, but handle it... */ if (*tx_fifo != 0) { svm_fifo_segment_free_fifo (fifo_segment, *tx_fifo, FIFO_SEGMENT_TX_FREELIST); *tx_fifo = 0; } return -1; } if (*tx_fifo == 0) { if (*rx_fifo != 0) { svm_fifo_segment_free_fifo (fifo_segment, *rx_fifo, FIFO_SEGMENT_RX_FREELIST); *rx_fifo = 0; } return -1; } return 0; } int segment_manager_alloc_session_fifos (segment_manager_t * sm, svm_fifo_t ** rx_fifo, svm_fifo_t ** tx_fifo, u32 * fifo_segment_index) { svm_fifo_segment_private_t *fifo_segment = 0; int alloc_fail = 1, rv = 0, new_fs_index; segment_manager_properties_t *props; u8 added_a_segment = 0; u32 sm_index; props = segment_manager_properties_get (sm); /* * Find the first free segment to allocate the fifos in */ /* *INDENT-OFF* */ segment_manager_foreach_segment_w_lock (fifo_segment, sm, ({ alloc_fail = segment_manager_try_alloc_fifos (fifo_segment, props->rx_fifo_size, props->tx_fifo_size, rx_fifo, tx_fifo); /* Exit with lock held, drop it after notifying app */ if (!alloc_fail) goto alloc_success; })); /* *INDENT-ON* */ alloc_check: if (!alloc_fail) { alloc_success: ASSERT (rx_fifo && tx_fifo); sm_index = segment_manager_index (sm); (*tx_fifo)->segment_manager = sm_index; (*rx_fifo)->segment_manager = sm_index; *fifo_segment_index = segment_manager_segment_index (sm, fifo_segment); if (added_a_segment) rv = application_add_segment_notify (sm->app_index, &fifo_segment->ssvm); /* Drop the lock after app is notified */ segment_manager_segment_reader_unlock (sm); return rv; } /* * Allocation failed, see if we can add a new segment */ if (props->add_segment) { if (added_a_segment) { clib_warning ("Added a segment, still can't allocate a fifo"); segment_manager_segment_reader_unlock (sm); return SESSION_ERROR_NEW_SEG_NO_SPACE; } if ((new_fs_index = segment_manager_add_segment (sm, 0)) < 0) { clib_warning ("Failed to add new segment"); return SESSION_ERROR_SEG_CREATE; } fifo_segment = segment_manager_get_segment_w_lock (sm, new_fs_index); alloc_fail = segment_manager_try_alloc_fifos (fifo_segment, props->rx_fifo_size, props->tx_fifo_size, rx_fifo, tx_fifo); added_a_segment = 1; goto alloc_check; } else { clib_warning ("Can't add new seg and no space to allocate fifos!"); return SESSION_ERROR_NO_SPACE; } } void segment_manager_dealloc_fifos (u32 segment_index, svm_fifo_t * rx_fifo, svm_fifo_t * tx_fifo) { svm_fifo_segment_private_t *fifo_segment; segment_manager_t *sm; /* It's possible to have no segment manager if the session was removed * as result of a detach. */ if (!(sm = segment_manager_get_if_valid (rx_fifo->segment_manager))) return; fifo_segment = segment_manager_get_segment_w_lock (sm, segment_index); svm_fifo_segment_free_fifo (fifo_segment, rx_fifo, FIFO_SEGMENT_RX_FREELIST); svm_fifo_segment_free_fifo (fifo_segment, tx_fifo, FIFO_SEGMENT_TX_FREELIST); /* * Try to remove svm segment if it has no fifos. This can be done only if * the segment is not the first in the segment manager or if it is first * and it is not protected. Moreover, if the segment is first and the app * has detached from the segment manager, remove the segment manager. */ if (!svm_fifo_segment_has_fifos (fifo_segment)) { segment_manager_segment_reader_unlock (sm); /* Remove segment if it holds no fifos or first but not protected */ if (segment_index != 0 || !sm->first_is_protected) segment_manager_lock_and_del_segment (sm, segment_index); /* Remove segment manager if no sessions and detached from app */ if (segment_manager_app_detached (sm) && !segment_manager_has_fifos (sm)) segment_manager_del (sm); } else segment_manager_segment_reader_unlock (sm); } /** * Allocates shm queue in the first segment * * Must be called with lock held */ svm_queue_t * segment_manager_alloc_queue (svm_fifo_segment_private_t * segment, u32 queue_size) { ssvm_shared_header_t *sh; svm_queue_t *q; void *oldheap; sh = segment->ssvm.sh; oldheap = ssvm_push_heap (sh); q = svm_queue_init (queue_size, sizeof (session_fifo_event_t), 0 /* consumer pid */ , 0 /* signal when queue non-empty */ ); ssvm_pop_heap (oldheap); return q; } /** * Frees shm queue allocated in the first segment */ void segment_manager_dealloc_queue (segment_manager_t * sm, svm_queue_t * q) { svm_fifo_segment_private_t *segment; ssvm_shared_header_t *sh; void *oldheap; ASSERT (!pool_is_free_index (sm->segments, 0)); segment = segment_manager_get_segment_w_lock (sm, 0); sh = segment->ssvm.sh; oldheap = ssvm_push_heap (sh); svm_queue_free (q); ssvm_pop_heap (oldheap); segment_manager_segment_reader_unlock (sm); } /* * Init segment vm address allocator */ void segment_manager_main_init (segment_manager_main_init_args_t * a) { segment_manager_main_t *sm = &segment_manager_main; clib_valloc_chunk_t _ip, *ip = &_ip; ip->baseva = a->baseva; ip->size = a->size; clib_valloc_init (&sm->va_allocator, ip, 1 /* lock */ ); } static clib_error_t * segment_manager_show_fn (vlib_main_t * vm, unformat_input_t * input, vlib_cli_command_t * cmd) { segment_manager_main_t *smm = &segment_manager_main; svm_fifo_segment_private_t *seg; segment_manager_t *sm; u8 show_segments = 0, verbose = 0; uword address; u64 size; u32 active_fifos; u32 free_fifos; while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { if (unformat (input, "segments")) show_segments = 1; else if (unformat (input, "verbose")) verbose = 1; else return clib_error_return (0, "unknown input `%U'", format_unformat_error, input); } vlib_cli_output (vm, "%d segment managers allocated", pool_elts (smm->segment_managers)); if (verbose && pool_elts (smm->segment_managers)) { vlib_cli_output (vm, "%-10s%=15s%=12s", "Index", "App Index", "Segments"); /* *INDENT-OFF* */ pool_foreach (sm, smm->segment_managers, ({ vlib_cli_output (vm, "%-10d%=15d%=12d", segment_manager_index(sm), sm->app_index, pool_elts (sm->segments)); })); /* *INDENT-ON* */ } if (show_segments) { vlib_cli_output (vm, "%-15s%15s%15s%15s%15s%15s", "Name", "Type", "HeapSize (M)", "ActiveFifos", "FreeFifos", "Address"); /* *INDENT-OFF* */ pool_foreach (sm, smm->segment_managers, ({ segment_manager_foreach_segment_w_lock (seg, sm, ({ svm_fifo_segment_info (seg, &address, &size); active_fifos = svm_fifo_segment_num_fifos (seg); free_fifos = svm_fifo_segment_num_free_fifos (seg, ~0 /* size */); vlib_cli_output (vm, "%-15v%15U%15llu%15u%15u%15llx", ssvm_name (&seg->ssvm), format_svm_fifo_segment_type, seg, size >> 20ULL, active_fifos, free_fifos, address); if (verbose) vlib_cli_output (vm, "%U", format_svm_fifo_segment, seg, verbose); })); })); /* *INDENT-ON* */ } return 0; } /* *INDENT-OFF* */ VLIB_CLI_COMMAND (segment_manager_show_command, static) = { .path = "show segment-manager", .short_help = "show segment-manager [segments][verbose]", .function = segment_manager_show_fn, }; /* *INDENT-ON* */ /* * fd.io coding-style-patch-verification: ON * * Local Variables: * eval: (c-set-style "gnu") * End: */