aboutsummaryrefslogtreecommitdiffstats
path: root/vnet
diff options
context:
space:
mode:
authorDamjan Marion <damarion@cisco.com>2016-08-16 18:56:08 +0200
committerDave Barach <openvpp@barachs.net>2016-08-16 22:56:03 +0000
commit256656735c49983a5a2ca94015facca717094730 (patch)
tree0b7b88edd92123e6715796de2e9b09f511f3352a /vnet
parent66ea26b1bc7bbc8d54a3498dbd3d0919c4712fa8 (diff)
Coding standards cleanup in vnet/vnet/devices/dpdk, partially fixes VPP-248
Change-Id: Ie3ee4ccc126c5db9020779cbd06b0d1da99c8411 Signed-off-by: Damjan Marion <damarion@cisco.com>
Diffstat (limited to 'vnet')
-rw-r--r--vnet/vnet/devices/dpdk/cli.c1027
-rw-r--r--vnet/vnet/devices/dpdk/device.c1311
-rw-r--r--vnet/vnet/devices/dpdk/dpdk.h329
-rw-r--r--vnet/vnet/devices/dpdk/dpdk_priv.h175
-rw-r--r--vnet/vnet/devices/dpdk/format.c539
-rw-r--r--vnet/vnet/devices/dpdk/init.c1621
-rw-r--r--vnet/vnet/devices/dpdk/node.c704
-rw-r--r--vnet/vnet/devices/dpdk/vhost_user.c2075
8 files changed, 4146 insertions, 3635 deletions
diff --git a/vnet/vnet/devices/dpdk/cli.c b/vnet/vnet/devices/dpdk/cli.c
index a40dcc7c9ad..f257a8bb892 100644
--- a/vnet/vnet/devices/dpdk/cli.c
+++ b/vnet/vnet/devices/dpdk/cli.c
@@ -27,199 +27,204 @@
static clib_error_t *
pcap_trace_command_fn (vlib_main_t * vm,
- unformat_input_t * input,
- vlib_cli_command_t * cmd)
+ unformat_input_t * input, vlib_cli_command_t * cmd)
{
- dpdk_main_t * dm = &dpdk_main;
- u8 * filename;
+ dpdk_main_t *dm = &dpdk_main;
+ u8 *filename;
u32 max;
int matched = 0;
- clib_error_t * error = 0;
+ clib_error_t *error = 0;
while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT)
{
if (unformat (input, "on"))
- {
- if (dm->tx_pcap_enable == 0)
- {
- if (dm->pcap_filename == 0)
- dm->pcap_filename = format (0, "/tmp/vpe.pcap%c", 0);
-
- memset (&dm->pcap_main, 0, sizeof (dm->pcap_main));
- dm->pcap_main.file_name = (char *) dm->pcap_filename;
- dm->pcap_main.n_packets_to_capture = 100;
- if (dm->pcap_pkts_to_capture)
- dm->pcap_main.n_packets_to_capture = dm->pcap_pkts_to_capture;
-
- dm->pcap_main.packet_type = PCAP_PACKET_TYPE_ethernet;
- dm->tx_pcap_enable = 1;
- matched = 1;
- vlib_cli_output (vm, "pcap tx capture on...");
- }
- else
- {
- vlib_cli_output (vm, "pcap tx capture already on...");
- }
- matched = 1;
- }
+ {
+ if (dm->tx_pcap_enable == 0)
+ {
+ if (dm->pcap_filename == 0)
+ dm->pcap_filename = format (0, "/tmp/vpe.pcap%c", 0);
+
+ memset (&dm->pcap_main, 0, sizeof (dm->pcap_main));
+ dm->pcap_main.file_name = (char *) dm->pcap_filename;
+ dm->pcap_main.n_packets_to_capture = 100;
+ if (dm->pcap_pkts_to_capture)
+ dm->pcap_main.n_packets_to_capture = dm->pcap_pkts_to_capture;
+
+ dm->pcap_main.packet_type = PCAP_PACKET_TYPE_ethernet;
+ dm->tx_pcap_enable = 1;
+ matched = 1;
+ vlib_cli_output (vm, "pcap tx capture on...");
+ }
+ else
+ {
+ vlib_cli_output (vm, "pcap tx capture already on...");
+ }
+ matched = 1;
+ }
else if (unformat (input, "off"))
- {
- if (dm->tx_pcap_enable)
- {
- vlib_cli_output (vm, "captured %d pkts...",
- dm->pcap_main.n_packets_captured+1);
- if (dm->pcap_main.n_packets_captured)
- {
- dm->pcap_main.n_packets_to_capture =
- dm->pcap_main.n_packets_captured;
- error = pcap_write (&dm->pcap_main);
- if (error)
- clib_error_report (error);
- else
- vlib_cli_output (vm, "saved to %s...", dm->pcap_filename);
- }
- }
- else
- {
- vlib_cli_output (vm, "pcap tx capture already off...");
- }
-
- dm->tx_pcap_enable = 0;
- matched = 1;
- }
+ {
+ if (dm->tx_pcap_enable)
+ {
+ vlib_cli_output (vm, "captured %d pkts...",
+ dm->pcap_main.n_packets_captured + 1);
+ if (dm->pcap_main.n_packets_captured)
+ {
+ dm->pcap_main.n_packets_to_capture =
+ dm->pcap_main.n_packets_captured;
+ error = pcap_write (&dm->pcap_main);
+ if (error)
+ clib_error_report (error);
+ else
+ vlib_cli_output (vm, "saved to %s...", dm->pcap_filename);
+ }
+ }
+ else
+ {
+ vlib_cli_output (vm, "pcap tx capture already off...");
+ }
+
+ dm->tx_pcap_enable = 0;
+ matched = 1;
+ }
else if (unformat (input, "max %d", &max))
- {
- dm->pcap_pkts_to_capture = max;
- matched = 1;
- }
+ {
+ dm->pcap_pkts_to_capture = max;
+ matched = 1;
+ }
else if (unformat (input, "intfc %U",
- unformat_vnet_sw_interface, dm->vnet_main,
- &dm->pcap_sw_if_index))
- matched = 1;
+ unformat_vnet_sw_interface, dm->vnet_main,
+ &dm->pcap_sw_if_index))
+ matched = 1;
else if (unformat (input, "intfc any"))
- {
- dm->pcap_sw_if_index = 0;
- matched = 1;
- }
+ {
+ dm->pcap_sw_if_index = 0;
+ matched = 1;
+ }
else if (unformat (input, "file %s", &filename))
- {
- u8 * chroot_filename;
- /* Brain-police user path input */
- if (strstr((char *)filename, "..") || index((char *)filename, '/'))
- {
- vlib_cli_output (vm, "illegal characters in filename '%s'",
- filename);
- continue;
- }
-
- chroot_filename = format (0, "/tmp/%s%c", filename, 0);
- vec_free (filename);
-
- if (dm->pcap_filename)
- vec_free (dm->pcap_filename);
- vec_add1 (filename, 0);
- dm->pcap_filename = chroot_filename;
- matched = 1;
- }
+ {
+ u8 *chroot_filename;
+ /* Brain-police user path input */
+ if (strstr ((char *) filename, "..")
+ || index ((char *) filename, '/'))
+ {
+ vlib_cli_output (vm, "illegal characters in filename '%s'",
+ filename);
+ continue;
+ }
+
+ chroot_filename = format (0, "/tmp/%s%c", filename, 0);
+ vec_free (filename);
+
+ if (dm->pcap_filename)
+ vec_free (dm->pcap_filename);
+ vec_add1 (filename, 0);
+ dm->pcap_filename = chroot_filename;
+ matched = 1;
+ }
else if (unformat (input, "status"))
- {
- if (dm->tx_pcap_enable == 0)
- {
- vlib_cli_output (vm, "pcap tx capture is off...");
- continue;
- }
-
- vlib_cli_output (vm, "pcap tx capture: %d of %d pkts...",
- dm->pcap_main.n_packets_captured,
- dm->pcap_main.n_packets_to_capture);
- matched = 1;
- }
+ {
+ if (dm->tx_pcap_enable == 0)
+ {
+ vlib_cli_output (vm, "pcap tx capture is off...");
+ continue;
+ }
+
+ vlib_cli_output (vm, "pcap tx capture: %d of %d pkts...",
+ dm->pcap_main.n_packets_captured,
+ dm->pcap_main.n_packets_to_capture);
+ matched = 1;
+ }
else
- break;
+ break;
}
if (matched == 0)
return clib_error_return (0, "unknown input `%U'",
- format_unformat_error, input);
+ format_unformat_error, input);
return 0;
}
+/* *INDENT-OFF* */
VLIB_CLI_COMMAND (pcap_trace_command, static) = {
.path = "pcap tx trace",
.short_help =
"pcap tx trace on off max <nn> intfc <intfc> file <name> status",
.function = pcap_trace_command_fn,
};
+/* *INDENT-ON* */
static clib_error_t *
show_dpdk_buffer (vlib_main_t * vm, unformat_input_t * input,
- vlib_cli_command_t * cmd)
+ vlib_cli_command_t * cmd)
{
- struct rte_mempool * rmp;
+ struct rte_mempool *rmp;
int i;
- for(i = 0; i < vec_len(vm->buffer_main->pktmbuf_pools); i++)
+ for (i = 0; i < vec_len (vm->buffer_main->pktmbuf_pools); i++)
{
rmp = vm->buffer_main->pktmbuf_pools[i];
if (rmp)
- {
+ {
#if RTE_VERSION >= RTE_VERSION_NUM(16, 7, 0, 0)
- unsigned count = rte_mempool_avail_count(rmp);
- unsigned free_count = rte_mempool_in_use_count(rmp);
+ unsigned count = rte_mempool_avail_count (rmp);
+ unsigned free_count = rte_mempool_in_use_count (rmp);
#else
- unsigned count = rte_mempool_count(rmp);
- unsigned free_count = rte_mempool_free_count(rmp);
+ unsigned count = rte_mempool_count (rmp);
+ unsigned free_count = rte_mempool_free_count (rmp);
#endif
- vlib_cli_output(vm, "name=\"%s\" available = %7d allocated = %7d total = %7d\n",
- rmp->name, (u32)count, (u32)free_count,
- (u32)(count+free_count));
- }
+ vlib_cli_output (vm,
+ "name=\"%s\" available = %7d allocated = %7d total = %7d\n",
+ rmp->name, (u32) count, (u32) free_count,
+ (u32) (count + free_count));
+ }
else
- {
- vlib_cli_output(vm, "rte_mempool is NULL (!)\n");
- }
+ {
+ vlib_cli_output (vm, "rte_mempool is NULL (!)\n");
+ }
}
return 0;
}
+/* *INDENT-OFF* */
VLIB_CLI_COMMAND (cmd_show_dpdk_bufferr,static) = {
.path = "show dpdk buffer",
.short_help = "show dpdk buffer state",
.function = show_dpdk_buffer,
.is_mp_safe = 1,
};
+/* *INDENT-ON* */
static clib_error_t *
test_dpdk_buffer (vlib_main_t * vm, unformat_input_t * input,
- vlib_cli_command_t * cmd)
+ vlib_cli_command_t * cmd)
{
- static u32 * allocated_buffers;
+ static u32 *allocated_buffers;
u32 n_alloc = 0;
u32 n_free = 0;
u32 first, actual_alloc;
- while (unformat_check_input(input) != UNFORMAT_END_OF_INPUT)
+ while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT)
{
if (unformat (input, "allocate %d", &n_alloc))
- ;
+ ;
else if (unformat (input, "free %d", &n_free))
- ;
+ ;
else
- break;
+ break;
}
if (n_free)
{
if (vec_len (allocated_buffers) < n_free)
- return clib_error_return (0, "Can't free %d, only %d allocated",
- n_free, vec_len (allocated_buffers));
+ return clib_error_return (0, "Can't free %d, only %d allocated",
+ n_free, vec_len (allocated_buffers));
- first = vec_len(allocated_buffers) - n_free;
+ first = vec_len (allocated_buffers) - n_free;
vlib_buffer_free (vm, allocated_buffers + first, n_free);
_vec_len (allocated_buffers) = first;
}
@@ -227,186 +232,199 @@ test_dpdk_buffer (vlib_main_t * vm, unformat_input_t * input,
{
first = vec_len (allocated_buffers);
vec_validate (allocated_buffers,
- vec_len (allocated_buffers) + n_alloc - 1);
+ vec_len (allocated_buffers) + n_alloc - 1);
actual_alloc = vlib_buffer_alloc (vm, allocated_buffers + first,
- n_alloc);
+ n_alloc);
_vec_len (allocated_buffers) = first + actual_alloc;
if (actual_alloc < n_alloc)
- vlib_cli_output (vm, "WARNING: only allocated %d buffers",
- actual_alloc);
+ vlib_cli_output (vm, "WARNING: only allocated %d buffers",
+ actual_alloc);
}
vlib_cli_output (vm, "Currently %d buffers allocated",
- vec_len (allocated_buffers));
+ vec_len (allocated_buffers));
- if (allocated_buffers && vec_len(allocated_buffers) == 0)
- vec_free(allocated_buffers);
+ if (allocated_buffers && vec_len (allocated_buffers) == 0)
+ vec_free (allocated_buffers);
return 0;
}
+/* *INDENT-OFF* */
VLIB_CLI_COMMAND (cmd_test_dpdk_buffer,static) = {
.path = "test dpdk buffer",
.short_help = "test dpdk buffer [allocate <nn>][free <nn>]",
.function = test_dpdk_buffer,
.is_mp_safe = 1,
};
+/* *INDENT-ON* */
static void
show_dpdk_device_stats (vlib_main_t * vm, dpdk_device_t * xd)
{
- vlib_cli_output(vm,
- "device_index %d\n"
- " last_burst_sz %d\n"
- " max_burst_sz %d\n"
- " full_frames_cnt %u\n"
- " consec_full_frames_cnt %u\n"
- " congestion_cnt %d\n"
- " last_poll_time %llu\n"
- " max_poll_delay %llu\n"
- " discard_cnt %u\n"
- " total_packet_cnt %u\n",
- xd->device_index,
- xd->efd_agent.last_burst_sz,
- xd->efd_agent.max_burst_sz,
- xd->efd_agent.full_frames_cnt,
- xd->efd_agent.consec_full_frames_cnt,
- xd->efd_agent.congestion_cnt,
- xd->efd_agent.last_poll_time,
- xd->efd_agent.max_poll_delay,
- xd->efd_agent.discard_cnt,
- xd->efd_agent.total_packet_cnt);
-
- u32 device_queue_sz = rte_eth_rx_queue_count(xd->device_index,
- 0 /* queue_id */);
- vlib_cli_output(vm,
- " device_queue_sz %u\n",
- device_queue_sz);
+ vlib_cli_output (vm,
+ "device_index %d\n"
+ " last_burst_sz %d\n"
+ " max_burst_sz %d\n"
+ " full_frames_cnt %u\n"
+ " consec_full_frames_cnt %u\n"
+ " congestion_cnt %d\n"
+ " last_poll_time %llu\n"
+ " max_poll_delay %llu\n"
+ " discard_cnt %u\n"
+ " total_packet_cnt %u\n",
+ xd->device_index,
+ xd->efd_agent.last_burst_sz,
+ xd->efd_agent.max_burst_sz,
+ xd->efd_agent.full_frames_cnt,
+ xd->efd_agent.consec_full_frames_cnt,
+ xd->efd_agent.congestion_cnt,
+ xd->efd_agent.last_poll_time,
+ xd->efd_agent.max_poll_delay,
+ xd->efd_agent.discard_cnt, xd->efd_agent.total_packet_cnt);
+
+ u32 device_queue_sz = rte_eth_rx_queue_count (xd->device_index,
+ 0 /* queue_id */ );
+ vlib_cli_output (vm, " device_queue_sz %u\n", device_queue_sz);
}
static void
show_efd_config (vlib_main_t * vm)
{
- vlib_thread_main_t * tm = vlib_get_thread_main();
- dpdk_main_t * dm = &dpdk_main;
-
- vlib_cli_output(vm,
- "dpdk: (0x%04x) enabled:%d monitor:%d drop_all:%d\n"
- " dpdk_queue_hi_thresh %d\n"
- " consec_full_frames_hi_thresh %d\n"
- "---------\n"
- "worker: (0x%04x) enabled:%d monitor:%d\n"
- " worker_queue_hi_thresh %d\n",
- dm->efd.enabled,
- ((dm->efd.enabled & DPDK_EFD_DISCARD_ENABLED) ? 1:0),
- ((dm->efd.enabled & DPDK_EFD_MONITOR_ENABLED) ? 1:0),
- ((dm->efd.enabled & DPDK_EFD_DROPALL_ENABLED) ? 1:0),
- dm->efd.queue_hi_thresh,
- dm->efd.consec_full_frames_hi_thresh,
- tm->efd.enabled,
- ((tm->efd.enabled & VLIB_EFD_DISCARD_ENABLED) ? 1:0),
- ((dm->efd.enabled & VLIB_EFD_MONITOR_ENABLED) ? 1:0),
- tm->efd.queue_hi_thresh);
- vlib_cli_output(vm,
- "---------\n"
- "ip_prec_bitmap 0x%02x\n"
- "mpls_exp_bitmap 0x%02x\n"
- "vlan_cos_bitmap 0x%02x\n",
- tm->efd.ip_prec_bitmap,
- tm->efd.mpls_exp_bitmap,
- tm->efd.vlan_cos_bitmap);
+ vlib_thread_main_t *tm = vlib_get_thread_main ();
+ dpdk_main_t *dm = &dpdk_main;
+
+ vlib_cli_output (vm,
+ "dpdk: (0x%04x) enabled:%d monitor:%d drop_all:%d\n"
+ " dpdk_queue_hi_thresh %d\n"
+ " consec_full_frames_hi_thresh %d\n"
+ "---------\n"
+ "worker: (0x%04x) enabled:%d monitor:%d\n"
+ " worker_queue_hi_thresh %d\n",
+ dm->efd.enabled,
+ ((dm->efd.enabled & DPDK_EFD_DISCARD_ENABLED) ? 1 : 0),
+ ((dm->efd.enabled & DPDK_EFD_MONITOR_ENABLED) ? 1 : 0),
+ ((dm->efd.enabled & DPDK_EFD_DROPALL_ENABLED) ? 1 : 0),
+ dm->efd.queue_hi_thresh,
+ dm->efd.consec_full_frames_hi_thresh,
+ tm->efd.enabled,
+ ((tm->efd.enabled & VLIB_EFD_DISCARD_ENABLED) ? 1 : 0),
+ ((dm->efd.enabled & VLIB_EFD_MONITOR_ENABLED) ? 1 : 0),
+ tm->efd.queue_hi_thresh);
+ vlib_cli_output (vm,
+ "---------\n"
+ "ip_prec_bitmap 0x%02x\n"
+ "mpls_exp_bitmap 0x%02x\n"
+ "vlan_cos_bitmap 0x%02x\n",
+ tm->efd.ip_prec_bitmap,
+ tm->efd.mpls_exp_bitmap, tm->efd.vlan_cos_bitmap);
}
static clib_error_t *
show_efd (vlib_main_t * vm,
- unformat_input_t * input,
- vlib_cli_command_t * cmd)
+ unformat_input_t * input, vlib_cli_command_t * cmd)
{
- if (unformat(input, "config")) {
- show_efd_config(vm);
- } else if (unformat(input, "dpdk")) {
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_t * xd;
- u32 device_id = ~0;
-
- (void) unformat(input, "device %d", &device_id);
- vec_foreach (xd, dm->devices) {
- if ((xd->device_index == device_id) || (device_id == ~0)) {
+ if (unformat (input, "config"))
+ {
+ show_efd_config (vm);
+ }
+ else if (unformat (input, "dpdk"))
+ {
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_t *xd;
+ u32 device_id = ~0;
+
+ (void) unformat (input, "device %d", &device_id);
+ /* *INDENT-OFF* */
+ vec_foreach (xd, dm->devices)
+ {
+ if ((xd->device_index == device_id) || (device_id == ~0))
+ {
show_dpdk_device_stats(vm, xd);
- }
- }
- } else if (unformat(input, "worker")) {
- vlib_thread_main_t * tm = vlib_get_thread_main();
- vlib_frame_queue_t *fq;
- vlib_thread_registration_t * tr;
- int thread_id;
- u32 num_workers = 0;
- u32 first_worker_index = 0;
- uword * p;
-
- p = hash_get_mem (tm->thread_registrations_by_name, "workers");
- ASSERT (p);
- tr = (vlib_thread_registration_t *) p[0];
- if (tr)
- {
- num_workers = tr->count;
- first_worker_index = tr->first_index;
+ }
}
-
- vlib_cli_output(vm,
- "num_workers %d\n"
- "first_worker_index %d\n"
- "vlib_frame_queues[%d]:\n",
- num_workers,
- first_worker_index,
- tm->n_vlib_mains);
-
- for (thread_id = 0; thread_id < tm->n_vlib_mains; thread_id++) {
- fq = vlib_frame_queues[thread_id];
- if (fq) {
- vlib_cli_output(vm,
- "%2d: frames_queued %u\n"
- " frames_queued_hint %u\n"
- " enqueue_full_events %u\n"
- " enqueue_efd_discards %u\n",
- thread_id,
- (fq->tail - fq->head),
- (fq->tail - fq->head_hint),
- fq->enqueue_full_events,
- fq->enqueue_efd_discards);
- }
- }
- } else if (unformat(input, "help")) {
- vlib_cli_output(vm, "Usage: show efd config | "
- "dpdk [device <id>] | worker\n");
- } else {
- show_efd_config(vm);
+ /* *INDENT-ON* */
+ }
+ else if (unformat (input, "worker"))
+ {
+ vlib_thread_main_t *tm = vlib_get_thread_main ();
+ vlib_frame_queue_t *fq;
+ vlib_thread_registration_t *tr;
+ int thread_id;
+ u32 num_workers = 0;
+ u32 first_worker_index = 0;
+ uword *p;
+
+ p = hash_get_mem (tm->thread_registrations_by_name, "workers");
+ ASSERT (p);
+ tr = (vlib_thread_registration_t *) p[0];
+ if (tr)
+ {
+ num_workers = tr->count;
+ first_worker_index = tr->first_index;
+ }
+
+ vlib_cli_output (vm,
+ "num_workers %d\n"
+ "first_worker_index %d\n"
+ "vlib_frame_queues[%d]:\n",
+ num_workers, first_worker_index, tm->n_vlib_mains);
+
+ for (thread_id = 0; thread_id < tm->n_vlib_mains; thread_id++)
+ {
+ fq = vlib_frame_queues[thread_id];
+ if (fq)
+ {
+ vlib_cli_output (vm,
+ "%2d: frames_queued %u\n"
+ " frames_queued_hint %u\n"
+ " enqueue_full_events %u\n"
+ " enqueue_efd_discards %u\n",
+ thread_id,
+ (fq->tail - fq->head),
+ (fq->tail - fq->head_hint),
+ fq->enqueue_full_events,
+ fq->enqueue_efd_discards);
+ }
+ }
+ }
+ else if (unformat (input, "help"))
+ {
+ vlib_cli_output (vm, "Usage: show efd config | "
+ "dpdk [device <id>] | worker\n");
+ }
+ else
+ {
+ show_efd_config (vm);
}
- return 0;
+ return 0;
}
+/* *INDENT-OFF* */
VLIB_CLI_COMMAND (show_efd_command, static) = {
.path = "show efd",
.short_help = "Show efd [device <id>] | [config]",
.function = show_efd,
};
+/* *INDENT-ON* */
static clib_error_t *
clear_efd (vlib_main_t * vm,
- unformat_input_t * input,
- vlib_cli_command_t * cmd)
+ unformat_input_t * input, vlib_cli_command_t * cmd)
{
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_t * xd;
- vlib_thread_main_t * tm = vlib_get_thread_main();
- vlib_frame_queue_t *fq;
- int thread_id;
-
- vec_foreach (xd, dm->devices) {
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_t *xd;
+ vlib_thread_main_t *tm = vlib_get_thread_main ();
+ vlib_frame_queue_t *fq;
+ int thread_id;
+
+ /* *INDENT-OFF* */
+ vec_foreach (xd, dm->devices)
+ {
xd->efd_agent.last_burst_sz = 0;
xd->efd_agent.max_burst_sz = 0;
xd->efd_agent.full_frames_cnt = 0;
@@ -416,211 +434,280 @@ clear_efd (vlib_main_t * vm,
xd->efd_agent.max_poll_delay = 0;
xd->efd_agent.discard_cnt = 0;
xd->efd_agent.total_packet_cnt = 0;
- }
+ }
+ /* *INDENT-ON* */
- for (thread_id = 0; thread_id < tm->n_vlib_mains; thread_id++) {
- fq = vlib_frame_queues[thread_id];
- if (fq) {
- fq->enqueue_full_events = 0;
- fq->enqueue_efd_discards = 0;
- }
+ for (thread_id = 0; thread_id < tm->n_vlib_mains; thread_id++)
+ {
+ fq = vlib_frame_queues[thread_id];
+ if (fq)
+ {
+ fq->enqueue_full_events = 0;
+ fq->enqueue_efd_discards = 0;
+ }
}
- return 0;
+ return 0;
}
+/* *INDENT-OFF* */
VLIB_CLI_COMMAND (clear_efd_command,static) = {
.path = "clear efd",
.short_help = "Clear early-fast-discard counters",
.function = clear_efd,
};
+/* *INDENT-ON* */
static clib_error_t *
-parse_op_and_prec (vlib_main_t *vm, unformat_input_t *input,
- vlib_cli_command_t *cmd,
- char *prec_type, u8 *prec_bitmap)
+parse_op_and_prec (vlib_main_t * vm, unformat_input_t * input,
+ vlib_cli_command_t * cmd,
+ char *prec_type, u8 * prec_bitmap)
{
- clib_error_t * error = NULL;
- u8 op = 0;
- u8 prec = 0;
-
- if (unformat(input, "ge")) {
- op = EFD_OPERATION_GREATER_OR_EQUAL;
- } else if (unformat(input, "lt")) {
- op = EFD_OPERATION_LESS_THAN;
- } else if (unformat(input, "help")) {
- vlib_cli_output(vm,
- "enter operation [ge | lt] and precedence <0-7>)");
- return (error);
- } else {
- return clib_error_return(0, "unknown input `%U'",
- format_unformat_error, input);
+ clib_error_t *error = NULL;
+ u8 op = 0;
+ u8 prec = 0;
+
+ if (unformat (input, "ge"))
+ {
+ op = EFD_OPERATION_GREATER_OR_EQUAL;
+ }
+ else if (unformat (input, "lt"))
+ {
+ op = EFD_OPERATION_LESS_THAN;
+ }
+ else if (unformat (input, "help"))
+ {
+ vlib_cli_output (vm, "enter operation [ge | lt] and precedence <0-7>)");
+ return (error);
+ }
+ else
+ {
+ return clib_error_return (0, "unknown input `%U'",
+ format_unformat_error, input);
}
- if (unformat (input, "%u", &prec)) {
- if (prec > 7) {
- return clib_error_return(0, "precedence %d is out of range <0-7>",
- prec);
- }
- } else {
- return clib_error_return(0, "unknown input `%U'",
- format_unformat_error, input);
+ if (unformat (input, "%u", &prec))
+ {
+ if (prec > 7)
+ {
+ return clib_error_return (0, "precedence %d is out of range <0-7>",
+ prec);
+ }
+ }
+ else
+ {
+ return clib_error_return (0, "unknown input `%U'",
+ format_unformat_error, input);
}
- set_efd_bitmap(prec_bitmap, prec, op);
+ set_efd_bitmap (prec_bitmap, prec, op);
- vlib_cli_output(vm,
- "EFD will be set for %s precedence %s%u%s.",
- prec_type,
- (op == EFD_OPERATION_LESS_THAN) ? "less than " : "",
- prec,
- (op == EFD_OPERATION_GREATER_OR_EQUAL) ? " and greater" : "");
+ vlib_cli_output (vm,
+ "EFD will be set for %s precedence %s%u%s.",
+ prec_type,
+ (op == EFD_OPERATION_LESS_THAN) ? "less than " : "",
+ prec,
+ (op ==
+ EFD_OPERATION_GREATER_OR_EQUAL) ? " and greater" : "");
- return (error);
+ return (error);
}
static clib_error_t *
-set_efd (vlib_main_t *vm, unformat_input_t *input,
- vlib_cli_command_t *cmd)
+set_efd (vlib_main_t * vm, unformat_input_t * input, vlib_cli_command_t * cmd)
{
- dpdk_main_t * dm = &dpdk_main;
- vlib_thread_main_t * tm = vlib_get_thread_main();
- clib_error_t * error = NULL;
- vlib_node_runtime_t * rt = vlib_node_get_runtime (vm, dpdk_input_node.index);
-
- if (unformat(input, "enable")) {
- if (unformat(input, "dpdk")) {
- dm->efd.enabled |= DPDK_EFD_DISCARD_ENABLED;
- } else if (unformat(input, "worker")) {
- tm->efd.enabled |= VLIB_EFD_DISCARD_ENABLED;
- } else if (unformat(input, "monitor")) {
- dm->efd.enabled |= DPDK_EFD_MONITOR_ENABLED;
- tm->efd.enabled |= VLIB_EFD_MONITOR_ENABLED;
- } else if (unformat(input, "drop_all")) {
- dm->efd.enabled |= DPDK_EFD_DROPALL_ENABLED;
- } else if (unformat(input, "default")) {
- dm->efd.enabled = (DPDK_EFD_DISCARD_ENABLED |
- DPDK_EFD_MONITOR_ENABLED);
- tm->efd.enabled = (VLIB_EFD_DISCARD_ENABLED |
- VLIB_EFD_MONITOR_ENABLED);
- } else {
- return clib_error_return(0, "Usage: set efd enable [dpdk | "
- "worker | monitor | drop_all | default]");
- }
- } else if (unformat(input, "disable")) {
- if (unformat(input, "dpdk")) {
- dm->efd.enabled &= ~DPDK_EFD_DISCARD_ENABLED;
- } else if (unformat(input, "worker")) {
- tm->efd.enabled &= ~VLIB_EFD_DISCARD_ENABLED;
- } else if (unformat(input, "monitor")) {
- dm->efd.enabled &= ~DPDK_EFD_MONITOR_ENABLED;
- tm->efd.enabled &= ~VLIB_EFD_MONITOR_ENABLED;
- } else if (unformat(input, "drop_all")) {
- dm->efd.enabled &= ~DPDK_EFD_DROPALL_ENABLED;
- } else if (unformat(input, "all")) {
- dm->efd.enabled = 0;
- tm->efd.enabled = 0;
- } else {
- return clib_error_return(0, "Usage: set efd disable [dpdk | "
- "worker | monitor | drop_all | all]");
- }
- } else if (unformat(input, "worker_queue_hi_thresh")) {
- u32 mark;
- if (unformat (input, "%u", &mark)) {
- tm->efd.queue_hi_thresh = mark;
- } else {
- return clib_error_return(0, "unknown input `%U'",
- format_unformat_error, input);
- }
- } else if (unformat(input, "dpdk_device_hi_thresh")) {
- u32 thresh;
- if (unformat (input, "%u", &thresh)) {
- dm->efd.queue_hi_thresh = thresh;
- } else {
- return clib_error_return(0, "unknown input `%U'",
- format_unformat_error, input);
- }
- } else if (unformat(input, "consec_full_frames_hi_thresh")) {
- u32 thresh;
- if (unformat (input, "%u", &thresh)) {
- dm->efd.consec_full_frames_hi_thresh = thresh;
- } else {
- return clib_error_return(0, "unknown input `%U'",
- format_unformat_error, input);
- }
- } else if (unformat(input, "ip-prec")) {
- return (parse_op_and_prec(vm, input, cmd,
- "ip", &tm->efd.ip_prec_bitmap));
- } else if (unformat(input, "mpls-exp")) {
- return (parse_op_and_prec(vm, input, cmd,
- "mpls", &tm->efd.mpls_exp_bitmap));
- } else if (unformat(input, "vlan-cos")) {
- return (parse_op_and_prec(vm, input, cmd,
- "vlan", &tm->efd.vlan_cos_bitmap));
- } else if (unformat(input, "help")) {
- vlib_cli_output(vm,
- "Usage:\n"
- " set efd enable <dpdk | worker | monitor | drop_all | default> |\n"
- " set efd disable <dpdk | worker | monitor | drop_all | all> |\n"
- " set efd <ip-prec | mpls-exp | vlan-cos> <ge | lt> <0-7>\n"
- " set efd worker_queue_hi_thresh <0-32> |\n"
- " set efd dpdk_device_hi_thresh <0-%d> |\n"
- " set efd consec_full_frames_hi_thresh <count> |\n",
- DPDK_NB_RX_DESC_10GE);
- } else {
- return clib_error_return(0, "unknown input `%U'",
- format_unformat_error, input);
+ dpdk_main_t *dm = &dpdk_main;
+ vlib_thread_main_t *tm = vlib_get_thread_main ();
+ clib_error_t *error = NULL;
+ vlib_node_runtime_t *rt = vlib_node_get_runtime (vm, dpdk_input_node.index);
+
+ if (unformat (input, "enable"))
+ {
+ if (unformat (input, "dpdk"))
+ {
+ dm->efd.enabled |= DPDK_EFD_DISCARD_ENABLED;
+ }
+ else if (unformat (input, "worker"))
+ {
+ tm->efd.enabled |= VLIB_EFD_DISCARD_ENABLED;
+ }
+ else if (unformat (input, "monitor"))
+ {
+ dm->efd.enabled |= DPDK_EFD_MONITOR_ENABLED;
+ tm->efd.enabled |= VLIB_EFD_MONITOR_ENABLED;
+ }
+ else if (unformat (input, "drop_all"))
+ {
+ dm->efd.enabled |= DPDK_EFD_DROPALL_ENABLED;
+ }
+ else if (unformat (input, "default"))
+ {
+ dm->efd.enabled = (DPDK_EFD_DISCARD_ENABLED |
+ DPDK_EFD_MONITOR_ENABLED);
+ tm->efd.enabled = (VLIB_EFD_DISCARD_ENABLED |
+ VLIB_EFD_MONITOR_ENABLED);
+ }
+ else
+ {
+ return clib_error_return (0, "Usage: set efd enable [dpdk | "
+ "worker | monitor | drop_all | default]");
+ }
+ }
+ else if (unformat (input, "disable"))
+ {
+ if (unformat (input, "dpdk"))
+ {
+ dm->efd.enabled &= ~DPDK_EFD_DISCARD_ENABLED;
+ }
+ else if (unformat (input, "worker"))
+ {
+ tm->efd.enabled &= ~VLIB_EFD_DISCARD_ENABLED;
+ }
+ else if (unformat (input, "monitor"))
+ {
+ dm->efd.enabled &= ~DPDK_EFD_MONITOR_ENABLED;
+ tm->efd.enabled &= ~VLIB_EFD_MONITOR_ENABLED;
+ }
+ else if (unformat (input, "drop_all"))
+ {
+ dm->efd.enabled &= ~DPDK_EFD_DROPALL_ENABLED;
+ }
+ else if (unformat (input, "all"))
+ {
+ dm->efd.enabled = 0;
+ tm->efd.enabled = 0;
+ }
+ else
+ {
+ return clib_error_return (0, "Usage: set efd disable [dpdk | "
+ "worker | monitor | drop_all | all]");
+ }
+ }
+ else if (unformat (input, "worker_queue_hi_thresh"))
+ {
+ u32 mark;
+ if (unformat (input, "%u", &mark))
+ {
+ tm->efd.queue_hi_thresh = mark;
+ }
+ else
+ {
+ return clib_error_return (0, "unknown input `%U'",
+ format_unformat_error, input);
+ }
+ }
+ else if (unformat (input, "dpdk_device_hi_thresh"))
+ {
+ u32 thresh;
+ if (unformat (input, "%u", &thresh))
+ {
+ dm->efd.queue_hi_thresh = thresh;
+ }
+ else
+ {
+ return clib_error_return (0, "unknown input `%U'",
+ format_unformat_error, input);
+ }
+ }
+ else if (unformat (input, "consec_full_frames_hi_thresh"))
+ {
+ u32 thresh;
+ if (unformat (input, "%u", &thresh))
+ {
+ dm->efd.consec_full_frames_hi_thresh = thresh;
+ }
+ else
+ {
+ return clib_error_return (0, "unknown input `%U'",
+ format_unformat_error, input);
+ }
+ }
+ else if (unformat (input, "ip-prec"))
+ {
+ return (parse_op_and_prec (vm, input, cmd,
+ "ip", &tm->efd.ip_prec_bitmap));
+ }
+ else if (unformat (input, "mpls-exp"))
+ {
+ return (parse_op_and_prec (vm, input, cmd,
+ "mpls", &tm->efd.mpls_exp_bitmap));
+ }
+ else if (unformat (input, "vlan-cos"))
+ {
+ return (parse_op_and_prec (vm, input, cmd,
+ "vlan", &tm->efd.vlan_cos_bitmap));
+ }
+ else if (unformat (input, "help"))
+ {
+ vlib_cli_output (vm,
+ "Usage:\n"
+ " set efd enable <dpdk | worker | monitor | drop_all | default> |\n"
+ " set efd disable <dpdk | worker | monitor | drop_all | all> |\n"
+ " set efd <ip-prec | mpls-exp | vlan-cos> <ge | lt> <0-7>\n"
+ " set efd worker_queue_hi_thresh <0-32> |\n"
+ " set efd dpdk_device_hi_thresh <0-%d> |\n"
+ " set efd consec_full_frames_hi_thresh <count> |\n",
+ DPDK_NB_RX_DESC_10GE);
+ }
+ else
+ {
+ return clib_error_return (0, "unknown input `%U'",
+ format_unformat_error, input);
}
- if (dm->efd.enabled)
- rt->function = dpdk_input_efd_multiarch_select();
- else if (dm->use_rss)
- rt->function = dpdk_input_rss_multiarch_select();
- else
- rt->function = dpdk_input_multiarch_select();
+ if (dm->efd.enabled)
+ rt->function = dpdk_input_efd_multiarch_select ();
+ else if (dm->use_rss)
+ rt->function = dpdk_input_rss_multiarch_select ();
+ else
+ rt->function = dpdk_input_multiarch_select ();
- return error;
+ return error;
}
+/* *INDENT-OFF* */
VLIB_CLI_COMMAND (cmd_set_efd,static) = {
.path = "set efd",
.short_help = "set early-fast-discard commands",
.function = set_efd,
};
+/* *INDENT-ON* */
static clib_error_t *
-set_dpdk_if_desc (vlib_main_t *vm, unformat_input_t *input,
- vlib_cli_command_t *cmd)
+set_dpdk_if_desc (vlib_main_t * vm, unformat_input_t * input,
+ vlib_cli_command_t * cmd)
{
- unformat_input_t _line_input, * line_input = &_line_input;
- dpdk_main_t * dm = &dpdk_main;
- vnet_hw_interface_t * hw;
- dpdk_device_t * xd;
- u32 hw_if_index = (u32) ~0;
- u32 nb_rx_desc = (u32) ~0;
- u32 nb_tx_desc = (u32) ~0;
- clib_error_t * rv;
-
- if (! unformat_user (input, unformat_line_input, line_input))
+ unformat_input_t _line_input, *line_input = &_line_input;
+ dpdk_main_t *dm = &dpdk_main;
+ vnet_hw_interface_t *hw;
+ dpdk_device_t *xd;
+ u32 hw_if_index = (u32) ~ 0;
+ u32 nb_rx_desc = (u32) ~ 0;
+ u32 nb_tx_desc = (u32) ~ 0;
+ clib_error_t *rv;
+
+ if (!unformat_user (input, unformat_line_input, line_input))
return 0;
- while (unformat_check_input (line_input) != UNFORMAT_END_OF_INPUT) {
- if (unformat (line_input, "%U", unformat_vnet_hw_interface, dm->vnet_main,
- &hw_if_index))
- ;
- else if (unformat (line_input, "tx %d", &nb_tx_desc))
- ;
- else if (unformat (line_input, "rx %d", &nb_rx_desc))
- ;
- else
- return clib_error_return (0, "parse error: '%U'",
- format_unformat_error, line_input);
- }
+ while (unformat_check_input (line_input) != UNFORMAT_END_OF_INPUT)
+ {
+ if (unformat
+ (line_input, "%U", unformat_vnet_hw_interface, dm->vnet_main,
+ &hw_if_index))
+ ;
+ else if (unformat (line_input, "tx %d", &nb_tx_desc))
+ ;
+ else if (unformat (line_input, "rx %d", &nb_rx_desc))
+ ;
+ else
+ return clib_error_return (0, "parse error: '%U'",
+ format_unformat_error, line_input);
+ }
unformat_free (line_input);
- if (hw_if_index == (u32) ~0)
+ if (hw_if_index == (u32) ~ 0)
return clib_error_return (0, "please specify valid interface name");
hw = vnet_get_hw_interface (dm->vnet_main, hw_if_index);
@@ -628,69 +715,75 @@ set_dpdk_if_desc (vlib_main_t *vm, unformat_input_t *input,
if (xd->dev_type != VNET_DPDK_DEV_ETH)
return clib_error_return (0, "number of descriptors can be set only for "
- "physical devices");
+ "physical devices");
- if ((nb_rx_desc == (u32) ~0 || nb_rx_desc == xd->nb_rx_desc) &&
- (nb_tx_desc == (u32) ~0 || nb_tx_desc == xd->nb_tx_desc))
+ if ((nb_rx_desc == (u32) ~ 0 || nb_rx_desc == xd->nb_rx_desc) &&
+ (nb_tx_desc == (u32) ~ 0 || nb_tx_desc == xd->nb_tx_desc))
return clib_error_return (0, "nothing changed");
- if (nb_rx_desc != (u32) ~0)
- xd->nb_rx_desc = nb_rx_desc;
+ if (nb_rx_desc != (u32) ~ 0)
+ xd->nb_rx_desc = nb_rx_desc;
- if (nb_tx_desc != (u32) ~0)
- xd->nb_rx_desc = nb_rx_desc;
+ if (nb_tx_desc != (u32) ~ 0)
+ xd->nb_rx_desc = nb_rx_desc;
- rv = dpdk_port_setup(dm, xd);
+ rv = dpdk_port_setup (dm, xd);
return rv;
}
+/* *INDENT-OFF* */
VLIB_CLI_COMMAND (cmd_set_dpdk_if_desc,static) = {
.path = "set dpdk interface descriptors",
.short_help = "set dpdk interface descriptors <if-name> [rx <n>] [tx <n>]",
.function = set_dpdk_if_desc,
};
+/* *INDENT-ON* */
static clib_error_t *
-show_dpdk_if_placement (vlib_main_t *vm, unformat_input_t *input,
- vlib_cli_command_t *cmd)
+show_dpdk_if_placement (vlib_main_t * vm, unformat_input_t * input,
+ vlib_cli_command_t * cmd)
{
- vlib_thread_main_t * tm = vlib_get_thread_main();
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_and_queue_t * dq;
+ vlib_thread_main_t *tm = vlib_get_thread_main ();
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_and_queue_t *dq;
int cpu;
if (tm->n_vlib_mains == 1)
- vlib_cli_output(vm, "All interfaces are handled by main thread");
+ vlib_cli_output (vm, "All interfaces are handled by main thread");
- for(cpu = 0; cpu < vec_len(dm->devices_by_cpu); cpu++)
+ for (cpu = 0; cpu < vec_len (dm->devices_by_cpu); cpu++)
{
- if (vec_len(dm->devices_by_cpu[cpu]))
- vlib_cli_output(vm, "Thread %u (%s at lcore %u):", cpu,
- vlib_worker_threads[cpu].name,
- vlib_worker_threads[cpu].dpdk_lcore_id);
+ if (vec_len (dm->devices_by_cpu[cpu]))
+ vlib_cli_output (vm, "Thread %u (%s at lcore %u):", cpu,
+ vlib_worker_threads[cpu].name,
+ vlib_worker_threads[cpu].dpdk_lcore_id);
+ /* *INDENT-OFF* */
vec_foreach(dq, dm->devices_by_cpu[cpu])
{
u32 hw_if_index = dm->devices[dq->device].vlib_hw_if_index;
vnet_hw_interface_t * hi = vnet_get_hw_interface(dm->vnet_main, hw_if_index);
vlib_cli_output(vm, " %v queue %u", hi->name, dq->queue_id);
}
+ /* *INDENT-ON* */
}
return 0;
}
+/* *INDENT-OFF* */
VLIB_CLI_COMMAND (cmd_show_dpdk_if_placement,static) = {
.path = "show dpdk interface placement",
.short_help = "show dpdk interface placement",
.function = show_dpdk_if_placement,
};
+/* *INDENT-ON* */
static int
-dpdk_device_queue_sort(void * a1, void * a2)
+dpdk_device_queue_sort (void *a1, void *a2)
{
- dpdk_device_and_queue_t * dq1 = a1;
- dpdk_device_and_queue_t * dq2 = a2;
+ dpdk_device_and_queue_t *dq1 = a1;
+ dpdk_device_and_queue_t *dq2 = a2;
if (dq1->device > dq2->device)
return 1;
@@ -705,38 +798,40 @@ dpdk_device_queue_sort(void * a1, void * a2)
}
static clib_error_t *
-set_dpdk_if_placement (vlib_main_t *vm, unformat_input_t *input,
- vlib_cli_command_t *cmd)
+set_dpdk_if_placement (vlib_main_t * vm, unformat_input_t * input,
+ vlib_cli_command_t * cmd)
{
- unformat_input_t _line_input, * line_input = &_line_input;
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_and_queue_t * dq;
- vnet_hw_interface_t * hw;
- dpdk_device_t * xd;
- u32 hw_if_index = (u32) ~0;
+ unformat_input_t _line_input, *line_input = &_line_input;
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_and_queue_t *dq;
+ vnet_hw_interface_t *hw;
+ dpdk_device_t *xd;
+ u32 hw_if_index = (u32) ~ 0;
u32 queue = (u32) 0;
- u32 cpu = (u32) ~0;
+ u32 cpu = (u32) ~ 0;
int i;
- if (! unformat_user (input, unformat_line_input, line_input))
+ if (!unformat_user (input, unformat_line_input, line_input))
return 0;
- while (unformat_check_input (line_input) != UNFORMAT_END_OF_INPUT) {
- if (unformat (line_input, "%U", unformat_vnet_hw_interface, dm->vnet_main,
- &hw_if_index))
- ;
- else if (unformat (line_input, "queue %d", &queue))
- ;
- else if (unformat (line_input, "thread %d", &cpu))
- ;
- else
- return clib_error_return (0, "parse error: '%U'",
- format_unformat_error, line_input);
- }
+ while (unformat_check_input (line_input) != UNFORMAT_END_OF_INPUT)
+ {
+ if (unformat
+ (line_input, "%U", unformat_vnet_hw_interface, dm->vnet_main,
+ &hw_if_index))
+ ;
+ else if (unformat (line_input, "queue %d", &queue))
+ ;
+ else if (unformat (line_input, "thread %d", &cpu))
+ ;
+ else
+ return clib_error_return (0, "parse error: '%U'",
+ format_unformat_error, line_input);
+ }
unformat_free (line_input);
- if (hw_if_index == (u32) ~0)
+ if (hw_if_index == (u32) ~ 0)
return clib_error_return (0, "please specify valid interface name");
if (cpu < dm->input_cpu_first_index ||
@@ -746,8 +841,9 @@ set_dpdk_if_placement (vlib_main_t *vm, unformat_input_t *input,
hw = vnet_get_hw_interface (dm->vnet_main, hw_if_index);
xd = vec_elt_at_index (dm->devices, hw->dev_instance);
- for(i = 0; i < vec_len(dm->devices_by_cpu); i++)
+ for (i = 0; i < vec_len (dm->devices_by_cpu); i++)
{
+ /* *INDENT-OFF* */
vec_foreach(dq, dm->devices_by_cpu[i])
{
if (hw_if_index == dm->devices[dq->device].vlib_hw_if_index &&
@@ -780,16 +876,19 @@ set_dpdk_if_placement (vlib_main_t *vm, unformat_input_t *input,
return 0;
}
}
+ /* *INDENT-ON* */
}
return clib_error_return (0, "not found");
}
+/* *INDENT-OFF* */
VLIB_CLI_COMMAND (cmd_set_dpdk_if_placement,static) = {
.path = "set dpdk interface placement",
.short_help = "set dpdk interface placement <if-name> [queue <n>] thread <n>",
.function = set_dpdk_if_placement,
};
+/* *INDENT-ON* */
clib_error_t *
dpdk_cli_init (vlib_main_t * vm)
@@ -798,3 +897,11 @@ dpdk_cli_init (vlib_main_t * vm)
}
VLIB_INIT_FUNCTION (dpdk_cli_init);
+
+/*
+ * fd.io coding-style-patch-verification: ON
+ *
+ * Local Variables:
+ * eval: (c-set-style "gnu")
+ * End:
+ */
diff --git a/vnet/vnet/devices/dpdk/device.c b/vnet/vnet/devices/dpdk/device.c
index 61774f0f3e2..f00b80b6def 100644
--- a/vnet/vnet/devices/dpdk/device.c
+++ b/vnet/vnet/devices/dpdk/device.c
@@ -30,84 +30,92 @@
_(PKT_DROP, "Tx packet drops (dpdk tx failure)") \
_(REPL_FAIL, "Tx packet drops (replication failure)")
-typedef enum {
+typedef enum
+{
#define _(f,s) DPDK_TX_FUNC_ERROR_##f,
foreach_dpdk_tx_func_error
#undef _
- DPDK_TX_FUNC_N_ERROR,
+ DPDK_TX_FUNC_N_ERROR,
} dpdk_tx_func_error_t;
-static char * dpdk_tx_func_error_strings[] = {
+static char *dpdk_tx_func_error_strings[] = {
#define _(n,s) s,
- foreach_dpdk_tx_func_error
+ foreach_dpdk_tx_func_error
#undef _
};
clib_error_t *
-dpdk_set_mac_address (vnet_hw_interface_t * hi, char * address)
+dpdk_set_mac_address (vnet_hw_interface_t * hi, char *address)
{
- int error;
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_t * xd = vec_elt_at_index (dm->devices, hi->dev_instance);
-
- error=rte_eth_dev_default_mac_addr_set(xd->device_index,
- (struct ether_addr *) address);
-
- if (error) {
- return clib_error_return (0, "mac address set failed: %d", error);
- } else {
- return NULL;
- }
+ int error;
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_t *xd = vec_elt_at_index (dm->devices, hi->dev_instance);
+
+ error = rte_eth_dev_default_mac_addr_set (xd->device_index,
+ (struct ether_addr *) address);
+
+ if (error)
+ {
+ return clib_error_return (0, "mac address set failed: %d", error);
+ }
+ else
+ {
+ return NULL;
+ }
}
clib_error_t *
dpdk_set_mc_filter (vnet_hw_interface_t * hi,
- struct ether_addr mc_addr_vec[], int naddr)
+ struct ether_addr mc_addr_vec[], int naddr)
{
int error;
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_t * xd = vec_elt_at_index (dm->devices, hi->dev_instance);
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_t *xd = vec_elt_at_index (dm->devices, hi->dev_instance);
- error=rte_eth_dev_set_mc_addr_list(xd->device_index, mc_addr_vec, naddr);
+ error = rte_eth_dev_set_mc_addr_list (xd->device_index, mc_addr_vec, naddr);
- if (error) {
- return clib_error_return (0, "mc addr list failed: %d", error);
- } else {
- return NULL;
- }
+ if (error)
+ {
+ return clib_error_return (0, "mc addr list failed: %d", error);
+ }
+ else
+ {
+ return NULL;
+ }
}
-struct rte_mbuf * dpdk_replicate_packet_mb (vlib_buffer_t * b)
+struct rte_mbuf *
+dpdk_replicate_packet_mb (vlib_buffer_t * b)
{
- vlib_main_t * vm = vlib_get_main();
- vlib_buffer_main_t * bm = vm->buffer_main;
- struct rte_mbuf * first_mb = 0, * new_mb, * pkt_mb, ** prev_mb_next = 0;
+ vlib_main_t *vm = vlib_get_main ();
+ vlib_buffer_main_t *bm = vm->buffer_main;
+ struct rte_mbuf *first_mb = 0, *new_mb, *pkt_mb, **prev_mb_next = 0;
u8 nb_segs, nb_segs_left;
u32 copy_bytes;
- unsigned socket_id = rte_socket_id();
+ unsigned socket_id = rte_socket_id ();
ASSERT (bm->pktmbuf_pools[socket_id]);
- pkt_mb = rte_mbuf_from_vlib_buffer(b);
+ pkt_mb = rte_mbuf_from_vlib_buffer (b);
nb_segs = pkt_mb->nb_segs;
for (nb_segs_left = nb_segs; nb_segs_left; nb_segs_left--)
{
- if (PREDICT_FALSE(pkt_mb == 0))
+ if (PREDICT_FALSE (pkt_mb == 0))
{
clib_warning ("Missing %d mbuf chain segment(s): "
"(nb_segs = %d, nb_segs_left = %d)!",
nb_segs - nb_segs_left, nb_segs, nb_segs_left);
if (first_mb)
- rte_pktmbuf_free(first_mb);
+ rte_pktmbuf_free (first_mb);
return NULL;
}
new_mb = rte_pktmbuf_alloc (bm->pktmbuf_pools[socket_id]);
- if (PREDICT_FALSE(new_mb == 0))
+ if (PREDICT_FALSE (new_mb == 0))
{
if (first_mb)
- rte_pktmbuf_free(first_mb);
+ rte_pktmbuf_free (first_mb);
return NULL;
}
-
+
/*
* Copy packet info into 1st segment.
*/
@@ -117,7 +125,7 @@ struct rte_mbuf * dpdk_replicate_packet_mb (vlib_buffer_t * b)
rte_pktmbuf_pkt_len (first_mb) = pkt_mb->pkt_len;
first_mb->nb_segs = pkt_mb->nb_segs;
first_mb->port = pkt_mb->port;
-#ifdef DAW_FIXME // TX Offload support TBD
+#ifdef DAW_FIXME // TX Offload support TBD
first_mb->vlan_macip = pkt_mb->vlan_macip;
first_mb->hash = pkt_mb->hash;
first_mb->ol_flags = pkt_mb->ol_flags
@@ -125,58 +133,59 @@ struct rte_mbuf * dpdk_replicate_packet_mb (vlib_buffer_t * b)
}
else
{
- ASSERT(prev_mb_next != 0);
+ ASSERT (prev_mb_next != 0);
*prev_mb_next = new_mb;
}
-
+
/*
* Copy packet segment data into new mbuf segment.
*/
rte_pktmbuf_data_len (new_mb) = pkt_mb->data_len;
copy_bytes = pkt_mb->data_len + RTE_PKTMBUF_HEADROOM;
- ASSERT(copy_bytes <= pkt_mb->buf_len);
- clib_memcpy(new_mb->buf_addr, pkt_mb->buf_addr, copy_bytes);
+ ASSERT (copy_bytes <= pkt_mb->buf_len);
+ clib_memcpy (new_mb->buf_addr, pkt_mb->buf_addr, copy_bytes);
prev_mb_next = &new_mb->next;
pkt_mb = pkt_mb->next;
}
- ASSERT(pkt_mb == 0);
- __rte_mbuf_sanity_check(first_mb, 1);
+ ASSERT (pkt_mb == 0);
+ __rte_mbuf_sanity_check (first_mb, 1);
return first_mb;
}
-struct rte_mbuf * dpdk_zerocopy_replicate_packet_mb (vlib_buffer_t * b)
+struct rte_mbuf *
+dpdk_zerocopy_replicate_packet_mb (vlib_buffer_t * b)
{
- vlib_main_t * vm = vlib_get_main();
- vlib_buffer_main_t * bm = vm->buffer_main;
- struct rte_mbuf * first_mb = 0, * new_mb, * pkt_mb, ** prev_mb_next = 0;
+ vlib_main_t *vm = vlib_get_main ();
+ vlib_buffer_main_t *bm = vm->buffer_main;
+ struct rte_mbuf *first_mb = 0, *new_mb, *pkt_mb, **prev_mb_next = 0;
u8 nb_segs, nb_segs_left;
- unsigned socket_id = rte_socket_id();
+ unsigned socket_id = rte_socket_id ();
ASSERT (bm->pktmbuf_pools[socket_id]);
- pkt_mb = rte_mbuf_from_vlib_buffer(b);
+ pkt_mb = rte_mbuf_from_vlib_buffer (b);
nb_segs = pkt_mb->nb_segs;
for (nb_segs_left = nb_segs; nb_segs_left; nb_segs_left--)
{
- if (PREDICT_FALSE(pkt_mb == 0))
+ if (PREDICT_FALSE (pkt_mb == 0))
{
clib_warning ("Missing %d mbuf chain segment(s): "
"(nb_segs = %d, nb_segs_left = %d)!",
nb_segs - nb_segs_left, nb_segs, nb_segs_left);
if (first_mb)
- rte_pktmbuf_free(first_mb);
+ rte_pktmbuf_free (first_mb);
return NULL;
}
- new_mb = rte_pktmbuf_clone(pkt_mb, bm->pktmbuf_pools[socket_id]);
- if (PREDICT_FALSE(new_mb == 0))
+ new_mb = rte_pktmbuf_clone (pkt_mb, bm->pktmbuf_pools[socket_id]);
+ if (PREDICT_FALSE (new_mb == 0))
{
if (first_mb)
- rte_pktmbuf_free(first_mb);
+ rte_pktmbuf_free (first_mb);
return NULL;
}
-
+
/*
* Copy packet info into 1st segment.
*/
@@ -186,7 +195,7 @@ struct rte_mbuf * dpdk_zerocopy_replicate_packet_mb (vlib_buffer_t * b)
rte_pktmbuf_pkt_len (first_mb) = pkt_mb->pkt_len;
first_mb->nb_segs = pkt_mb->nb_segs;
first_mb->port = pkt_mb->port;
-#ifdef DAW_FIXME // TX Offload support TBD
+#ifdef DAW_FIXME // TX Offload support TBD
first_mb->vlan_macip = pkt_mb->vlan_macip;
first_mb->hash = pkt_mb->hash;
first_mb->ol_flags = pkt_mb->ol_flags
@@ -194,10 +203,10 @@ struct rte_mbuf * dpdk_zerocopy_replicate_packet_mb (vlib_buffer_t * b)
}
else
{
- ASSERT(prev_mb_next != 0);
+ ASSERT (prev_mb_next != 0);
*prev_mb_next = new_mb;
}
-
+
/*
* Copy packet segment data into new mbuf segment.
*/
@@ -207,8 +216,8 @@ struct rte_mbuf * dpdk_zerocopy_replicate_packet_mb (vlib_buffer_t * b)
pkt_mb = pkt_mb->next;
}
- ASSERT(pkt_mb == 0);
- __rte_mbuf_sanity_check(first_mb, 1);
+ ASSERT (pkt_mb == 0);
+ __rte_mbuf_sanity_check (first_mb, 1);
return first_mb;
@@ -219,48 +228,47 @@ static void
dpdk_tx_trace_buffer (dpdk_main_t * dm,
vlib_node_runtime_t * node,
dpdk_device_t * xd,
- u16 queue_id,
- u32 buffer_index,
- vlib_buffer_t * buffer)
+ u16 queue_id, u32 buffer_index, vlib_buffer_t * buffer)
{
- vlib_main_t * vm = vlib_get_main();
- dpdk_tx_dma_trace_t * t0;
- struct rte_mbuf * mb;
+ vlib_main_t *vm = vlib_get_main ();
+ dpdk_tx_dma_trace_t *t0;
+ struct rte_mbuf *mb;
- mb = rte_mbuf_from_vlib_buffer(buffer);
+ mb = rte_mbuf_from_vlib_buffer (buffer);
t0 = vlib_add_trace (vm, node, buffer, sizeof (t0[0]));
t0->queue_index = queue_id;
t0->device_index = xd->device_index;
t0->buffer_index = buffer_index;
clib_memcpy (&t0->mb, mb, sizeof (t0->mb));
- clib_memcpy (&t0->buffer, buffer, sizeof (buffer[0]) - sizeof (buffer->pre_data));
+ clib_memcpy (&t0->buffer, buffer,
+ sizeof (buffer[0]) - sizeof (buffer->pre_data));
clib_memcpy (t0->buffer.pre_data, buffer->data + buffer->current_data,
- sizeof (t0->buffer.pre_data));
+ sizeof (t0->buffer.pre_data));
}
/*
* This function calls the dpdk's tx_burst function to transmit the packets
* on the tx_vector. It manages a lock per-device if the device does not
- * support multiple queues. It returns the number of packets untransmitted
- * on the tx_vector. If all packets are transmitted (the normal case), the
+ * support multiple queues. It returns the number of packets untransmitted
+ * on the tx_vector. If all packets are transmitted (the normal case), the
* function returns 0.
- *
- * The tx_burst function may not be able to transmit all packets because the
+ *
+ * The tx_burst function may not be able to transmit all packets because the
* dpdk ring is full. If a flowcontrol callback function has been configured
- * then the function simply returns. If no callback has been configured, the
- * function will retry calling tx_burst with the remaining packets. This will
+ * then the function simply returns. If no callback has been configured, the
+ * function will retry calling tx_burst with the remaining packets. This will
* continue until all packets are transmitted or tx_burst indicates no packets
* could be transmitted. (The caller can drop the remaining packets.)
*
* The function assumes there is at least one packet on the tx_vector.
*/
static_always_inline
-u32 tx_burst_vector_internal (vlib_main_t * vm,
- dpdk_device_t * xd,
- struct rte_mbuf ** tx_vector)
+ u32 tx_burst_vector_internal (vlib_main_t * vm,
+ dpdk_device_t * xd,
+ struct rte_mbuf **tx_vector)
{
- dpdk_main_t * dm = &dpdk_main;
+ dpdk_main_t *dm = &dpdk_main;
u32 n_packets;
u32 tx_head;
u32 tx_tail;
@@ -269,7 +277,7 @@ u32 tx_burst_vector_internal (vlib_main_t * vm,
int queue_id;
tx_ring_hdr_t *ring;
- ring = vec_header(tx_vector, sizeof(*ring));
+ ring = vec_header (tx_vector, sizeof (*ring));
n_packets = ring->tx_head - ring->tx_tail;
@@ -279,7 +287,7 @@ u32 tx_burst_vector_internal (vlib_main_t * vm,
* Ensure rte_eth_tx_burst is not called with 0 packets, which can lead to
* unpredictable results.
*/
- ASSERT(n_packets > 0);
+ ASSERT (n_packets > 0);
/*
* Check for tx_vector overflow. If this fails it is a system configuration
@@ -288,229 +296,255 @@ u32 tx_burst_vector_internal (vlib_main_t * vm,
* a bit because it decreases the probability of having to issue two tx_burst
* calls due to a ring wrap.
*/
- ASSERT(n_packets < DPDK_TX_RING_SIZE);
+ ASSERT (n_packets < DPDK_TX_RING_SIZE);
/*
* If there is no flowcontrol callback, there is only temporary buffering
* on the tx_vector and so the tail should always be 0.
*/
- ASSERT(dm->flowcontrol_callback || ring->tx_tail == 0);
+ ASSERT (dm->flowcontrol_callback || ring->tx_tail == 0);
/*
- * If there is a flowcontrol callback, don't retry any incomplete tx_bursts.
+ * If there is a flowcontrol callback, don't retry any incomplete tx_bursts.
* Apply backpressure instead. If there is no callback, keep retrying until
- * a tx_burst sends no packets. n_retry of 255 essentially means no retry
+ * a tx_burst sends no packets. n_retry of 255 essentially means no retry
* limit.
*/
n_retry = dm->flowcontrol_callback ? 0 : 255;
queue_id = vm->cpu_index;
- do {
+ do
+ {
/* start the burst at the tail */
tx_tail = ring->tx_tail % DPDK_TX_RING_SIZE;
- /*
+ /*
* This device only supports one TX queue,
* and we're running multi-threaded...
*/
- if (PREDICT_FALSE(xd->dev_type != VNET_DPDK_DEV_VHOST_USER &&
- xd->lockp != 0))
- {
- queue_id = queue_id % xd->tx_q_used;
- while (__sync_lock_test_and_set (xd->lockp[queue_id], 1))
- /* zzzz */
- queue_id = (queue_id + 1) % xd->tx_q_used;
- }
-
- if (PREDICT_TRUE(xd->dev_type == VNET_DPDK_DEV_ETH))
- {
- if (PREDICT_TRUE(tx_head > tx_tail))
- {
- /* no wrap, transmit in one burst */
- rv = rte_eth_tx_burst(xd->device_index,
- (uint16_t) queue_id,
- &tx_vector[tx_tail],
- (uint16_t) (tx_head-tx_tail));
- }
- else
- {
- /*
- * This can only happen if there is a flowcontrol callback.
- * We need to split the transmit into two calls: one for
- * the packets up to the wrap point, and one to continue
- * at the start of the ring.
- * Transmit pkts up to the wrap point.
- */
- rv = rte_eth_tx_burst(xd->device_index,
- (uint16_t) queue_id,
- &tx_vector[tx_tail],
- (uint16_t) (DPDK_TX_RING_SIZE - tx_tail));
-
- /*
- * If we transmitted everything we wanted, then allow 1 retry
- * so we can try to transmit the rest. If we didn't transmit
- * everything, stop now.
- */
- n_retry = (rv == DPDK_TX_RING_SIZE - tx_tail) ? 1 : 0;
- }
- }
+ if (PREDICT_FALSE (xd->dev_type != VNET_DPDK_DEV_VHOST_USER &&
+ xd->lockp != 0))
+ {
+ queue_id = queue_id % xd->tx_q_used;
+ while (__sync_lock_test_and_set (xd->lockp[queue_id], 1))
+ /* zzzz */
+ queue_id = (queue_id + 1) % xd->tx_q_used;
+ }
+
+ if (PREDICT_TRUE (xd->dev_type == VNET_DPDK_DEV_ETH))
+ {
+ if (PREDICT_TRUE (tx_head > tx_tail))
+ {
+ /* no wrap, transmit in one burst */
+ rv = rte_eth_tx_burst (xd->device_index,
+ (uint16_t) queue_id,
+ &tx_vector[tx_tail],
+ (uint16_t) (tx_head - tx_tail));
+ }
+ else
+ {
+ /*
+ * This can only happen if there is a flowcontrol callback.
+ * We need to split the transmit into two calls: one for
+ * the packets up to the wrap point, and one to continue
+ * at the start of the ring.
+ * Transmit pkts up to the wrap point.
+ */
+ rv = rte_eth_tx_burst (xd->device_index,
+ (uint16_t) queue_id,
+ &tx_vector[tx_tail],
+ (uint16_t) (DPDK_TX_RING_SIZE -
+ tx_tail));
+
+ /*
+ * If we transmitted everything we wanted, then allow 1 retry
+ * so we can try to transmit the rest. If we didn't transmit
+ * everything, stop now.
+ */
+ n_retry = (rv == DPDK_TX_RING_SIZE - tx_tail) ? 1 : 0;
+ }
+ }
#if DPDK_VHOST_USER
else if (xd->dev_type == VNET_DPDK_DEV_VHOST_USER)
- {
- u32 offset = 0;
- if (xd->need_txlock) {
- queue_id = 0;
- while (__sync_lock_test_and_set (xd->lockp[queue_id], 1));
- }
- else {
- dpdk_device_and_queue_t * dq;
- vec_foreach (dq, dm->devices_by_cpu[vm->cpu_index])
- {
- if (xd->device_index == dq->device)
- break;
- }
- assert (dq);
- offset = dq->queue_id * VIRTIO_QNUM;
- }
- if (PREDICT_TRUE(tx_head > tx_tail))
- {
- int i; u32 bytes = 0;
- struct rte_mbuf **pkts = &tx_vector[tx_tail];
- for (i = 0; i < (tx_head - tx_tail); i++) {
- struct rte_mbuf *buff = pkts[i];
- bytes += rte_pktmbuf_data_len(buff);
- }
-
- /* no wrap, transmit in one burst */
- rv = rte_vhost_enqueue_burst(&xd->vu_vhost_dev, offset + VIRTIO_RXQ,
- &tx_vector[tx_tail],
- (uint16_t) (tx_head-tx_tail));
- if (PREDICT_TRUE(rv > 0))
- {
- dpdk_vu_vring *vring = &(xd->vu_intf->vrings[offset + VIRTIO_TXQ]);
- vring->packets += rv;
- vring->bytes += bytes;
-
- if (dpdk_vhost_user_want_interrupt(xd, offset + VIRTIO_RXQ)) {
- vring = &(xd->vu_intf->vrings[offset + VIRTIO_RXQ]);
- vring->n_since_last_int += rv;
-
- f64 now = vlib_time_now (vm);
- if (vring->int_deadline < now ||
- vring->n_since_last_int > dm->conf->vhost_coalesce_frames)
- dpdk_vhost_user_send_interrupt(vm, xd, offset + VIRTIO_RXQ);
- }
-
- int c = rv;
- while(c--)
- rte_pktmbuf_free (tx_vector[tx_tail+c]);
- }
- }
- else
- {
- /*
- * If we transmitted everything we wanted, then allow 1 retry
- * so we can try to transmit the rest. If we didn't transmit
- * everything, stop now.
- */
- int i; u32 bytes = 0;
- struct rte_mbuf **pkts = &tx_vector[tx_tail];
- for (i = 0; i < (DPDK_TX_RING_SIZE - tx_tail); i++) {
- struct rte_mbuf *buff = pkts[i];
- bytes += rte_pktmbuf_data_len(buff);
- }
- rv = rte_vhost_enqueue_burst(&xd->vu_vhost_dev, offset + VIRTIO_RXQ,
- &tx_vector[tx_tail],
- (uint16_t) (DPDK_TX_RING_SIZE - tx_tail));
-
- if (PREDICT_TRUE(rv > 0))
- {
- dpdk_vu_vring *vring = &(xd->vu_intf->vrings[offset + VIRTIO_TXQ]);
- vring->packets += rv;
- vring->bytes += bytes;
-
- if (dpdk_vhost_user_want_interrupt(xd, offset + VIRTIO_RXQ)) {
- vring = &(xd->vu_intf->vrings[offset + VIRTIO_RXQ]);
- vring->n_since_last_int += rv;
-
- f64 now = vlib_time_now (vm);
- if (vring->int_deadline < now ||
- vring->n_since_last_int > dm->conf->vhost_coalesce_frames)
- dpdk_vhost_user_send_interrupt(vm, xd, offset + VIRTIO_RXQ);
- }
-
- int c = rv;
- while(c--)
- rte_pktmbuf_free (tx_vector[tx_tail+c]);
- }
-
- n_retry = (rv == DPDK_TX_RING_SIZE - tx_tail) ? 1 : 0;
- }
-
- if (xd->need_txlock)
- *xd->lockp[queue_id] = 0;
- }
+ {
+ u32 offset = 0;
+ if (xd->need_txlock)
+ {
+ queue_id = 0;
+ while (__sync_lock_test_and_set (xd->lockp[queue_id], 1));
+ }
+ else
+ {
+ dpdk_device_and_queue_t *dq;
+ vec_foreach (dq, dm->devices_by_cpu[vm->cpu_index])
+ {
+ if (xd->device_index == dq->device)
+ break;
+ }
+ assert (dq);
+ offset = dq->queue_id * VIRTIO_QNUM;
+ }
+ if (PREDICT_TRUE (tx_head > tx_tail))
+ {
+ int i;
+ u32 bytes = 0;
+ struct rte_mbuf **pkts = &tx_vector[tx_tail];
+ for (i = 0; i < (tx_head - tx_tail); i++)
+ {
+ struct rte_mbuf *buff = pkts[i];
+ bytes += rte_pktmbuf_data_len (buff);
+ }
+
+ /* no wrap, transmit in one burst */
+ rv =
+ rte_vhost_enqueue_burst (&xd->vu_vhost_dev,
+ offset + VIRTIO_RXQ,
+ &tx_vector[tx_tail],
+ (uint16_t) (tx_head - tx_tail));
+ if (PREDICT_TRUE (rv > 0))
+ {
+ dpdk_vu_vring *vring =
+ &(xd->vu_intf->vrings[offset + VIRTIO_TXQ]);
+ vring->packets += rv;
+ vring->bytes += bytes;
+
+ if (dpdk_vhost_user_want_interrupt
+ (xd, offset + VIRTIO_RXQ))
+ {
+ vring = &(xd->vu_intf->vrings[offset + VIRTIO_RXQ]);
+ vring->n_since_last_int += rv;
+
+ f64 now = vlib_time_now (vm);
+ if (vring->int_deadline < now ||
+ vring->n_since_last_int >
+ dm->conf->vhost_coalesce_frames)
+ dpdk_vhost_user_send_interrupt (vm, xd,
+ offset + VIRTIO_RXQ);
+ }
+
+ int c = rv;
+ while (c--)
+ rte_pktmbuf_free (tx_vector[tx_tail + c]);
+ }
+ }
+ else
+ {
+ /*
+ * If we transmitted everything we wanted, then allow 1 retry
+ * so we can try to transmit the rest. If we didn't transmit
+ * everything, stop now.
+ */
+ int i;
+ u32 bytes = 0;
+ struct rte_mbuf **pkts = &tx_vector[tx_tail];
+ for (i = 0; i < (DPDK_TX_RING_SIZE - tx_tail); i++)
+ {
+ struct rte_mbuf *buff = pkts[i];
+ bytes += rte_pktmbuf_data_len (buff);
+ }
+ rv =
+ rte_vhost_enqueue_burst (&xd->vu_vhost_dev,
+ offset + VIRTIO_RXQ,
+ &tx_vector[tx_tail],
+ (uint16_t) (DPDK_TX_RING_SIZE -
+ tx_tail));
+
+ if (PREDICT_TRUE (rv > 0))
+ {
+ dpdk_vu_vring *vring =
+ &(xd->vu_intf->vrings[offset + VIRTIO_TXQ]);
+ vring->packets += rv;
+ vring->bytes += bytes;
+
+ if (dpdk_vhost_user_want_interrupt
+ (xd, offset + VIRTIO_RXQ))
+ {
+ vring = &(xd->vu_intf->vrings[offset + VIRTIO_RXQ]);
+ vring->n_since_last_int += rv;
+
+ f64 now = vlib_time_now (vm);
+ if (vring->int_deadline < now ||
+ vring->n_since_last_int >
+ dm->conf->vhost_coalesce_frames)
+ dpdk_vhost_user_send_interrupt (vm, xd,
+ offset + VIRTIO_RXQ);
+ }
+
+ int c = rv;
+ while (c--)
+ rte_pktmbuf_free (tx_vector[tx_tail + c]);
+ }
+
+ n_retry = (rv == DPDK_TX_RING_SIZE - tx_tail) ? 1 : 0;
+ }
+
+ if (xd->need_txlock)
+ *xd->lockp[queue_id] = 0;
+ }
#endif
#if RTE_LIBRTE_KNI
else if (xd->dev_type == VNET_DPDK_DEV_KNI)
- {
- if (PREDICT_TRUE(tx_head > tx_tail))
- {
- /* no wrap, transmit in one burst */
- rv = rte_kni_tx_burst(xd->kni,
- &tx_vector[tx_tail],
- (uint16_t) (tx_head-tx_tail));
- }
- else
- {
- /*
- * This can only happen if there is a flowcontrol callback.
- * We need to split the transmit into two calls: one for
- * the packets up to the wrap point, and one to continue
- * at the start of the ring.
- * Transmit pkts up to the wrap point.
- */
- rv = rte_kni_tx_burst(xd->kni,
- &tx_vector[tx_tail],
- (uint16_t) (DPDK_TX_RING_SIZE - tx_tail));
-
- /*
- * If we transmitted everything we wanted, then allow 1 retry
- * so we can try to transmit the rest. If we didn't transmit
- * everything, stop now.
- */
- n_retry = (rv == DPDK_TX_RING_SIZE - tx_tail) ? 1 : 0;
- }
- }
+ {
+ if (PREDICT_TRUE (tx_head > tx_tail))
+ {
+ /* no wrap, transmit in one burst */
+ rv = rte_kni_tx_burst (xd->kni,
+ &tx_vector[tx_tail],
+ (uint16_t) (tx_head - tx_tail));
+ }
+ else
+ {
+ /*
+ * This can only happen if there is a flowcontrol callback.
+ * We need to split the transmit into two calls: one for
+ * the packets up to the wrap point, and one to continue
+ * at the start of the ring.
+ * Transmit pkts up to the wrap point.
+ */
+ rv = rte_kni_tx_burst (xd->kni,
+ &tx_vector[tx_tail],
+ (uint16_t) (DPDK_TX_RING_SIZE -
+ tx_tail));
+
+ /*
+ * If we transmitted everything we wanted, then allow 1 retry
+ * so we can try to transmit the rest. If we didn't transmit
+ * everything, stop now.
+ */
+ n_retry = (rv == DPDK_TX_RING_SIZE - tx_tail) ? 1 : 0;
+ }
+ }
#endif
else
- {
- ASSERT(0);
- rv = 0;
- }
-
- if (PREDICT_FALSE(xd->dev_type != VNET_DPDK_DEV_VHOST_USER &&
- xd->lockp != 0))
- *xd->lockp[queue_id] = 0;
-
- if (PREDICT_FALSE(rv < 0))
- {
- // emit non-fatal message, bump counter
- vnet_main_t * vnm = dm->vnet_main;
- vnet_interface_main_t * im = &vnm->interface_main;
- u32 node_index;
-
- node_index = vec_elt_at_index(im->hw_interfaces,
- xd->vlib_hw_if_index)->tx_node_index;
-
- vlib_error_count (vm, node_index, DPDK_TX_FUNC_ERROR_BAD_RETVAL, 1);
- clib_warning ("rte_eth_tx_burst[%d]: error %d", xd->device_index, rv);
- return n_packets; // untransmitted packets
- }
- ring->tx_tail += (u16)rv;
+ {
+ ASSERT (0);
+ rv = 0;
+ }
+
+ if (PREDICT_FALSE (xd->dev_type != VNET_DPDK_DEV_VHOST_USER &&
+ xd->lockp != 0))
+ *xd->lockp[queue_id] = 0;
+
+ if (PREDICT_FALSE (rv < 0))
+ {
+ // emit non-fatal message, bump counter
+ vnet_main_t *vnm = dm->vnet_main;
+ vnet_interface_main_t *im = &vnm->interface_main;
+ u32 node_index;
+
+ node_index = vec_elt_at_index (im->hw_interfaces,
+ xd->vlib_hw_if_index)->tx_node_index;
+
+ vlib_error_count (vm, node_index, DPDK_TX_FUNC_ERROR_BAD_RETVAL, 1);
+ clib_warning ("rte_eth_tx_burst[%d]: error %d", xd->device_index,
+ rv);
+ return n_packets; // untransmitted packets
+ }
+ ring->tx_tail += (u16) rv;
n_packets -= (uint16_t) rv;
- } while (rv && n_packets && (n_retry>0));
+ }
+ while (rv && n_packets && (n_retry > 0));
return n_packets;
}
@@ -518,20 +552,21 @@ u32 tx_burst_vector_internal (vlib_main_t * vm,
/*
* This function transmits any packets on the interface's tx_vector and returns
- * the number of packets untransmitted on the tx_vector. If the tx_vector is
- * empty the function simply returns 0.
+ * the number of packets untransmitted on the tx_vector. If the tx_vector is
+ * empty the function simply returns 0.
*
* It is intended to be called by a traffic manager which has flowed-off an
* interface to see if the interface can be flowed-on again.
*/
-u32 dpdk_interface_tx_vector (vlib_main_t * vm, u32 dev_instance)
+u32
+dpdk_interface_tx_vector (vlib_main_t * vm, u32 dev_instance)
{
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_t * xd;
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_t *xd;
int queue_id;
- struct rte_mbuf ** tx_vector;
+ struct rte_mbuf **tx_vector;
tx_ring_hdr_t *ring;
-
+
/* param is dev_instance and not hw_if_index to save another lookup */
xd = vec_elt_at_index (dm->devices, dev_instance);
@@ -539,8 +574,8 @@ u32 dpdk_interface_tx_vector (vlib_main_t * vm, u32 dev_instance)
tx_vector = xd->tx_vectors[queue_id];
/* If no packets on the ring, don't bother calling tx function */
- ring = vec_header(tx_vector, sizeof(*ring));
- if (ring->tx_head == ring->tx_tail)
+ ring = vec_header (tx_vector, sizeof (*ring));
+ if (ring->tx_head == ring->tx_tail)
{
return 0;
}
@@ -550,14 +585,14 @@ u32 dpdk_interface_tx_vector (vlib_main_t * vm, u32 dev_instance)
/*
* Transmits the packets on the frame to the interface associated with the
- * node. It first copies packets on the frame to a tx_vector containing the
- * rte_mbuf pointers. It then passes this vector to tx_burst_vector_internal
+ * node. It first copies packets on the frame to a tx_vector containing the
+ * rte_mbuf pointers. It then passes this vector to tx_burst_vector_internal
* which calls the dpdk tx_burst function.
*
* The tx_vector is treated slightly differently depending on whether or
* not a flowcontrol callback function has been configured. If there is no
* callback, the tx_vector is a temporary array of rte_mbuf packet pointers.
- * Its entries are written and consumed before the function exits.
+ * Its entries are written and consumed before the function exits.
*
* If there is a callback then the transmit is being invoked in the presence
* of a traffic manager. Here the tx_vector is treated like a ring of rte_mbuf
@@ -569,16 +604,15 @@ u32 dpdk_interface_tx_vector (vlib_main_t * vm, u32 dev_instance)
*/
static uword
dpdk_interface_tx (vlib_main_t * vm,
- vlib_node_runtime_t * node,
- vlib_frame_t * f)
+ vlib_node_runtime_t * node, vlib_frame_t * f)
{
- dpdk_main_t * dm = &dpdk_main;
- vnet_interface_output_runtime_t * rd = (void *) node->runtime_data;
- dpdk_device_t * xd = vec_elt_at_index (dm->devices, rd->dev_instance);
+ dpdk_main_t *dm = &dpdk_main;
+ vnet_interface_output_runtime_t *rd = (void *) node->runtime_data;
+ dpdk_device_t *xd = vec_elt_at_index (dm->devices, rd->dev_instance);
u32 n_packets = f->n_vectors;
u32 n_left;
- u32 * from;
- struct rte_mbuf ** tx_vector;
+ u32 *from;
+ struct rte_mbuf **tx_vector;
int i;
int queue_id;
u32 my_cpu;
@@ -591,42 +625,42 @@ dpdk_interface_tx (vlib_main_t * vm,
queue_id = my_cpu;
tx_vector = xd->tx_vectors[queue_id];
- ring = vec_header(tx_vector, sizeof(*ring));
+ ring = vec_header (tx_vector, sizeof (*ring));
n_on_ring = ring->tx_head - ring->tx_tail;
from = vlib_frame_vector_args (f);
- ASSERT(n_packets <= VLIB_FRAME_SIZE);
+ ASSERT (n_packets <= VLIB_FRAME_SIZE);
- if (PREDICT_FALSE(n_on_ring + n_packets > DPDK_TX_RING_SIZE))
+ if (PREDICT_FALSE (n_on_ring + n_packets > DPDK_TX_RING_SIZE))
{
/*
- * Overflowing the ring should never happen.
+ * Overflowing the ring should never happen.
* If it does then drop the whole frame.
*/
vlib_error_count (vm, node->node_index, DPDK_TX_FUNC_ERROR_RING_FULL,
- n_packets);
-
- while (n_packets--)
- {
- u32 bi0 = from[n_packets];
- vlib_buffer_t *b0 = vlib_get_buffer (vm, bi0);
- struct rte_mbuf *mb0 = rte_mbuf_from_vlib_buffer(b0);
- rte_pktmbuf_free (mb0);
- }
+ n_packets);
+
+ while (n_packets--)
+ {
+ u32 bi0 = from[n_packets];
+ vlib_buffer_t *b0 = vlib_get_buffer (vm, bi0);
+ struct rte_mbuf *mb0 = rte_mbuf_from_vlib_buffer (b0);
+ rte_pktmbuf_free (mb0);
+ }
return n_on_ring;
}
- if (PREDICT_FALSE(dm->tx_pcap_enable))
+ if (PREDICT_FALSE (dm->tx_pcap_enable))
{
n_left = n_packets;
while (n_left > 0)
- {
+ {
u32 bi0 = from[0];
- vlib_buffer_t * b0 = vlib_get_buffer (vm, bi0);
+ vlib_buffer_t *b0 = vlib_get_buffer (vm, bi0);
if (dm->pcap_sw_if_index == 0 ||
- dm->pcap_sw_if_index == vnet_buffer(b0)->sw_if_index [VLIB_TX])
- pcap_add_buffer (&dm->pcap_main, vm, bi0, 512);
+ dm->pcap_sw_if_index == vnet_buffer (b0)->sw_if_index[VLIB_TX])
+ pcap_add_buffer (&dm->pcap_main, vm, bi0, 512);
from++;
n_left--;
}
@@ -640,10 +674,10 @@ dpdk_interface_tx (vlib_main_t * vm,
{
u32 bi0, bi1;
u32 pi0, pi1;
- struct rte_mbuf * mb0, * mb1;
- struct rte_mbuf * prefmb0, * prefmb1;
- vlib_buffer_t * b0, * b1;
- vlib_buffer_t * pref0, * pref1;
+ struct rte_mbuf *mb0, *mb1;
+ struct rte_mbuf *prefmb0, *prefmb1;
+ vlib_buffer_t *b0, *b1;
+ vlib_buffer_t *pref0, *pref1;
i16 delta0, delta1;
u16 new_data_len0, new_data_len1;
u16 new_pkt_len0, new_pkt_len1;
@@ -654,33 +688,32 @@ dpdk_interface_tx (vlib_main_t * vm,
pref0 = vlib_get_buffer (vm, pi0);
pref1 = vlib_get_buffer (vm, pi1);
- prefmb0 = rte_mbuf_from_vlib_buffer(pref0);
- prefmb1 = rte_mbuf_from_vlib_buffer(pref1);
+ prefmb0 = rte_mbuf_from_vlib_buffer (pref0);
+ prefmb1 = rte_mbuf_from_vlib_buffer (pref1);
- CLIB_PREFETCH(prefmb0, CLIB_CACHE_LINE_BYTES, LOAD);
- CLIB_PREFETCH(pref0, CLIB_CACHE_LINE_BYTES, LOAD);
- CLIB_PREFETCH(prefmb1, CLIB_CACHE_LINE_BYTES, LOAD);
- CLIB_PREFETCH(pref1, CLIB_CACHE_LINE_BYTES, LOAD);
+ CLIB_PREFETCH (prefmb0, CLIB_CACHE_LINE_BYTES, LOAD);
+ CLIB_PREFETCH (pref0, CLIB_CACHE_LINE_BYTES, LOAD);
+ CLIB_PREFETCH (prefmb1, CLIB_CACHE_LINE_BYTES, LOAD);
+ CLIB_PREFETCH (pref1, CLIB_CACHE_LINE_BYTES, LOAD);
bi0 = from[0];
bi1 = from[1];
from += 2;
-
+
b0 = vlib_get_buffer (vm, bi0);
b1 = vlib_get_buffer (vm, bi1);
- mb0 = rte_mbuf_from_vlib_buffer(b0);
- mb1 = rte_mbuf_from_vlib_buffer(b1);
+ mb0 = rte_mbuf_from_vlib_buffer (b0);
+ mb1 = rte_mbuf_from_vlib_buffer (b1);
any_clone = (b0->flags & VLIB_BUFFER_RECYCLE)
- | (b1->flags & VLIB_BUFFER_RECYCLE);
- if (PREDICT_FALSE(any_clone != 0))
- {
- if (PREDICT_FALSE
- ((b0->flags & VLIB_BUFFER_RECYCLE) != 0))
+ | (b1->flags & VLIB_BUFFER_RECYCLE);
+ if (PREDICT_FALSE (any_clone != 0))
+ {
+ if (PREDICT_FALSE ((b0->flags & VLIB_BUFFER_RECYCLE) != 0))
{
- struct rte_mbuf * mb0_new = dpdk_replicate_packet_mb (b0);
- if (PREDICT_FALSE(mb0_new == 0))
+ struct rte_mbuf *mb0_new = dpdk_replicate_packet_mb (b0);
+ if (PREDICT_FALSE (mb0_new == 0))
{
vlib_error_count (vm, node->node_index,
DPDK_TX_FUNC_ERROR_REPL_FAIL, 1);
@@ -690,11 +723,10 @@ dpdk_interface_tx (vlib_main_t * vm,
mb0 = mb0_new;
vec_add1 (dm->recycle[my_cpu], bi0);
}
- if (PREDICT_FALSE
- ((b1->flags & VLIB_BUFFER_RECYCLE) != 0))
+ if (PREDICT_FALSE ((b1->flags & VLIB_BUFFER_RECYCLE) != 0))
{
- struct rte_mbuf * mb1_new = dpdk_replicate_packet_mb (b1);
- if (PREDICT_FALSE(mb1_new == 0))
+ struct rte_mbuf *mb1_new = dpdk_replicate_packet_mb (b1);
+ if (PREDICT_FALSE (mb1_new == 0))
{
vlib_error_count (vm, node->node_index,
DPDK_TX_FUNC_ERROR_REPL_FAIL, 1);
@@ -706,15 +738,15 @@ dpdk_interface_tx (vlib_main_t * vm,
}
}
- delta0 = PREDICT_FALSE(b0->flags & VLIB_BUFFER_REPL_FAIL) ? 0 :
+ delta0 = PREDICT_FALSE (b0->flags & VLIB_BUFFER_REPL_FAIL) ? 0 :
vlib_buffer_length_in_chain (vm, b0) - (i16) mb0->pkt_len;
- delta1 = PREDICT_FALSE(b1->flags & VLIB_BUFFER_REPL_FAIL) ? 0 :
+ delta1 = PREDICT_FALSE (b1->flags & VLIB_BUFFER_REPL_FAIL) ? 0 :
vlib_buffer_length_in_chain (vm, b1) - (i16) mb1->pkt_len;
-
- new_data_len0 = (u16)((i16) mb0->data_len + delta0);
- new_data_len1 = (u16)((i16) mb1->data_len + delta1);
- new_pkt_len0 = (u16)((i16) mb0->pkt_len + delta0);
- new_pkt_len1 = (u16)((i16) mb1->pkt_len + delta1);
+
+ new_data_len0 = (u16) ((i16) mb0->data_len + delta0);
+ new_data_len1 = (u16) ((i16) mb1->data_len + delta1);
+ new_pkt_len0 = (u16) ((i16) mb0->pkt_len + delta0);
+ new_pkt_len1 = (u16) ((i16) mb1->pkt_len + delta1);
b0->current_length = new_data_len0;
b1->current_length = new_data_len1;
@@ -723,62 +755,62 @@ dpdk_interface_tx (vlib_main_t * vm,
mb0->pkt_len = new_pkt_len0;
mb1->pkt_len = new_pkt_len1;
- mb0->data_off = (PREDICT_FALSE(b0->flags & VLIB_BUFFER_REPL_FAIL)) ?
- mb0->data_off : (u16)(RTE_PKTMBUF_HEADROOM + b0->current_data);
- mb1->data_off = (PREDICT_FALSE(b1->flags & VLIB_BUFFER_REPL_FAIL)) ?
- mb1->data_off : (u16)(RTE_PKTMBUF_HEADROOM + b1->current_data);
+ mb0->data_off = (PREDICT_FALSE (b0->flags & VLIB_BUFFER_REPL_FAIL)) ?
+ mb0->data_off : (u16) (RTE_PKTMBUF_HEADROOM + b0->current_data);
+ mb1->data_off = (PREDICT_FALSE (b1->flags & VLIB_BUFFER_REPL_FAIL)) ?
+ mb1->data_off : (u16) (RTE_PKTMBUF_HEADROOM + b1->current_data);
- if (PREDICT_FALSE(node->flags & VLIB_NODE_FLAG_TRACE))
+ if (PREDICT_FALSE (node->flags & VLIB_NODE_FLAG_TRACE))
{
- if (b0->flags & VLIB_BUFFER_IS_TRACED)
+ if (b0->flags & VLIB_BUFFER_IS_TRACED)
dpdk_tx_trace_buffer (dm, node, xd, queue_id, bi0, b0);
- if (b1->flags & VLIB_BUFFER_IS_TRACED)
+ if (b1->flags & VLIB_BUFFER_IS_TRACED)
dpdk_tx_trace_buffer (dm, node, xd, queue_id, bi1, b1);
}
- if (PREDICT_TRUE(any_clone == 0))
- {
+ if (PREDICT_TRUE (any_clone == 0))
+ {
tx_vector[i % DPDK_TX_RING_SIZE] = mb0;
- i++;
+ i++;
tx_vector[i % DPDK_TX_RING_SIZE] = mb1;
- i++;
- }
+ i++;
+ }
else
- {
- /* cloning was done, need to check for failure */
- if (PREDICT_TRUE((b0->flags & VLIB_BUFFER_REPL_FAIL) == 0))
- {
+ {
+ /* cloning was done, need to check for failure */
+ if (PREDICT_TRUE ((b0->flags & VLIB_BUFFER_REPL_FAIL) == 0))
+ {
tx_vector[i % DPDK_TX_RING_SIZE] = mb0;
- i++;
- }
- if (PREDICT_TRUE((b1->flags & VLIB_BUFFER_REPL_FAIL) == 0))
- {
+ i++;
+ }
+ if (PREDICT_TRUE ((b1->flags & VLIB_BUFFER_REPL_FAIL) == 0))
+ {
tx_vector[i % DPDK_TX_RING_SIZE] = mb1;
- i++;
- }
- }
-
+ i++;
+ }
+ }
+
n_left -= 2;
}
while (n_left > 0)
{
u32 bi0;
- struct rte_mbuf * mb0;
- vlib_buffer_t * b0;
+ struct rte_mbuf *mb0;
+ vlib_buffer_t *b0;
i16 delta0;
u16 new_data_len0;
u16 new_pkt_len0;
bi0 = from[0];
from++;
-
+
b0 = vlib_get_buffer (vm, bi0);
- mb0 = rte_mbuf_from_vlib_buffer(b0);
- if (PREDICT_FALSE((b0->flags & VLIB_BUFFER_RECYCLE) != 0))
+ mb0 = rte_mbuf_from_vlib_buffer (b0);
+ if (PREDICT_FALSE ((b0->flags & VLIB_BUFFER_RECYCLE) != 0))
{
- struct rte_mbuf * mb0_new = dpdk_replicate_packet_mb (b0);
- if (PREDICT_FALSE(mb0_new == 0))
+ struct rte_mbuf *mb0_new = dpdk_replicate_packet_mb (b0);
+ if (PREDICT_FALSE (mb0_new == 0))
{
vlib_error_count (vm, node->node_index,
DPDK_TX_FUNC_ERROR_REPL_FAIL, 1);
@@ -789,27 +821,27 @@ dpdk_interface_tx (vlib_main_t * vm,
vec_add1 (dm->recycle[my_cpu], bi0);
}
- delta0 = PREDICT_FALSE(b0->flags & VLIB_BUFFER_REPL_FAIL) ? 0 :
+ delta0 = PREDICT_FALSE (b0->flags & VLIB_BUFFER_REPL_FAIL) ? 0 :
vlib_buffer_length_in_chain (vm, b0) - (i16) mb0->pkt_len;
-
- new_data_len0 = (u16)((i16) mb0->data_len + delta0);
- new_pkt_len0 = (u16)((i16) mb0->pkt_len + delta0);
-
+
+ new_data_len0 = (u16) ((i16) mb0->data_len + delta0);
+ new_pkt_len0 = (u16) ((i16) mb0->pkt_len + delta0);
+
b0->current_length = new_data_len0;
mb0->data_len = new_data_len0;
mb0->pkt_len = new_pkt_len0;
- mb0->data_off = (PREDICT_FALSE(b0->flags & VLIB_BUFFER_REPL_FAIL)) ?
- mb0->data_off : (u16)(RTE_PKTMBUF_HEADROOM + b0->current_data);
+ mb0->data_off = (PREDICT_FALSE (b0->flags & VLIB_BUFFER_REPL_FAIL)) ?
+ mb0->data_off : (u16) (RTE_PKTMBUF_HEADROOM + b0->current_data);
- if (PREDICT_FALSE(node->flags & VLIB_NODE_FLAG_TRACE))
+ if (PREDICT_FALSE (node->flags & VLIB_NODE_FLAG_TRACE))
if (b0->flags & VLIB_BUFFER_IS_TRACED)
dpdk_tx_trace_buffer (dm, node, xd, queue_id, bi0, b0);
- if (PREDICT_TRUE((b0->flags & VLIB_BUFFER_REPL_FAIL) == 0))
- {
+ if (PREDICT_TRUE ((b0->flags & VLIB_BUFFER_REPL_FAIL) == 0))
+ {
tx_vector[i % DPDK_TX_RING_SIZE] = mb0;
- i++;
- }
+ i++;
+ }
n_left--;
}
@@ -824,86 +856,92 @@ dpdk_interface_tx (vlib_main_t * vm,
* tx_pkts is the number of packets successfully transmitted
* This is the number originally on ring minus the number remaining on ring
*/
- tx_pkts = n_on_ring - n_packets;
+ tx_pkts = n_on_ring - n_packets;
- if (PREDICT_FALSE(dm->flowcontrol_callback != 0))
+ if (PREDICT_FALSE (dm->flowcontrol_callback != 0))
{
- if (PREDICT_FALSE(n_packets))
- {
- /* Callback may want to enable flowcontrol */
- dm->flowcontrol_callback(vm, xd->vlib_hw_if_index, ring->tx_head - ring->tx_tail);
- }
- else
- {
- /* Reset head/tail to avoid unnecessary wrap */
- ring->tx_head = 0;
- ring->tx_tail = 0;
- }
+ if (PREDICT_FALSE (n_packets))
+ {
+ /* Callback may want to enable flowcontrol */
+ dm->flowcontrol_callback (vm, xd->vlib_hw_if_index,
+ ring->tx_head - ring->tx_tail);
+ }
+ else
+ {
+ /* Reset head/tail to avoid unnecessary wrap */
+ ring->tx_head = 0;
+ ring->tx_tail = 0;
+ }
}
- else
+ else
{
/* If there is no callback then drop any non-transmitted packets */
- if (PREDICT_FALSE(n_packets))
- {
- vlib_simple_counter_main_t * cm;
- vnet_main_t * vnm = vnet_get_main();
+ if (PREDICT_FALSE (n_packets))
+ {
+ vlib_simple_counter_main_t *cm;
+ vnet_main_t *vnm = vnet_get_main ();
- cm = vec_elt_at_index (vnm->interface_main.sw_if_counters,
- VNET_INTERFACE_COUNTER_TX_ERROR);
+ cm = vec_elt_at_index (vnm->interface_main.sw_if_counters,
+ VNET_INTERFACE_COUNTER_TX_ERROR);
- vlib_increment_simple_counter (cm, my_cpu, xd->vlib_sw_if_index, n_packets);
+ vlib_increment_simple_counter (cm, my_cpu, xd->vlib_sw_if_index,
+ n_packets);
- vlib_error_count (vm, node->node_index, DPDK_TX_FUNC_ERROR_PKT_DROP,
+ vlib_error_count (vm, node->node_index, DPDK_TX_FUNC_ERROR_PKT_DROP,
n_packets);
- while (n_packets--)
- rte_pktmbuf_free (tx_vector[ring->tx_tail + n_packets]);
- }
+ while (n_packets--)
+ rte_pktmbuf_free (tx_vector[ring->tx_tail + n_packets]);
+ }
- /* Reset head/tail to avoid unnecessary wrap */
+ /* Reset head/tail to avoid unnecessary wrap */
ring->tx_head = 0;
ring->tx_tail = 0;
}
/* Recycle replicated buffers */
- if (PREDICT_FALSE(vec_len(dm->recycle[my_cpu])))
+ if (PREDICT_FALSE (vec_len (dm->recycle[my_cpu])))
{
- vlib_buffer_free (vm, dm->recycle[my_cpu], vec_len(dm->recycle[my_cpu]));
- _vec_len(dm->recycle[my_cpu]) = 0;
+ vlib_buffer_free (vm, dm->recycle[my_cpu],
+ vec_len (dm->recycle[my_cpu]));
+ _vec_len (dm->recycle[my_cpu]) = 0;
}
- ASSERT(ring->tx_head >= ring->tx_tail);
+ ASSERT (ring->tx_head >= ring->tx_tail);
return tx_pkts;
}
-static int dpdk_device_renumber (vnet_hw_interface_t * hi,
- u32 new_dev_instance)
+static int
+dpdk_device_renumber (vnet_hw_interface_t * hi, u32 new_dev_instance)
{
#if DPDK_VHOST_USER
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_t * xd = vec_elt_at_index (dm->devices, hi->dev_instance);
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_t *xd = vec_elt_at_index (dm->devices, hi->dev_instance);
- if (!xd || xd->dev_type != VNET_DPDK_DEV_VHOST_USER) {
- clib_warning("cannot renumber non-vhost-user interface (sw_if_index: %d)",
- hi->sw_if_index);
- return 0;
- }
+ if (!xd || xd->dev_type != VNET_DPDK_DEV_VHOST_USER)
+ {
+ clib_warning
+ ("cannot renumber non-vhost-user interface (sw_if_index: %d)",
+ hi->sw_if_index);
+ return 0;
+ }
xd->vu_if_id = new_dev_instance;
#endif
return 0;
}
-static void dpdk_clear_hw_interface_counters (u32 instance)
+static void
+dpdk_clear_hw_interface_counters (u32 instance)
{
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_t * xd = vec_elt_at_index (dm->devices, instance);
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_t *xd = vec_elt_at_index (dm->devices, instance);
/*
- * DAW-FIXME: VMXNET3 device stop/start doesn't work,
+ * DAW-FIXME: VMXNET3 device stop/start doesn't work,
* therefore fake the stop in the dpdk driver by
- * silently dropping all of the incoming pkts instead of
+ * silently dropping all of the incoming pkts instead of
* stopping the driver / hardware.
*/
if (xd->admin_up != 0xff)
@@ -914,10 +952,10 @@ static void dpdk_clear_hw_interface_counters (u32 instance)
*/
dpdk_update_counters (xd, vlib_time_now (dm->vlib_main));
- clib_memcpy (&xd->last_cleared_stats, &xd->stats, sizeof(xd->stats));
+ clib_memcpy (&xd->last_cleared_stats, &xd->stats, sizeof (xd->stats));
clib_memcpy (xd->last_cleared_xstats, xd->xstats,
- vec_len(xd->last_cleared_xstats) *
- sizeof(xd->last_cleared_xstats[0]));
+ vec_len (xd->last_cleared_xstats) *
+ sizeof (xd->last_cleared_xstats[0]));
}
else
{
@@ -926,60 +964,68 @@ static void dpdk_clear_hw_interface_counters (u32 instance)
* so we're only calling xstats_reset() here.
*/
rte_eth_xstats_reset (xd->device_index);
- memset (&xd->stats, 0, sizeof(xd->stats));
+ memset (&xd->stats, 0, sizeof (xd->stats));
memset (&xd->last_stats, 0, sizeof (xd->last_stats));
}
#if DPDK_VHOST_USER
- if (PREDICT_FALSE(xd->dev_type == VNET_DPDK_DEV_VHOST_USER)) {
- int i;
- for (i = 0; i < xd->rx_q_used * VIRTIO_QNUM; i++) {
- xd->vu_intf->vrings[i].packets = 0;
- xd->vu_intf->vrings[i].bytes = 0;
+ if (PREDICT_FALSE (xd->dev_type == VNET_DPDK_DEV_VHOST_USER))
+ {
+ int i;
+ for (i = 0; i < xd->rx_q_used * VIRTIO_QNUM; i++)
+ {
+ xd->vu_intf->vrings[i].packets = 0;
+ xd->vu_intf->vrings[i].bytes = 0;
+ }
}
- }
#endif
}
#ifdef RTE_LIBRTE_KNI
static int
-kni_config_network_if(u8 port_id, u8 if_up)
+kni_config_network_if (u8 port_id, u8 if_up)
{
- vnet_main_t * vnm = vnet_get_main();
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_t * xd;
+ vnet_main_t *vnm = vnet_get_main ();
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_t *xd;
uword *p;
p = hash_get (dm->dpdk_device_by_kni_port_id, port_id);
- if (p == 0) {
- clib_warning("unknown interface");
- return 0;
- } else {
- xd = vec_elt_at_index (dm->devices, p[0]);
- }
+ if (p == 0)
+ {
+ clib_warning ("unknown interface");
+ return 0;
+ }
+ else
+ {
+ xd = vec_elt_at_index (dm->devices, p[0]);
+ }
vnet_hw_interface_set_flags (vnm, xd->vlib_hw_if_index,
- if_up ? VNET_HW_INTERFACE_FLAG_LINK_UP |
- ETH_LINK_FULL_DUPLEX : 0);
+ if_up ? VNET_HW_INTERFACE_FLAG_LINK_UP |
+ ETH_LINK_FULL_DUPLEX : 0);
return 0;
}
static int
-kni_change_mtu(u8 port_id, unsigned new_mtu)
+kni_change_mtu (u8 port_id, unsigned new_mtu)
{
- vnet_main_t * vnm = vnet_get_main();
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_t * xd;
+ vnet_main_t *vnm = vnet_get_main ();
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_t *xd;
uword *p;
- vnet_hw_interface_t * hif;
+ vnet_hw_interface_t *hif;
p = hash_get (dm->dpdk_device_by_kni_port_id, port_id);
- if (p == 0) {
- clib_warning("unknown interface");
- return 0;
- } else {
- xd = vec_elt_at_index (dm->devices, p[0]);
- }
+ if (p == 0)
+ {
+ clib_warning ("unknown interface");
+ return 0;
+ }
+ else
+ {
+ xd = vec_elt_at_index (dm->devices, p[0]);
+ }
hif = vnet_get_hw_interface (vnm, xd->vlib_hw_if_index);
hif->max_packet_bytes = new_mtu;
@@ -991,68 +1037,69 @@ kni_change_mtu(u8 port_id, unsigned new_mtu)
static clib_error_t *
dpdk_interface_admin_up_down (vnet_main_t * vnm, u32 hw_if_index, u32 flags)
{
- vnet_hw_interface_t * hif = vnet_get_hw_interface (vnm, hw_if_index);
+ vnet_hw_interface_t *hif = vnet_get_hw_interface (vnm, hw_if_index);
uword is_up = (flags & VNET_SW_INTERFACE_FLAG_ADMIN_UP) != 0;
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_t * xd = vec_elt_at_index (dm->devices, hif->dev_instance);
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_t *xd = vec_elt_at_index (dm->devices, hif->dev_instance);
int rv = 0;
#ifdef RTE_LIBRTE_KNI
if (xd->dev_type == VNET_DPDK_DEV_KNI)
- {
+ {
if (is_up)
- {
- struct rte_kni_conf conf;
- struct rte_kni_ops ops;
- vlib_main_t * vm = vlib_get_main();
- vlib_buffer_main_t * bm = vm->buffer_main;
- memset(&conf, 0, sizeof(conf));
- snprintf(conf.name, RTE_KNI_NAMESIZE, "vpp%u", xd->kni_port_id);
- conf.mbuf_size = VLIB_BUFFER_DATA_SIZE;
- memset(&ops, 0, sizeof(ops));
- ops.port_id = xd->kni_port_id;
- ops.change_mtu = kni_change_mtu;
- ops.config_network_if = kni_config_network_if;
-
- xd->kni = rte_kni_alloc(bm->pktmbuf_pools[rte_socket_id()], &conf, &ops);
- if (!xd->kni)
- {
- clib_warning("failed to allocate kni interface");
- }
- else
- {
- hif->max_packet_bytes = 1500; /* kni interface default value */
- xd->admin_up = 1;
- }
- }
+ {
+ struct rte_kni_conf conf;
+ struct rte_kni_ops ops;
+ vlib_main_t *vm = vlib_get_main ();
+ vlib_buffer_main_t *bm = vm->buffer_main;
+ memset (&conf, 0, sizeof (conf));
+ snprintf (conf.name, RTE_KNI_NAMESIZE, "vpp%u", xd->kni_port_id);
+ conf.mbuf_size = VLIB_BUFFER_DATA_SIZE;
+ memset (&ops, 0, sizeof (ops));
+ ops.port_id = xd->kni_port_id;
+ ops.change_mtu = kni_change_mtu;
+ ops.config_network_if = kni_config_network_if;
+
+ xd->kni =
+ rte_kni_alloc (bm->pktmbuf_pools[rte_socket_id ()], &conf, &ops);
+ if (!xd->kni)
+ {
+ clib_warning ("failed to allocate kni interface");
+ }
+ else
+ {
+ hif->max_packet_bytes = 1500; /* kni interface default value */
+ xd->admin_up = 1;
+ }
+ }
else
- {
- xd->admin_up = 0;
- int kni_rv;
-
- kni_rv = rte_kni_release(xd->kni);
- if (kni_rv < 0)
- clib_warning ("rte_kni_release returned %d", kni_rv);
- }
+ {
+ xd->admin_up = 0;
+ int kni_rv;
+
+ kni_rv = rte_kni_release (xd->kni);
+ if (kni_rv < 0)
+ clib_warning ("rte_kni_release returned %d", kni_rv);
+ }
return 0;
- }
+ }
#endif
#if DPDK_VHOST_USER
if (xd->dev_type == VNET_DPDK_DEV_VHOST_USER)
{
if (is_up)
- {
- if (xd->vu_is_running)
- vnet_hw_interface_set_flags (vnm, xd->vlib_hw_if_index,
- VNET_HW_INTERFACE_FLAG_LINK_UP |
- ETH_LINK_FULL_DUPLEX );
- xd->admin_up = 1;
- }
+ {
+ if (xd->vu_is_running)
+ vnet_hw_interface_set_flags (vnm, xd->vlib_hw_if_index,
+ VNET_HW_INTERFACE_FLAG_LINK_UP |
+ ETH_LINK_FULL_DUPLEX);
+ xd->admin_up = 1;
+ }
else
- {
- vnet_hw_interface_set_flags (vnm, xd->vlib_hw_if_index, 0);
- xd->admin_up = 0;
- }
+ {
+ vnet_hw_interface_set_flags (vnm, xd->vlib_hw_if_index, 0);
+ xd->admin_up = 0;
+ }
return 0;
}
@@ -1064,18 +1111,18 @@ dpdk_interface_admin_up_down (vnet_main_t * vnm, u32 hw_if_index, u32 flags)
f64 now = vlib_time_now (dm->vlib_main);
/*
- * DAW-FIXME: VMXNET3 device stop/start doesn't work,
+ * DAW-FIXME: VMXNET3 device stop/start doesn't work,
* therefore fake the stop in the dpdk driver by
- * silently dropping all of the incoming pkts instead of
+ * silently dropping all of the incoming pkts instead of
* stopping the driver / hardware.
*/
if (xd->admin_up == 0)
rv = rte_eth_dev_start (xd->device_index);
if (xd->promisc)
- rte_eth_promiscuous_enable(xd->device_index);
+ rte_eth_promiscuous_enable (xd->device_index);
else
- rte_eth_promiscuous_disable(xd->device_index);
+ rte_eth_promiscuous_disable (xd->device_index);
rte_eth_allmulticast_enable (xd->device_index);
xd->admin_up = 1;
@@ -1091,38 +1138,37 @@ dpdk_interface_admin_up_down (vnet_main_t * vnm, u32 hw_if_index, u32 flags)
* stopping the driver / hardware.
*/
if (xd->pmd != VNET_DPDK_PMD_VMXNET3)
- xd->admin_up = 0;
+ xd->admin_up = 0;
else
- xd->admin_up = ~0;
+ xd->admin_up = ~0;
rte_eth_allmulticast_disable (xd->device_index);
vnet_hw_interface_set_flags (vnm, xd->vlib_hw_if_index, 0);
/*
- * DAW-FIXME: VMXNET3 device stop/start doesn't work,
+ * DAW-FIXME: VMXNET3 device stop/start doesn't work,
* therefore fake the stop in the dpdk driver by
- * silently dropping all of the incoming pkts instead of
+ * silently dropping all of the incoming pkts instead of
* stopping the driver / hardware.
*/
if (xd->pmd != VNET_DPDK_PMD_VMXNET3)
- rte_eth_dev_stop (xd->device_index);
+ rte_eth_dev_stop (xd->device_index);
/* For bonded interface, stop slave links */
- if (xd->pmd == VNET_DPDK_PMD_BOND)
- {
- u8 slink[16];
- int nlink = rte_eth_bond_slaves_get(xd->device_index, slink, 16);
- while (nlink >=1)
+ if (xd->pmd == VNET_DPDK_PMD_BOND)
+ {
+ u8 slink[16];
+ int nlink = rte_eth_bond_slaves_get (xd->device_index, slink, 16);
+ while (nlink >= 1)
{
u8 dpdk_port = slink[--nlink];
rte_eth_dev_stop (dpdk_port);
}
- }
+ }
}
if (rv < 0)
- clib_warning ("rte_eth_dev_%s error: %d", is_up ? "start" : "stop",
- rv);
+ clib_warning ("rte_eth_dev_%s error: %d", is_up ? "start" : "stop", rv);
return /* no error */ 0;
}
@@ -1131,76 +1177,83 @@ dpdk_interface_admin_up_down (vnet_main_t * vnm, u32 hw_if_index, u32 flags)
* Dynamically redirect all pkts from a specific interface
* to the specified node
*/
-static void dpdk_set_interface_next_node (vnet_main_t *vnm, u32 hw_if_index,
- u32 node_index)
+static void
+dpdk_set_interface_next_node (vnet_main_t * vnm, u32 hw_if_index,
+ u32 node_index)
{
- dpdk_main_t * xm = &dpdk_main;
+ dpdk_main_t *xm = &dpdk_main;
vnet_hw_interface_t *hw = vnet_get_hw_interface (vnm, hw_if_index);
- dpdk_device_t * xd = vec_elt_at_index (xm->devices, hw->dev_instance);
-
+ dpdk_device_t *xd = vec_elt_at_index (xm->devices, hw->dev_instance);
+
/* Shut off redirection */
if (node_index == ~0)
{
xd->per_interface_next_index = node_index;
return;
}
-
- xd->per_interface_next_index =
+
+ xd->per_interface_next_index =
vlib_node_add_next (xm->vlib_main, dpdk_input_node.index, node_index);
}
static clib_error_t *
dpdk_subif_add_del_function (vnet_main_t * vnm,
- u32 hw_if_index,
- struct vnet_sw_interface_t * st,
- int is_add)
+ u32 hw_if_index,
+ struct vnet_sw_interface_t *st, int is_add)
{
- dpdk_main_t * xm = &dpdk_main;
+ dpdk_main_t *xm = &dpdk_main;
vnet_hw_interface_t *hw = vnet_get_hw_interface (vnm, hw_if_index);
- dpdk_device_t * xd = vec_elt_at_index (xm->devices, hw->dev_instance);
- vnet_sw_interface_t * t = (vnet_sw_interface_t *) st;
+ dpdk_device_t *xd = vec_elt_at_index (xm->devices, hw->dev_instance);
+ vnet_sw_interface_t *t = (vnet_sw_interface_t *) st;
int r, vlan_offload;
u32 prev_subifs = xd->vlan_subifs;
- if (is_add) xd->vlan_subifs++;
- else if (xd->vlan_subifs) xd->vlan_subifs--;
+ if (is_add)
+ xd->vlan_subifs++;
+ else if (xd->vlan_subifs)
+ xd->vlan_subifs--;
if (xd->dev_type != VNET_DPDK_DEV_ETH)
- return 0;
+ return 0;
/* currently we program VLANS only for IXGBE VF and I40E VF */
- if ((xd->pmd != VNET_DPDK_PMD_IXGBEVF) &&
- (xd->pmd != VNET_DPDK_PMD_I40EVF))
- return 0;
+ if ((xd->pmd != VNET_DPDK_PMD_IXGBEVF) && (xd->pmd != VNET_DPDK_PMD_I40EVF))
+ return 0;
if (t->sub.eth.flags.no_tags == 1)
- return 0;
+ return 0;
- if ((t->sub.eth.flags.one_tag != 1) || (t->sub.eth.flags.exact_match != 1 )) {
- xd->vlan_subifs = prev_subifs;
- return clib_error_return (0, "unsupported VLAN setup");
- }
+ if ((t->sub.eth.flags.one_tag != 1) || (t->sub.eth.flags.exact_match != 1))
+ {
+ xd->vlan_subifs = prev_subifs;
+ return clib_error_return (0, "unsupported VLAN setup");
+ }
- vlan_offload = rte_eth_dev_get_vlan_offload(xd->device_index);
+ vlan_offload = rte_eth_dev_get_vlan_offload (xd->device_index);
vlan_offload |= ETH_VLAN_FILTER_OFFLOAD;
- if ((r = rte_eth_dev_set_vlan_offload(xd->device_index, vlan_offload))) {
- xd->vlan_subifs = prev_subifs;
- return clib_error_return (0, "rte_eth_dev_set_vlan_offload[%d]: err %d",
- xd->device_index, r);
- }
+ if ((r = rte_eth_dev_set_vlan_offload (xd->device_index, vlan_offload)))
+ {
+ xd->vlan_subifs = prev_subifs;
+ return clib_error_return (0, "rte_eth_dev_set_vlan_offload[%d]: err %d",
+ xd->device_index, r);
+ }
- if ((r = rte_eth_dev_vlan_filter(xd->device_index, t->sub.eth.outer_vlan_id, is_add))) {
- xd->vlan_subifs = prev_subifs;
- return clib_error_return (0, "rte_eth_dev_vlan_filter[%d]: err %d",
- xd->device_index, r);
- }
+ if ((r =
+ rte_eth_dev_vlan_filter (xd->device_index, t->sub.eth.outer_vlan_id,
+ is_add)))
+ {
+ xd->vlan_subifs = prev_subifs;
+ return clib_error_return (0, "rte_eth_dev_vlan_filter[%d]: err %d",
+ xd->device_index, r);
+ }
return 0;
}
+/* *INDENT-OFF* */
VNET_DEVICE_CLASS (dpdk_device_class) = {
.name = "dpdk",
.tx_function = dpdk_interface_tx,
@@ -1217,11 +1270,12 @@ VNET_DEVICE_CLASS (dpdk_device_class) = {
.name_renumber = dpdk_device_renumber,
};
-VLIB_DEVICE_TX_FUNCTION_MULTIARCH (dpdk_device_class,
- dpdk_interface_tx)
+VLIB_DEVICE_TX_FUNCTION_MULTIARCH (dpdk_device_class, dpdk_interface_tx)
+/* *INDENT-ON* */
-void dpdk_set_flowcontrol_callback (vlib_main_t *vm,
- dpdk_flowcontrol_callback_t callback)
+void
+dpdk_set_flowcontrol_callback (vlib_main_t * vm,
+ dpdk_flowcontrol_callback_t callback)
{
dpdk_main.flowcontrol_callback = callback;
}
@@ -1229,72 +1283,79 @@ void dpdk_set_flowcontrol_callback (vlib_main_t *vm,
#define UP_DOWN_FLAG_EVENT 1
-u32 dpdk_get_admin_up_down_in_progress (void)
+u32
+dpdk_get_admin_up_down_in_progress (void)
{
return dpdk_main.admin_up_down_in_progress;
}
uword
admin_up_down_process (vlib_main_t * vm,
- vlib_node_runtime_t * rt,
- vlib_frame_t * f)
+ vlib_node_runtime_t * rt, vlib_frame_t * f)
{
- clib_error_t * error = 0;
+ clib_error_t *error = 0;
uword event_type;
uword *event_data = 0;
u32 sw_if_index;
u32 flags;
- while (1)
- {
+ while (1)
+ {
vlib_process_wait_for_event (vm);
event_type = vlib_process_get_events (vm, &event_data);
dpdk_main.admin_up_down_in_progress = 1;
- switch (event_type) {
- case UP_DOWN_FLAG_EVENT:
- {
- if (vec_len(event_data) == 2) {
- sw_if_index = event_data[0];
- flags = event_data[1];
- error = vnet_sw_interface_set_flags (vnet_get_main(), sw_if_index, flags);
- clib_error_report(error);
- }
- }
- break;
- }
+ switch (event_type)
+ {
+ case UP_DOWN_FLAG_EVENT:
+ {
+ if (vec_len (event_data) == 2)
+ {
+ sw_if_index = event_data[0];
+ flags = event_data[1];
+ error =
+ vnet_sw_interface_set_flags (vnet_get_main (), sw_if_index,
+ flags);
+ clib_error_report (error);
+ }
+ }
+ break;
+ }
vec_reset_length (event_data);
dpdk_main.admin_up_down_in_progress = 0;
}
- return 0; /* or not */
+ return 0; /* or not */
}
+/* *INDENT-OFF* */
VLIB_REGISTER_NODE (admin_up_down_process_node,static) = {
.function = admin_up_down_process,
.type = VLIB_NODE_TYPE_PROCESS,
.name = "admin-up-down-process",
.process_log2_n_stack_bytes = 17, // 256KB
};
+/* *INDENT-ON* */
/*
- * Asynchronously invoke vnet_sw_interface_set_flags via the admin_up_down
- * process. Useful for avoiding long blocking delays (>150ms) in the dpdk
+ * Asynchronously invoke vnet_sw_interface_set_flags via the admin_up_down
+ * process. Useful for avoiding long blocking delays (>150ms) in the dpdk
* drivers.
* WARNING: when posting this event, no other interface-related calls should
* be made (e.g. vnet_create_sw_interface()) while the event is being
- * processed (admin_up_down_in_progress). This is required in order to avoid
+ * processed (admin_up_down_in_progress). This is required in order to avoid
* race conditions in manipulating interface data structures.
*/
-void post_sw_interface_set_flags (vlib_main_t *vm, u32 sw_if_index, u32 flags)
+void
+post_sw_interface_set_flags (vlib_main_t * vm, u32 sw_if_index, u32 flags)
{
- uword * d = vlib_process_signal_event_data
- (vm, admin_up_down_process_node.index,
- UP_DOWN_FLAG_EVENT, 2, sizeof(u32));
+ uword *d = vlib_process_signal_event_data
+ (vm, admin_up_down_process_node.index,
+ UP_DOWN_FLAG_EVENT, 2, sizeof (u32));
d[0] = sw_if_index;
d[1] = flags;
}
@@ -1302,33 +1363,37 @@ void post_sw_interface_set_flags (vlib_main_t *vm, u32 sw_if_index, u32 flags)
/*
* Return a copy of the DPDK port stats in dest.
*/
-clib_error_t*
-dpdk_get_hw_interface_stats (u32 hw_if_index, struct rte_eth_stats* dest)
+clib_error_t *
+dpdk_get_hw_interface_stats (u32 hw_if_index, struct rte_eth_stats *dest)
{
- dpdk_main_t * dm = &dpdk_main;
- vnet_main_t * vnm = vnet_get_main();
- vnet_hw_interface_t * hi = vnet_get_hw_interface (vnm, hw_if_index);
- dpdk_device_t * xd = vec_elt_at_index (dm->devices, hi->dev_instance);
-
- if (!dest) {
- return clib_error_return (0, "Missing or NULL argument");
- }
- if (!xd) {
- return clib_error_return (0, "Unable to get DPDK device from HW interface");
- }
+ dpdk_main_t *dm = &dpdk_main;
+ vnet_main_t *vnm = vnet_get_main ();
+ vnet_hw_interface_t *hi = vnet_get_hw_interface (vnm, hw_if_index);
+ dpdk_device_t *xd = vec_elt_at_index (dm->devices, hi->dev_instance);
+
+ if (!dest)
+ {
+ return clib_error_return (0, "Missing or NULL argument");
+ }
+ if (!xd)
+ {
+ return clib_error_return (0,
+ "Unable to get DPDK device from HW interface");
+ }
dpdk_update_counters (xd, vlib_time_now (dm->vlib_main));
- clib_memcpy(dest, &xd->stats, sizeof(xd->stats));
+ clib_memcpy (dest, &xd->stats, sizeof (xd->stats));
return (0);
}
/*
* Return the number of dpdk mbufs
*/
-u32 dpdk_num_mbufs (void)
+u32
+dpdk_num_mbufs (void)
{
- dpdk_main_t * dm = &dpdk_main;
+ dpdk_main_t *dm = &dpdk_main;
return dm->conf->num_mbufs;
}
@@ -1336,10 +1401,11 @@ u32 dpdk_num_mbufs (void)
/*
* Return the pmd type for a given hardware interface
*/
-dpdk_pmd_t dpdk_get_pmd_type (vnet_hw_interface_t *hi)
+dpdk_pmd_t
+dpdk_get_pmd_type (vnet_hw_interface_t * hi)
{
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_t * xd;
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_t *xd;
assert (hi);
@@ -1353,16 +1419,25 @@ dpdk_pmd_t dpdk_get_pmd_type (vnet_hw_interface_t *hi)
/*
* Return the cpu socket for a given hardware interface
*/
-i8 dpdk_get_cpu_socket (vnet_hw_interface_t *hi)
+i8
+dpdk_get_cpu_socket (vnet_hw_interface_t * hi)
{
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_t * xd;
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_t *xd;
assert (hi);
- xd = vec_elt_at_index(dm->devices, hi->dev_instance);
+ xd = vec_elt_at_index (dm->devices, hi->dev_instance);
assert (xd);
return xd->cpu_socket;
}
+
+/*
+ * fd.io coding-style-patch-verification: ON
+ *
+ * Local Variables:
+ * eval: (c-set-style "gnu")
+ * End:
+ */
diff --git a/vnet/vnet/devices/dpdk/dpdk.h b/vnet/vnet/devices/dpdk/dpdk.h
index fba32317b8c..90a50dafa43 100644
--- a/vnet/vnet/devices/dpdk/dpdk.h
+++ b/vnet/vnet/devices/dpdk/dpdk.h
@@ -75,11 +75,12 @@ extern vnet_device_class_t dpdk_device_class;
extern vlib_node_registration_t dpdk_input_node;
extern vlib_node_registration_t handoff_dispatch_node;
-typedef enum {
- VNET_DPDK_DEV_ETH = 1, /* Standard DPDK PMD driver */
- VNET_DPDK_DEV_KNI, /* Kernel NIC Interface */
+typedef enum
+{
+ VNET_DPDK_DEV_ETH = 1, /* Standard DPDK PMD driver */
+ VNET_DPDK_DEV_KNI, /* Kernel NIC Interface */
VNET_DPDK_DEV_VHOST_USER,
- VNET_DPDK_DEV_UNKNOWN, /* must be last */
+ VNET_DPDK_DEV_UNKNOWN, /* must be last */
} dpdk_device_type_t;
#define foreach_dpdk_pmd \
@@ -100,15 +101,17 @@ typedef enum {
_ ("rte_cxgbe_pmd", CXGBE) \
_ ("rte_dpaa2_dpni", DPAA2)
-typedef enum {
+typedef enum
+{
VNET_DPDK_PMD_NONE,
#define _(s,f) VNET_DPDK_PMD_##f,
foreach_dpdk_pmd
#undef _
- VNET_DPDK_PMD_UNKNOWN, /* must be last */
+ VNET_DPDK_PMD_UNKNOWN, /* must be last */
} dpdk_pmd_t;
-typedef enum {
+typedef enum
+{
VNET_DPDK_PORT_TYPE_ETH_1G,
VNET_DPDK_PORT_TYPE_ETH_10G,
VNET_DPDK_PORT_TYPE_ETH_40G,
@@ -118,14 +121,16 @@ typedef enum {
VNET_DPDK_PORT_TYPE_UNKNOWN,
} dpdk_port_type_t;
-typedef struct {
+typedef struct
+{
f64 deadline;
- vlib_frame_t * frame;
+ vlib_frame_t *frame;
} dpdk_frame_t;
#define DPDK_EFD_MAX_DISCARD_RATE 10
-typedef struct {
+typedef struct
+{
u16 last_burst_sz;
u16 max_burst_sz;
u32 full_frames_cnt;
@@ -138,7 +143,8 @@ typedef struct {
} dpdk_efd_agent_t;
#if DPDK_VHOST_USER
-typedef struct {
+typedef struct
+{
int callfd;
int kickfd;
int errfd;
@@ -150,7 +156,8 @@ typedef struct {
u64 bytes;
} dpdk_vu_vring;
-typedef struct {
+typedef struct
+{
u32 is_up;
u32 unix_fd;
u32 unix_file_index;
@@ -169,22 +176,23 @@ typedef struct {
} dpdk_vu_intf_t;
#endif
-typedef void (*dpdk_flowcontrol_callback_t) (vlib_main_t *vm,
- u32 hw_if_index,
- u32 n_packets);
+typedef void (*dpdk_flowcontrol_callback_t) (vlib_main_t * vm,
+ u32 hw_if_index, u32 n_packets);
/*
* The header for the tx_vector in dpdk_device_t.
* Head and tail are indexes into the tx_vector and are of type
* u64 so they never overflow.
*/
-typedef struct {
+typedef struct
+{
u64 tx_head;
u64 tx_tail;
} tx_ring_hdr_t;
-typedef struct {
- CLIB_CACHE_LINE_ALIGN_MARK(cacheline0);
+typedef struct
+{
+ CLIB_CACHE_LINE_ALIGN_MARK (cacheline0);
volatile u32 **lockp;
/* Instance ID */
@@ -197,11 +205,11 @@ typedef struct {
u32 per_interface_next_index;
/* dpdk rte_mbuf rx and tx vectors, VLIB_FRAME_SIZE */
- struct rte_mbuf *** tx_vectors; /* one per worker thread */
- struct rte_mbuf *** rx_vectors;
+ struct rte_mbuf ***tx_vectors; /* one per worker thread */
+ struct rte_mbuf ***rx_vectors;
/* vector of traced contexts, per device */
- u32 * d_trace_buffers;
+ u32 *d_trace_buffers;
/* number of sub-interfaces */
u16 vlan_subifs;
@@ -213,16 +221,16 @@ typedef struct {
u8 admin_up;
u8 promisc;
- CLIB_CACHE_LINE_ALIGN_MARK(cacheline1);
+ CLIB_CACHE_LINE_ALIGN_MARK (cacheline1);
- u8 * interface_name_suffix;
+ u8 *interface_name_suffix;
/* PMD related */
u16 tx_q_used;
u16 rx_q_used;
u16 nb_rx_desc;
u16 nb_tx_desc;
- u16 * cpu_socket_id_by_queue;
+ u16 *cpu_socket_id_by_queue;
struct rte_eth_conf port_conf;
struct rte_eth_txconf tx_conf;
@@ -233,7 +241,7 @@ typedef struct {
#if DPDK_VHOST_USER
/* vhost-user related */
u32 vu_if_id;
- struct virtio_net vu_vhost_dev;
+ struct virtio_net vu_vhost_dev;
u32 vu_is_running;
dpdk_vu_intf_t *vu_intf;
#endif
@@ -248,36 +256,38 @@ typedef struct {
struct rte_eth_stats last_stats;
struct rte_eth_stats last_cleared_stats;
#if RTE_VERSION >= RTE_VERSION_NUM(16, 7, 0, 0)
- struct rte_eth_xstat * xstats;
- struct rte_eth_xstat * last_cleared_xstats;
+ struct rte_eth_xstat *xstats;
+ struct rte_eth_xstat *last_cleared_xstats;
#else
- struct rte_eth_xstats * xstats;
- struct rte_eth_xstats * last_cleared_xstats;
+ struct rte_eth_xstats *xstats;
+ struct rte_eth_xstats *last_cleared_xstats;
#endif
f64 time_last_stats_update;
dpdk_port_type_t port_type;
dpdk_efd_agent_t efd_agent;
- u8 need_txlock; /* Used by VNET_DPDK_DEV_VHOST_USER */
+ u8 need_txlock; /* Used by VNET_DPDK_DEV_VHOST_USER */
} dpdk_device_t;
#define DPDK_TX_RING_SIZE (4 * 1024)
#define DPDK_STATS_POLL_INTERVAL (10.0)
-#define DPDK_MIN_STATS_POLL_INTERVAL (0.001) /* 1msec */
+#define DPDK_MIN_STATS_POLL_INTERVAL (0.001) /* 1msec */
#define DPDK_LINK_POLL_INTERVAL (3.0)
-#define DPDK_MIN_LINK_POLL_INTERVAL (0.001) /* 1msec */
+#define DPDK_MIN_LINK_POLL_INTERVAL (0.001) /* 1msec */
-typedef struct {
- CLIB_CACHE_LINE_ALIGN_MARK(cacheline0);
+typedef struct
+{
+ CLIB_CACHE_LINE_ALIGN_MARK (cacheline0);
/* total input packet counter */
u64 aggregate_rx_packets;
} dpdk_worker_t;
-typedef struct {
+typedef struct
+{
u32 device;
u16 queue_id;
} dpdk_device_and_queue_t;
@@ -291,7 +301,8 @@ typedef struct {
#define DPDK_EFD_DEFAULT_DEVICE_QUEUE_HI_THRESH_PCT 90
#define DPDK_EFD_DEFAULT_CONSEC_FULL_FRAMES_HI_THRESH 6
-typedef struct dpdk_efd_t {
+typedef struct dpdk_efd_t
+{
u16 enabled;
u16 queue_hi_thresh;
u16 consec_full_frames_hi_thresh;
@@ -305,10 +316,11 @@ typedef struct dpdk_efd_t {
_ (num_tx_desc) \
_ (rss_fn)
-typedef struct {
- vlib_pci_addr_t pci_addr;
- u8 is_blacklisted;
- u8 vlan_strip_offload;
+typedef struct
+{
+ vlib_pci_addr_t pci_addr;
+ u8 is_blacklisted;
+ u8 vlan_strip_offload;
#define DPDK_DEVICE_VLAN_STRIP_DEFAULT 0
#define DPDK_DEVICE_VLAN_STRIP_OFF 1
#define DPDK_DEVICE_VLAN_STRIP_ON 2
@@ -319,12 +331,13 @@ typedef struct {
clib_bitmap_t * workers;
} dpdk_device_config_t;
-typedef struct {
+typedef struct
+{
/* Config stuff */
- u8 ** eal_init_args;
- u8 * eal_init_args_str;
- u8 * uio_driver_name;
+ u8 **eal_init_args;
+ u8 *eal_init_args_str;
+ u8 *uio_driver_name;
u8 no_multi_seg;
u8 enable_tcp_udp_checksum;
@@ -334,7 +347,7 @@ typedef struct {
u32 coremask;
u32 nchannels;
u32 num_mbufs;
- u8 num_kni;/* while kni_init allows u32, port_id in callback fn is only u8 */
+ u8 num_kni; /* while kni_init allows u32, port_id in callback fn is only u8 */
/*
* format interface names ala xxxEthernet%d/%d/%d instead of
@@ -351,21 +364,22 @@ typedef struct {
/* per-device config */
dpdk_device_config_t default_devconf;
- dpdk_device_config_t * dev_confs;
- uword * device_config_index_by_pci_addr;
+ dpdk_device_config_t *dev_confs;
+ uword *device_config_index_by_pci_addr;
} dpdk_config_main_t;
dpdk_config_main_t dpdk_config_main;
-typedef struct {
+typedef struct
+{
/* Devices */
- dpdk_device_t * devices;
- dpdk_device_and_queue_t ** devices_by_cpu;
+ dpdk_device_t *devices;
+ dpdk_device_and_queue_t **devices_by_cpu;
/* per-thread recycle lists */
- u32 ** recycle;
+ u32 **recycle;
/* buffer flags template, configurable to enable/disable tcp / udp cksum */
u32 buffer_flags_template;
@@ -377,7 +391,7 @@ typedef struct {
u32 vlib_buffer_free_list_index;
/* dpdk worker "threads" */
- dpdk_worker_t * workers;
+ dpdk_worker_t *workers;
/* Ethernet input node index */
@@ -386,15 +400,15 @@ typedef struct {
/* pcap tracing [only works if (CLIB_DEBUG > 0)] */
int tx_pcap_enable;
pcap_main_t pcap_main;
- u8 * pcap_filename;
+ u8 *pcap_filename;
u32 pcap_sw_if_index;
u32 pcap_pkts_to_capture;
/* hashes */
- uword * dpdk_device_by_kni_port_id;
- uword * vu_sw_if_index_by_listener_fd;
- uword * vu_sw_if_index_by_sock_fd;
- u32 * vu_inactive_interfaces_device_index;
+ uword *dpdk_device_by_kni_port_id;
+ uword *vu_sw_if_index_by_listener_fd;
+ uword *vu_sw_if_index_by_sock_fd;
+ u32 *vu_inactive_interfaces_device_index;
u32 next_vu_if_id;
@@ -421,14 +435,15 @@ typedef struct {
u32 poll_sleep;
/* convenience */
- vlib_main_t * vlib_main;
- vnet_main_t * vnet_main;
- dpdk_config_main_t * conf;
+ vlib_main_t *vlib_main;
+ vnet_main_t *vnet_main;
+ dpdk_config_main_t *conf;
} dpdk_main_t;
dpdk_main_t dpdk_main;
-typedef enum {
+typedef enum
+{
DPDK_RX_NEXT_IP4_INPUT,
DPDK_RX_NEXT_IP6_INPUT,
DPDK_RX_NEXT_MPLS_INPUT,
@@ -437,7 +452,8 @@ typedef enum {
DPDK_RX_N_NEXT,
} dpdk_rx_next_t;
-typedef struct {
+typedef struct
+{
u32 buffer_index;
u16 device_index;
u8 queue_index;
@@ -446,37 +462,38 @@ typedef struct {
vlib_buffer_t buffer;
} dpdk_tx_dma_trace_t;
-typedef struct {
+typedef struct
+{
u32 buffer_index;
u16 device_index;
u16 queue_index;
struct rte_mbuf mb;
- vlib_buffer_t buffer; /* Copy of VLIB buffer; pkt data stored in pre_data. */
- u8 data[256]; /* First 256 data bytes, used for hexdump */
+ vlib_buffer_t buffer; /* Copy of VLIB buffer; pkt data stored in pre_data. */
+ u8 data[256]; /* First 256 data bytes, used for hexdump */
} dpdk_rx_dma_trace_t;
void vnet_buffer_needs_dpdk_mb (vlib_buffer_t * b);
void dpdk_set_next_node (dpdk_rx_next_t, char *);
-clib_error_t * dpdk_set_mac_address (vnet_hw_interface_t * hi, char * address);
+clib_error_t *dpdk_set_mac_address (vnet_hw_interface_t * hi, char *address);
-clib_error_t * dpdk_set_mc_filter (vnet_hw_interface_t * hi,
- struct ether_addr mc_addr_vec[], int naddr);
+clib_error_t *dpdk_set_mc_filter (vnet_hw_interface_t * hi,
+ struct ether_addr mc_addr_vec[], int naddr);
void dpdk_thread_input (dpdk_main_t * dm, dpdk_device_t * xd);
-clib_error_t * dpdk_port_setup (dpdk_main_t * dm, dpdk_device_t * xd);
+clib_error_t *dpdk_port_setup (dpdk_main_t * dm, dpdk_device_t * xd);
-void dpdk_set_flowcontrol_callback (vlib_main_t *vm,
- dpdk_flowcontrol_callback_t callback);
+void dpdk_set_flowcontrol_callback (vlib_main_t * vm,
+ dpdk_flowcontrol_callback_t callback);
u32 dpdk_interface_tx_vector (vlib_main_t * vm, u32 dev_instance);
-void set_efd_bitmap (u8 *bitmap, u32 value, u32 op);
+void set_efd_bitmap (u8 * bitmap, u32 value, u32 op);
-struct rte_mbuf * dpdk_replicate_packet_mb (vlib_buffer_t * b);
-struct rte_mbuf * dpdk_zerocopy_replicate_packet_mb (vlib_buffer_t * b);
+struct rte_mbuf *dpdk_replicate_packet_mb (vlib_buffer_t * b);
+struct rte_mbuf *dpdk_zerocopy_replicate_packet_mb (vlib_buffer_t * b);
#define foreach_dpdk_error \
_(NONE, "no error") \
@@ -492,113 +509,113 @@ struct rte_mbuf * dpdk_zerocopy_replicate_packet_mb (vlib_buffer_t * b);
_(MPLS_EFD_DROP_PKTS, "MPLS Early Fast Discard rx drops") \
_(VLAN_EFD_DROP_PKTS, "VLAN Early Fast Discard rx drops")
-typedef enum {
+typedef enum
+{
#define _(f,s) DPDK_ERROR_##f,
foreach_dpdk_error
#undef _
- DPDK_N_ERROR,
+ DPDK_N_ERROR,
} dpdk_error_t;
/*
* Increment EFD drop counter
*/
-static_always_inline
-void increment_efd_drop_counter (vlib_main_t * vm, u32 counter_index, u32 count)
+static_always_inline void
+increment_efd_drop_counter (vlib_main_t * vm, u32 counter_index, u32 count)
{
- vlib_node_t *my_n;
+ vlib_node_t *my_n;
- my_n = vlib_get_node (vm, dpdk_input_node.index);
- vm->error_main.counters[my_n->error_heap_index+counter_index] += count;
+ my_n = vlib_get_node (vm, dpdk_input_node.index);
+ vm->error_main.counters[my_n->error_heap_index + counter_index] += count;
}
int dpdk_set_stat_poll_interval (f64 interval);
int dpdk_set_link_state_poll_interval (f64 interval);
void dpdk_update_link_state (dpdk_device_t * xd, f64 now);
-void dpdk_device_lock_init(dpdk_device_t * xd);
-void dpdk_device_lock_free(dpdk_device_t * xd);
-void dpdk_efd_update_counters(dpdk_device_t *xd, u32 n_buffers, u16 enabled);
-u32 is_efd_discardable(vlib_thread_main_t *tm,
- vlib_buffer_t * b0,
- struct rte_mbuf *mb);
+void dpdk_device_lock_init (dpdk_device_t * xd);
+void dpdk_device_lock_free (dpdk_device_t * xd);
+void dpdk_efd_update_counters (dpdk_device_t * xd, u32 n_buffers,
+ u16 enabled);
+u32 is_efd_discardable (vlib_thread_main_t * tm, vlib_buffer_t * b0,
+ struct rte_mbuf *mb);
#if DPDK_VHOST_USER
/* dpdk vhost-user interrupt management */
-u8 dpdk_vhost_user_want_interrupt (dpdk_device_t *xd, int idx);
+u8 dpdk_vhost_user_want_interrupt (dpdk_device_t * xd, int idx);
void dpdk_vhost_user_send_interrupt (vlib_main_t * vm, dpdk_device_t * xd,
- int idx);
+ int idx);
#endif
-static inline u64 vnet_get_aggregate_rx_packets (void)
+static inline u64
+vnet_get_aggregate_rx_packets (void)
{
- dpdk_main_t * dm = &dpdk_main;
- u64 sum = 0;
- dpdk_worker_t * dw;
+ dpdk_main_t *dm = &dpdk_main;
+ u64 sum = 0;
+ dpdk_worker_t *dw;
- vec_foreach(dw, dm->workers)
- sum += dw->aggregate_rx_packets;
+ vec_foreach (dw, dm->workers) sum += dw->aggregate_rx_packets;
- return sum;
+ return sum;
}
void dpdk_rx_trace (dpdk_main_t * dm,
- vlib_node_runtime_t * node,
- dpdk_device_t * xd,
- u16 queue_id,
- u32 * buffers,
- uword n_buffers);
+ vlib_node_runtime_t * node,
+ dpdk_device_t * xd,
+ u16 queue_id, u32 * buffers, uword n_buffers);
#define EFD_OPERATION_LESS_THAN 0
#define EFD_OPERATION_GREATER_OR_EQUAL 1
-void efd_config(u32 enabled,
- u32 ip_prec, u32 ip_op,
- u32 mpls_exp, u32 mpls_op,
- u32 vlan_cos, u32 vlan_op);
+void efd_config (u32 enabled,
+ u32 ip_prec, u32 ip_op,
+ u32 mpls_exp, u32 mpls_op, u32 vlan_cos, u32 vlan_op);
-void post_sw_interface_set_flags (vlib_main_t *vm, u32 sw_if_index, u32 flags);
+void post_sw_interface_set_flags (vlib_main_t * vm, u32 sw_if_index,
+ u32 flags);
#if DPDK_VHOST_USER
typedef struct vhost_user_memory vhost_user_memory_t;
void dpdk_vhost_user_process_init (void **ctx);
void dpdk_vhost_user_process_cleanup (void *ctx);
-uword dpdk_vhost_user_process_if (vlib_main_t *vm, dpdk_device_t *xd, void *ctx);
+uword dpdk_vhost_user_process_if (vlib_main_t * vm, dpdk_device_t * xd,
+ void *ctx);
// vhost-user calls
int dpdk_vhost_user_create_if (vnet_main_t * vnm, vlib_main_t * vm,
- const char * sock_filename,
- u8 is_server,
- u32 * sw_if_index,
- u64 feature_mask,
- u8 renumber, u32 custom_dev_instance,
- u8 *hwaddr);
+ const char *sock_filename,
+ u8 is_server,
+ u32 * sw_if_index,
+ u64 feature_mask,
+ u8 renumber, u32 custom_dev_instance,
+ u8 * hwaddr);
int dpdk_vhost_user_modify_if (vnet_main_t * vnm, vlib_main_t * vm,
- const char * sock_filename,
- u8 is_server,
- u32 sw_if_index,
- u64 feature_mask,
- u8 renumber, u32 custom_dev_instance);
+ const char *sock_filename,
+ u8 is_server,
+ u32 sw_if_index,
+ u64 feature_mask,
+ u8 renumber, u32 custom_dev_instance);
int dpdk_vhost_user_delete_if (vnet_main_t * vnm, vlib_main_t * vm,
- u32 sw_if_index);
+ u32 sw_if_index);
int dpdk_vhost_user_dump_ifs (vnet_main_t * vnm, vlib_main_t * vm,
- vhost_user_intf_details_t **out_vuids);
+ vhost_user_intf_details_t ** out_vuids);
#endif
u32 dpdk_get_admin_up_down_in_progress (void);
u32 dpdk_num_mbufs (void);
-dpdk_pmd_t dpdk_get_pmd_type (vnet_hw_interface_t *hi);
+dpdk_pmd_t dpdk_get_pmd_type (vnet_hw_interface_t * hi);
-i8 dpdk_get_cpu_socket (vnet_hw_interface_t *hi);
+i8 dpdk_get_cpu_socket (vnet_hw_interface_t * hi);
-void * dpdk_input_multiarch_select();
-void * dpdk_input_rss_multiarch_select();
-void * dpdk_input_efd_multiarch_select();
+void *dpdk_input_multiarch_select ();
+void *dpdk_input_rss_multiarch_select ();
+void *dpdk_input_efd_multiarch_select ();
-clib_error_t*
-dpdk_get_hw_interface_stats (u32 hw_if_index, struct rte_eth_stats* dest);
+clib_error_t *dpdk_get_hw_interface_stats (u32 hw_if_index,
+ struct rte_eth_stats *dest);
format_function_t format_dpdk_device_name;
format_function_t format_dpdk_device;
@@ -607,11 +624,11 @@ format_function_t format_dpdk_rx_dma_trace;
format_function_t format_dpdk_rte_mbuf;
format_function_t format_dpdk_rx_rte_mbuf;
unformat_function_t unformat_socket_mem;
-clib_error_t * unformat_rss_fn(unformat_input_t * input, uword * rss_fn);
+clib_error_t *unformat_rss_fn (unformat_input_t * input, uword * rss_fn);
static inline void
-dpdk_pmd_constructor_init()
+dpdk_pmd_constructor_init ()
{
/* Add references to DPDK Driver Constructor functions to get the dynamic
* loader to pull in the driver library & run the constructors.
@@ -626,58 +643,43 @@ dpdk_pmd_constructor_init()
#ifdef RTE_LIBRTE_EM_PMD
_(em_pmd_drv)
#endif
-
#ifdef RTE_LIBRTE_IGB_PMD
- _(pmd_igb_drv)
+ _(pmd_igb_drv)
#endif
-
#ifdef RTE_LIBRTE_IXGBE_PMD
- _(rte_ixgbe_driver)
+ _(rte_ixgbe_driver)
#endif
-
#ifdef RTE_LIBRTE_I40E_PMD
- _(rte_i40e_driver)
- _(rte_i40evf_driver)
+ _(rte_i40e_driver) _(rte_i40evf_driver)
#endif
-
#ifdef RTE_LIBRTE_FM10K_PMD
- _(rte_fm10k_driver)
+ _(rte_fm10k_driver)
#endif
-
#ifdef RTE_LIBRTE_VIRTIO_PMD
- _(rte_virtio_driver)
+ _(rte_virtio_driver)
#endif
-
#ifdef RTE_LIBRTE_VMXNET3_PMD
- _(rte_vmxnet3_driver)
+ _(rte_vmxnet3_driver)
#endif
-
#ifdef RTE_LIBRTE_VICE_PMD
- _(rte_vice_driver)
+ _(rte_vice_driver)
#endif
-
#ifdef RTE_LIBRTE_ENIC_PMD
- _(rte_enic_driver)
+ _(rte_enic_driver)
#endif
-
#ifdef RTE_LIBRTE_PMD_AF_PACKET
- _(pmd_af_packet_drv)
+ _(pmd_af_packet_drv)
#endif
-
#ifdef RTE_LIBRTE_CXGBE_PMD
- _(rte_cxgbe_driver)
+ _(rte_cxgbe_driver)
#endif
-
#ifdef RTE_LIBRTE_PMD_BOND
- _(bond_drv)
+ _(bond_drv)
#endif
-
#ifdef RTE_LIBRTE_DPAA2_PMD
- _(pmd_dpaa2_drv)
+ _(pmd_dpaa2_drv)
#endif
-
#undef _
-
/*
* At the moment, the ThunderX NIC driver doesn't have
* an entry point named "devinitfn_rte_xxx_driver"
@@ -688,17 +690,22 @@ dpdk_pmd_constructor_init()
__attribute__((unused)) void (* volatile pf)(void); \
pf = d; \
} while(0);
-
#ifdef RTE_LIBRTE_THUNDERVNIC_PMD
- _(rte_nicvf_pmd_init)
+ _(rte_nicvf_pmd_init)
#endif
#undef _
-
}
uword
admin_up_down_process (vlib_main_t * vm,
- vlib_node_runtime_t * rt,
- vlib_frame_t * f);
+ vlib_node_runtime_t * rt, vlib_frame_t * f);
#endif /* __included_dpdk_h__ */
+
+/*
+ * fd.io coding-style-patch-verification: ON
+ *
+ * Local Variables:
+ * eval: (c-set-style "gnu")
+ * End:
+ */
diff --git a/vnet/vnet/devices/dpdk/dpdk_priv.h b/vnet/vnet/devices/dpdk/dpdk_priv.h
index d4d14222de3..4ecb1b82502 100644
--- a/vnet/vnet/devices/dpdk/dpdk_priv.h
+++ b/vnet/vnet/devices/dpdk/dpdk_priv.h
@@ -56,121 +56,24 @@ _(proc-type) \
_(file-prefix) \
_(vdev)
-static inline u32
-dpdk_rx_burst ( dpdk_main_t * dm, dpdk_device_t * xd, u16 queue_id)
-{
- u32 n_buffers;
- u32 n_left;
- u32 n_this_chunk;
-
- n_left = VLIB_FRAME_SIZE;
- n_buffers = 0;
-
- if (PREDICT_TRUE(xd->dev_type == VNET_DPDK_DEV_ETH))
- {
- while (n_left)
- {
- n_this_chunk = rte_eth_rx_burst (xd->device_index, queue_id,
- xd->rx_vectors[queue_id] + n_buffers, n_left);
- n_buffers += n_this_chunk;
- n_left -= n_this_chunk;
-
- /* Empirically, DPDK r1.8 produces vectors w/ 32 or fewer elts */
- if (n_this_chunk < 32)
- break;
- }
- }
-#if DPDK_VHOST_USER
- else if (xd->dev_type == VNET_DPDK_DEV_VHOST_USER)
- {
- vlib_main_t * vm = vlib_get_main();
- vlib_buffer_main_t * bm = vm->buffer_main;
- unsigned socket_id = rte_socket_id();
- u32 offset = 0;
-
- offset = queue_id * VIRTIO_QNUM;
-
- struct vhost_virtqueue *vq =
- xd->vu_vhost_dev.virtqueue[offset + VIRTIO_TXQ];
-
- if (PREDICT_FALSE(!vq->enabled))
- return 0;
-
- struct rte_mbuf **pkts = xd->rx_vectors[queue_id];
- while (n_left) {
- n_this_chunk = rte_vhost_dequeue_burst(&xd->vu_vhost_dev,
- offset + VIRTIO_TXQ,
- bm->pktmbuf_pools[socket_id],
- pkts + n_buffers,
- n_left);
- n_buffers += n_this_chunk;
- n_left -= n_this_chunk;
- if (n_this_chunk == 0)
- break;
- }
-
- int i; u32 bytes = 0;
- for (i = 0; i < n_buffers; i++) {
- struct rte_mbuf *buff = pkts[i];
- bytes += rte_pktmbuf_data_len(buff);
- }
-
- f64 now = vlib_time_now (vm);
-
- dpdk_vu_vring *vring = NULL;
- /* send pending interrupts if needed */
- if (dpdk_vhost_user_want_interrupt(xd, offset + VIRTIO_TXQ)) {
- vring = &(xd->vu_intf->vrings[offset + VIRTIO_TXQ]);
- vring->n_since_last_int += n_buffers;
-
- if ((vring->n_since_last_int && (vring->int_deadline < now))
- || (vring->n_since_last_int > dm->conf->vhost_coalesce_frames))
- dpdk_vhost_user_send_interrupt(vm, xd, offset + VIRTIO_TXQ);
- }
-
- vring = &(xd->vu_intf->vrings[offset + VIRTIO_RXQ]);
- vring->packets += n_buffers;
- vring->bytes += bytes;
-
- if (dpdk_vhost_user_want_interrupt(xd, offset + VIRTIO_RXQ)) {
- if (vring->n_since_last_int && (vring->int_deadline < now))
- dpdk_vhost_user_send_interrupt(vm, xd, offset + VIRTIO_RXQ);
- }
-
- }
-#endif
-#ifdef RTE_LIBRTE_KNI
- else if (xd->dev_type == VNET_DPDK_DEV_KNI)
- {
- n_buffers = rte_kni_rx_burst(xd->kni, xd->rx_vectors[queue_id], VLIB_FRAME_SIZE);
- rte_kni_handle_request(xd->kni);
- }
-#endif
- else
- {
- ASSERT(0);
- }
-
- return n_buffers;
-}
-
-
static inline void
dpdk_get_xstats (dpdk_device_t * xd)
{
int len;
- if ((len = rte_eth_xstats_get(xd->device_index, NULL, 0)) > 0)
+ if ((len = rte_eth_xstats_get (xd->device_index, NULL, 0)) > 0)
{
- vec_validate(xd->xstats, len - 1);
- vec_validate(xd->last_cleared_xstats, len - 1);
+ vec_validate (xd->xstats, len - 1);
+ vec_validate (xd->last_cleared_xstats, len - 1);
- len = rte_eth_xstats_get(xd->device_index, xd->xstats, vec_len(xd->xstats));
+ len =
+ rte_eth_xstats_get (xd->device_index, xd->xstats,
+ vec_len (xd->xstats));
- ASSERT(vec_len(xd->xstats) == len);
- ASSERT(vec_len(xd->last_cleared_xstats) == len);
+ ASSERT (vec_len (xd->xstats) == len);
+ ASSERT (vec_len (xd->last_cleared_xstats) == len);
- _vec_len(xd->xstats) = len;
- _vec_len(xd->last_cleared_xstats) = len;
+ _vec_len (xd->xstats) = len;
+ _vec_len (xd->last_cleared_xstats) = len;
}
}
@@ -179,9 +82,9 @@ dpdk_get_xstats (dpdk_device_t * xd)
static inline void
dpdk_update_counters (dpdk_device_t * xd, f64 now)
{
- vlib_simple_counter_main_t * cm;
- vnet_main_t * vnm = vnet_get_main();
- u32 my_cpu = os_get_cpu_number();
+ vlib_simple_counter_main_t *cm;
+ vnet_main_t *vnm = vnet_get_main ();
+ u32 my_cpu = os_get_cpu_number ();
u64 rxerrors, last_rxerrors;
/* only update counters for PMD interfaces */
@@ -202,37 +105,45 @@ dpdk_update_counters (dpdk_device_t * xd, f64 now)
/* maybe bump interface rx no buffer counter */
if (PREDICT_FALSE (xd->stats.rx_nombuf != xd->last_stats.rx_nombuf))
- {
- cm = vec_elt_at_index (vnm->interface_main.sw_if_counters,
- VNET_INTERFACE_COUNTER_RX_NO_BUF);
+ {
+ cm = vec_elt_at_index (vnm->interface_main.sw_if_counters,
+ VNET_INTERFACE_COUNTER_RX_NO_BUF);
- vlib_increment_simple_counter (cm, my_cpu, xd->vlib_sw_if_index,
- xd->stats.rx_nombuf -
- xd->last_stats.rx_nombuf);
- }
+ vlib_increment_simple_counter (cm, my_cpu, xd->vlib_sw_if_index,
+ xd->stats.rx_nombuf -
+ xd->last_stats.rx_nombuf);
+ }
/* missed pkt counter */
if (PREDICT_FALSE (xd->stats.imissed != xd->last_stats.imissed))
- {
- cm = vec_elt_at_index (vnm->interface_main.sw_if_counters,
- VNET_INTERFACE_COUNTER_RX_MISS);
-
- vlib_increment_simple_counter (cm, my_cpu, xd->vlib_sw_if_index,
- xd->stats.imissed -
- xd->last_stats.imissed);
- }
+ {
+ cm = vec_elt_at_index (vnm->interface_main.sw_if_counters,
+ VNET_INTERFACE_COUNTER_RX_MISS);
+
+ vlib_increment_simple_counter (cm, my_cpu, xd->vlib_sw_if_index,
+ xd->stats.imissed -
+ xd->last_stats.imissed);
+ }
rxerrors = xd->stats.ierrors;
last_rxerrors = xd->last_stats.ierrors;
if (PREDICT_FALSE (rxerrors != last_rxerrors))
- {
- cm = vec_elt_at_index (vnm->interface_main.sw_if_counters,
- VNET_INTERFACE_COUNTER_RX_ERROR);
+ {
+ cm = vec_elt_at_index (vnm->interface_main.sw_if_counters,
+ VNET_INTERFACE_COUNTER_RX_ERROR);
- vlib_increment_simple_counter (cm, my_cpu, xd->vlib_sw_if_index,
- rxerrors - last_rxerrors);
- }
+ vlib_increment_simple_counter (cm, my_cpu, xd->vlib_sw_if_index,
+ rxerrors - last_rxerrors);
+ }
}
- dpdk_get_xstats(xd);
+ dpdk_get_xstats (xd);
}
+
+/*
+ * fd.io coding-style-patch-verification: ON
+ *
+ * Local Variables:
+ * eval: (c-set-style "gnu")
+ * End:
+ */
diff --git a/vnet/vnet/devices/dpdk/format.c b/vnet/vnet/devices/dpdk/format.c
index 075f35777b5..1f401a9e01b 100644
--- a/vnet/vnet/devices/dpdk/format.c
+++ b/vnet/vnet/devices/dpdk/format.c
@@ -150,14 +150,15 @@
#endif /* RTE_LIBRTE_MBUF_EXT_RX_OLFLAGS */
-u8 * format_dpdk_device_name (u8 * s, va_list * args)
+u8 *
+format_dpdk_device_name (u8 * s, va_list * args)
{
- dpdk_main_t * dm = &dpdk_main;
+ dpdk_main_t *dm = &dpdk_main;
char *devname_format;
char *device_name;
u32 i = va_arg (*args, u32);
struct rte_eth_dev_info dev_info;
- u8 * ret;
+ u8 *ret;
if (dm->conf->interface_name_format_decimal)
devname_format = "%s%d/%d/%d";
@@ -165,14 +166,17 @@ u8 * format_dpdk_device_name (u8 * s, va_list * args)
devname_format = "%s%x/%x/%x";
#ifdef RTE_LIBRTE_KNI
- if (dm->devices[i].dev_type == VNET_DPDK_DEV_KNI) {
- return format(s, "kni%d", dm->devices[i].kni_port_id);
- } else
+ if (dm->devices[i].dev_type == VNET_DPDK_DEV_KNI)
+ {
+ return format (s, "kni%d", dm->devices[i].kni_port_id);
+ }
+ else
#endif
#if DPDK_VHOST_USER
- if (dm->devices[i].dev_type == VNET_DPDK_DEV_VHOST_USER) {
- return format(s, "VirtualEthernet0/0/%d", dm->devices[i].vu_if_id);
- }
+ if (dm->devices[i].dev_type == VNET_DPDK_DEV_VHOST_USER)
+ {
+ return format (s, "VirtualEthernet0/0/%d", dm->devices[i].vu_if_id);
+ }
#endif
switch (dm->devices[i].port_type)
{
@@ -189,15 +193,15 @@ u8 * format_dpdk_device_name (u8 * s, va_list * args)
break;
case VNET_DPDK_PORT_TYPE_ETH_BOND:
- return format(s, "BondEthernet%d", dm->devices[i].device_index);
+ return format (s, "BondEthernet%d", dm->devices[i].device_index);
case VNET_DPDK_PORT_TYPE_ETH_SWITCH:
device_name = "EthernetSwitch";
break;
case VNET_DPDK_PORT_TYPE_AF_PACKET:
- rte_eth_dev_info_get(i, &dev_info);
- return format(s, "af_packet%d", dm->devices[i].af_packet_port_id);
+ rte_eth_dev_info_get (i, &dev_info);
+ return format (s, "af_packet%d", dm->devices[i].af_packet_port_id);
default:
case VNET_DPDK_PORT_TYPE_UNKNOWN:
@@ -205,105 +209,110 @@ u8 * format_dpdk_device_name (u8 * s, va_list * args)
break;
}
- rte_eth_dev_info_get(i, &dev_info);
+ rte_eth_dev_info_get (i, &dev_info);
ret = format (s, devname_format, device_name, dev_info.pci_dev->addr.bus,
- dev_info.pci_dev->addr.devid,
- dev_info.pci_dev->addr.function);
+ dev_info.pci_dev->addr.devid,
+ dev_info.pci_dev->addr.function);
if (dm->devices[i].interface_name_suffix)
return format (ret, "/%s", dm->devices[i].interface_name_suffix);
return ret;
}
-static u8 * format_dpdk_device_type (u8 * s, va_list * args)
+static u8 *
+format_dpdk_device_type (u8 * s, va_list * args)
{
- dpdk_main_t * dm = &dpdk_main;
+ dpdk_main_t *dm = &dpdk_main;
char *dev_type;
u32 i = va_arg (*args, u32);
- if (dm->devices[i].dev_type == VNET_DPDK_DEV_KNI) {
- return format(s, "Kernel NIC Interface");
- } else if (dm->devices[i].dev_type == VNET_DPDK_DEV_VHOST_USER) {
- return format(s, "vhost-user interface");
- }
+ if (dm->devices[i].dev_type == VNET_DPDK_DEV_KNI)
+ {
+ return format (s, "Kernel NIC Interface");
+ }
+ else if (dm->devices[i].dev_type == VNET_DPDK_DEV_VHOST_USER)
+ {
+ return format (s, "vhost-user interface");
+ }
switch (dm->devices[i].pmd)
{
case VNET_DPDK_PMD_E1000EM:
- dev_type = "Intel 82540EM (e1000)";
- break;
+ dev_type = "Intel 82540EM (e1000)";
+ break;
case VNET_DPDK_PMD_IGB:
- dev_type = "Intel e1000";
- break;
+ dev_type = "Intel e1000";
+ break;
case VNET_DPDK_PMD_I40E:
- dev_type = "Intel X710/XL710 Family";
- break;
+ dev_type = "Intel X710/XL710 Family";
+ break;
case VNET_DPDK_PMD_I40EVF:
- dev_type = "Intel X710/XL710 Family VF";
- break;
+ dev_type = "Intel X710/XL710 Family VF";
+ break;
case VNET_DPDK_PMD_FM10K:
- dev_type = "Intel FM10000 Family Ethernet Switch";
- break;
+ dev_type = "Intel FM10000 Family Ethernet Switch";
+ break;
case VNET_DPDK_PMD_IGBVF:
- dev_type = "Intel e1000 VF";
- break;
+ dev_type = "Intel e1000 VF";
+ break;
case VNET_DPDK_PMD_VIRTIO:
- dev_type = "Red Hat Virtio";
- break;
+ dev_type = "Red Hat Virtio";
+ break;
case VNET_DPDK_PMD_IXGBEVF:
- dev_type = "Intel 82599 VF";
- break;
+ dev_type = "Intel 82599 VF";
+ break;
case VNET_DPDK_PMD_IXGBE:
- dev_type = "Intel 82599";
- break;
+ dev_type = "Intel 82599";
+ break;
case VNET_DPDK_PMD_ENIC:
- dev_type = "Cisco VIC";
- break;
+ dev_type = "Cisco VIC";
+ break;
case VNET_DPDK_PMD_CXGBE:
- dev_type = "Chelsio T4/T5";
- break;
+ dev_type = "Chelsio T4/T5";
+ break;
case VNET_DPDK_PMD_VMXNET3:
- dev_type = "VMware VMXNET3";
- break;
+ dev_type = "VMware VMXNET3";
+ break;
case VNET_DPDK_PMD_AF_PACKET:
- dev_type = "af_packet";
- break;
+ dev_type = "af_packet";
+ break;
case VNET_DPDK_PMD_BOND:
- dev_type = "Ethernet Bonding";
- break;
+ dev_type = "Ethernet Bonding";
+ break;
case VNET_DPDK_PMD_DPAA2:
- dev_type = "NXP DPAA2 Mac";
- break;
+ dev_type = "NXP DPAA2 Mac";
+ break;
default:
case VNET_DPDK_PMD_UNKNOWN:
- dev_type = "### UNKNOWN ###";
- break;
+ dev_type = "### UNKNOWN ###";
+ break;
}
return format (s, dev_type);
}
-static u8 * format_dpdk_link_status (u8 * s, va_list * args)
+static u8 *
+format_dpdk_link_status (u8 * s, va_list * args)
{
- dpdk_device_t * xd = va_arg (*args, dpdk_device_t *);
- struct rte_eth_link * l = &xd->link;
- vnet_main_t * vnm = vnet_get_main();
- vnet_hw_interface_t * hi = vnet_get_hw_interface (vnm, xd->vlib_hw_if_index);
+ dpdk_device_t *xd = va_arg (*args, dpdk_device_t *);
+ struct rte_eth_link *l = &xd->link;
+ vnet_main_t *vnm = vnet_get_main ();
+ vnet_hw_interface_t *hi = vnet_get_hw_interface (vnm, xd->vlib_hw_if_index);
s = format (s, "%s ", l->link_status ? "up" : "down");
if (l->link_status)
@@ -311,7 +320,7 @@ static u8 * format_dpdk_link_status (u8 * s, va_list * args)
u32 promisc = rte_eth_promiscuous_get (xd->device_index);
s = format (s, "%s duplex ", (l->link_duplex == ETH_LINK_FULL_DUPLEX) ?
- "full" : "half");
+ "full" : "half");
s = format (s, "speed %u mtu %d %s\n", l->link_speed,
hi->max_packet_bytes, promisc ? " promisc" : "");
}
@@ -331,56 +340,54 @@ if (bitmap & v) { \
s = format(s, "%s ", str); \
}
-static u8 * format_dpdk_rss_hf_name(u8 * s, va_list * args)
+static u8 *
+format_dpdk_rss_hf_name (u8 * s, va_list * args)
{
u64 bitmap = va_arg (*args, u64);
int next_split = _line_len;
int indent = format_get_indent (s);
if (!bitmap)
- return format(s, "none");
+ return format (s, "none");
- foreach_dpdk_rss_hf
-
- return s;
+ foreach_dpdk_rss_hf return s;
}
-static u8 * format_dpdk_rx_offload_caps(u8 * s, va_list * args)
+static u8 *
+format_dpdk_rx_offload_caps (u8 * s, va_list * args)
{
u32 bitmap = va_arg (*args, u32);
int next_split = _line_len;
int indent = format_get_indent (s);
if (!bitmap)
- return format(s, "none");
-
- foreach_dpdk_rx_offload_caps
+ return format (s, "none");
- return s;
+ foreach_dpdk_rx_offload_caps return s;
}
-static u8 * format_dpdk_tx_offload_caps(u8 * s, va_list * args)
+static u8 *
+format_dpdk_tx_offload_caps (u8 * s, va_list * args)
{
u32 bitmap = va_arg (*args, u32);
int next_split = _line_len;
int indent = format_get_indent (s);
if (!bitmap)
- return format(s, "none");
-
- foreach_dpdk_tx_offload_caps
+ return format (s, "none");
- return s;
+ foreach_dpdk_tx_offload_caps return s;
}
#undef _line_len
#undef _
-u8 * format_dpdk_device (u8 * s, va_list * args)
+u8 *
+format_dpdk_device (u8 * s, va_list * args)
{
u32 dev_instance = va_arg (*args, u32);
int verbose = va_arg (*args, int);
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_t * xd = vec_elt_at_index (dm->devices, dev_instance);
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_t *xd = vec_elt_at_index (dm->devices, dev_instance);
uword indent = format_get_indent (s);
f64 now = vlib_time_now (dm->vlib_main);
struct rte_eth_dev_info di;
@@ -390,77 +397,78 @@ u8 * format_dpdk_device (u8 * s, va_list * args)
s = format (s, "%U\n%Ucarrier %U",
format_dpdk_device_type, xd->device_index,
- format_white_space, indent + 2,
- format_dpdk_link_status, xd);
+ format_white_space, indent + 2, format_dpdk_link_status, xd);
- rte_eth_dev_info_get(xd->device_index, &di);
+ rte_eth_dev_info_get (xd->device_index, &di);
if (verbose > 1 && xd->dev_type == VNET_DPDK_DEV_ETH)
{
- struct rte_pci_device * pci;
+ struct rte_pci_device *pci;
struct rte_eth_rss_conf rss_conf;
int vlan_off;
int retval;
rss_conf.rss_key = 0;
- retval = rte_eth_dev_rss_hash_conf_get(xd->device_index, &rss_conf);
+ retval = rte_eth_dev_rss_hash_conf_get (xd->device_index, &rss_conf);
if (retval < 0)
- clib_warning ("rte_eth_dev_rss_hash_conf_get returned %d", retval);
+ clib_warning ("rte_eth_dev_rss_hash_conf_get returned %d", retval);
pci = di.pci_dev;
if (pci)
- s = format(s, "%Upci id: device %04x:%04x subsystem %04x:%04x\n"
- "%Upci address: %04x:%02x:%02x.%02x\n",
- format_white_space, indent + 2,
- pci->id.vendor_id, pci->id.device_id,
- pci->id.subsystem_vendor_id,
- pci->id.subsystem_device_id,
- format_white_space, indent + 2,
- pci->addr.domain, pci->addr.bus,
- pci->addr.devid, pci->addr.function);
- s = format(s, "%Umax rx packet len: %d\n",
- format_white_space, indent + 2, di.max_rx_pktlen);
- s = format(s, "%Umax num of queues: rx %d tx %d\n",
- format_white_space, indent + 2, di.max_rx_queues, di.max_tx_queues);
- s = format(s, "%Upromiscuous: unicast %s all-multicast %s\n",
- format_white_space, indent + 2,
- rte_eth_promiscuous_get(xd->device_index) ? "on" : "off",
- rte_eth_promiscuous_get(xd->device_index) ? "on" : "off");
- vlan_off = rte_eth_dev_get_vlan_offload(xd->device_index);
- s = format(s, "%Uvlan offload: strip %s filter %s qinq %s\n",
- format_white_space, indent + 2,
- vlan_off & ETH_VLAN_STRIP_OFFLOAD ? "on" : "off",
- vlan_off & ETH_VLAN_FILTER_OFFLOAD ? "on" : "off",
- vlan_off & ETH_VLAN_EXTEND_OFFLOAD ? "on" : "off");
- s = format(s, "%Urx offload caps: %U\n",
- format_white_space, indent + 2,
- format_dpdk_rx_offload_caps, di.rx_offload_capa);
- s = format(s, "%Utx offload caps: %U\n",
- format_white_space, indent + 2,
- format_dpdk_tx_offload_caps, di.tx_offload_capa);
- s = format(s, "%Urss active: %U\n"
- "%Urss supported: %U\n",
- format_white_space, indent + 2,
- format_dpdk_rss_hf_name, rss_conf.rss_hf,
- format_white_space, indent + 2,
- format_dpdk_rss_hf_name, di.flow_type_rss_offloads);
+ s =
+ format (s,
+ "%Upci id: device %04x:%04x subsystem %04x:%04x\n"
+ "%Upci address: %04x:%02x:%02x.%02x\n",
+ format_white_space, indent + 2, pci->id.vendor_id,
+ pci->id.device_id, pci->id.subsystem_vendor_id,
+ pci->id.subsystem_device_id, format_white_space, indent + 2,
+ pci->addr.domain, pci->addr.bus, pci->addr.devid,
+ pci->addr.function);
+ s =
+ format (s, "%Umax rx packet len: %d\n", format_white_space,
+ indent + 2, di.max_rx_pktlen);
+ s =
+ format (s, "%Umax num of queues: rx %d tx %d\n", format_white_space,
+ indent + 2, di.max_rx_queues, di.max_tx_queues);
+ s =
+ format (s, "%Upromiscuous: unicast %s all-multicast %s\n",
+ format_white_space, indent + 2,
+ rte_eth_promiscuous_get (xd->device_index) ? "on" : "off",
+ rte_eth_promiscuous_get (xd->device_index) ? "on" : "off");
+ vlan_off = rte_eth_dev_get_vlan_offload (xd->device_index);
+ s = format (s, "%Uvlan offload: strip %s filter %s qinq %s\n",
+ format_white_space, indent + 2,
+ vlan_off & ETH_VLAN_STRIP_OFFLOAD ? "on" : "off",
+ vlan_off & ETH_VLAN_FILTER_OFFLOAD ? "on" : "off",
+ vlan_off & ETH_VLAN_EXTEND_OFFLOAD ? "on" : "off");
+ s = format (s, "%Urx offload caps: %U\n",
+ format_white_space, indent + 2,
+ format_dpdk_rx_offload_caps, di.rx_offload_capa);
+ s = format (s, "%Utx offload caps: %U\n",
+ format_white_space, indent + 2,
+ format_dpdk_tx_offload_caps, di.tx_offload_capa);
+ s = format (s, "%Urss active: %U\n"
+ "%Urss supported: %U\n",
+ format_white_space, indent + 2,
+ format_dpdk_rss_hf_name, rss_conf.rss_hf,
+ format_white_space, indent + 2,
+ format_dpdk_rss_hf_name, di.flow_type_rss_offloads);
}
- if (verbose && xd->dev_type == VNET_DPDK_DEV_VHOST_USER) {
- s = format(s, "%Uqueue size (max): rx %d (%d) tx %d (%d)\n",
- format_white_space, indent + 2,
- xd->rx_q_used, xd->rx_q_used,
- xd->tx_q_used, xd->tx_q_used);
+ if (verbose && xd->dev_type == VNET_DPDK_DEV_VHOST_USER)
+ {
+ s = format (s, "%Uqueue size (max): rx %d (%d) tx %d (%d)\n",
+ format_white_space, indent + 2,
+ xd->rx_q_used, xd->rx_q_used, xd->tx_q_used, xd->tx_q_used);
}
s = format (s, "%Urx queues %d, rx desc %d, tx queues %d, tx desc %d\n",
- format_white_space, indent + 2,
- xd->rx_q_used, xd->nb_rx_desc,
- xd->tx_q_used, xd->nb_tx_desc);
+ format_white_space, indent + 2,
+ xd->rx_q_used, xd->nb_rx_desc, xd->tx_q_used, xd->nb_tx_desc);
if (xd->cpu_socket > -1)
s = format (s, "%Ucpu socket %d\n",
- format_white_space, indent + 2, xd->cpu_socket);
+ format_white_space, indent + 2, xd->cpu_socket);
/* $$$ MIB counters */
{
@@ -476,20 +484,21 @@ u8 * format_dpdk_device (u8 * s, va_list * args)
#undef _
}
- u8 * xs = 0;
+ u8 *xs = 0;
u32 i = 0;
#if RTE_VERSION < RTE_VERSION_NUM(16, 7, 0, 0)
- struct rte_eth_xstats * xstat, * last_xstat;
+ struct rte_eth_xstats *xstat, *last_xstat;
#else
- struct rte_eth_xstat * xstat, * last_xstat;
- struct rte_eth_xstat_name * xstat_names = 0;
+ struct rte_eth_xstat *xstat, *last_xstat;
+ struct rte_eth_xstat_name *xstat_names = 0;
int len = rte_eth_xstats_get_names (xd->device_index, NULL, 0);
vec_validate (xstat_names, len - 1);
rte_eth_xstats_get_names (xd->device_index, xstat_names, len);
#endif
- ASSERT(vec_len(xd->xstats) == vec_len(xd->last_cleared_xstats));
+ ASSERT (vec_len (xd->xstats) == vec_len (xd->last_cleared_xstats));
+ /* *INDENT-OFF* */
vec_foreach_index(i, xd->xstats)
{
u64 delta = 0;
@@ -511,70 +520,81 @@ u8 * format_dpdk_device (u8 * s, va_list * args)
vec_free(name);
}
}
+ /* *INDENT-ON* */
#if RTE_VERSION >= RTE_VERSION_NUM(16, 7, 0, 0)
vec_free (xstat_names);
#endif
#if DPDK_VHOST_USER
- if (verbose && xd->dev_type == VNET_DPDK_DEV_VHOST_USER) {
- int i;
- for (i = 0; i < xd->rx_q_used * VIRTIO_QNUM; i++) {
- u8 * name;
- if (verbose == 2 || xd->vu_intf->vrings[i].packets) {
- if (i & 1) {
- name = format(NULL, "tx q%d packets", i >> 1);
- } else {
- name = format(NULL, "rx q%d packets", i >> 1);
- }
- xs = format(xs, "\n%U%-38U%16Ld",
- format_white_space, indent + 4,
- format_c_identifier, name, xd->vu_intf->vrings[i].packets);
- vec_free(name);
-
- if (i & 1) {
- name = format(NULL, "tx q%d bytes", i >> 1);
- } else {
- name = format(NULL, "rx q%d bytes", i >> 1);
- }
- xs = format(xs, "\n%U%-38U%16Ld",
- format_white_space, indent + 4,
- format_c_identifier, name, xd->vu_intf->vrings[i].bytes);
- vec_free(name);
- }
- }
+ if (verbose && xd->dev_type == VNET_DPDK_DEV_VHOST_USER)
+ {
+ int i;
+ for (i = 0; i < xd->rx_q_used * VIRTIO_QNUM; i++)
+ {
+ u8 *name;
+ if (verbose == 2 || xd->vu_intf->vrings[i].packets)
+ {
+ if (i & 1)
+ {
+ name = format (NULL, "tx q%d packets", i >> 1);
+ }
+ else
+ {
+ name = format (NULL, "rx q%d packets", i >> 1);
+ }
+ xs = format (xs, "\n%U%-38U%16Ld",
+ format_white_space, indent + 4,
+ format_c_identifier, name,
+ xd->vu_intf->vrings[i].packets);
+ vec_free (name);
+
+ if (i & 1)
+ {
+ name = format (NULL, "tx q%d bytes", i >> 1);
+ }
+ else
+ {
+ name = format (NULL, "rx q%d bytes", i >> 1);
+ }
+ xs = format (xs, "\n%U%-38U%16Ld",
+ format_white_space, indent + 4,
+ format_c_identifier, name,
+ xd->vu_intf->vrings[i].bytes);
+ vec_free (name);
+ }
+ }
}
#endif
if (xs)
{
- s = format(s, "\n%Uextended stats:%v",
- format_white_space, indent + 2, xs);
- vec_free(xs);
+ s = format (s, "\n%Uextended stats:%v",
+ format_white_space, indent + 2, xs);
+ vec_free (xs);
}
return s;
}
-u8 * format_dpdk_tx_dma_trace (u8 * s, va_list * va)
+u8 *
+format_dpdk_tx_dma_trace (u8 * s, va_list * va)
{
CLIB_UNUSED (vlib_main_t * vm) = va_arg (*va, vlib_main_t *);
CLIB_UNUSED (vlib_node_t * node) = va_arg (*va, vlib_node_t *);
- CLIB_UNUSED (vnet_main_t * vnm) = vnet_get_main();
- dpdk_tx_dma_trace_t * t = va_arg (*va, dpdk_tx_dma_trace_t *);
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_t * xd = vec_elt_at_index (dm->devices, t->device_index);
+ CLIB_UNUSED (vnet_main_t * vnm) = vnet_get_main ();
+ dpdk_tx_dma_trace_t *t = va_arg (*va, dpdk_tx_dma_trace_t *);
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_t *xd = vec_elt_at_index (dm->devices, t->device_index);
uword indent = format_get_indent (s);
- vnet_sw_interface_t * sw = vnet_get_sw_interface (vnm, xd->vlib_sw_if_index);
+ vnet_sw_interface_t *sw = vnet_get_sw_interface (vnm, xd->vlib_sw_if_index);
s = format (s, "%U tx queue %d",
- format_vnet_sw_interface_name, vnm, sw,
- t->queue_index);
+ format_vnet_sw_interface_name, vnm, sw, t->queue_index);
s = format (s, "\n%Ubuffer 0x%x: %U",
format_white_space, indent,
- t->buffer_index,
- format_vlib_buffer, &t->buffer);
+ t->buffer_index, format_vlib_buffer, &t->buffer);
s = format (s, "\n%U%U", format_white_space, indent,
format_ethernet_header_with_length, t->buffer.pre_data,
@@ -583,26 +603,25 @@ u8 * format_dpdk_tx_dma_trace (u8 * s, va_list * va)
return s;
}
-u8 * format_dpdk_rx_dma_trace (u8 * s, va_list * va)
+u8 *
+format_dpdk_rx_dma_trace (u8 * s, va_list * va)
{
CLIB_UNUSED (vlib_main_t * vm) = va_arg (*va, vlib_main_t *);
CLIB_UNUSED (vlib_node_t * node) = va_arg (*va, vlib_node_t *);
- CLIB_UNUSED (vnet_main_t * vnm) = vnet_get_main();
- dpdk_rx_dma_trace_t * t = va_arg (*va, dpdk_rx_dma_trace_t *);
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_t * xd = vec_elt_at_index (dm->devices, t->device_index);
- format_function_t * f;
+ CLIB_UNUSED (vnet_main_t * vnm) = vnet_get_main ();
+ dpdk_rx_dma_trace_t *t = va_arg (*va, dpdk_rx_dma_trace_t *);
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_t *xd = vec_elt_at_index (dm->devices, t->device_index);
+ format_function_t *f;
uword indent = format_get_indent (s);
- vnet_sw_interface_t * sw = vnet_get_sw_interface (vnm, xd->vlib_sw_if_index);
+ vnet_sw_interface_t *sw = vnet_get_sw_interface (vnm, xd->vlib_sw_if_index);
s = format (s, "%U rx queue %d",
- format_vnet_sw_interface_name, vnm, sw,
- t->queue_index);
+ format_vnet_sw_interface_name, vnm, sw, t->queue_index);
s = format (s, "\n%Ubuffer 0x%x: %U",
format_white_space, indent,
- t->buffer_index,
- format_vlib_buffer, &t->buffer);
+ t->buffer_index, format_vlib_buffer, &t->buffer);
#ifdef RTE_LIBRTE_MBUF_EXT_RX_OLFLAGS
s = format (s, "\n%U%U",
@@ -616,10 +635,11 @@ u8 * format_dpdk_rx_dma_trace (u8 * s, va_list * va)
if (vm->trace_main.verbose)
{
s = format (s, "\n%UPacket Dump%s", format_white_space, indent + 2,
- t->mb.data_len > sizeof(t->data) ? " (truncated)": "");
+ t->mb.data_len > sizeof (t->data) ? " (truncated)" : "");
s = format (s, "\n%U%U", format_white_space, indent + 4,
format_hexdump, &t->data,
- t->mb.data_len > sizeof(t->data) ? sizeof(t->data) : t->mb.data_len);
+ t->mb.data_len >
+ sizeof (t->data) ? sizeof (t->data) : t->mb.data_len);
}
f = node->format_buffer;
if (!f)
@@ -631,10 +651,11 @@ u8 * format_dpdk_rx_dma_trace (u8 * s, va_list * va)
}
-static inline u8 * format_dpdk_pkt_types (u8 * s, va_list * va)
+static inline u8 *
+format_dpdk_pkt_types (u8 * s, va_list * va)
{
u32 *pkt_types = va_arg (*va, u32 *);
- uword indent __attribute__((unused)) = format_get_indent (s) + 2;
+ uword indent __attribute__ ((unused)) = format_get_indent (s) + 2;
if (!*pkt_types)
return s;
@@ -649,13 +670,12 @@ static inline u8 * format_dpdk_pkt_types (u8 * s, va_list * va)
}
foreach_dpdk_pkt_type
-
#undef _
-
- return s;
+ return s;
}
-static inline u8 * format_dpdk_pkt_offload_flags (u8 * s, va_list * va)
+static inline u8 *
+format_dpdk_pkt_offload_flags (u8 * s, va_list * va)
{
u64 *ol_flags = va_arg (*va, u64 *);
uword indent = format_get_indent (s) + 2;
@@ -673,64 +693,68 @@ static inline u8 * format_dpdk_pkt_offload_flags (u8 * s, va_list * va)
}
foreach_dpdk_pkt_offload_flag
-
#undef _
-
- return s;
+ return s;
}
-u8 * format_dpdk_rte_mbuf_vlan (u8 * s, va_list * va)
+u8 *
+format_dpdk_rte_mbuf_vlan (u8 * s, va_list * va)
{
- ethernet_vlan_header_tv_t * vlan_hdr = va_arg (*va, ethernet_vlan_header_tv_t *);
+ ethernet_vlan_header_tv_t *vlan_hdr =
+ va_arg (*va, ethernet_vlan_header_tv_t *);
- if (clib_net_to_host_u16(vlan_hdr->type) == ETHERNET_TYPE_DOT1AD) {
- s = format (s, "%U 802.1q vlan ",
- format_ethernet_vlan_tci,
- clib_net_to_host_u16(vlan_hdr->priority_cfi_and_id));
- vlan_hdr++;
+ if (clib_net_to_host_u16 (vlan_hdr->type) == ETHERNET_TYPE_DOT1AD)
+ {
+ s = format (s, "%U 802.1q vlan ",
+ format_ethernet_vlan_tci,
+ clib_net_to_host_u16 (vlan_hdr->priority_cfi_and_id));
+ vlan_hdr++;
}
- s = format (s, "%U",
- format_ethernet_vlan_tci,
- clib_net_to_host_u16(vlan_hdr->priority_cfi_and_id));
+ s = format (s, "%U",
+ format_ethernet_vlan_tci,
+ clib_net_to_host_u16 (vlan_hdr->priority_cfi_and_id));
- return s;
+ return s;
}
-u8 * format_dpdk_rte_mbuf (u8 * s, va_list * va)
+u8 *
+format_dpdk_rte_mbuf (u8 * s, va_list * va)
{
- struct rte_mbuf * mb = va_arg (*va, struct rte_mbuf *);
+ struct rte_mbuf *mb = va_arg (*va, struct rte_mbuf *);
ethernet_header_t *eth_hdr = va_arg (*va, ethernet_header_t *);
uword indent = format_get_indent (s) + 2;
s = format (s, "PKT MBUF: port %d, nb_segs %d, pkt_len %d"
- "\n%Ubuf_len %d, data_len %d, ol_flags 0x%x, data_off %d, phys_addr 0x%x"
- "\n%Upacket_type 0x%x",
- mb->port, mb->nb_segs, mb->pkt_len,
- format_white_space, indent,
- mb->buf_len, mb->data_len, mb->ol_flags, mb->data_off, mb->buf_physaddr,
- format_white_space, indent,
- mb->packet_type);
+ "\n%Ubuf_len %d, data_len %d, ol_flags 0x%x, data_off %d, phys_addr 0x%x"
+ "\n%Upacket_type 0x%x",
+ mb->port, mb->nb_segs, mb->pkt_len,
+ format_white_space, indent,
+ mb->buf_len, mb->data_len, mb->ol_flags, mb->data_off,
+ mb->buf_physaddr, format_white_space, indent, mb->packet_type);
if (mb->ol_flags)
s = format (s, "\n%U%U", format_white_space, indent,
- format_dpdk_pkt_offload_flags, &mb->ol_flags);
+ format_dpdk_pkt_offload_flags, &mb->ol_flags);
- if (mb->ol_flags & PKT_RX_VLAN_PKT) {
- ethernet_vlan_header_tv_t * vlan_hdr = ((ethernet_vlan_header_tv_t *)&(eth_hdr->type));
- s = format (s, " %U", format_dpdk_rte_mbuf_vlan, vlan_hdr);
- }
+ if (mb->ol_flags & PKT_RX_VLAN_PKT)
+ {
+ ethernet_vlan_header_tv_t *vlan_hdr =
+ ((ethernet_vlan_header_tv_t *) & (eth_hdr->type));
+ s = format (s, " %U", format_dpdk_rte_mbuf_vlan, vlan_hdr);
+ }
if (mb->packet_type)
s = format (s, "\n%U%U", format_white_space, indent,
- format_dpdk_pkt_types, &mb->packet_type);
+ format_dpdk_pkt_types, &mb->packet_type);
return s;
}
#ifdef RTE_LIBRTE_MBUF_EXT_RX_OLFLAGS
-static inline u8 * format_dpdk_pkt_rx_offload_flags (u8 * s, va_list * va)
+static inline u8 *
+format_dpdk_pkt_rx_offload_flags (u8 * s, va_list * va)
{
u16 *ol_flags = va_arg (*va, u16 *);
uword indent = format_get_indent (s) + 2;
@@ -748,15 +772,14 @@ static inline u8 * format_dpdk_pkt_rx_offload_flags (u8 * s, va_list * va)
}
foreach_dpdk_pkt_ext_offload_flag
-
#undef _
-
- return s;
+ return s;
}
-u8 * format_dpdk_rx_rte_mbuf (u8 * s, va_list * va)
+u8 *
+format_dpdk_rx_rte_mbuf (u8 * s, va_list * va)
{
- struct rte_mbuf * mb = va_arg (*va, struct rte_mbuf *);
+ struct rte_mbuf *mb = va_arg (*va, struct rte_mbuf *);
ethernet_header_t *eth_hdr = va_arg (*args, ethernet_header_t *);
uword indent = format_get_indent (s) + 2;
@@ -765,26 +788,27 @@ u8 * format_dpdk_rx_rte_mbuf (u8 * s, va_list * va)
* are only valid for the 1st mbuf segment.
*/
s = format (s, "PKT MBUF: port %d, nb_segs %d, pkt_len %d"
- "\n%Ubuf_len %d, data_len %d, ol_flags 0x%x"
- "\n%Upacket_type 0x%x",
- mb->port, mb->nb_segs, mb->pkt_len,
- format_white_space, indent,
- mb->buf_len, mb->data_len, mb->ol_flags,
- format_white_space, indent,
- mb->packet_type);
+ "\n%Ubuf_len %d, data_len %d, ol_flags 0x%x"
+ "\n%Upacket_type 0x%x",
+ mb->port, mb->nb_segs, mb->pkt_len,
+ format_white_space, indent,
+ mb->buf_len, mb->data_len, mb->ol_flags,
+ format_white_space, indent, mb->packet_type);
if (mb->ol_flags)
s = format (s, "\n%U%U", format_white_space, indent,
- format_dpdk_pkt_rx_offload_flags, &mb->ol_flags);
+ format_dpdk_pkt_rx_offload_flags, &mb->ol_flags);
- if (mb->ol_flags & PKT_RX_VLAN_PKT) {
- ethernet_vlan_header_tv_t * vlan_hdr = ((ethernet_vlan_header_tv_t *)&(eth_hdr->type));
- s = format (s, " %U", format_dpdk_rte_mbuf_vlan, vlan_hdr);
- }
+ if (mb->ol_flags & PKT_RX_VLAN_PKT)
+ {
+ ethernet_vlan_header_tv_t *vlan_hdr =
+ ((ethernet_vlan_header_tv_t *) & (eth_hdr->type));
+ s = format (s, " %U", format_dpdk_rte_mbuf_vlan, vlan_hdr);
+ }
if (mb->packet_type)
s = format (s, "\n%U%U", format_white_space, indent,
- format_dpdk_pkt_types, &mb->packet_type);
+ format_dpdk_pkt_types, &mb->packet_type);
return s;
}
@@ -793,23 +817,23 @@ u8 * format_dpdk_rx_rte_mbuf (u8 * s, va_list * va)
uword
unformat_socket_mem (unformat_input_t * input, va_list * va)
{
- uword ** r = va_arg (* va, uword **);
+ uword **r = va_arg (*va, uword **);
int i = 0;
u32 mem;
while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT)
{
if (unformat (input, ","))
- hash_set (*r, i, 1024);
+ hash_set (*r, i, 1024);
else if (unformat (input, "%u,", &mem))
- hash_set (*r, i, mem);
+ hash_set (*r, i, mem);
else if (unformat (input, "%u", &mem))
- hash_set (*r, i, mem);
+ hash_set (*r, i, mem);
else
- {
- unformat_put_input (input);
- goto done;
- }
+ {
+ unformat_put_input (input);
+ goto done;
+ }
i++;
}
@@ -823,7 +847,7 @@ unformat_rss_fn (unformat_input_t * input, uword * rss_fn)
while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT)
{
if (0)
- ;
+ ;
#undef _
#define _(f, s) \
else if (unformat (input, s)) \
@@ -831,12 +855,19 @@ unformat_rss_fn (unformat_input_t * input, uword * rss_fn)
foreach_dpdk_rss_hf
#undef _
-
- else
- {
- return clib_error_return (0, "unknown input `%U'",
- format_unformat_error, input);
- }
+ else
+ {
+ return clib_error_return (0, "unknown input `%U'",
+ format_unformat_error, input);
+ }
}
return 0;
}
+
+/*
+ * fd.io coding-style-patch-verification: ON
+ *
+ * Local Variables:
+ * eval: (c-set-style "gnu")
+ * End:
+ */
diff --git a/vnet/vnet/devices/dpdk/init.c b/vnet/vnet/devices/dpdk/init.c
index 239ae3bbed6..705c0371b97 100644
--- a/vnet/vnet/devices/dpdk/init.c
+++ b/vnet/vnet/devices/dpdk/init.c
@@ -50,110 +50,115 @@ void *vlib_weakly_linked_functions[] = {
static struct rte_eth_conf port_conf_template = {
.rxmode = {
- .split_hdr_size = 0,
- .header_split = 0, /**< Header Split disabled */
- .hw_ip_checksum = 0, /**< IP checksum offload disabled */
- .hw_vlan_filter = 0, /**< VLAN filtering disabled */
- .hw_strip_crc = 0, /**< CRC stripped by hardware */
- },
+ .split_hdr_size = 0,
+ .header_split = 0, /**< Header Split disabled */
+ .hw_ip_checksum = 0, /**< IP checksum offload disabled */
+ .hw_vlan_filter = 0, /**< VLAN filtering disabled */
+ .hw_strip_crc = 0, /**< CRC stripped by hardware */
+ },
.txmode = {
- .mq_mode = ETH_MQ_TX_NONE,
- },
+ .mq_mode = ETH_MQ_TX_NONE,
+ },
};
clib_error_t *
dpdk_port_setup (dpdk_main_t * dm, dpdk_device_t * xd)
{
- vlib_main_t * vm = vlib_get_main();
- vlib_buffer_main_t * bm = vm->buffer_main;
+ vlib_main_t *vm = vlib_get_main ();
+ vlib_buffer_main_t *bm = vm->buffer_main;
int rv;
int j;
- ASSERT(os_get_cpu_number() == 0);
+ ASSERT (os_get_cpu_number () == 0);
- if (xd->admin_up) {
- vnet_hw_interface_set_flags (dm->vnet_main, xd->vlib_hw_if_index, 0);
- rte_eth_dev_stop (xd->device_index);
- }
+ if (xd->admin_up)
+ {
+ vnet_hw_interface_set_flags (dm->vnet_main, xd->vlib_hw_if_index, 0);
+ rte_eth_dev_stop (xd->device_index);
+ }
rv = rte_eth_dev_configure (xd->device_index, xd->rx_q_used,
- xd->tx_q_used, &xd->port_conf);
+ xd->tx_q_used, &xd->port_conf);
if (rv < 0)
return clib_error_return (0, "rte_eth_dev_configure[%d]: err %d",
- xd->device_index, rv);
+ xd->device_index, rv);
/* Set up one TX-queue per worker thread */
for (j = 0; j < xd->tx_q_used; j++)
{
- rv = rte_eth_tx_queue_setup(xd->device_index, j, xd->nb_tx_desc,
- xd->cpu_socket, &xd->tx_conf);
+ rv = rte_eth_tx_queue_setup (xd->device_index, j, xd->nb_tx_desc,
+ xd->cpu_socket, &xd->tx_conf);
/* retry with any other CPU socket */
if (rv < 0)
- rv = rte_eth_tx_queue_setup(xd->device_index, j, xd->nb_tx_desc,
- SOCKET_ID_ANY, &xd->tx_conf);
+ rv = rte_eth_tx_queue_setup (xd->device_index, j, xd->nb_tx_desc,
+ SOCKET_ID_ANY, &xd->tx_conf);
if (rv < 0)
- break;
+ break;
}
- if (rv < 0)
- return clib_error_return (0, "rte_eth_tx_queue_setup[%d]: err %d",
- xd->device_index, rv);
+ if (rv < 0)
+ return clib_error_return (0, "rte_eth_tx_queue_setup[%d]: err %d",
+ xd->device_index, rv);
for (j = 0; j < xd->rx_q_used; j++)
{
- rv = rte_eth_rx_queue_setup(xd->device_index, j, xd->nb_rx_desc,
- xd->cpu_socket, 0,
- bm->pktmbuf_pools[xd->cpu_socket_id_by_queue[j]]);
+ rv = rte_eth_rx_queue_setup (xd->device_index, j, xd->nb_rx_desc,
+ xd->cpu_socket, 0,
+ bm->
+ pktmbuf_pools[xd->cpu_socket_id_by_queue
+ [j]]);
/* retry with any other CPU socket */
if (rv < 0)
- rv = rte_eth_rx_queue_setup(xd->device_index, j, xd->nb_rx_desc,
- SOCKET_ID_ANY, 0,
- bm->pktmbuf_pools[xd->cpu_socket_id_by_queue[j]]);
+ rv = rte_eth_rx_queue_setup (xd->device_index, j, xd->nb_rx_desc,
+ SOCKET_ID_ANY, 0,
+ bm->
+ pktmbuf_pools[xd->cpu_socket_id_by_queue
+ [j]]);
if (rv < 0)
- return clib_error_return (0, "rte_eth_rx_queue_setup[%d]: err %d",
- xd->device_index, rv);
+ return clib_error_return (0, "rte_eth_rx_queue_setup[%d]: err %d",
+ xd->device_index, rv);
}
- if (xd->admin_up) {
+ if (xd->admin_up)
+ {
int rv;
rv = rte_eth_dev_start (xd->device_index);
if (rv < 0)
- clib_warning ("rte_eth_dev_start %d returned %d",
- xd->device_index, rv);
- }
+ clib_warning ("rte_eth_dev_start %d returned %d",
+ xd->device_index, rv);
+ }
return 0;
}
-static u32 dpdk_flag_change (vnet_main_t * vnm,
- vnet_hw_interface_t * hi,
- u32 flags)
+static u32
+dpdk_flag_change (vnet_main_t * vnm, vnet_hw_interface_t * hi, u32 flags)
{
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_t * xd = vec_elt_at_index (dm->devices, hi->dev_instance);
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_t *xd = vec_elt_at_index (dm->devices, hi->dev_instance);
u32 old = 0;
- if (ETHERNET_INTERFACE_FLAG_CONFIG_PROMISC(flags))
+ if (ETHERNET_INTERFACE_FLAG_CONFIG_PROMISC (flags))
{
old = xd->promisc;
xd->promisc = flags & ETHERNET_INTERFACE_FLAG_ACCEPT_ALL;
-
+
if (xd->admin_up)
{
if (xd->promisc)
- rte_eth_promiscuous_enable(xd->device_index);
+ rte_eth_promiscuous_enable (xd->device_index);
else
- rte_eth_promiscuous_disable(xd->device_index);
+ rte_eth_promiscuous_disable (xd->device_index);
}
}
- else if (ETHERNET_INTERFACE_FLAG_CONFIG_MTU(flags))
+ else if (ETHERNET_INTERFACE_FLAG_CONFIG_MTU (flags))
{
/*
* DAW-FIXME: The Cisco VIC firmware does not provide an api for a
- * driver to dynamically change the mtu. If/when the
+ * driver to dynamically change the mtu. If/when the
* VIC firmware gets fixed, then this should be removed.
*/
if (xd->pmd == VNET_DPDK_PMD_ENIC)
@@ -163,69 +168,66 @@ static u32 dpdk_flag_change (vnet_main_t * vnm,
/*
* Restore mtu to what has been set by CIMC in the firmware cfg.
*/
- rte_eth_dev_info_get(xd->device_index, &dev_info);
+ rte_eth_dev_info_get (xd->device_index, &dev_info);
hi->max_packet_bytes = dev_info.max_rx_pktlen;
- vlib_cli_output (vlib_get_main(),
+ vlib_cli_output (vlib_get_main (),
"Cisco VIC mtu can only be changed "
"using CIMC then rebooting the server!");
}
else
{
int rv;
-
+
xd->port_conf.rxmode.max_rx_pkt_len = hi->max_packet_bytes;
if (xd->admin_up)
rte_eth_dev_stop (xd->device_index);
- rv = rte_eth_dev_configure
- (xd->device_index,
- xd->rx_q_used,
- xd->tx_q_used,
- &xd->port_conf);
+ rv = rte_eth_dev_configure
+ (xd->device_index, xd->rx_q_used, xd->tx_q_used, &xd->port_conf);
if (rv < 0)
- vlib_cli_output (vlib_get_main(),
+ vlib_cli_output (vlib_get_main (),
"rte_eth_dev_configure[%d]: err %d",
xd->device_index, rv);
- rte_eth_dev_set_mtu(xd->device_index, hi->max_packet_bytes);
+ rte_eth_dev_set_mtu (xd->device_index, hi->max_packet_bytes);
if (xd->admin_up)
- {
- int rv = rte_eth_dev_start (xd->device_index);
- if (rv < 0)
- clib_warning ("rte_eth_dev_start %d returned %d",
- xd->device_index, rv);
- }
+ {
+ int rv = rte_eth_dev_start (xd->device_index);
+ if (rv < 0)
+ clib_warning ("rte_eth_dev_start %d returned %d",
+ xd->device_index, rv);
+ }
}
}
return old;
}
void
-dpdk_device_lock_init(dpdk_device_t * xd)
+dpdk_device_lock_init (dpdk_device_t * xd)
{
int q;
- vec_validate(xd->lockp, xd->tx_q_used - 1);
+ vec_validate (xd->lockp, xd->tx_q_used - 1);
for (q = 0; q < xd->tx_q_used; q++)
{
xd->lockp[q] = clib_mem_alloc_aligned (CLIB_CACHE_LINE_BYTES,
- CLIB_CACHE_LINE_BYTES);
+ CLIB_CACHE_LINE_BYTES);
memset ((void *) xd->lockp[q], 0, CLIB_CACHE_LINE_BYTES);
- }
+ }
xd->need_txlock = 1;
}
void
-dpdk_device_lock_free(dpdk_device_t * xd)
+dpdk_device_lock_free (dpdk_device_t * xd)
{
int q;
- for (q = 0; q < vec_len(xd->lockp); q++)
- clib_mem_free((void *) xd->lockp[q]);
- vec_free(xd->lockp);
+ for (q = 0; q < vec_len (xd->lockp); q++)
+ clib_mem_free ((void *) xd->lockp[q]);
+ vec_free (xd->lockp);
xd->lockp = 0;
xd->need_txlock = 0;
}
@@ -236,17 +238,17 @@ dpdk_lib_init (dpdk_main_t * dm)
u32 nports;
u32 nb_desc = 0;
int i;
- clib_error_t * error;
- vlib_main_t * vm = vlib_get_main();
- vlib_thread_main_t * tm = vlib_get_thread_main();
- vlib_node_runtime_t * rt;
- vnet_sw_interface_t * sw;
- vnet_hw_interface_t * hi;
- dpdk_device_t * xd;
+ clib_error_t *error;
+ vlib_main_t *vm = vlib_get_main ();
+ vlib_thread_main_t *tm = vlib_get_thread_main ();
+ vlib_node_runtime_t *rt;
+ vnet_sw_interface_t *sw;
+ vnet_hw_interface_t *hi;
+ dpdk_device_t *xd;
vlib_pci_addr_t last_pci_addr;
u32 last_pci_addr_port = 0;
- vlib_thread_registration_t * tr;
- uword * p;
+ vlib_thread_registration_t *tr;
+ uword *p;
u32 next_cpu = 0;
u8 af_packet_port_id = 0;
@@ -256,7 +258,7 @@ dpdk_lib_init (dpdk_main_t * dm)
dm->input_cpu_count = 1;
rt = vlib_node_get_runtime (vm, dpdk_input_node.index);
- rt->function = dpdk_input_multiarch_select();
+ rt->function = dpdk_input_multiarch_select ();
/* find out which cpus will be used for input */
p = hash_get_mem (tm->thread_registrations_by_name, "workers");
@@ -269,13 +271,13 @@ dpdk_lib_init (dpdk_main_t * dm)
}
vec_validate_aligned (dm->devices_by_cpu, tm->n_vlib_mains - 1,
- CLIB_CACHE_LINE_BYTES);
+ CLIB_CACHE_LINE_BYTES);
vec_validate_aligned (dm->workers, tm->n_vlib_mains - 1,
- CLIB_CACHE_LINE_BYTES);
+ CLIB_CACHE_LINE_BYTES);
- nports = rte_eth_dev_count();
- if (nports < 1)
+ nports = rte_eth_dev_count ();
+ if (nports < 1)
{
clib_warning ("DPDK drivers found no ports...");
}
@@ -283,13 +285,14 @@ dpdk_lib_init (dpdk_main_t * dm)
if (CLIB_DEBUG > 0)
clib_warning ("DPDK drivers found %d ports...", nports);
- /*
+ /*
* All buffers are all allocated from the same rte_mempool.
* Thus they all have the same number of data bytes.
*/
- dm->vlib_buffer_free_list_index =
- vlib_buffer_get_or_create_free_list (
- vm, VLIB_BUFFER_DEFAULT_FREE_LIST_BYTES, "dpdk rx");
+ dm->vlib_buffer_free_list_index =
+ vlib_buffer_get_or_create_free_list (vm,
+ VLIB_BUFFER_DEFAULT_FREE_LIST_BYTES,
+ "dpdk rx");
if (dm->conf->enable_tcp_udp_checksum)
dm->buffer_flags_template &= ~(IP_BUFFER_L4_CHECKSUM_CORRECT
@@ -301,21 +304,23 @@ dpdk_lib_init (dpdk_main_t * dm)
u8 vlan_strip = 0;
int j;
struct rte_eth_dev_info dev_info;
- clib_error_t * rv;
+ clib_error_t *rv;
struct rte_eth_link l;
- dpdk_device_config_t * devconf = 0;
+ dpdk_device_config_t *devconf = 0;
vlib_pci_addr_t pci_addr;
- uword * p = 0;
+ uword *p = 0;
- rte_eth_dev_info_get(i, &dev_info);
- if (dev_info.pci_dev) /* bonded interface has no pci info */
- {
+ rte_eth_dev_info_get (i, &dev_info);
+ if (dev_info.pci_dev) /* bonded interface has no pci info */
+ {
pci_addr.domain = dev_info.pci_dev->addr.domain;
pci_addr.bus = dev_info.pci_dev->addr.bus;
pci_addr.slot = dev_info.pci_dev->addr.devid;
pci_addr.function = dev_info.pci_dev->addr.function;
- p = hash_get (dm->conf->device_config_index_by_pci_addr, pci_addr.as_u32);
- }
+ p =
+ hash_get (dm->conf->device_config_index_by_pci_addr,
+ pci_addr.as_u32);
+ }
if (p)
devconf = pool_elt_at_index (dm->conf->dev_confs, p[0]);
@@ -326,15 +331,16 @@ dpdk_lib_init (dpdk_main_t * dm)
vec_add2_aligned (dm->devices, xd, 1, CLIB_CACHE_LINE_BYTES);
xd->nb_rx_desc = DPDK_NB_RX_DESC_DEFAULT;
xd->nb_tx_desc = DPDK_NB_TX_DESC_DEFAULT;
- xd->cpu_socket = (i8) rte_eth_dev_socket_id(i);
+ xd->cpu_socket = (i8) rte_eth_dev_socket_id (i);
/* Handle interface naming for devices with multiple ports sharing same PCI ID */
if (dev_info.pci_dev)
{
- struct rte_eth_dev_info di = {0};
+ struct rte_eth_dev_info di = { 0 };
rte_eth_dev_info_get (i + 1, &di);
if (di.pci_dev && pci_addr.as_u32 != last_pci_addr.as_u32 &&
- memcmp(&dev_info.pci_dev->addr, &di.pci_dev->addr, sizeof(struct rte_pci_addr)) == 0)
+ memcmp (&dev_info.pci_dev->addr, &di.pci_dev->addr,
+ sizeof (struct rte_pci_addr)) == 0)
{
xd->interface_name_suffix = format (0, "0");
last_pci_addr.as_u32 = pci_addr.as_u32;
@@ -342,7 +348,8 @@ dpdk_lib_init (dpdk_main_t * dm)
}
else if (pci_addr.as_u32 == last_pci_addr.as_u32)
{
- xd->interface_name_suffix = format (0, "%u", i - last_pci_addr_port);
+ xd->interface_name_suffix =
+ format (0, "%u", i - last_pci_addr_port);
}
else
{
@@ -352,187 +359,195 @@ dpdk_lib_init (dpdk_main_t * dm)
else
last_pci_addr.as_u32 = ~0;
- clib_memcpy(&xd->tx_conf, &dev_info.default_txconf,
- sizeof(struct rte_eth_txconf));
+ clib_memcpy (&xd->tx_conf, &dev_info.default_txconf,
+ sizeof (struct rte_eth_txconf));
if (dm->conf->no_multi_seg)
- {
- xd->tx_conf.txq_flags |= ETH_TXQ_FLAGS_NOMULTSEGS;
- port_conf_template.rxmode.jumbo_frame = 0;
- }
+ {
+ xd->tx_conf.txq_flags |= ETH_TXQ_FLAGS_NOMULTSEGS;
+ port_conf_template.rxmode.jumbo_frame = 0;
+ }
else
- {
- xd->tx_conf.txq_flags &= ~ETH_TXQ_FLAGS_NOMULTSEGS;
- port_conf_template.rxmode.jumbo_frame = 1;
- }
+ {
+ xd->tx_conf.txq_flags &= ~ETH_TXQ_FLAGS_NOMULTSEGS;
+ port_conf_template.rxmode.jumbo_frame = 1;
+ }
- clib_memcpy(&xd->port_conf, &port_conf_template, sizeof(struct rte_eth_conf));
+ clib_memcpy (&xd->port_conf, &port_conf_template,
+ sizeof (struct rte_eth_conf));
- xd->tx_q_used = clib_min(dev_info.max_tx_queues, tm->n_vlib_mains);
+ xd->tx_q_used = clib_min (dev_info.max_tx_queues, tm->n_vlib_mains);
- if (devconf->num_tx_queues > 0 && devconf->num_tx_queues < xd->tx_q_used)
- xd->tx_q_used = clib_min(xd->tx_q_used, devconf->num_tx_queues);
+ if (devconf->num_tx_queues > 0
+ && devconf->num_tx_queues < xd->tx_q_used)
+ xd->tx_q_used = clib_min (xd->tx_q_used, devconf->num_tx_queues);
if (devconf->num_rx_queues > 1 && dm->use_rss == 0)
{
- rt->function = dpdk_input_rss_multiarch_select();
+ rt->function = dpdk_input_rss_multiarch_select ();
dm->use_rss = 1;
}
- if (devconf->num_rx_queues > 1 && dev_info.max_rx_queues >= devconf->num_rx_queues)
- {
- xd->rx_q_used = devconf->num_rx_queues;
- xd->port_conf.rxmode.mq_mode = ETH_MQ_RX_RSS;
- if (devconf->rss_fn == 0)
- xd->port_conf.rx_adv_conf.rss_conf.rss_hf = ETH_RSS_IP | ETH_RSS_UDP | ETH_RSS_TCP;
- else
- xd->port_conf.rx_adv_conf.rss_conf.rss_hf = devconf->rss_fn;
- }
+ if (devconf->num_rx_queues > 1
+ && dev_info.max_rx_queues >= devconf->num_rx_queues)
+ {
+ xd->rx_q_used = devconf->num_rx_queues;
+ xd->port_conf.rxmode.mq_mode = ETH_MQ_RX_RSS;
+ if (devconf->rss_fn == 0)
+ xd->port_conf.rx_adv_conf.rss_conf.rss_hf =
+ ETH_RSS_IP | ETH_RSS_UDP | ETH_RSS_TCP;
+ else
+ xd->port_conf.rx_adv_conf.rss_conf.rss_hf = devconf->rss_fn;
+ }
else
- xd->rx_q_used = 1;
+ xd->rx_q_used = 1;
xd->dev_type = VNET_DPDK_DEV_ETH;
/* workaround for drivers not setting driver_name */
if ((!dev_info.driver_name) && (dev_info.pci_dev))
- dev_info.driver_name = dev_info.pci_dev->driver->name;
- ASSERT(dev_info.driver_name);
+ dev_info.driver_name = dev_info.pci_dev->driver->name;
+ ASSERT (dev_info.driver_name);
- if (!xd->pmd) {
+ if (!xd->pmd)
+ {
#define _(s,f) else if (dev_info.driver_name && \
!strcmp(dev_info.driver_name, s)) \
xd->pmd = VNET_DPDK_PMD_##f;
- if (0)
- ;
- foreach_dpdk_pmd
+ if (0)
+ ;
+ foreach_dpdk_pmd
#undef _
- else
- xd->pmd = VNET_DPDK_PMD_UNKNOWN;
-
-
- switch (xd->pmd) {
- /* 1G adapters */
- case VNET_DPDK_PMD_E1000EM:
- case VNET_DPDK_PMD_IGB:
- case VNET_DPDK_PMD_IGBVF:
- xd->port_type = VNET_DPDK_PORT_TYPE_ETH_1G;
- break;
-
- /* 10G adapters */
- case VNET_DPDK_PMD_IXGBE:
- case VNET_DPDK_PMD_IXGBEVF:
- case VNET_DPDK_PMD_THUNDERX:
- xd->port_type = VNET_DPDK_PORT_TYPE_ETH_10G;
- xd->nb_rx_desc = DPDK_NB_RX_DESC_10GE;
- xd->nb_tx_desc = DPDK_NB_TX_DESC_10GE;
- break;
- case VNET_DPDK_PMD_DPAA2:
- xd->port_type = VNET_DPDK_PORT_TYPE_ETH_10G;
- break;
+ else
+ xd->pmd = VNET_DPDK_PMD_UNKNOWN;
- /* Cisco VIC */
- case VNET_DPDK_PMD_ENIC:
- rte_eth_link_get_nowait(i, &l);
- xd->nb_rx_desc = DPDK_NB_RX_DESC_ENIC;
- if (l.link_speed == 40000)
- {
- xd->port_type = VNET_DPDK_PORT_TYPE_ETH_40G;
- xd->nb_tx_desc = DPDK_NB_TX_DESC_40GE;
- }
- else
- {
- xd->port_type = VNET_DPDK_PORT_TYPE_ETH_10G;
- xd->nb_tx_desc = DPDK_NB_TX_DESC_10GE;
- }
- break;
-
- /* Intel Fortville */
- case VNET_DPDK_PMD_I40E:
- case VNET_DPDK_PMD_I40EVF:
- xd->port_type = VNET_DPDK_PORT_TYPE_ETH_40G;
- xd->nb_rx_desc = DPDK_NB_RX_DESC_40GE;
- xd->nb_tx_desc = DPDK_NB_TX_DESC_40GE;
-
- switch (dev_info.pci_dev->id.device_id) {
- case I40E_DEV_ID_10G_BASE_T:
- case I40E_DEV_ID_SFP_XL710:
- xd->port_type = VNET_DPDK_PORT_TYPE_ETH_10G;
- break;
- case I40E_DEV_ID_QSFP_A:
- case I40E_DEV_ID_QSFP_B:
- case I40E_DEV_ID_QSFP_C:
- xd->port_type = VNET_DPDK_PORT_TYPE_ETH_40G;
- break;
- case I40E_DEV_ID_VF:
- rte_eth_link_get_nowait(i, &l);
- xd->port_type = l.link_speed == 10000 ?
- VNET_DPDK_PORT_TYPE_ETH_10G : VNET_DPDK_PORT_TYPE_ETH_40G;
- break;
- default:
- xd->port_type = VNET_DPDK_PORT_TYPE_UNKNOWN;
- }
- break;
-
- case VNET_DPDK_PMD_CXGBE:
- switch (dev_info.pci_dev->id.device_id) {
- case 0x540d: /* T580-CR */
- case 0x5410: /* T580-LP-cr */
- xd->nb_rx_desc = DPDK_NB_RX_DESC_40GE;
- xd->nb_tx_desc = DPDK_NB_TX_DESC_40GE;
- xd->port_type = VNET_DPDK_PORT_TYPE_ETH_40G;
- break;
- case 0x5403: /* T540-CR */
- xd->nb_rx_desc = DPDK_NB_RX_DESC_10GE;
- xd->nb_tx_desc = DPDK_NB_TX_DESC_10GE;
- xd->port_type = VNET_DPDK_PORT_TYPE_ETH_10G;
- break;
- default:
- xd->nb_rx_desc = DPDK_NB_RX_DESC_10GE;
- xd->nb_tx_desc = DPDK_NB_TX_DESC_10GE;
- xd->port_type = VNET_DPDK_PORT_TYPE_UNKNOWN;
- }
- break;
- /* Intel Red Rock Canyon */
- case VNET_DPDK_PMD_FM10K:
- xd->port_type = VNET_DPDK_PORT_TYPE_ETH_SWITCH;
- xd->nb_rx_desc = DPDK_NB_RX_DESC_40GE;
- xd->nb_tx_desc = DPDK_NB_TX_DESC_40GE;
- break;
-
- /* virtio */
- case VNET_DPDK_PMD_VIRTIO:
- xd->port_type = VNET_DPDK_PORT_TYPE_ETH_1G;
- xd->nb_rx_desc = DPDK_NB_RX_DESC_VIRTIO;
- xd->nb_tx_desc = DPDK_NB_TX_DESC_VIRTIO;
- break;
-
- /* vmxnet3 */
- case VNET_DPDK_PMD_VMXNET3:
- xd->port_type = VNET_DPDK_PORT_TYPE_ETH_1G;
- xd->tx_conf.txq_flags |= ETH_TXQ_FLAGS_NOMULTSEGS;
- break;
-
- case VNET_DPDK_PMD_AF_PACKET:
- xd->port_type = VNET_DPDK_PORT_TYPE_AF_PACKET;
- xd->af_packet_port_id = af_packet_port_id++;
- break;
-
- case VNET_DPDK_PMD_BOND:
- xd->port_type = VNET_DPDK_PORT_TYPE_ETH_BOND;
- break;
-
- default:
- xd->port_type = VNET_DPDK_PORT_TYPE_UNKNOWN;
- }
+ switch (xd->pmd)
+ {
+ /* 1G adapters */
+ case VNET_DPDK_PMD_E1000EM:
+ case VNET_DPDK_PMD_IGB:
+ case VNET_DPDK_PMD_IGBVF:
+ xd->port_type = VNET_DPDK_PORT_TYPE_ETH_1G;
+ break;
+
+ /* 10G adapters */
+ case VNET_DPDK_PMD_IXGBE:
+ case VNET_DPDK_PMD_IXGBEVF:
+ case VNET_DPDK_PMD_THUNDERX:
+ xd->port_type = VNET_DPDK_PORT_TYPE_ETH_10G;
+ xd->nb_rx_desc = DPDK_NB_RX_DESC_10GE;
+ xd->nb_tx_desc = DPDK_NB_TX_DESC_10GE;
+ break;
+ case VNET_DPDK_PMD_DPAA2:
+ xd->port_type = VNET_DPDK_PORT_TYPE_ETH_10G;
+ break;
+
+ /* Cisco VIC */
+ case VNET_DPDK_PMD_ENIC:
+ rte_eth_link_get_nowait (i, &l);
+ xd->nb_rx_desc = DPDK_NB_RX_DESC_ENIC;
+ if (l.link_speed == 40000)
+ {
+ xd->port_type = VNET_DPDK_PORT_TYPE_ETH_40G;
+ xd->nb_tx_desc = DPDK_NB_TX_DESC_40GE;
+ }
+ else
+ {
+ xd->port_type = VNET_DPDK_PORT_TYPE_ETH_10G;
+ xd->nb_tx_desc = DPDK_NB_TX_DESC_10GE;
+ }
+ break;
+
+ /* Intel Fortville */
+ case VNET_DPDK_PMD_I40E:
+ case VNET_DPDK_PMD_I40EVF:
+ xd->port_type = VNET_DPDK_PORT_TYPE_ETH_40G;
+ xd->nb_rx_desc = DPDK_NB_RX_DESC_40GE;
+ xd->nb_tx_desc = DPDK_NB_TX_DESC_40GE;
+
+ switch (dev_info.pci_dev->id.device_id)
+ {
+ case I40E_DEV_ID_10G_BASE_T:
+ case I40E_DEV_ID_SFP_XL710:
+ xd->port_type = VNET_DPDK_PORT_TYPE_ETH_10G;
+ break;
+ case I40E_DEV_ID_QSFP_A:
+ case I40E_DEV_ID_QSFP_B:
+ case I40E_DEV_ID_QSFP_C:
+ xd->port_type = VNET_DPDK_PORT_TYPE_ETH_40G;
+ break;
+ case I40E_DEV_ID_VF:
+ rte_eth_link_get_nowait (i, &l);
+ xd->port_type = l.link_speed == 10000 ?
+ VNET_DPDK_PORT_TYPE_ETH_10G : VNET_DPDK_PORT_TYPE_ETH_40G;
+ break;
+ default:
+ xd->port_type = VNET_DPDK_PORT_TYPE_UNKNOWN;
+ }
+ break;
+
+ case VNET_DPDK_PMD_CXGBE:
+ switch (dev_info.pci_dev->id.device_id)
+ {
+ case 0x540d: /* T580-CR */
+ case 0x5410: /* T580-LP-cr */
+ xd->nb_rx_desc = DPDK_NB_RX_DESC_40GE;
+ xd->nb_tx_desc = DPDK_NB_TX_DESC_40GE;
+ xd->port_type = VNET_DPDK_PORT_TYPE_ETH_40G;
+ break;
+ case 0x5403: /* T540-CR */
+ xd->nb_rx_desc = DPDK_NB_RX_DESC_10GE;
+ xd->nb_tx_desc = DPDK_NB_TX_DESC_10GE;
+ xd->port_type = VNET_DPDK_PORT_TYPE_ETH_10G;
+ break;
+ default:
+ xd->nb_rx_desc = DPDK_NB_RX_DESC_10GE;
+ xd->nb_tx_desc = DPDK_NB_TX_DESC_10GE;
+ xd->port_type = VNET_DPDK_PORT_TYPE_UNKNOWN;
+ }
+ break;
+
+ /* Intel Red Rock Canyon */
+ case VNET_DPDK_PMD_FM10K:
+ xd->port_type = VNET_DPDK_PORT_TYPE_ETH_SWITCH;
+ xd->nb_rx_desc = DPDK_NB_RX_DESC_40GE;
+ xd->nb_tx_desc = DPDK_NB_TX_DESC_40GE;
+ break;
+
+ /* virtio */
+ case VNET_DPDK_PMD_VIRTIO:
+ xd->port_type = VNET_DPDK_PORT_TYPE_ETH_1G;
+ xd->nb_rx_desc = DPDK_NB_RX_DESC_VIRTIO;
+ xd->nb_tx_desc = DPDK_NB_TX_DESC_VIRTIO;
+ break;
+
+ /* vmxnet3 */
+ case VNET_DPDK_PMD_VMXNET3:
+ xd->port_type = VNET_DPDK_PORT_TYPE_ETH_1G;
+ xd->tx_conf.txq_flags |= ETH_TXQ_FLAGS_NOMULTSEGS;
+ break;
+
+ case VNET_DPDK_PMD_AF_PACKET:
+ xd->port_type = VNET_DPDK_PORT_TYPE_AF_PACKET;
+ xd->af_packet_port_id = af_packet_port_id++;
+ break;
+
+ case VNET_DPDK_PMD_BOND:
+ xd->port_type = VNET_DPDK_PORT_TYPE_ETH_BOND;
+ break;
+
+ default:
+ xd->port_type = VNET_DPDK_PORT_TYPE_UNKNOWN;
+ }
- if (devconf->num_rx_desc)
- xd->nb_rx_desc = devconf->num_rx_desc;
+ if (devconf->num_rx_desc)
+ xd->nb_rx_desc = devconf->num_rx_desc;
- if (devconf->num_tx_desc)
- xd->nb_tx_desc = devconf->num_tx_desc;
- }
+ if (devconf->num_tx_desc)
+ xd->nb_tx_desc = devconf->num_tx_desc;
+ }
/*
* Ensure default mtu is not > the mtu read from the hardware.
@@ -540,72 +555,73 @@ dpdk_lib_init (dpdk_main_t * dm)
* not be available.
*/
if (ETHERNET_MAX_PACKET_BYTES > dev_info.max_rx_pktlen)
- {
- /*
- * This device does not support the platforms's max frame
- * size. Use it's advertised mru instead.
- */
- xd->port_conf.rxmode.max_rx_pkt_len = dev_info.max_rx_pktlen;
- }
+ {
+ /*
+ * This device does not support the platforms's max frame
+ * size. Use it's advertised mru instead.
+ */
+ xd->port_conf.rxmode.max_rx_pkt_len = dev_info.max_rx_pktlen;
+ }
else
- {
- xd->port_conf.rxmode.max_rx_pkt_len = ETHERNET_MAX_PACKET_BYTES;
-
- /*
- * Some platforms do not account for Ethernet FCS (4 bytes) in
- * MTU calculations. To interop with them increase mru but only
- * if the device's settings can support it.
- */
- if ((dev_info.max_rx_pktlen >= (ETHERNET_MAX_PACKET_BYTES + 4)) &&
- xd->port_conf.rxmode.hw_strip_crc)
- {
- /*
- * Allow additional 4 bytes (for Ethernet FCS). These bytes are
- * stripped by h/w and so will not consume any buffer memory.
- */
- xd->port_conf.rxmode.max_rx_pkt_len += 4;
- }
- }
+ {
+ xd->port_conf.rxmode.max_rx_pkt_len = ETHERNET_MAX_PACKET_BYTES;
-#if RTE_VERSION < RTE_VERSION_NUM(16, 4, 0, 0)
+ /*
+ * Some platforms do not account for Ethernet FCS (4 bytes) in
+ * MTU calculations. To interop with them increase mru but only
+ * if the device's settings can support it.
+ */
+ if ((dev_info.max_rx_pktlen >= (ETHERNET_MAX_PACKET_BYTES + 4)) &&
+ xd->port_conf.rxmode.hw_strip_crc)
+ {
+ /*
+ * Allow additional 4 bytes (for Ethernet FCS). These bytes are
+ * stripped by h/w and so will not consume any buffer memory.
+ */
+ xd->port_conf.rxmode.max_rx_pkt_len += 4;
+ }
+ }
+
+#if RTE_VERSION < RTE_VERSION_NUM(16, 4, 0, 0)
/*
* Older VMXNET3 driver doesn't support jumbo / multi-buffer pkts
*/
if (xd->pmd == VNET_DPDK_PMD_VMXNET3)
- {
- xd->port_conf.rxmode.max_rx_pkt_len = 1518;
- xd->port_conf.rxmode.jumbo_frame = 0;
- }
+ {
+ xd->port_conf.rxmode.max_rx_pkt_len = 1518;
+ xd->port_conf.rxmode.jumbo_frame = 0;
+ }
#endif
if (xd->pmd == VNET_DPDK_PMD_AF_PACKET)
- {
- f64 now = vlib_time_now(vm);
- u32 rnd;
- rnd = (u32) (now * 1e6);
- rnd = random_u32 (&rnd);
- clib_memcpy (addr+2, &rnd, sizeof(rnd));
- addr[0] = 2;
- addr[1] = 0xfe;
- }
+ {
+ f64 now = vlib_time_now (vm);
+ u32 rnd;
+ rnd = (u32) (now * 1e6);
+ rnd = random_u32 (&rnd);
+ clib_memcpy (addr + 2, &rnd, sizeof (rnd));
+ addr[0] = 2;
+ addr[1] = 0xfe;
+ }
else
- rte_eth_macaddr_get(i,(struct ether_addr *)addr);
+ rte_eth_macaddr_get (i, (struct ether_addr *) addr);
if (xd->tx_q_used < tm->n_vlib_mains)
- dpdk_device_lock_init(xd);
+ dpdk_device_lock_init (xd);
xd->device_index = xd - dm->devices;
- ASSERT(i == xd->device_index);
+ ASSERT (i == xd->device_index);
xd->per_interface_next_index = ~0;
/* assign interface to input thread */
- dpdk_device_and_queue_t * dq;
+ dpdk_device_and_queue_t *dq;
int q;
if (devconf->workers)
{
int i;
q = 0;
+ /* *INDENT-OFF* */
clib_bitmap_foreach (i, devconf->workers, ({
int cpu = dm->input_cpu_first_index + i;
unsigned lcore = vlib_worker_threads[cpu].dpdk_lcore_id;
@@ -615,6 +631,7 @@ dpdk_lib_init (dpdk_main_t * dm)
dq->device = xd->device_index;
dq->queue_id = q++;
}));
+ /* *INDENT-ON* */
}
else
for (q = 0; q < xd->rx_q_used; q++)
@@ -626,13 +643,13 @@ dpdk_lib_init (dpdk_main_t * dm)
* numa node for worker thread handling this queue
* needed for taking buffers from the right mempool
*/
- vec_validate(xd->cpu_socket_id_by_queue, q);
- xd->cpu_socket_id_by_queue[q] = rte_lcore_to_socket_id(lcore);
+ vec_validate (xd->cpu_socket_id_by_queue, q);
+ xd->cpu_socket_id_by_queue[q] = rte_lcore_to_socket_id (lcore);
/*
* construct vector of (device,queue) pairs for each worker thread
*/
- vec_add2(dm->devices_by_cpu[cpu], dq, 1);
+ vec_add2 (dm->devices_by_cpu[cpu], dq, 1);
dq->device = xd->device_index;
dq->queue_id = q;
@@ -642,48 +659,45 @@ dpdk_lib_init (dpdk_main_t * dm)
}
vec_validate_aligned (xd->tx_vectors, tm->n_vlib_mains,
- CLIB_CACHE_LINE_BYTES);
+ CLIB_CACHE_LINE_BYTES);
for (j = 0; j < tm->n_vlib_mains; j++)
- {
- vec_validate_ha (xd->tx_vectors[j], DPDK_TX_RING_SIZE,
- sizeof(tx_ring_hdr_t), CLIB_CACHE_LINE_BYTES);
- vec_reset_length (xd->tx_vectors[j]);
- }
+ {
+ vec_validate_ha (xd->tx_vectors[j], DPDK_TX_RING_SIZE,
+ sizeof (tx_ring_hdr_t), CLIB_CACHE_LINE_BYTES);
+ vec_reset_length (xd->tx_vectors[j]);
+ }
vec_validate_aligned (xd->rx_vectors, xd->rx_q_used,
- CLIB_CACHE_LINE_BYTES);
- for (j = 0; j< xd->rx_q_used; j++)
- {
- vec_validate_aligned (xd->rx_vectors[j], VLIB_FRAME_SIZE-1,
- CLIB_CACHE_LINE_BYTES);
- vec_reset_length (xd->rx_vectors[j]);
- }
+ CLIB_CACHE_LINE_BYTES);
+ for (j = 0; j < xd->rx_q_used; j++)
+ {
+ vec_validate_aligned (xd->rx_vectors[j], VLIB_FRAME_SIZE - 1,
+ CLIB_CACHE_LINE_BYTES);
+ vec_reset_length (xd->rx_vectors[j]);
+ }
- rv = dpdk_port_setup(dm, xd);
+ rv = dpdk_port_setup (dm, xd);
if (rv)
- return rv;
+ return rv;
/* count the number of descriptors used for this device */
nb_desc += xd->nb_rx_desc + xd->nb_tx_desc * xd->tx_q_used;
error = ethernet_register_interface
- (dm->vnet_main,
- dpdk_device_class.index,
- xd->device_index,
- /* ethernet address */ addr,
- &xd->vlib_hw_if_index,
- dpdk_flag_change);
+ (dm->vnet_main, dpdk_device_class.index, xd->device_index,
+ /* ethernet address */ addr,
+ &xd->vlib_hw_if_index, dpdk_flag_change);
if (error)
- return error;
-
+ return error;
+
sw = vnet_get_hw_sw_interface (dm->vnet_main, xd->vlib_hw_if_index);
xd->vlib_sw_if_index = sw->sw_if_index;
hi = vnet_get_hw_interface (dm->vnet_main, xd->vlib_hw_if_index);
/*
* DAW-FIXME: The Cisco VIC firmware does not provide an api for a
- * driver to dynamically change the mtu. If/when the
+ * driver to dynamically change the mtu. If/when the
* VIC firmware gets fixed, then this should be removed.
*/
if (xd->pmd == VNET_DPDK_PMD_ENIC)
@@ -693,117 +707,115 @@ dpdk_lib_init (dpdk_main_t * dm)
*/
hi->max_packet_bytes = dev_info.max_rx_pktlen;
if (devconf->vlan_strip_offload != DPDK_DEVICE_VLAN_STRIP_OFF)
- vlan_strip = 1; /* remove vlan tag from VIC port by default */
+ vlan_strip = 1; /* remove vlan tag from VIC port by default */
else
- clib_warning("VLAN strip disabled for interface\n");
+ clib_warning ("VLAN strip disabled for interface\n");
}
else if (devconf->vlan_strip_offload == DPDK_DEVICE_VLAN_STRIP_ON)
vlan_strip = 1;
if (vlan_strip)
- {
+ {
int vlan_off;
- vlan_off = rte_eth_dev_get_vlan_offload(xd->device_index);
+ vlan_off = rte_eth_dev_get_vlan_offload (xd->device_index);
vlan_off |= ETH_VLAN_STRIP_OFFLOAD;
- if (rte_eth_dev_set_vlan_offload(xd->device_index, vlan_off) == 0)
- clib_warning("VLAN strip enabled for interface\n");
+ if (rte_eth_dev_set_vlan_offload (xd->device_index, vlan_off) == 0)
+ clib_warning ("VLAN strip enabled for interface\n");
else
- clib_warning("VLAN strip cannot be supported by interface\n");
- }
+ clib_warning ("VLAN strip cannot be supported by interface\n");
+ }
-#if RTE_VERSION < RTE_VERSION_NUM(16, 4, 0, 0)
+#if RTE_VERSION < RTE_VERSION_NUM(16, 4, 0, 0)
/*
* Older VMXNET3 driver doesn't support jumbo / multi-buffer pkts
*/
else if (xd->pmd == VNET_DPDK_PMD_VMXNET3)
- hi->max_packet_bytes = 1518;
+ hi->max_packet_bytes = 1518;
#endif
- hi->max_l3_packet_bytes[VLIB_RX] = hi->max_l3_packet_bytes[VLIB_TX] =
- xd->port_conf.rxmode.max_rx_pkt_len - sizeof(ethernet_header_t);
+ hi->max_l3_packet_bytes[VLIB_RX] = hi->max_l3_packet_bytes[VLIB_TX] =
+ xd->port_conf.rxmode.max_rx_pkt_len - sizeof (ethernet_header_t);
- rte_eth_dev_set_mtu(xd->device_index, hi->max_packet_bytes);
+ rte_eth_dev_set_mtu (xd->device_index, hi->max_packet_bytes);
}
#ifdef RTE_LIBRTE_KNI
- if (dm->conf->num_kni) {
- clib_warning("Initializing KNI interfaces...");
- rte_kni_init(dm->conf->num_kni);
- for (i = 0; i < dm->conf->num_kni; i++)
+ if (dm->conf->num_kni)
{
- u8 addr[6];
- int j;
-
- /* Create vnet interface */
- vec_add2_aligned (dm->devices, xd, 1, CLIB_CACHE_LINE_BYTES);
- xd->dev_type = VNET_DPDK_DEV_KNI;
-
- xd->device_index = xd - dm->devices;
- ASSERT(nports + i == xd->device_index);
- xd->per_interface_next_index = ~0;
- xd->kni_port_id = i;
- xd->cpu_socket = -1;
- hash_set (dm->dpdk_device_by_kni_port_id, i, xd - dm->devices);
- xd->rx_q_used = 1;
-
- /* assign interface to input thread */
- dpdk_device_and_queue_t * dq;
- vec_add2(dm->devices_by_cpu[dm->input_cpu_first_index], dq, 1);
- dq->device = xd->device_index;
- dq->queue_id = 0;
-
- vec_validate_aligned (xd->tx_vectors, tm->n_vlib_mains,
- CLIB_CACHE_LINE_BYTES);
- for (j = 0; j < tm->n_vlib_mains; j++)
- {
- vec_validate_ha (xd->tx_vectors[j], DPDK_TX_RING_SIZE,
- sizeof(tx_ring_hdr_t), CLIB_CACHE_LINE_BYTES);
- vec_reset_length (xd->tx_vectors[j]);
- }
+ clib_warning ("Initializing KNI interfaces...");
+ rte_kni_init (dm->conf->num_kni);
+ for (i = 0; i < dm->conf->num_kni; i++)
+ {
+ u8 addr[6];
+ int j;
+
+ /* Create vnet interface */
+ vec_add2_aligned (dm->devices, xd, 1, CLIB_CACHE_LINE_BYTES);
+ xd->dev_type = VNET_DPDK_DEV_KNI;
+
+ xd->device_index = xd - dm->devices;
+ ASSERT (nports + i == xd->device_index);
+ xd->per_interface_next_index = ~0;
+ xd->kni_port_id = i;
+ xd->cpu_socket = -1;
+ hash_set (dm->dpdk_device_by_kni_port_id, i, xd - dm->devices);
+ xd->rx_q_used = 1;
+
+ /* assign interface to input thread */
+ dpdk_device_and_queue_t *dq;
+ vec_add2 (dm->devices_by_cpu[dm->input_cpu_first_index], dq, 1);
+ dq->device = xd->device_index;
+ dq->queue_id = 0;
+
+ vec_validate_aligned (xd->tx_vectors, tm->n_vlib_mains,
+ CLIB_CACHE_LINE_BYTES);
+ for (j = 0; j < tm->n_vlib_mains; j++)
+ {
+ vec_validate_ha (xd->tx_vectors[j], DPDK_TX_RING_SIZE,
+ sizeof (tx_ring_hdr_t), CLIB_CACHE_LINE_BYTES);
+ vec_reset_length (xd->tx_vectors[j]);
+ }
- vec_validate_aligned (xd->rx_vectors, xd->rx_q_used,
- CLIB_CACHE_LINE_BYTES);
- for (j = 0; j< xd->rx_q_used; j++)
- {
- vec_validate_aligned (xd->rx_vectors[j], VLIB_FRAME_SIZE-1,
- CLIB_CACHE_LINE_BYTES);
- vec_reset_length (xd->rx_vectors[j]);
- }
+ vec_validate_aligned (xd->rx_vectors, xd->rx_q_used,
+ CLIB_CACHE_LINE_BYTES);
+ for (j = 0; j < xd->rx_q_used; j++)
+ {
+ vec_validate_aligned (xd->rx_vectors[j], VLIB_FRAME_SIZE - 1,
+ CLIB_CACHE_LINE_BYTES);
+ vec_reset_length (xd->rx_vectors[j]);
+ }
- /* FIXME Set up one TX-queue per worker thread */
+ /* FIXME Set up one TX-queue per worker thread */
- {
- f64 now = vlib_time_now(vm);
- u32 rnd;
- rnd = (u32) (now * 1e6);
- rnd = random_u32 (&rnd);
-
- clib_memcpy (addr+2, &rnd, sizeof(rnd));
- addr[0] = 2;
- addr[1] = 0xfe;
- }
+ {
+ f64 now = vlib_time_now (vm);
+ u32 rnd;
+ rnd = (u32) (now * 1e6);
+ rnd = random_u32 (&rnd);
+
+ clib_memcpy (addr + 2, &rnd, sizeof (rnd));
+ addr[0] = 2;
+ addr[1] = 0xfe;
+ }
- error = ethernet_register_interface
- (dm->vnet_main,
- dpdk_device_class.index,
- xd->device_index,
- /* ethernet address */ addr,
- &xd->vlib_hw_if_index,
- dpdk_flag_change);
+ error = ethernet_register_interface
+ (dm->vnet_main, dpdk_device_class.index, xd->device_index,
+ /* ethernet address */ addr,
+ &xd->vlib_hw_if_index, dpdk_flag_change);
- if (error)
- return error;
+ if (error)
+ return error;
- sw = vnet_get_hw_sw_interface (dm->vnet_main, xd->vlib_hw_if_index);
- xd->vlib_sw_if_index = sw->sw_if_index;
- hi = vnet_get_hw_interface (dm->vnet_main, xd->vlib_hw_if_index);
+ sw = vnet_get_hw_sw_interface (dm->vnet_main, xd->vlib_hw_if_index);
+ xd->vlib_sw_if_index = sw->sw_if_index;
+ hi = vnet_get_hw_interface (dm->vnet_main, xd->vlib_hw_if_index);
+ }
}
- }
#endif
- if (nb_desc > dm->conf->num_mbufs)
+ if (nb_desc > dm->conf->num_mbufs)
clib_warning ("%d mbufs allocated but total rx/tx ring size is %d\n",
- dm->conf->num_mbufs, nb_desc);
+ dm->conf->num_mbufs, nb_desc);
/* init next vhost-user if index */
dm->next_vu_if_id = 0;
@@ -814,13 +826,14 @@ dpdk_lib_init (dpdk_main_t * dm)
static void
dpdk_bind_devices_to_uio (dpdk_config_main_t * conf)
{
- vlib_pci_main_t * pm = &pci_main;
- clib_error_t * error;
- vlib_pci_device_t * d;
- pci_config_header_t * c;
- u8 * pci_addr = 0;
+ vlib_pci_main_t *pm = &pci_main;
+ clib_error_t *error;
+ vlib_pci_device_t *d;
+ pci_config_header_t *c;
+ u8 *pci_addr = 0;
int num_whitelisted = vec_len (conf->dev_confs);
+ /* *INDENT-OFF* */
pool_foreach (d, pm->pci_devs, ({
dpdk_device_config_t * devconf = 0;
c = &d->config0.header;
@@ -878,15 +891,17 @@ dpdk_bind_devices_to_uio (dpdk_config_main_t * conf)
clib_error_report (error);
}
}));
+ /* *INDENT-ON* */
vec_free (pci_addr);
}
static clib_error_t *
-dpdk_device_config (dpdk_config_main_t * conf, vlib_pci_addr_t pci_addr, unformat_input_t * input, u8 is_default)
+dpdk_device_config (dpdk_config_main_t * conf, vlib_pci_addr_t pci_addr,
+ unformat_input_t * input, u8 is_default)
{
- clib_error_t * error = 0;
- uword * p;
- dpdk_device_config_t * devconf;
+ clib_error_t *error = 0;
+ uword *p;
+ dpdk_device_config_t *devconf;
unformat_input_t sub_input;
if (is_default)
@@ -900,11 +915,13 @@ dpdk_device_config (dpdk_config_main_t * conf, vlib_pci_addr_t pci_addr, unforma
if (!p)
{
pool_get (conf->dev_confs, devconf);
- hash_set (conf->device_config_index_by_pci_addr, pci_addr.as_u32, devconf - conf->dev_confs);
+ hash_set (conf->device_config_index_by_pci_addr, pci_addr.as_u32,
+ devconf - conf->dev_confs);
}
else
- return clib_error_return(0, "duplicate configuration for PCI address %U",
- format_vlib_pci_addr, &pci_addr);
+ return clib_error_return (0,
+ "duplicate configuration for PCI address %U",
+ format_vlib_pci_addr, &pci_addr);
}
devconf->pci_addr.as_u32 = pci_addr.as_u32;
@@ -925,16 +942,18 @@ dpdk_device_config (dpdk_config_main_t * conf, vlib_pci_addr_t pci_addr, unforma
else if (unformat (input, "workers %U", unformat_bitmap_list,
&devconf->workers))
;
- else if (unformat (input, "rss %U", unformat_vlib_cli_sub_input, &sub_input))
- {
- error = unformat_rss_fn(&sub_input, &devconf->rss_fn);
- if (error)
- break;
- }
+ else
+ if (unformat
+ (input, "rss %U", unformat_vlib_cli_sub_input, &sub_input))
+ {
+ error = unformat_rss_fn (&sub_input, &devconf->rss_fn);
+ if (error)
+ break;
+ }
else if (unformat (input, "vlan-strip-offload off"))
- devconf->vlan_strip_offload = DPDK_DEVICE_VLAN_STRIP_OFF;
+ devconf->vlan_strip_offload = DPDK_DEVICE_VLAN_STRIP_OFF;
else if (unformat (input, "vlan-strip-offload on"))
- devconf->vlan_strip_offload = DPDK_DEVICE_VLAN_STRIP_ON;
+ devconf->vlan_strip_offload = DPDK_DEVICE_VLAN_STRIP_ON;
else
{
error = clib_error_return (0, "unknown input `%U'",
@@ -947,12 +966,15 @@ dpdk_device_config (dpdk_config_main_t * conf, vlib_pci_addr_t pci_addr, unforma
return error;
if (devconf->workers && devconf->num_rx_queues == 0)
- devconf->num_rx_queues = clib_bitmap_count_set_bits(devconf->workers);
+ devconf->num_rx_queues = clib_bitmap_count_set_bits (devconf->workers);
else if (devconf->workers &&
- clib_bitmap_count_set_bits(devconf->workers) != devconf->num_rx_queues)
- error = clib_error_return (0, "%U: number of worker threadds must be "
- "equal to number of rx queues",
- format_vlib_pci_addr, &pci_addr);
+ clib_bitmap_count_set_bits (devconf->workers) !=
+ devconf->num_rx_queues)
+ error =
+ clib_error_return (0,
+ "%U: number of worker threadds must be "
+ "equal to number of rx queues", format_vlib_pci_addr,
+ &pci_addr);
return error;
}
@@ -960,15 +982,15 @@ dpdk_device_config (dpdk_config_main_t * conf, vlib_pci_addr_t pci_addr, unforma
static clib_error_t *
dpdk_config (vlib_main_t * vm, unformat_input_t * input)
{
- clib_error_t * error = 0;
- dpdk_main_t * dm = &dpdk_main;
- dpdk_config_main_t * conf = &dpdk_config_main;
- vlib_thread_main_t * tm = vlib_get_thread_main();
- dpdk_device_config_t * devconf;
+ clib_error_t *error = 0;
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_config_main_t *conf = &dpdk_config_main;
+ vlib_thread_main_t *tm = vlib_get_thread_main ();
+ dpdk_device_config_t *devconf;
vlib_pci_addr_t pci_addr;
unformat_input_t sub_input;
- u8 * s, * tmp = 0;
- u8 * rte_cmd = 0, * ethname = 0;
+ u8 *s, *tmp = 0;
+ u8 *rte_cmd = 0, *ethname = 0;
u32 log_level;
int ret, i;
int num_whitelisted = 0;
@@ -976,41 +998,45 @@ dpdk_config (vlib_main_t * vm, unformat_input_t * input)
u8 no_huge = 0;
u8 huge_dir = 0;
u8 file_prefix = 0;
- u8 * socket_mem = 0;
+ u8 *socket_mem = 0;
conf->device_config_index_by_pci_addr = hash_create (0, sizeof (uword));
// MATT-FIXME: inverted virtio-vhost logic to use virtio by default
conf->use_virtio_vhost = 1;
- while (unformat_check_input(input) != UNFORMAT_END_OF_INPUT)
+ while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT)
{
/* Prime the pump */
if (unformat (input, "no-hugetlb"))
- {
- vec_add1 (conf->eal_init_args, (u8 *) "no-huge");
- no_huge = 1;
- }
+ {
+ vec_add1 (conf->eal_init_args, (u8 *) "no-huge");
+ no_huge = 1;
+ }
else if (unformat (input, "enable-tcp-udp-checksum"))
conf->enable_tcp_udp_checksum = 1;
else if (unformat (input, "decimal-interface-names"))
- conf->interface_name_format_decimal = 1;
+ conf->interface_name_format_decimal = 1;
else if (unformat (input, "no-multi-seg"))
- conf->no_multi_seg = 1;
+ conf->no_multi_seg = 1;
else if (unformat (input, "dev default %U", unformat_vlib_cli_sub_input,
&sub_input))
{
- error = dpdk_device_config (conf, (vlib_pci_addr_t) (u32) ~1, &sub_input, 1);
+ error =
+ dpdk_device_config (conf, (vlib_pci_addr_t) (u32) ~ 1, &sub_input,
+ 1);
if (error)
return error;
}
- else if (unformat (input, "dev %U %U", unformat_vlib_pci_addr, &pci_addr,
- unformat_vlib_cli_sub_input, &sub_input))
+ else
+ if (unformat
+ (input, "dev %U %U", unformat_vlib_pci_addr, &pci_addr,
+ unformat_vlib_cli_sub_input, &sub_input))
{
error = dpdk_device_config (conf, pci_addr, &sub_input, 0);
@@ -1029,19 +1055,25 @@ dpdk_config (vlib_main_t * vm, unformat_input_t * input)
num_whitelisted++;
}
else if (unformat (input, "num-mbufs %d", &conf->num_mbufs))
- ;
+ ;
else if (unformat (input, "kni %d", &conf->num_kni))
- ;
+ ;
else if (unformat (input, "uio-driver %s", &conf->uio_driver_name))
;
else if (unformat (input, "socket-mem %s", &socket_mem))
;
- else if (unformat (input, "vhost-user-coalesce-frames %d", &conf->vhost_coalesce_frames))
- ;
- else if (unformat (input, "vhost-user-coalesce-time %f", &conf->vhost_coalesce_time))
- ;
+ else
+ if (unformat
+ (input, "vhost-user-coalesce-frames %d",
+ &conf->vhost_coalesce_frames))
+ ;
+ else
+ if (unformat
+ (input, "vhost-user-coalesce-time %f",
+ &conf->vhost_coalesce_time))
+ ;
else if (unformat (input, "enable-vhost-user"))
- conf->use_virtio_vhost = 0;
+ conf->use_virtio_vhost = 0;
else if (unformat (input, "no-pci"))
{
no_pci = 1;
@@ -1049,7 +1081,7 @@ dpdk_config (vlib_main_t * vm, unformat_input_t * input)
vec_add1 (conf->eal_init_args, tmp);
}
else if (unformat (input, "poll-sleep %d", &dm->poll_sleep))
- ;
+ ;
#define _(a) \
else if (unformat(input, #a)) \
@@ -1059,7 +1091,6 @@ dpdk_config (vlib_main_t * vm, unformat_input_t * input)
}
foreach_eal_double_hyphen_predicate_arg
#undef _
-
#define _(a) \
else if (unformat(input, #a " %s", &s)) \
{ \
@@ -1074,7 +1105,6 @@ dpdk_config (vlib_main_t * vm, unformat_input_t * input)
}
foreach_eal_double_hyphen_arg
#undef _
-
#define _(a,b) \
else if (unformat(input, #a " %s", &s)) \
{ \
@@ -1083,9 +1113,8 @@ dpdk_config (vlib_main_t * vm, unformat_input_t * input)
vec_add1 (s, 0); \
vec_add1 (conf->eal_init_args, s); \
}
- foreach_eal_single_hyphen_arg
+ foreach_eal_single_hyphen_arg
#undef _
-
#define _(a,b) \
else if (unformat(input, #a " %s", &s)) \
{ \
@@ -1095,18 +1124,17 @@ dpdk_config (vlib_main_t * vm, unformat_input_t * input)
vec_add1 (conf->eal_init_args, s); \
conf->a##_set_manually = 1; \
}
- foreach_eal_single_hyphen_mandatory_arg
+ foreach_eal_single_hyphen_mandatory_arg
#undef _
+ else if (unformat (input, "default"))
+ ;
- else if (unformat(input, "default"))
- ;
-
- else
- {
- error = clib_error_return (0, "unknown input `%U'",
- format_unformat_error, input);
- goto done;
- }
+ else
+ {
+ error = clib_error_return (0, "unknown input `%U'",
+ format_unformat_error, input);
+ goto done;
+ }
}
if (!conf->uio_driver_name)
@@ -1117,20 +1145,20 @@ dpdk_config (vlib_main_t * vm, unformat_input_t * input)
*/
if (!no_huge && !huge_dir)
{
- u32 x, * mem_by_socket = 0;
+ u32 x, *mem_by_socket = 0;
uword c = 0;
u8 use_1g = 1;
u8 use_2m = 1;
u8 less_than_1g = 1;
int rv;
- umount(DEFAULT_HUGE_DIR);
+ umount (DEFAULT_HUGE_DIR);
/* Process "socket-mem" parameter value */
if (vec_len (socket_mem))
{
unformat_input_t in;
- unformat_init_vector(&in, socket_mem);
+ unformat_init_vector (&in, socket_mem);
while (unformat_check_input (&in) != UNFORMAT_END_OF_INPUT)
{
if (unformat (&in, "%u,", &x))
@@ -1142,26 +1170,29 @@ dpdk_config (vlib_main_t * vm, unformat_input_t * input)
else
break;
- vec_add1(mem_by_socket, x);
+ vec_add1 (mem_by_socket, x);
if (x > 1023)
less_than_1g = 0;
}
- /* Note: unformat_free vec_frees(in.buffer), aka socket_mem... */
- unformat_free(&in);
- socket_mem = 0;
+ /* Note: unformat_free vec_frees(in.buffer), aka socket_mem... */
+ unformat_free (&in);
+ socket_mem = 0;
}
else
{
+ /* *INDENT-OFF* */
clib_bitmap_foreach (c, tm->cpu_socket_bitmap, (
{
vec_validate(mem_by_socket, c);
mem_by_socket[c] = 256; /* default per-socket mem */
}
));
+ /* *INDENT-ON* */
}
/* check if available enough 1GB pages for each socket */
+ /* *INDENT-OFF* */
clib_bitmap_foreach (c, tm->cpu_socket_bitmap, (
{
u32 pages_avail, page_size, mem;
@@ -1216,69 +1247,70 @@ dpdk_config (vlib_main_t * vm, unformat_input_t * input)
vec_free(p);
vec_free(path);
}));
+ /* *INDENT-ON* */
if (mem_by_socket == 0)
- {
- error = clib_error_return (0, "mem_by_socket NULL");
- goto done;
- }
+ {
+ error = clib_error_return (0, "mem_by_socket NULL");
+ goto done;
+ }
_vec_len (mem_by_socket) = c + 1;
/* regenerate socket_mem string */
vec_foreach_index (x, mem_by_socket)
socket_mem = format (socket_mem, "%s%u",
- socket_mem ? "," : "",
- mem_by_socket[x]);
+ socket_mem ? "," : "", mem_by_socket[x]);
socket_mem = format (socket_mem, "%c", 0);
vec_free (mem_by_socket);
- rv = mkdir(VPP_RUN_DIR, 0755);
+ rv = mkdir (VPP_RUN_DIR, 0755);
if (rv && errno != EEXIST)
- {
- error = clib_error_return (0, "mkdir '%s' failed errno %d",
- VPP_RUN_DIR, errno);
- goto done;
- }
+ {
+ error = clib_error_return (0, "mkdir '%s' failed errno %d",
+ VPP_RUN_DIR, errno);
+ goto done;
+ }
- rv = mkdir(DEFAULT_HUGE_DIR, 0755);
+ rv = mkdir (DEFAULT_HUGE_DIR, 0755);
if (rv && errno != EEXIST)
- {
- error = clib_error_return (0, "mkdir '%s' failed errno %d",
- DEFAULT_HUGE_DIR, errno);
- goto done;
- }
+ {
+ error = clib_error_return (0, "mkdir '%s' failed errno %d",
+ DEFAULT_HUGE_DIR, errno);
+ goto done;
+ }
if (use_1g && !(less_than_1g && use_2m))
- {
- rv = mount("none", DEFAULT_HUGE_DIR, "hugetlbfs", 0, "pagesize=1G");
- }
+ {
+ rv =
+ mount ("none", DEFAULT_HUGE_DIR, "hugetlbfs", 0, "pagesize=1G");
+ }
else if (use_2m)
- {
- rv = mount("none", DEFAULT_HUGE_DIR, "hugetlbfs", 0, NULL);
- }
+ {
+ rv = mount ("none", DEFAULT_HUGE_DIR, "hugetlbfs", 0, NULL);
+ }
else
- {
- return clib_error_return (0, "not enough free huge pages");
- }
+ {
+ return clib_error_return (0, "not enough free huge pages");
+ }
if (rv)
- {
- error = clib_error_return (0, "mount failed %d", errno);
- goto done;
- }
+ {
+ error = clib_error_return (0, "mount failed %d", errno);
+ goto done;
+ }
tmp = format (0, "--huge-dir%c", 0);
vec_add1 (conf->eal_init_args, tmp);
tmp = format (0, "%s%c", DEFAULT_HUGE_DIR, 0);
vec_add1 (conf->eal_init_args, tmp);
if (!file_prefix)
- {
- tmp = format (0, "--file-prefix%c", 0);
- vec_add1 (conf->eal_init_args, tmp);
- tmp = format (0, "vpp%c", 0);
- vec_add1 (conf->eal_init_args, tmp);
- }
+ {
+ tmp = format (0, "--file-prefix%c", 0);
+ vec_add1 (conf->eal_init_args, tmp);
+ tmp = format (0, "vpp%c", 0);
+ vec_add1 (conf->eal_init_args, tmp);
+ }
}
vec_free (rte_cmd);
@@ -1290,24 +1322,24 @@ dpdk_config (vlib_main_t * vm, unformat_input_t * input)
/* I'll bet that -c and -n must be the first and second args... */
if (!conf->coremask_set_manually)
{
- vlib_thread_registration_t * tr;
- uword * coremask = 0;
+ vlib_thread_registration_t *tr;
+ uword *coremask = 0;
int i;
/* main thread core */
- coremask = clib_bitmap_set(coremask, tm->main_lcore, 1);
+ coremask = clib_bitmap_set (coremask, tm->main_lcore, 1);
for (i = 0; i < vec_len (tm->registrations); i++)
- {
- tr = tm->registrations[i];
- coremask = clib_bitmap_or(coremask, tr->coremask);
- }
+ {
+ tr = tm->registrations[i];
+ coremask = clib_bitmap_or (coremask, tr->coremask);
+ }
vec_insert (conf->eal_init_args, 2, 1);
conf->eal_init_args[1] = (u8 *) "-c";
tmp = format (0, "%U%c", format_bitmap_hex, coremask, 0);
conf->eal_init_args[2] = tmp;
- clib_bitmap_free(coremask);
+ clib_bitmap_free (coremask);
}
if (!conf->nchannels_set_manually)
@@ -1318,13 +1350,14 @@ dpdk_config (vlib_main_t * vm, unformat_input_t * input)
conf->eal_init_args[4] = tmp;
}
- if (no_pci == 0 && geteuid() == 0)
- dpdk_bind_devices_to_uio(conf);
+ if (no_pci == 0 && geteuid () == 0)
+ dpdk_bind_devices_to_uio (conf);
#define _(x) \
if (devconf->x == 0 && conf->default_devconf.x > 0) \
devconf->x = conf->default_devconf.x ;
+ /* *INDENT-OFF* */
pool_foreach (devconf, conf->dev_confs, ({
/* default per-device config items */
@@ -1346,6 +1379,7 @@ dpdk_config (vlib_main_t * vm, unformat_input_t * input)
vec_add1 (conf->eal_init_args, tmp);
}
}));
+ /* *INDENT-ON* */
#undef _
@@ -1363,7 +1397,7 @@ dpdk_config (vlib_main_t * vm, unformat_input_t * input)
/* NULL terminate the "argv" vector, in case of stupidity */
vec_add1 (conf->eal_init_args, 0);
- _vec_len(conf->eal_init_args) -= 1;
+ _vec_len (conf->eal_init_args) -= 1;
/* Set up DPDK eal and packet mbuf pool early. */
@@ -1374,188 +1408,197 @@ dpdk_config (vlib_main_t * vm, unformat_input_t * input)
vm = vlib_get_main ();
/* make copy of args as rte_eal_init tends to mess up with arg array */
- for (i = 1; i < vec_len(conf->eal_init_args); i++)
- conf->eal_init_args_str = format(conf->eal_init_args_str, "%s ",
- conf->eal_init_args[i]);
+ for (i = 1; i < vec_len (conf->eal_init_args); i++)
+ conf->eal_init_args_str = format (conf->eal_init_args_str, "%s ",
+ conf->eal_init_args[i]);
- ret = rte_eal_init(vec_len(conf->eal_init_args), (char **) conf->eal_init_args);
+ ret =
+ rte_eal_init (vec_len (conf->eal_init_args),
+ (char **) conf->eal_init_args);
/* lazy umount hugepages */
- umount2(DEFAULT_HUGE_DIR, MNT_DETACH);
+ umount2 (DEFAULT_HUGE_DIR, MNT_DETACH);
if (ret < 0)
return clib_error_return (0, "rte_eal_init returned %d", ret);
/* Dump the physical memory layout prior to creating the mbuf_pool */
- fprintf(stdout, "DPDK physical memory layout:\n");
- rte_dump_physmem_layout(stdout);
+ fprintf (stdout, "DPDK physical memory layout:\n");
+ rte_dump_physmem_layout (stdout);
/* main thread 1st */
- error = vlib_buffer_pool_create(vm, conf->num_mbufs, rte_socket_id());
+ error = vlib_buffer_pool_create (vm, conf->num_mbufs, rte_socket_id ());
if (error)
return error;
for (i = 0; i < RTE_MAX_LCORE; i++)
{
- error = vlib_buffer_pool_create(vm, conf->num_mbufs,
- rte_lcore_to_socket_id(i));
+ error = vlib_buffer_pool_create (vm, conf->num_mbufs,
+ rte_lcore_to_socket_id (i));
if (error)
- return error;
+ return error;
}
- done:
+done:
return error;
}
VLIB_CONFIG_FUNCTION (dpdk_config, "dpdk");
-void dpdk_update_link_state (dpdk_device_t * xd, f64 now)
+void
+dpdk_update_link_state (dpdk_device_t * xd, f64 now)
{
- vnet_main_t * vnm = vnet_get_main();
- struct rte_eth_link prev_link = xd->link;
- u32 hw_flags = 0;
- u8 hw_flags_chg = 0;
+ vnet_main_t *vnm = vnet_get_main ();
+ struct rte_eth_link prev_link = xd->link;
+ u32 hw_flags = 0;
+ u8 hw_flags_chg = 0;
- /* only update link state for PMD interfaces */
- if (xd->dev_type != VNET_DPDK_DEV_ETH)
- return;
+ /* only update link state for PMD interfaces */
+ if (xd->dev_type != VNET_DPDK_DEV_ETH)
+ return;
- xd->time_last_link_update = now ? now : xd->time_last_link_update;
- memset(&xd->link, 0, sizeof(xd->link));
- rte_eth_link_get_nowait (xd->device_index, &xd->link);
+ xd->time_last_link_update = now ? now : xd->time_last_link_update;
+ memset (&xd->link, 0, sizeof (xd->link));
+ rte_eth_link_get_nowait (xd->device_index, &xd->link);
- if (LINK_STATE_ELOGS)
+ if (LINK_STATE_ELOGS)
+ {
+ vlib_main_t *vm = vlib_get_main ();
+ ELOG_TYPE_DECLARE (e) =
{
- vlib_main_t * vm = vlib_get_main();
- ELOG_TYPE_DECLARE(e) = {
- .format =
- "update-link-state: sw_if_index %d, admin_up %d,"
- "old link_state %d new link_state %d",
- .format_args = "i4i1i1i1",
- };
-
- struct { u32 sw_if_index; u8 admin_up;
- u8 old_link_state; u8 new_link_state;} *ed;
- ed = ELOG_DATA (&vm->elog_main, e);
- ed->sw_if_index = xd->vlib_sw_if_index;
- ed->admin_up = xd->admin_up;
- ed->old_link_state = (u8)
- vnet_hw_interface_is_link_up (vnm, xd->vlib_hw_if_index);
- ed->new_link_state = (u8) xd->link.link_status;
- }
+ .format =
+ "update-link-state: sw_if_index %d, admin_up %d,"
+ "old link_state %d new link_state %d",.format_args = "i4i1i1i1",};
- if ((xd->admin_up == 1) &&
- ((xd->link.link_status != 0) ^
- vnet_hw_interface_is_link_up (vnm, xd->vlib_hw_if_index)))
+ struct
{
- hw_flags_chg = 1;
- hw_flags |= (xd->link.link_status ?
- VNET_HW_INTERFACE_FLAG_LINK_UP: 0);
- }
+ u32 sw_if_index;
+ u8 admin_up;
+ u8 old_link_state;
+ u8 new_link_state;
+ } *ed;
+ ed = ELOG_DATA (&vm->elog_main, e);
+ ed->sw_if_index = xd->vlib_sw_if_index;
+ ed->admin_up = xd->admin_up;
+ ed->old_link_state = (u8)
+ vnet_hw_interface_is_link_up (vnm, xd->vlib_hw_if_index);
+ ed->new_link_state = (u8) xd->link.link_status;
+ }
- if (hw_flags_chg || (xd->link.link_duplex != prev_link.link_duplex))
- {
- hw_flags_chg = 1;
- switch (xd->link.link_duplex)
- {
- case ETH_LINK_HALF_DUPLEX:
- hw_flags |= VNET_HW_INTERFACE_FLAG_HALF_DUPLEX;
- break;
- case ETH_LINK_FULL_DUPLEX:
- hw_flags |= VNET_HW_INTERFACE_FLAG_FULL_DUPLEX;
- break;
- default:
- break;
- }
- }
+ if ((xd->admin_up == 1) &&
+ ((xd->link.link_status != 0) ^
+ vnet_hw_interface_is_link_up (vnm, xd->vlib_hw_if_index)))
+ {
+ hw_flags_chg = 1;
+ hw_flags |= (xd->link.link_status ? VNET_HW_INTERFACE_FLAG_LINK_UP : 0);
+ }
+
+ if (hw_flags_chg || (xd->link.link_duplex != prev_link.link_duplex))
+ {
+ hw_flags_chg = 1;
+ switch (xd->link.link_duplex)
+ {
+ case ETH_LINK_HALF_DUPLEX:
+ hw_flags |= VNET_HW_INTERFACE_FLAG_HALF_DUPLEX;
+ break;
+ case ETH_LINK_FULL_DUPLEX:
+ hw_flags |= VNET_HW_INTERFACE_FLAG_FULL_DUPLEX;
+ break;
+ default:
+ break;
+ }
+ }
#if RTE_VERSION >= RTE_VERSION_NUM(16, 4, 0, 0)
- if (hw_flags_chg || (xd->link.link_speed != prev_link.link_speed))
- {
- hw_flags_chg = 1;
- switch (xd->link.link_speed)
- {
- case ETH_SPEED_NUM_10M:
- hw_flags |= VNET_HW_INTERFACE_FLAG_SPEED_10M;
- break;
- case ETH_SPEED_NUM_100M:
- hw_flags |= VNET_HW_INTERFACE_FLAG_SPEED_100M;
- break;
- case ETH_SPEED_NUM_1G:
- hw_flags |= VNET_HW_INTERFACE_FLAG_SPEED_1G;
- break;
- case ETH_SPEED_NUM_10G:
- hw_flags |= VNET_HW_INTERFACE_FLAG_SPEED_10G;
- break;
- case ETH_SPEED_NUM_40G:
- hw_flags |= VNET_HW_INTERFACE_FLAG_SPEED_40G;
- break;
- case 0:
- break;
- default:
- clib_warning("unknown link speed %d", xd->link.link_speed);
- break;
- }
- }
+ if (hw_flags_chg || (xd->link.link_speed != prev_link.link_speed))
+ {
+ hw_flags_chg = 1;
+ switch (xd->link.link_speed)
+ {
+ case ETH_SPEED_NUM_10M:
+ hw_flags |= VNET_HW_INTERFACE_FLAG_SPEED_10M;
+ break;
+ case ETH_SPEED_NUM_100M:
+ hw_flags |= VNET_HW_INTERFACE_FLAG_SPEED_100M;
+ break;
+ case ETH_SPEED_NUM_1G:
+ hw_flags |= VNET_HW_INTERFACE_FLAG_SPEED_1G;
+ break;
+ case ETH_SPEED_NUM_10G:
+ hw_flags |= VNET_HW_INTERFACE_FLAG_SPEED_10G;
+ break;
+ case ETH_SPEED_NUM_40G:
+ hw_flags |= VNET_HW_INTERFACE_FLAG_SPEED_40G;
+ break;
+ case 0:
+ break;
+ default:
+ clib_warning ("unknown link speed %d", xd->link.link_speed);
+ break;
+ }
+ }
#else
- if (hw_flags_chg || (xd->link.link_speed != prev_link.link_speed))
- {
- hw_flags_chg = 1;
- switch (xd->link.link_speed)
- {
- case ETH_LINK_SPEED_10:
- hw_flags |= VNET_HW_INTERFACE_FLAG_SPEED_10M;
- break;
- case ETH_LINK_SPEED_100:
- hw_flags |= VNET_HW_INTERFACE_FLAG_SPEED_100M;
- break;
- case ETH_LINK_SPEED_1000:
- hw_flags |= VNET_HW_INTERFACE_FLAG_SPEED_1G;
- break;
- case ETH_LINK_SPEED_10000:
- hw_flags |= VNET_HW_INTERFACE_FLAG_SPEED_10G;
- break;
- case ETH_LINK_SPEED_40G:
- hw_flags |= VNET_HW_INTERFACE_FLAG_SPEED_40G;
- break;
- case 0:
- break;
- default:
- clib_warning("unknown link speed %d", xd->link.link_speed);
- break;
- }
- }
+ if (hw_flags_chg || (xd->link.link_speed != prev_link.link_speed))
+ {
+ hw_flags_chg = 1;
+ switch (xd->link.link_speed)
+ {
+ case ETH_LINK_SPEED_10:
+ hw_flags |= VNET_HW_INTERFACE_FLAG_SPEED_10M;
+ break;
+ case ETH_LINK_SPEED_100:
+ hw_flags |= VNET_HW_INTERFACE_FLAG_SPEED_100M;
+ break;
+ case ETH_LINK_SPEED_1000:
+ hw_flags |= VNET_HW_INTERFACE_FLAG_SPEED_1G;
+ break;
+ case ETH_LINK_SPEED_10000:
+ hw_flags |= VNET_HW_INTERFACE_FLAG_SPEED_10G;
+ break;
+ case ETH_LINK_SPEED_40G:
+ hw_flags |= VNET_HW_INTERFACE_FLAG_SPEED_40G;
+ break;
+ case 0:
+ break;
+ default:
+ clib_warning ("unknown link speed %d", xd->link.link_speed);
+ break;
+ }
+ }
#endif
- if (hw_flags_chg)
- {
- if (LINK_STATE_ELOGS)
- {
- vlib_main_t * vm = vlib_get_main();
-
- ELOG_TYPE_DECLARE(e) = {
- .format = "update-link-state: sw_if_index %d, new flags %d",
- .format_args = "i4i4",
- };
-
- struct { u32 sw_if_index; u32 flags; } *ed;
- ed = ELOG_DATA (&vm->elog_main, e);
- ed->sw_if_index = xd->vlib_sw_if_index;
- ed->flags = hw_flags;
- }
- vnet_hw_interface_set_flags (vnm, xd->vlib_hw_if_index, hw_flags);
- }
+ if (hw_flags_chg)
+ {
+ if (LINK_STATE_ELOGS)
+ {
+ vlib_main_t *vm = vlib_get_main ();
+
+ ELOG_TYPE_DECLARE (e) =
+ {
+ .format =
+ "update-link-state: sw_if_index %d, new flags %d",.format_args
+ = "i4i4",};
+
+ struct
+ {
+ u32 sw_if_index;
+ u32 flags;
+ } *ed;
+ ed = ELOG_DATA (&vm->elog_main, e);
+ ed->sw_if_index = xd->vlib_sw_if_index;
+ ed->flags = hw_flags;
+ }
+ vnet_hw_interface_set_flags (vnm, xd->vlib_hw_if_index, hw_flags);
+ }
}
static uword
-dpdk_process (vlib_main_t * vm,
- vlib_node_runtime_t * rt,
- vlib_frame_t * f)
+dpdk_process (vlib_main_t * vm, vlib_node_runtime_t * rt, vlib_frame_t * f)
{
- clib_error_t * error;
- vnet_main_t * vnm = vnet_get_main();
- dpdk_main_t * dm = &dpdk_main;
- ethernet_main_t * em = &ethernet_main;
- dpdk_device_t * xd;
- vlib_thread_main_t * tm = vlib_get_thread_main();
+ clib_error_t *error;
+ vnet_main_t *vnm = vnet_get_main ();
+ dpdk_main_t *dm = &dpdk_main;
+ ethernet_main_t *em = &ethernet_main;
+ dpdk_device_t *xd;
+ vlib_thread_main_t *tm = vlib_get_thread_main ();
#if DPDK_VHOST_USER
void *vu_state;
#endif
@@ -1563,100 +1606,117 @@ dpdk_process (vlib_main_t * vm,
error = dpdk_lib_init (dm);
- /*
+ /*
* Turn on the input node if we found some devices to drive
* and we're not running worker threads or i/o threads
*/
- if (error == 0 && vec_len(dm->devices) > 0)
+ if (error == 0 && vec_len (dm->devices) > 0)
{
- if (tm->n_vlib_mains == 1)
- vlib_node_set_state (vm, dpdk_input_node.index,
- VLIB_NODE_STATE_POLLING);
- else
- for (i=0; i < tm->n_vlib_mains; i++)
- if (vec_len(dm->devices_by_cpu[i]) > 0)
- vlib_node_set_state (vlib_mains[i], dpdk_input_node.index,
- VLIB_NODE_STATE_POLLING);
+ if (tm->n_vlib_mains == 1)
+ vlib_node_set_state (vm, dpdk_input_node.index,
+ VLIB_NODE_STATE_POLLING);
+ else
+ for (i = 0; i < tm->n_vlib_mains; i++)
+ if (vec_len (dm->devices_by_cpu[i]) > 0)
+ vlib_node_set_state (vlib_mains[i], dpdk_input_node.index,
+ VLIB_NODE_STATE_POLLING);
}
if (error)
clib_error_report (error);
#if DPDK_VHOST_USER
- dpdk_vhost_user_process_init(&vu_state);
+ dpdk_vhost_user_process_init (&vu_state);
#endif
tm->worker_thread_release = 1;
f64 now = vlib_time_now (vm);
vec_foreach (xd, dm->devices)
- {
- dpdk_update_link_state (xd, now);
- }
+ {
+ dpdk_update_link_state (xd, now);
+ }
-{ /*
- * Extra set up for bond interfaces:
- * 1. Setup MACs for bond interfaces and their slave links which was set
- * in dpdk_port_setup() but needs to be done again here to take effect.
- * 2. Set up info for bond interface related CLI support.
- */
- int nports = rte_eth_dev_count();
- if (nports > 0) {
- for (i = 0; i < nports; i++) {
- struct rte_eth_dev_info dev_info;
- rte_eth_dev_info_get(i, &dev_info);
- if (!dev_info.driver_name)
+ {
+ /*
+ * Extra set up for bond interfaces:
+ * 1. Setup MACs for bond interfaces and their slave links which was set
+ * in dpdk_port_setup() but needs to be done again here to take effect.
+ * 2. Set up info for bond interface related CLI support.
+ */
+ int nports = rte_eth_dev_count ();
+ if (nports > 0)
+ {
+ for (i = 0; i < nports; i++)
+ {
+ struct rte_eth_dev_info dev_info;
+ rte_eth_dev_info_get (i, &dev_info);
+ if (!dev_info.driver_name)
dev_info.driver_name = dev_info.pci_dev->driver->name;
- ASSERT(dev_info.driver_name);
- if (strncmp(dev_info.driver_name, "rte_bond_pmd", 12) == 0) {
- u8 addr[6];
- u8 slink[16];
- int nlink = rte_eth_bond_slaves_get(i, slink, 16);
- if (nlink > 0) {
- vnet_hw_interface_t * bhi;
- ethernet_interface_t * bei;
- int rv;
-
- /* Get MAC of 1st slave link */
- rte_eth_macaddr_get(slink[0], (struct ether_addr *)addr);
- /* Set MAC of bounded interface to that of 1st slave link */
- rv = rte_eth_bond_mac_address_set(i, (struct ether_addr *)addr);
- if (rv < 0)
- clib_warning("Failed to set MAC address");
-
- /* Populate MAC of bonded interface in VPP hw tables */
- bhi = vnet_get_hw_interface(
- vnm, dm->devices[i].vlib_hw_if_index);
- bei = pool_elt_at_index(em->interfaces, bhi->hw_instance);
- clib_memcpy(bhi->hw_address, addr, 6);
- clib_memcpy(bei->address, addr, 6);
- /* Init l3 packet size allowed on bonded interface */
- bhi->max_l3_packet_bytes[VLIB_RX] =
- bhi->max_l3_packet_bytes[VLIB_TX] =
- ETHERNET_MAX_PACKET_BYTES - sizeof(ethernet_header_t);
- while (nlink >= 1) { /* for all slave links */
- int slave = slink[--nlink];
- dpdk_device_t * sdev = &dm->devices[slave];
- vnet_hw_interface_t * shi;
- vnet_sw_interface_t * ssi;
- /* Add MAC to all slave links except the first one */
- if (nlink) rte_eth_dev_mac_addr_add(
- slave, (struct ether_addr *)addr, 0);
- /* Set slaves bitmap for bonded interface */
- bhi->bond_info = clib_bitmap_set(
- bhi->bond_info, sdev->vlib_hw_if_index, 1);
- /* Set slave link flags on slave interface */
- shi = vnet_get_hw_interface(vnm, sdev->vlib_hw_if_index);
- ssi = vnet_get_sw_interface(vnm, sdev->vlib_sw_if_index);
- shi->bond_info = VNET_HW_INTERFACE_BOND_INFO_SLAVE;
- ssi->flags |= VNET_SW_INTERFACE_FLAG_BOND_SLAVE;
+ ASSERT (dev_info.driver_name);
+ if (strncmp (dev_info.driver_name, "rte_bond_pmd", 12) == 0)
+ {
+ u8 addr[6];
+ u8 slink[16];
+ int nlink = rte_eth_bond_slaves_get (i, slink, 16);
+ if (nlink > 0)
+ {
+ vnet_hw_interface_t *bhi;
+ ethernet_interface_t *bei;
+ int rv;
+
+ /* Get MAC of 1st slave link */
+ rte_eth_macaddr_get (slink[0],
+ (struct ether_addr *) addr);
+ /* Set MAC of bounded interface to that of 1st slave link */
+ rv =
+ rte_eth_bond_mac_address_set (i,
+ (struct ether_addr *)
+ addr);
+ if (rv < 0)
+ clib_warning ("Failed to set MAC address");
+
+ /* Populate MAC of bonded interface in VPP hw tables */
+ bhi =
+ vnet_get_hw_interface (vnm,
+ dm->devices[i].vlib_hw_if_index);
+ bei =
+ pool_elt_at_index (em->interfaces, bhi->hw_instance);
+ clib_memcpy (bhi->hw_address, addr, 6);
+ clib_memcpy (bei->address, addr, 6);
+ /* Init l3 packet size allowed on bonded interface */
+ bhi->max_l3_packet_bytes[VLIB_RX] =
+ bhi->max_l3_packet_bytes[VLIB_TX] =
+ ETHERNET_MAX_PACKET_BYTES - sizeof (ethernet_header_t);
+ while (nlink >= 1)
+ { /* for all slave links */
+ int slave = slink[--nlink];
+ dpdk_device_t *sdev = &dm->devices[slave];
+ vnet_hw_interface_t *shi;
+ vnet_sw_interface_t *ssi;
+ /* Add MAC to all slave links except the first one */
+ if (nlink)
+ rte_eth_dev_mac_addr_add (slave,
+ (struct ether_addr *)
+ addr, 0);
+ /* Set slaves bitmap for bonded interface */
+ bhi->bond_info =
+ clib_bitmap_set (bhi->bond_info,
+ sdev->vlib_hw_if_index, 1);
+ /* Set slave link flags on slave interface */
+ shi =
+ vnet_get_hw_interface (vnm, sdev->vlib_hw_if_index);
+ ssi =
+ vnet_get_sw_interface (vnm, sdev->vlib_sw_if_index);
+ shi->bond_info = VNET_HW_INTERFACE_BOND_INFO_SLAVE;
+ ssi->flags |= VNET_SW_INTERFACE_FLAG_BOND_SLAVE;
+ }
}
}
}
}
}
-}
while (1)
{
@@ -1664,58 +1724,62 @@ dpdk_process (vlib_main_t * vm,
* check each time through the loop in case intervals are changed
*/
f64 min_wait = dm->link_state_poll_interval < dm->stat_poll_interval ?
- dm->link_state_poll_interval : dm->stat_poll_interval;
+ dm->link_state_poll_interval : dm->stat_poll_interval;
vlib_process_wait_for_event_or_clock (vm, min_wait);
- if (dpdk_get_admin_up_down_in_progress())
- /* skip the poll if an admin up down is in progress (on any interface) */
- continue;
+ if (dpdk_get_admin_up_down_in_progress ())
+ /* skip the poll if an admin up down is in progress (on any interface) */
+ continue;
vec_foreach (xd, dm->devices)
- {
- f64 now = vlib_time_now (vm);
- if ((now - xd->time_last_stats_update) >= dm->stat_poll_interval)
- dpdk_update_counters (xd, now);
- if ((now - xd->time_last_link_update) >= dm->link_state_poll_interval)
- dpdk_update_link_state (xd, now);
+ {
+ f64 now = vlib_time_now (vm);
+ if ((now - xd->time_last_stats_update) >= dm->stat_poll_interval)
+ dpdk_update_counters (xd, now);
+ if ((now - xd->time_last_link_update) >= dm->link_state_poll_interval)
+ dpdk_update_link_state (xd, now);
#if DPDK_VHOST_USER
- if (xd->dev_type == VNET_DPDK_DEV_VHOST_USER)
- if (dpdk_vhost_user_process_if(vm, xd, vu_state) != 0)
- continue;
+ if (xd->dev_type == VNET_DPDK_DEV_VHOST_USER)
+ if (dpdk_vhost_user_process_if (vm, xd, vu_state) != 0)
+ continue;
#endif
- }
+ }
}
#if DPDK_VHOST_USER
- dpdk_vhost_user_process_cleanup(vu_state);
+ dpdk_vhost_user_process_cleanup (vu_state);
#endif
- return 0;
+ return 0;
}
+/* *INDENT-OFF* */
VLIB_REGISTER_NODE (dpdk_process_node,static) = {
.function = dpdk_process,
.type = VLIB_NODE_TYPE_PROCESS,
.name = "dpdk-process",
.process_log2_n_stack_bytes = 17,
};
+/* *INDENT-ON* */
-int dpdk_set_stat_poll_interval (f64 interval)
+int
+dpdk_set_stat_poll_interval (f64 interval)
{
if (interval < DPDK_MIN_STATS_POLL_INTERVAL)
- return (VNET_API_ERROR_INVALID_VALUE);
+ return (VNET_API_ERROR_INVALID_VALUE);
dpdk_main.stat_poll_interval = interval;
return 0;
}
-int dpdk_set_link_state_poll_interval (f64 interval)
+int
+dpdk_set_link_state_poll_interval (f64 interval)
{
if (interval < DPDK_MIN_LINK_POLL_INTERVAL)
- return (VNET_API_ERROR_INVALID_VALUE);
+ return (VNET_API_ERROR_INVALID_VALUE);
dpdk_main.link_state_poll_interval = interval;
@@ -1725,24 +1789,24 @@ int dpdk_set_link_state_poll_interval (f64 interval)
clib_error_t *
dpdk_init (vlib_main_t * vm)
{
- dpdk_main_t * dm = &dpdk_main;
- vlib_node_t * ei;
- clib_error_t * error = 0;
- vlib_thread_main_t * tm = vlib_get_thread_main();
+ dpdk_main_t *dm = &dpdk_main;
+ vlib_node_t *ei;
+ clib_error_t *error = 0;
+ vlib_thread_main_t *tm = vlib_get_thread_main ();
/* verify that structs are cacheline aligned */
- ASSERT(offsetof(dpdk_device_t, cacheline0) == 0);
- ASSERT(offsetof(dpdk_device_t, cacheline1) == CLIB_CACHE_LINE_BYTES);
- ASSERT(offsetof(dpdk_worker_t, cacheline0) == 0);
- ASSERT(offsetof(frame_queue_trace_t, cacheline0) == 0);
+ ASSERT (offsetof (dpdk_device_t, cacheline0) == 0);
+ ASSERT (offsetof (dpdk_device_t, cacheline1) == CLIB_CACHE_LINE_BYTES);
+ ASSERT (offsetof (dpdk_worker_t, cacheline0) == 0);
+ ASSERT (offsetof (frame_queue_trace_t, cacheline0) == 0);
dm->vlib_main = vm;
- dm->vnet_main = vnet_get_main();
+ dm->vnet_main = vnet_get_main ();
dm->conf = &dpdk_config_main;
ei = vlib_get_node_by_name (vm, (u8 *) "ethernet-input");
if (ei == 0)
- return clib_error_return (0, "ethernet-input node AWOL");
+ return clib_error_return (0, "ethernet-input node AWOL");
dm->ethernet_input_node_index = ei->index;
@@ -1760,19 +1824,18 @@ dpdk_init (vlib_main_t * vm)
/* initialize EFD (early fast discard) default settings */
dm->efd.enabled = DPDK_EFD_DISABLED;
dm->efd.queue_hi_thresh = ((DPDK_EFD_DEFAULT_DEVICE_QUEUE_HI_THRESH_PCT *
- DPDK_NB_RX_DESC_10GE)/100);
+ DPDK_NB_RX_DESC_10GE) / 100);
dm->efd.consec_full_frames_hi_thresh =
- DPDK_EFD_DEFAULT_CONSEC_FULL_FRAMES_HI_THRESH;
+ DPDK_EFD_DEFAULT_CONSEC_FULL_FRAMES_HI_THRESH;
/* vhost-user coalescence frames defaults */
dm->conf->vhost_coalesce_frames = 32;
dm->conf->vhost_coalesce_time = 1e-3;
/* Default vlib_buffer_t flags, DISABLES tcp/udp checksumming... */
- dm->buffer_flags_template =
- (VLIB_BUFFER_TOTAL_LENGTH_VALID
- | IP_BUFFER_L4_CHECKSUM_COMPUTED
- | IP_BUFFER_L4_CHECKSUM_CORRECT);
+ dm->buffer_flags_template =
+ (VLIB_BUFFER_TOTAL_LENGTH_VALID
+ | IP_BUFFER_L4_CHECKSUM_COMPUTED | IP_BUFFER_L4_CHECKSUM_CORRECT);
dm->stat_poll_interval = DPDK_STATS_POLL_INTERVAL;
dm->link_state_poll_interval = DPDK_LINK_POLL_INTERVAL;
@@ -1786,3 +1849,11 @@ dpdk_init (vlib_main_t * vm)
VLIB_INIT_FUNCTION (dpdk_init);
+
+/*
+ * fd.io coding-style-patch-verification: ON
+ *
+ * Local Variables:
+ * eval: (c-set-style "gnu")
+ * End:
+ */
diff --git a/vnet/vnet/devices/dpdk/node.c b/vnet/vnet/devices/dpdk/node.c
index 942274b0bf5..303b44e489a 100644
--- a/vnet/vnet/devices/dpdk/node.c
+++ b/vnet/vnet/devices/dpdk/node.c
@@ -49,82 +49,87 @@
*/
#define VMWARE_LENGTH_BUG_WORKAROUND 0
-static char * dpdk_error_strings[] = {
+static char *dpdk_error_strings[] = {
#define _(n,s) s,
- foreach_dpdk_error
+ foreach_dpdk_error
#undef _
};
always_inline int
-dpdk_mbuf_is_ip4(struct rte_mbuf *mb)
+dpdk_mbuf_is_ip4 (struct rte_mbuf *mb)
{
- return RTE_ETH_IS_IPV4_HDR(mb->packet_type) != 0;
+ return RTE_ETH_IS_IPV4_HDR (mb->packet_type) != 0;
}
always_inline int
-dpdk_mbuf_is_ip6(struct rte_mbuf *mb)
+dpdk_mbuf_is_ip6 (struct rte_mbuf *mb)
{
- return RTE_ETH_IS_IPV6_HDR(mb->packet_type) != 0;
+ return RTE_ETH_IS_IPV6_HDR (mb->packet_type) != 0;
}
always_inline int
-vlib_buffer_is_mpls(vlib_buffer_t * b)
+vlib_buffer_is_mpls (vlib_buffer_t * b)
{
ethernet_header_t *h = (ethernet_header_t *) b->data;
- return (h->type == clib_host_to_net_u16(ETHERNET_TYPE_MPLS_UNICAST));
+ return (h->type == clib_host_to_net_u16 (ETHERNET_TYPE_MPLS_UNICAST));
}
always_inline void
-dpdk_rx_next_and_error_from_mb_flags_x1 (dpdk_device_t *xd, struct rte_mbuf *mb,
- vlib_buffer_t *b0,
- u8 * next0, u8 * error0)
+dpdk_rx_next_and_error_from_mb_flags_x1 (dpdk_device_t * xd,
+ struct rte_mbuf *mb,
+ vlib_buffer_t * b0, u8 * next0,
+ u8 * error0)
{
u8 n0;
uint16_t mb_flags = mb->ol_flags;
- if (PREDICT_FALSE(mb_flags & (
+ if (PREDICT_FALSE (mb_flags & (
#ifdef RTE_LIBRTE_MBUF_EXT_RX_OLFLAGS
- PKT_EXT_RX_PKT_ERROR | PKT_EXT_RX_BAD_FCS |
+ PKT_EXT_RX_PKT_ERROR | PKT_EXT_RX_BAD_FCS |
#endif /* RTE_LIBRTE_MBUF_EXT_RX_OLFLAGS */
- PKT_RX_IP_CKSUM_BAD | PKT_RX_L4_CKSUM_BAD
- )))
+ PKT_RX_IP_CKSUM_BAD | PKT_RX_L4_CKSUM_BAD)))
{
- /* some error was flagged. determine the drop reason */
+ /* some error was flagged. determine the drop reason */
n0 = DPDK_RX_NEXT_DROP;
- *error0 =
+ *error0 =
#ifdef RTE_LIBRTE_MBUF_EXT_RX_OLFLAGS
- (mb_flags & PKT_EXT_RX_PKT_ERROR) ? DPDK_ERROR_RX_PACKET_ERROR :
- (mb_flags & PKT_EXT_RX_BAD_FCS) ? DPDK_ERROR_RX_BAD_FCS :
+ (mb_flags & PKT_EXT_RX_PKT_ERROR) ? DPDK_ERROR_RX_PACKET_ERROR :
+ (mb_flags & PKT_EXT_RX_BAD_FCS) ? DPDK_ERROR_RX_BAD_FCS :
#endif /* RTE_LIBRTE_MBUF_EXT_RX_OLFLAGS */
- (mb_flags & PKT_RX_IP_CKSUM_BAD) ? DPDK_ERROR_IP_CHECKSUM_ERROR :
- (mb_flags & PKT_RX_L4_CKSUM_BAD) ? DPDK_ERROR_L4_CHECKSUM_ERROR :
- DPDK_ERROR_NONE;
+ (mb_flags & PKT_RX_IP_CKSUM_BAD) ? DPDK_ERROR_IP_CHECKSUM_ERROR :
+ (mb_flags & PKT_RX_L4_CKSUM_BAD) ? DPDK_ERROR_L4_CHECKSUM_ERROR :
+ DPDK_ERROR_NONE;
}
else
{
*error0 = DPDK_ERROR_NONE;
- if (PREDICT_FALSE(xd->per_interface_next_index != ~0))
+ if (PREDICT_FALSE (xd->per_interface_next_index != ~0))
{
n0 = xd->per_interface_next_index;
b0->flags |= BUFFER_HANDOFF_NEXT_VALID;
- if (PREDICT_TRUE (dpdk_mbuf_is_ip4(mb)))
- vnet_buffer(b0)->handoff.next_index = HANDOFF_DISPATCH_NEXT_IP4_INPUT;
- else if (PREDICT_TRUE(dpdk_mbuf_is_ip6(mb)))
- vnet_buffer(b0)->handoff.next_index = HANDOFF_DISPATCH_NEXT_IP6_INPUT;
- else if (PREDICT_TRUE(vlib_buffer_is_mpls(b0)))
- vnet_buffer(b0)->handoff.next_index = HANDOFF_DISPATCH_NEXT_MPLS_INPUT;
+ if (PREDICT_TRUE (dpdk_mbuf_is_ip4 (mb)))
+ vnet_buffer (b0)->handoff.next_index =
+ HANDOFF_DISPATCH_NEXT_IP4_INPUT;
+ else if (PREDICT_TRUE (dpdk_mbuf_is_ip6 (mb)))
+ vnet_buffer (b0)->handoff.next_index =
+ HANDOFF_DISPATCH_NEXT_IP6_INPUT;
+ else if (PREDICT_TRUE (vlib_buffer_is_mpls (b0)))
+ vnet_buffer (b0)->handoff.next_index =
+ HANDOFF_DISPATCH_NEXT_MPLS_INPUT;
else
- vnet_buffer(b0)->handoff.next_index = HANDOFF_DISPATCH_NEXT_ETHERNET_INPUT;
+ vnet_buffer (b0)->handoff.next_index =
+ HANDOFF_DISPATCH_NEXT_ETHERNET_INPUT;
}
- else if (PREDICT_FALSE(xd->vlan_subifs || (mb_flags & PKT_RX_VLAN_PKT)))
+ else
+ if (PREDICT_FALSE (xd->vlan_subifs || (mb_flags & PKT_RX_VLAN_PKT)))
n0 = DPDK_RX_NEXT_ETHERNET_INPUT;
else
{
- if (PREDICT_TRUE (dpdk_mbuf_is_ip4(mb)))
+ if (PREDICT_TRUE (dpdk_mbuf_is_ip4 (mb)))
n0 = DPDK_RX_NEXT_IP4_INPUT;
- else if (PREDICT_TRUE(dpdk_mbuf_is_ip6(mb)))
+ else if (PREDICT_TRUE (dpdk_mbuf_is_ip6 (mb)))
n0 = DPDK_RX_NEXT_IP6_INPUT;
- else if (PREDICT_TRUE(vlib_buffer_is_mpls(b0)))
+ else if (PREDICT_TRUE (vlib_buffer_is_mpls (b0)))
n0 = DPDK_RX_NEXT_MPLS_INPUT;
else
n0 = DPDK_RX_NEXT_ETHERNET_INPUT;
@@ -133,15 +138,14 @@ dpdk_rx_next_and_error_from_mb_flags_x1 (dpdk_device_t *xd, struct rte_mbuf *mb,
*next0 = n0;
}
-void dpdk_rx_trace (dpdk_main_t * dm,
- vlib_node_runtime_t * node,
- dpdk_device_t * xd,
- u16 queue_id,
- u32 * buffers,
- uword n_buffers)
+void
+dpdk_rx_trace (dpdk_main_t * dm,
+ vlib_node_runtime_t * node,
+ dpdk_device_t * xd,
+ u16 queue_id, u32 * buffers, uword n_buffers)
{
- vlib_main_t * vm = vlib_get_main();
- u32 * b, n_left;
+ vlib_main_t *vm = vlib_get_main ();
+ u32 *b, n_left;
u8 next0;
n_left = n_buffers;
@@ -150,8 +154,8 @@ void dpdk_rx_trace (dpdk_main_t * dm,
while (n_left >= 1)
{
u32 bi0;
- vlib_buffer_t * b0;
- dpdk_rx_dma_trace_t * t0;
+ vlib_buffer_t *b0;
+ dpdk_rx_dma_trace_t *t0;
struct rte_mbuf *mb;
u8 error0;
@@ -159,9 +163,8 @@ void dpdk_rx_trace (dpdk_main_t * dm,
n_left -= 1;
b0 = vlib_get_buffer (vm, bi0);
- mb = rte_mbuf_from_vlib_buffer(b0);
- dpdk_rx_next_and_error_from_mb_flags_x1 (xd, mb, b0,
- &next0, &error0);
+ mb = rte_mbuf_from_vlib_buffer (b0);
+ dpdk_rx_next_and_error_from_mb_flags_x1 (xd, mb, b0, &next0, &error0);
vlib_trace_buffer (vm, node, next0, b0, /* follow_chain */ 0);
t0 = vlib_add_trace (vm, node, b0, sizeof (t0[0]));
t0->queue_index = queue_id;
@@ -170,7 +173,8 @@ void dpdk_rx_trace (dpdk_main_t * dm,
clib_memcpy (&t0->mb, mb, sizeof (t0->mb));
clib_memcpy (&t0->buffer, b0, sizeof (b0[0]) - sizeof (b0->pre_data));
- clib_memcpy (t0->buffer.pre_data, b0->data, sizeof (t0->buffer.pre_data));
+ clib_memcpy (t0->buffer.pre_data, b0->data,
+ sizeof (t0->buffer.pre_data));
clib_memcpy (&t0->data, mb->buf_addr + mb->data_off, sizeof (t0->data));
#ifdef RTE_LIBRTE_MBUF_EXT_RX_OLFLAGS
@@ -189,29 +193,28 @@ void dpdk_rx_trace (dpdk_main_t * dm,
* dpdk_efd_update_counters()
* Update EFD (early-fast-discard) counters
*/
-void dpdk_efd_update_counters (dpdk_device_t *xd,
- u32 n_buffers,
- u16 enabled)
+void
+dpdk_efd_update_counters (dpdk_device_t * xd, u32 n_buffers, u16 enabled)
{
if (enabled & DPDK_EFD_MONITOR_ENABLED)
{
- u64 now = clib_cpu_time_now();
+ u64 now = clib_cpu_time_now ();
if (xd->efd_agent.last_poll_time > 0)
- {
- u64 elapsed_time = (now - xd->efd_agent.last_poll_time);
- if (elapsed_time > xd->efd_agent.max_poll_delay)
- xd->efd_agent.max_poll_delay = elapsed_time;
- }
+ {
+ u64 elapsed_time = (now - xd->efd_agent.last_poll_time);
+ if (elapsed_time > xd->efd_agent.max_poll_delay)
+ xd->efd_agent.max_poll_delay = elapsed_time;
+ }
xd->efd_agent.last_poll_time = now;
}
-
+
xd->efd_agent.total_packet_cnt += n_buffers;
xd->efd_agent.last_burst_sz = n_buffers;
if (n_buffers > xd->efd_agent.max_burst_sz)
xd->efd_agent.max_burst_sz = n_buffers;
- if (PREDICT_FALSE(n_buffers == VLIB_FRAME_SIZE))
+ if (PREDICT_FALSE (n_buffers == VLIB_FRAME_SIZE))
{
xd->efd_agent.full_frames_cnt++;
xd->efd_agent.consec_full_frames_cnt++;
@@ -226,90 +229,194 @@ void dpdk_efd_update_counters (dpdk_device_t *xd,
* returns non zero DPDK error if packet meets early-fast-discard criteria,
* zero otherwise
*/
-u32 is_efd_discardable (vlib_thread_main_t *tm,
- vlib_buffer_t * b0,
- struct rte_mbuf *mb)
+u32
+is_efd_discardable (vlib_thread_main_t * tm,
+ vlib_buffer_t * b0, struct rte_mbuf *mb)
{
ethernet_header_t *eh = (ethernet_header_t *) b0->data;
- if (eh->type == clib_host_to_net_u16(ETHERNET_TYPE_IP4))
+ if (eh->type == clib_host_to_net_u16 (ETHERNET_TYPE_IP4))
{
ip4_header_t *ipv4 =
- (ip4_header_t *)&(b0->data[sizeof(ethernet_header_t)]);
+ (ip4_header_t *) & (b0->data[sizeof (ethernet_header_t)]);
u8 pkt_prec = (ipv4->tos >> 5);
-
+
return (tm->efd.ip_prec_bitmap & (1 << pkt_prec) ?
- DPDK_ERROR_IPV4_EFD_DROP_PKTS : DPDK_ERROR_NONE);
+ DPDK_ERROR_IPV4_EFD_DROP_PKTS : DPDK_ERROR_NONE);
}
- else if (eh->type == clib_net_to_host_u16(ETHERNET_TYPE_IP6))
+ else if (eh->type == clib_net_to_host_u16 (ETHERNET_TYPE_IP6))
{
ip6_header_t *ipv6 =
- (ip6_header_t *)&(b0->data[sizeof(ethernet_header_t)]);
+ (ip6_header_t *) & (b0->data[sizeof (ethernet_header_t)]);
u8 pkt_tclass =
- ((ipv6->ip_version_traffic_class_and_flow_label >> 20) & 0xff);
-
+ ((ipv6->ip_version_traffic_class_and_flow_label >> 20) & 0xff);
+
return (tm->efd.ip_prec_bitmap & (1 << pkt_tclass) ?
- DPDK_ERROR_IPV6_EFD_DROP_PKTS : DPDK_ERROR_NONE);
+ DPDK_ERROR_IPV6_EFD_DROP_PKTS : DPDK_ERROR_NONE);
}
- else if (eh->type == clib_net_to_host_u16(ETHERNET_TYPE_MPLS_UNICAST))
+ else if (eh->type == clib_net_to_host_u16 (ETHERNET_TYPE_MPLS_UNICAST))
{
mpls_unicast_header_t *mpls =
- (mpls_unicast_header_t *)&(b0->data[sizeof(ethernet_header_t)]);
+ (mpls_unicast_header_t *) & (b0->data[sizeof (ethernet_header_t)]);
u8 pkt_exp = ((mpls->label_exp_s_ttl >> 9) & 0x07);
return (tm->efd.mpls_exp_bitmap & (1 << pkt_exp) ?
- DPDK_ERROR_MPLS_EFD_DROP_PKTS : DPDK_ERROR_NONE);
+ DPDK_ERROR_MPLS_EFD_DROP_PKTS : DPDK_ERROR_NONE);
}
- else if ((eh->type == clib_net_to_host_u16(ETHERNET_TYPE_VLAN)) ||
- (eh->type == clib_net_to_host_u16(ETHERNET_TYPE_DOT1AD)))
+ else if ((eh->type == clib_net_to_host_u16 (ETHERNET_TYPE_VLAN)) ||
+ (eh->type == clib_net_to_host_u16 (ETHERNET_TYPE_DOT1AD)))
{
ethernet_vlan_header_t *vlan =
- (ethernet_vlan_header_t *)&(b0->data[sizeof(ethernet_header_t)]);
+ (ethernet_vlan_header_t *) & (b0->data[sizeof (ethernet_header_t)]);
u8 pkt_cos = ((vlan->priority_cfi_and_id >> 13) & 0x07);
return (tm->efd.vlan_cos_bitmap & (1 << pkt_cos) ?
- DPDK_ERROR_VLAN_EFD_DROP_PKTS : DPDK_ERROR_NONE);
+ DPDK_ERROR_VLAN_EFD_DROP_PKTS : DPDK_ERROR_NONE);
}
return DPDK_ERROR_NONE;
}
+static inline u32
+dpdk_rx_burst (dpdk_main_t * dm, dpdk_device_t * xd, u16 queue_id)
+{
+ u32 n_buffers;
+ u32 n_left;
+ u32 n_this_chunk;
+
+ n_left = VLIB_FRAME_SIZE;
+ n_buffers = 0;
+
+ if (PREDICT_TRUE (xd->dev_type == VNET_DPDK_DEV_ETH))
+ {
+ while (n_left)
+ {
+ n_this_chunk = rte_eth_rx_burst (xd->device_index, queue_id,
+ xd->rx_vectors[queue_id] +
+ n_buffers, n_left);
+ n_buffers += n_this_chunk;
+ n_left -= n_this_chunk;
+
+ /* Empirically, DPDK r1.8 produces vectors w/ 32 or fewer elts */
+ if (n_this_chunk < 32)
+ break;
+ }
+ }
+#if DPDK_VHOST_USER
+ else if (xd->dev_type == VNET_DPDK_DEV_VHOST_USER)
+ {
+ vlib_main_t *vm = vlib_get_main ();
+ vlib_buffer_main_t *bm = vm->buffer_main;
+ unsigned socket_id = rte_socket_id ();
+ u32 offset = 0;
+
+ offset = queue_id * VIRTIO_QNUM;
+
+ struct vhost_virtqueue *vq =
+ xd->vu_vhost_dev.virtqueue[offset + VIRTIO_TXQ];
+
+ if (PREDICT_FALSE (!vq->enabled))
+ return 0;
+
+ struct rte_mbuf **pkts = xd->rx_vectors[queue_id];
+ while (n_left)
+ {
+ n_this_chunk = rte_vhost_dequeue_burst (&xd->vu_vhost_dev,
+ offset + VIRTIO_TXQ,
+ bm->pktmbuf_pools
+ [socket_id],
+ pkts + n_buffers, n_left);
+ n_buffers += n_this_chunk;
+ n_left -= n_this_chunk;
+ if (n_this_chunk == 0)
+ break;
+ }
+
+ int i;
+ u32 bytes = 0;
+ for (i = 0; i < n_buffers; i++)
+ {
+ struct rte_mbuf *buff = pkts[i];
+ bytes += rte_pktmbuf_data_len (buff);
+ }
+
+ f64 now = vlib_time_now (vm);
+
+ dpdk_vu_vring *vring = NULL;
+ /* send pending interrupts if needed */
+ if (dpdk_vhost_user_want_interrupt (xd, offset + VIRTIO_TXQ))
+ {
+ vring = &(xd->vu_intf->vrings[offset + VIRTIO_TXQ]);
+ vring->n_since_last_int += n_buffers;
+
+ if ((vring->n_since_last_int && (vring->int_deadline < now))
+ || (vring->n_since_last_int > dm->conf->vhost_coalesce_frames))
+ dpdk_vhost_user_send_interrupt (vm, xd, offset + VIRTIO_TXQ);
+ }
+
+ vring = &(xd->vu_intf->vrings[offset + VIRTIO_RXQ]);
+ vring->packets += n_buffers;
+ vring->bytes += bytes;
+
+ if (dpdk_vhost_user_want_interrupt (xd, offset + VIRTIO_RXQ))
+ {
+ if (vring->n_since_last_int && (vring->int_deadline < now))
+ dpdk_vhost_user_send_interrupt (vm, xd, offset + VIRTIO_RXQ);
+ }
+
+ }
+#endif
+#ifdef RTE_LIBRTE_KNI
+ else if (xd->dev_type == VNET_DPDK_DEV_KNI)
+ {
+ n_buffers =
+ rte_kni_rx_burst (xd->kni, xd->rx_vectors[queue_id], VLIB_FRAME_SIZE);
+ rte_kni_handle_request (xd->kni);
+ }
+#endif
+ else
+ {
+ ASSERT (0);
+ }
+
+ return n_buffers;
+}
+
/*
* This function is used when there are no worker threads.
- * The main thread performs IO and forwards the packets.
+ * The main thread performs IO and forwards the packets.
*/
-static inline u32 dpdk_device_input ( dpdk_main_t * dm,
- dpdk_device_t * xd,
- vlib_node_runtime_t * node,
- u32 cpu_index,
- u16 queue_id,
- int use_efd)
+static inline u32
+dpdk_device_input (dpdk_main_t * dm,
+ dpdk_device_t * xd,
+ vlib_node_runtime_t * node,
+ u32 cpu_index, u16 queue_id, int use_efd)
{
u32 n_buffers;
u32 next_index = DPDK_RX_NEXT_ETHERNET_INPUT;
- u32 n_left_to_next, * to_next;
+ u32 n_left_to_next, *to_next;
u32 mb_index;
- vlib_main_t * vm = vlib_get_main();
+ vlib_main_t *vm = vlib_get_main ();
uword n_rx_bytes = 0;
- u32 n_trace, trace_cnt __attribute__((unused));
- vlib_buffer_free_list_t * fl;
+ u32 n_trace, trace_cnt __attribute__ ((unused));
+ vlib_buffer_free_list_t *fl;
u8 efd_discard_burst = 0;
u32 buffer_flags_template;
-
+
if (xd->admin_up == 0)
return 0;
- n_buffers = dpdk_rx_burst(dm, xd, queue_id);
+ n_buffers = dpdk_rx_burst (dm, xd, queue_id);
if (n_buffers == 0)
{
/* check if EFD (dpdk) is enabled */
- if (PREDICT_FALSE(use_efd && dm->efd.enabled))
- {
- /* reset a few stats */
- xd->efd_agent.last_poll_time = 0;
- xd->efd_agent.last_burst_sz = 0;
- }
+ if (PREDICT_FALSE (use_efd && dm->efd.enabled))
+ {
+ /* reset a few stats */
+ xd->efd_agent.last_poll_time = 0;
+ xd->efd_agent.last_burst_sz = 0;
+ }
return 0;
}
@@ -321,57 +428,57 @@ static inline u32 dpdk_device_input ( dpdk_main_t * dm,
fl = vlib_buffer_get_free_list (vm, VLIB_BUFFER_DEFAULT_FREE_LIST_INDEX);
/*
- * DAW-FIXME: VMXNET3 device stop/start doesn't work,
+ * DAW-FIXME: VMXNET3 device stop/start doesn't work,
* therefore fake the stop in the dpdk driver by
- * silently dropping all of the incoming pkts instead of
+ * silently dropping all of the incoming pkts instead of
* stopping the driver / hardware.
*/
- if (PREDICT_FALSE(xd->admin_up != 1))
+ if (PREDICT_FALSE (xd->admin_up != 1))
{
for (mb_index = 0; mb_index < n_buffers; mb_index++)
rte_pktmbuf_free (xd->rx_vectors[queue_id][mb_index]);
-
+
return 0;
}
/* Check for congestion if EFD (Early-Fast-Discard) is enabled
* in any mode (e.g. dpdk, monitor, or drop_all)
*/
- if (PREDICT_FALSE(use_efd && dm->efd.enabled))
+ if (PREDICT_FALSE (use_efd && dm->efd.enabled))
{
/* update EFD counters */
- dpdk_efd_update_counters(xd, n_buffers, dm->efd.enabled);
-
- if (PREDICT_FALSE(dm->efd.enabled & DPDK_EFD_DROPALL_ENABLED))
- {
- /* discard all received packets */
- for (mb_index = 0; mb_index < n_buffers; mb_index++)
- rte_pktmbuf_free(xd->rx_vectors[queue_id][mb_index]);
-
- xd->efd_agent.discard_cnt += n_buffers;
- increment_efd_drop_counter(vm,
- DPDK_ERROR_VLAN_EFD_DROP_PKTS,
- n_buffers);
-
- return 0;
- }
-
- if (PREDICT_FALSE(xd->efd_agent.consec_full_frames_cnt >=
- dm->efd.consec_full_frames_hi_thresh))
- {
- u32 device_queue_sz = rte_eth_rx_queue_count(xd->device_index,
- queue_id);
- if (device_queue_sz >= dm->efd.queue_hi_thresh)
- {
- /* dpdk device queue has reached the critical threshold */
- xd->efd_agent.congestion_cnt++;
-
- /* apply EFD to packets from the burst */
- efd_discard_burst = 1;
- }
- }
+ dpdk_efd_update_counters (xd, n_buffers, dm->efd.enabled);
+
+ if (PREDICT_FALSE (dm->efd.enabled & DPDK_EFD_DROPALL_ENABLED))
+ {
+ /* discard all received packets */
+ for (mb_index = 0; mb_index < n_buffers; mb_index++)
+ rte_pktmbuf_free (xd->rx_vectors[queue_id][mb_index]);
+
+ xd->efd_agent.discard_cnt += n_buffers;
+ increment_efd_drop_counter (vm,
+ DPDK_ERROR_VLAN_EFD_DROP_PKTS,
+ n_buffers);
+
+ return 0;
+ }
+
+ if (PREDICT_FALSE (xd->efd_agent.consec_full_frames_cnt >=
+ dm->efd.consec_full_frames_hi_thresh))
+ {
+ u32 device_queue_sz = rte_eth_rx_queue_count (xd->device_index,
+ queue_id);
+ if (device_queue_sz >= dm->efd.queue_hi_thresh)
+ {
+ /* dpdk device queue has reached the critical threshold */
+ xd->efd_agent.congestion_cnt++;
+
+ /* apply EFD to packets from the burst */
+ efd_discard_burst = 1;
+ }
+ }
}
-
+
mb_index = 0;
while (n_buffers > 0)
@@ -379,65 +486,63 @@ static inline u32 dpdk_device_input ( dpdk_main_t * dm,
u32 bi0;
u8 next0, error0;
u32 l3_offset0;
- vlib_buffer_t * b0, * b_seg, * b_chain = 0;
+ vlib_buffer_t *b0, *b_seg, *b_chain = 0;
u32 cntr_type;
vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
while (n_buffers > 0 && n_left_to_next > 0)
- {
+ {
u8 nb_seg = 1;
- struct rte_mbuf *mb = xd->rx_vectors[queue_id][mb_index];
+ struct rte_mbuf *mb = xd->rx_vectors[queue_id][mb_index];
struct rte_mbuf *mb_seg = mb->next;
- if (PREDICT_TRUE(n_buffers > 2))
- {
- struct rte_mbuf *pfmb = xd->rx_vectors[queue_id][mb_index+2];
- vlib_buffer_t *bp = vlib_buffer_from_rte_mbuf(pfmb);
- CLIB_PREFETCH (pfmb, CLIB_CACHE_LINE_BYTES, STORE);
- CLIB_PREFETCH (bp, CLIB_CACHE_LINE_BYTES, STORE);
- }
-
- ASSERT(mb);
-
- b0 = vlib_buffer_from_rte_mbuf(mb);
-
- /* check whether EFD is looking for packets to discard */
- if (PREDICT_FALSE(efd_discard_burst))
- {
- vlib_thread_main_t * tm = vlib_get_thread_main();
-
- if (PREDICT_TRUE(cntr_type = is_efd_discardable(tm, b0, mb)))
- {
- rte_pktmbuf_free(mb);
- xd->efd_agent.discard_cnt++;
- increment_efd_drop_counter(vm,
- cntr_type,
- 1);
- n_buffers--;
- mb_index++;
- continue;
- }
- }
-
- /* Prefetch one next segment if it exists. */
- if (PREDICT_FALSE(mb->nb_segs > 1))
- {
- struct rte_mbuf *pfmb = mb->next;
- vlib_buffer_t *bp = vlib_buffer_from_rte_mbuf(pfmb);
- CLIB_PREFETCH (pfmb, CLIB_CACHE_LINE_BYTES, LOAD);
- CLIB_PREFETCH (bp, CLIB_CACHE_LINE_BYTES, STORE);
+ if (PREDICT_TRUE (n_buffers > 2))
+ {
+ struct rte_mbuf *pfmb = xd->rx_vectors[queue_id][mb_index + 2];
+ vlib_buffer_t *bp = vlib_buffer_from_rte_mbuf (pfmb);
+ CLIB_PREFETCH (pfmb, CLIB_CACHE_LINE_BYTES, STORE);
+ CLIB_PREFETCH (bp, CLIB_CACHE_LINE_BYTES, STORE);
+ }
+
+ ASSERT (mb);
+
+ b0 = vlib_buffer_from_rte_mbuf (mb);
+
+ /* check whether EFD is looking for packets to discard */
+ if (PREDICT_FALSE (efd_discard_burst))
+ {
+ vlib_thread_main_t *tm = vlib_get_thread_main ();
+
+ if (PREDICT_TRUE (cntr_type = is_efd_discardable (tm, b0, mb)))
+ {
+ rte_pktmbuf_free (mb);
+ xd->efd_agent.discard_cnt++;
+ increment_efd_drop_counter (vm, cntr_type, 1);
+ n_buffers--;
+ mb_index++;
+ continue;
+ }
+ }
+
+ /* Prefetch one next segment if it exists. */
+ if (PREDICT_FALSE (mb->nb_segs > 1))
+ {
+ struct rte_mbuf *pfmb = mb->next;
+ vlib_buffer_t *bp = vlib_buffer_from_rte_mbuf (pfmb);
+ CLIB_PREFETCH (pfmb, CLIB_CACHE_LINE_BYTES, LOAD);
+ CLIB_PREFETCH (bp, CLIB_CACHE_LINE_BYTES, STORE);
b_chain = b0;
- }
+ }
- vlib_buffer_init_for_free_list (b0, fl);
-
- bi0 = vlib_get_buffer_index (vm, b0);
+ vlib_buffer_init_for_free_list (b0, fl);
+
+ bi0 = vlib_get_buffer_index (vm, b0);
+
+ to_next[0] = bi0;
+ to_next++;
+ n_left_to_next--;
- to_next[0] = bi0;
- to_next++;
- n_left_to_next--;
-
dpdk_rx_next_and_error_from_mb_flags_x1 (xd, mb, b0,
&next0, &error0);
#ifdef RTE_LIBRTE_MBUF_EXT_RX_OLFLAGS
@@ -446,7 +551,7 @@ static inline u32 dpdk_device_input ( dpdk_main_t * dm,
* is using them for RX flags (e.g. Cisco VIC Ethernet driver)
*/
- if (PREDICT_TRUE(trace_cnt == 0))
+ if (PREDICT_TRUE (trace_cnt == 0))
mb->ol_flags &= PKT_EXT_RX_CLR_TX_FLAGS_MASK;
else
trace_cnt--;
@@ -456,44 +561,44 @@ static inline u32 dpdk_device_input ( dpdk_main_t * dm,
l3_offset0 = ((next0 == DPDK_RX_NEXT_IP4_INPUT ||
next0 == DPDK_RX_NEXT_IP6_INPUT ||
- next0 == DPDK_RX_NEXT_MPLS_INPUT) ?
+ next0 == DPDK_RX_NEXT_MPLS_INPUT) ?
sizeof (ethernet_header_t) : 0);
- b0->current_data = l3_offset0;
- /* Some drivers like fm10k receive frames with
- mb->data_off > RTE_PKTMBUF_HEADROOM */
- b0->current_data += mb->data_off - RTE_PKTMBUF_HEADROOM;
- b0->current_length = mb->data_len - l3_offset0;
+ b0->current_data = l3_offset0;
+ /* Some drivers like fm10k receive frames with
+ mb->data_off > RTE_PKTMBUF_HEADROOM */
+ b0->current_data += mb->data_off - RTE_PKTMBUF_HEADROOM;
+ b0->current_length = mb->data_len - l3_offset0;
- b0->flags = buffer_flags_template;
+ b0->flags = buffer_flags_template;
- if (VMWARE_LENGTH_BUG_WORKAROUND)
- b0->current_length -= 4;
+ if (VMWARE_LENGTH_BUG_WORKAROUND)
+ b0->current_length -= 4;
- vnet_buffer(b0)->sw_if_index[VLIB_RX] = xd->vlib_sw_if_index;
- vnet_buffer(b0)->sw_if_index[VLIB_TX] = (u32)~0;
+ vnet_buffer (b0)->sw_if_index[VLIB_RX] = xd->vlib_sw_if_index;
+ vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0;
n_rx_bytes += mb->pkt_len;
- /* Process subsequent segments of multi-segment packets */
+ /* Process subsequent segments of multi-segment packets */
while ((mb->nb_segs > 1) && (nb_seg < mb->nb_segs))
{
- ASSERT(mb_seg != 0);
+ ASSERT (mb_seg != 0);
- b_seg = vlib_buffer_from_rte_mbuf(mb_seg);
+ b_seg = vlib_buffer_from_rte_mbuf (mb_seg);
vlib_buffer_init_for_free_list (b_seg, fl);
- ASSERT((b_seg->flags & VLIB_BUFFER_NEXT_PRESENT) == 0);
- ASSERT(b_seg->current_data == 0);
+ ASSERT ((b_seg->flags & VLIB_BUFFER_NEXT_PRESENT) == 0);
+ ASSERT (b_seg->current_data == 0);
- /*
- * The driver (e.g. virtio) may not put the packet data at the start
- * of the segment, so don't assume b_seg->current_data == 0 is correct.
- */
- b_seg->current_data = (mb_seg->buf_addr + mb_seg->data_off) - (void *)b_seg->data;
+ /*
+ * The driver (e.g. virtio) may not put the packet data at the start
+ * of the segment, so don't assume b_seg->current_data == 0 is correct.
+ */
+ b_seg->current_data =
+ (mb_seg->buf_addr + mb_seg->data_off) - (void *) b_seg->data;
b_seg->current_length = mb_seg->data_len;
- b0->total_length_not_including_first_buffer +=
- mb_seg->data_len;
+ b0->total_length_not_including_first_buffer += mb_seg->data_len;
b_chain->flags |= VLIB_BUFFER_NEXT_PRESENT;
b_chain->next_buffer = vlib_get_buffer_index (vm, b_seg);
@@ -501,61 +606,61 @@ static inline u32 dpdk_device_input ( dpdk_main_t * dm,
b_chain = b_seg;
mb_seg = mb_seg->next;
nb_seg++;
- }
-
- /*
- * Turn this on if you run into
- * "bad monkey" contexts, and you want to know exactly
- * which nodes they've visited... See main.c...
- */
- VLIB_BUFFER_TRACE_TRAJECTORY_INIT(b0);
-
- vlib_validate_buffer_enqueue_x1 (vm, node, next_index,
- to_next, n_left_to_next,
- bi0, next0);
- if (PREDICT_FALSE (n_trace > mb_index))
- vec_add1 (xd->d_trace_buffers, bi0);
- n_buffers--;
- mb_index++;
- }
+ }
+
+ /*
+ * Turn this on if you run into
+ * "bad monkey" contexts, and you want to know exactly
+ * which nodes they've visited... See main.c...
+ */
+ VLIB_BUFFER_TRACE_TRAJECTORY_INIT (b0);
+
+ vlib_validate_buffer_enqueue_x1 (vm, node, next_index,
+ to_next, n_left_to_next,
+ bi0, next0);
+ if (PREDICT_FALSE (n_trace > mb_index))
+ vec_add1 (xd->d_trace_buffers, bi0);
+ n_buffers--;
+ mb_index++;
+ }
vlib_put_next_frame (vm, node, next_index, n_left_to_next);
}
if (PREDICT_FALSE (vec_len (xd->d_trace_buffers) > 0))
{
dpdk_rx_trace (dm, node, xd, queue_id, xd->d_trace_buffers,
- vec_len (xd->d_trace_buffers));
- vlib_set_trace_count (vm, node, n_trace - vec_len (xd->d_trace_buffers));
+ vec_len (xd->d_trace_buffers));
+ vlib_set_trace_count (vm, node,
+ n_trace - vec_len (xd->d_trace_buffers));
}
-
- vlib_increment_combined_counter
- (vnet_get_main()->interface_main.combined_sw_if_counters
+
+ vlib_increment_combined_counter
+ (vnet_get_main ()->interface_main.combined_sw_if_counters
+ VNET_INTERFACE_COUNTER_RX,
- cpu_index,
- xd->vlib_sw_if_index,
- mb_index, n_rx_bytes);
+ cpu_index, xd->vlib_sw_if_index, mb_index, n_rx_bytes);
- dpdk_worker_t * dw = vec_elt_at_index(dm->workers, cpu_index);
+ dpdk_worker_t *dw = vec_elt_at_index (dm->workers, cpu_index);
dw->aggregate_rx_packets += mb_index;
return mb_index;
}
-static inline void poll_rate_limit(dpdk_main_t * dm)
+static inline void
+poll_rate_limit (dpdk_main_t * dm)
{
/* Limit the poll rate by sleeping for N msec between polls */
if (PREDICT_FALSE (dm->poll_sleep != 0))
- {
- struct timespec ts, tsrem;
+ {
+ struct timespec ts, tsrem;
- ts.tv_sec = 0;
- ts.tv_nsec = 1000*1000*dm->poll_sleep; /* 1ms */
+ ts.tv_sec = 0;
+ ts.tv_nsec = 1000 * 1000 * dm->poll_sleep; /* 1ms */
- while (nanosleep(&ts, &tsrem) < 0)
- {
- ts = tsrem;
- }
- }
+ while (nanosleep (&ts, &tsrem) < 0)
+ {
+ ts = tsrem;
+ }
+ }
}
/** \brief Main DPDK input node
@@ -595,88 +700,90 @@ static inline void poll_rate_limit(dpdk_main_t * dm)
<em>Next Nodes:</em>
- Static arcs to: error-drop, ethernet-input,
- ip4-input-no-checksum, ip6-input, mpls-gre-input
+ ip4-input-no-checksum, ip6-input, mpls-gre-input
- per-interface redirection, controlled by
<code>xd->per_interface_next_index</code>
*/
static uword
-dpdk_input (vlib_main_t * vm,
- vlib_node_runtime_t * node,
- vlib_frame_t * f)
+dpdk_input (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * f)
{
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_t * xd;
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_t *xd;
uword n_rx_packets = 0;
- dpdk_device_and_queue_t * dq;
- u32 cpu_index = os_get_cpu_number();
+ dpdk_device_and_queue_t *dq;
+ u32 cpu_index = os_get_cpu_number ();
/*
* Poll all devices on this cpu for input/interrupts.
*/
+ /* *INDENT-OFF* */
vec_foreach (dq, dm->devices_by_cpu[cpu_index])
{
xd = vec_elt_at_index(dm->devices, dq->device);
ASSERT(dq->queue_id == 0);
n_rx_packets += dpdk_device_input (dm, xd, node, cpu_index, 0, 0);
}
+ /* *INDENT-ON* */
- poll_rate_limit(dm);
+ poll_rate_limit (dm);
return n_rx_packets;
}
uword
dpdk_input_rss (vlib_main_t * vm,
- vlib_node_runtime_t * node,
- vlib_frame_t * f)
+ vlib_node_runtime_t * node, vlib_frame_t * f)
{
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_t * xd;
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_t *xd;
uword n_rx_packets = 0;
- dpdk_device_and_queue_t * dq;
- u32 cpu_index = os_get_cpu_number();
+ dpdk_device_and_queue_t *dq;
+ u32 cpu_index = os_get_cpu_number ();
/*
* Poll all devices on this cpu for input/interrupts.
*/
+ /* *INDENT-OFF* */
vec_foreach (dq, dm->devices_by_cpu[cpu_index])
{
xd = vec_elt_at_index(dm->devices, dq->device);
n_rx_packets += dpdk_device_input (dm, xd, node, cpu_index, dq->queue_id, 0);
}
+ /* *INDENT-ON* */
- poll_rate_limit(dm);
+ poll_rate_limit (dm);
return n_rx_packets;
}
uword
dpdk_input_efd (vlib_main_t * vm,
- vlib_node_runtime_t * node,
- vlib_frame_t * f)
+ vlib_node_runtime_t * node, vlib_frame_t * f)
{
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_t * xd;
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_t *xd;
uword n_rx_packets = 0;
- dpdk_device_and_queue_t * dq;
- u32 cpu_index = os_get_cpu_number();
+ dpdk_device_and_queue_t *dq;
+ u32 cpu_index = os_get_cpu_number ();
/*
* Poll all devices on this cpu for input/interrupts.
*/
+ /* *INDENT-OFF* */
vec_foreach (dq, dm->devices_by_cpu[cpu_index])
{
xd = vec_elt_at_index(dm->devices, dq->device);
n_rx_packets += dpdk_device_input (dm, xd, node, cpu_index, dq->queue_id, 1);
}
+ /* *INDENT-ON* */
- poll_rate_limit(dm);
+ poll_rate_limit (dm);
return n_rx_packets;
}
-
+/* *INDENT-OFF* */
VLIB_REGISTER_NODE (dpdk_input_node) = {
.function = dpdk_input,
.type = VLIB_NODE_TYPE_INPUT,
@@ -711,12 +818,14 @@ VLIB_NODE_FUNCTION_MULTIARCH_CLONE(dpdk_input_efd)
CLIB_MULTIARCH_SELECT_FN(dpdk_input);
CLIB_MULTIARCH_SELECT_FN(dpdk_input_rss);
CLIB_MULTIARCH_SELECT_FN(dpdk_input_efd);
+/* *INDENT-ON* */
/*
* Override the next nodes for the dpdk input nodes.
* Must be invoked prior to VLIB_INIT_FUNCTION calls.
*/
-void dpdk_set_next_node (dpdk_rx_next_t next, char *name)
+void
+dpdk_set_next_node (dpdk_rx_next_t next, char *name)
{
vlib_node_registration_t *r = &dpdk_input_node;
vlib_node_registration_t *r_handoff = &handoff_dispatch_node;
@@ -742,38 +851,49 @@ void dpdk_set_next_node (dpdk_rx_next_t next, char *name)
* Based on the operation type, set lower/upper bits for the given index value
*/
void
-set_efd_bitmap (u8 *bitmap, u32 value, u32 op)
+set_efd_bitmap (u8 * bitmap, u32 value, u32 op)
{
- int ix;
-
- *bitmap = 0;
- for (ix = 0; ix < 8; ix++) {
- if (((op == EFD_OPERATION_LESS_THAN) && (ix < value)) ||
- ((op == EFD_OPERATION_GREATER_OR_EQUAL) && (ix >= value))){
- (*bitmap) |= (1 << ix);
- }
+ int ix;
+
+ *bitmap = 0;
+ for (ix = 0; ix < 8; ix++)
+ {
+ if (((op == EFD_OPERATION_LESS_THAN) && (ix < value)) ||
+ ((op == EFD_OPERATION_GREATER_OR_EQUAL) && (ix >= value)))
+ {
+ (*bitmap) |= (1 << ix);
+ }
}
}
void
-efd_config (u32 enabled,
- u32 ip_prec, u32 ip_op,
- u32 mpls_exp, u32 mpls_op,
- u32 vlan_cos, u32 vlan_op)
+efd_config (u32 enabled,
+ u32 ip_prec, u32 ip_op,
+ u32 mpls_exp, u32 mpls_op, u32 vlan_cos, u32 vlan_op)
{
- vlib_thread_main_t * tm = vlib_get_thread_main();
- dpdk_main_t * dm = &dpdk_main;
-
- if (enabled) {
- tm->efd.enabled |= VLIB_EFD_DISCARD_ENABLED;
- dm->efd.enabled |= DPDK_EFD_DISCARD_ENABLED;
- } else {
- tm->efd.enabled &= ~VLIB_EFD_DISCARD_ENABLED;
- dm->efd.enabled &= ~DPDK_EFD_DISCARD_ENABLED;
- }
+ vlib_thread_main_t *tm = vlib_get_thread_main ();
+ dpdk_main_t *dm = &dpdk_main;
- set_efd_bitmap(&tm->efd.ip_prec_bitmap, ip_prec, ip_op);
- set_efd_bitmap(&tm->efd.mpls_exp_bitmap, mpls_exp, mpls_op);
- set_efd_bitmap(&tm->efd.vlan_cos_bitmap, vlan_cos, vlan_op);
+ if (enabled)
+ {
+ tm->efd.enabled |= VLIB_EFD_DISCARD_ENABLED;
+ dm->efd.enabled |= DPDK_EFD_DISCARD_ENABLED;
+ }
+ else
+ {
+ tm->efd.enabled &= ~VLIB_EFD_DISCARD_ENABLED;
+ dm->efd.enabled &= ~DPDK_EFD_DISCARD_ENABLED;
+ }
+ set_efd_bitmap (&tm->efd.ip_prec_bitmap, ip_prec, ip_op);
+ set_efd_bitmap (&tm->efd.mpls_exp_bitmap, mpls_exp, mpls_op);
+ set_efd_bitmap (&tm->efd.vlan_cos_bitmap, vlan_cos, vlan_op);
}
+
+/*
+ * fd.io coding-style-patch-verification: ON
+ *
+ * Local Variables:
+ * eval: (c-set-style "gnu")
+ * End:
+ */
diff --git a/vnet/vnet/devices/dpdk/vhost_user.c b/vnet/vnet/devices/dpdk/vhost_user.c
index 8335963d22b..c153e2ede78 100644
--- a/vnet/vnet/devices/dpdk/vhost_user.c
+++ b/vnet/vnet/devices/dpdk/vhost_user.c
@@ -41,7 +41,9 @@
#if DPDK_VHOST_USER
-static const char *vhost_message_str[] __attribute__((unused)) = {
+/* *INDENT-OFF* */
+static const char *vhost_message_str[] __attribute__ ((unused)) =
+{
[VHOST_USER_NONE] = "VHOST_USER_NONE",
[VHOST_USER_GET_FEATURES] = "VHOST_USER_GET_FEATURES",
[VHOST_USER_SET_FEATURES] = "VHOST_USER_SET_FEATURES",
@@ -56,21 +58,22 @@ static const char *vhost_message_str[] __attribute__((unused)) = {
[VHOST_USER_GET_VRING_BASE] = "VHOST_USER_GET_VRING_BASE",
[VHOST_USER_SET_VRING_KICK] = "VHOST_USER_SET_VRING_KICK",
[VHOST_USER_SET_VRING_CALL] = "VHOST_USER_SET_VRING_CALL",
- [VHOST_USER_SET_VRING_ERR] = "VHOST_USER_SET_VRING_ERR",
- [VHOST_USER_GET_PROTOCOL_FEATURES] = "VHOST_USER_GET_PROTOCOL_FEATURES",
- [VHOST_USER_SET_PROTOCOL_FEATURES] = "VHOST_USER_SET_PROTOCOL_FEATURES",
- [VHOST_USER_GET_QUEUE_NUM] = "VHOST_USER_GET_QUEUE_NUM",
- [VHOST_USER_SET_VRING_ENABLE] = "VHOST_USER_SET_VRING_ENABLE",
+ [VHOST_USER_SET_VRING_ERR] = "VHOST_USER_SET_VRING_ERR",
+ [VHOST_USER_GET_PROTOCOL_FEATURES] = "VHOST_USER_GET_PROTOCOL_FEATURES",
+ [VHOST_USER_SET_PROTOCOL_FEATURES] = "VHOST_USER_SET_PROTOCOL_FEATURES",
+ [VHOST_USER_GET_QUEUE_NUM] = "VHOST_USER_GET_QUEUE_NUM",
+ [VHOST_USER_SET_VRING_ENABLE] = "VHOST_USER_SET_VRING_ENABLE",
};
+/* *INDENT-ON* */
-static int dpdk_vhost_user_set_vring_enable(u32 hw_if_index,
- u8 idx, int enable);
+static int dpdk_vhost_user_set_vring_enable (u32 hw_if_index,
+ u8 idx, int enable);
/*
- * DPDK vhost-user functions
+ * DPDK vhost-user functions
*/
-/* portions taken from dpdk
+/* portions taken from dpdk
* BSD LICENSE
*
* Copyright(c) 2010-2014 Intel Corporation. All rights reserved.
@@ -105,34 +108,34 @@ static int dpdk_vhost_user_set_vring_enable(u32 hw_if_index,
static uword
-qva_to_vva(struct virtio_net *dev, uword qemu_va)
+qva_to_vva (struct virtio_net *dev, uword qemu_va)
{
struct virtio_memory_regions *region;
uword vhost_va = 0;
uint32_t regionidx = 0;
/* Find the region where the address lives. */
- for (regionidx = 0; regionidx < dev->mem->nregions; regionidx++) {
- region = &dev->mem->regions[regionidx];
- if ((qemu_va >= region->userspace_address) &&
- (qemu_va <= region->userspace_address +
- region->memory_size)) {
- vhost_va = qemu_va + region->guest_phys_address +
- region->address_offset -
- region->userspace_address;
- break;
+ for (regionidx = 0; regionidx < dev->mem->nregions; regionidx++)
+ {
+ region = &dev->mem->regions[regionidx];
+ if ((qemu_va >= region->userspace_address) &&
+ (qemu_va <= region->userspace_address + region->memory_size))
+ {
+ vhost_va = qemu_va + region->guest_phys_address +
+ region->address_offset - region->userspace_address;
+ break;
+ }
}
- }
return vhost_va;
}
static dpdk_device_t *
-dpdk_vhost_user_device_from_hw_if_index(u32 hw_if_index)
+dpdk_vhost_user_device_from_hw_if_index (u32 hw_if_index)
{
- vnet_main_t *vnm = vnet_get_main();
- dpdk_main_t * dm = &dpdk_main;
- vnet_hw_interface_t * hi = vnet_get_hw_interface (vnm, hw_if_index);
- dpdk_device_t * xd = vec_elt_at_index (dm->devices, hi->dev_instance);
+ vnet_main_t *vnm = vnet_get_main ();
+ dpdk_main_t *dm = &dpdk_main;
+ vnet_hw_interface_t *hi = vnet_get_hw_interface (vnm, hw_if_index);
+ dpdk_device_t *xd = vec_elt_at_index (dm->devices, hi->dev_instance);
if (xd->dev_type != VNET_DPDK_DEV_VHOST_USER)
return 0;
@@ -141,124 +144,143 @@ dpdk_vhost_user_device_from_hw_if_index(u32 hw_if_index)
}
static dpdk_device_t *
-dpdk_vhost_user_device_from_sw_if_index(u32 sw_if_index)
+dpdk_vhost_user_device_from_sw_if_index (u32 sw_if_index)
{
- vnet_main_t *vnm = vnet_get_main();
- vnet_sw_interface_t * sw = vnet_get_sw_interface (vnm, sw_if_index);
+ vnet_main_t *vnm = vnet_get_main ();
+ vnet_sw_interface_t *sw = vnet_get_sw_interface (vnm, sw_if_index);
ASSERT (sw->type == VNET_SW_INTERFACE_TYPE_HARDWARE);
- return dpdk_vhost_user_device_from_hw_if_index(sw->hw_if_index);
+ return dpdk_vhost_user_device_from_hw_if_index (sw->hw_if_index);
}
-static void stop_processing_packets(u32 hw_if_index, u8 idx)
+static void
+stop_processing_packets (u32 hw_if_index, u8 idx)
{
- dpdk_device_t *xd =
- dpdk_vhost_user_device_from_hw_if_index(hw_if_index);
- assert(xd);
+ dpdk_device_t *xd = dpdk_vhost_user_device_from_hw_if_index (hw_if_index);
+ assert (xd);
xd->vu_vhost_dev.virtqueue[idx]->enabled = 0;
}
-static void disable_interface(dpdk_device_t * xd)
+static void
+disable_interface (dpdk_device_t * xd)
{
u8 idx;
int numqs = xd->vu_vhost_dev.virt_qp_nb * VIRTIO_QNUM;
- for (idx = 0; idx < numqs; idx++)
+ for (idx = 0; idx < numqs; idx++)
xd->vu_vhost_dev.virtqueue[idx]->enabled = 0;
xd->vu_is_running = 0;
}
-static inline void * map_guest_mem(dpdk_device_t * xd, uword addr)
+static inline void *
+map_guest_mem (dpdk_device_t * xd, uword addr)
{
- dpdk_vu_intf_t * vui = xd->vu_intf;
- struct virtio_memory * mem = xd->vu_vhost_dev.mem;
+ dpdk_vu_intf_t *vui = xd->vu_intf;
+ struct virtio_memory *mem = xd->vu_vhost_dev.mem;
int i;
- for (i=0; i<mem->nregions; i++) {
- if ((mem->regions[i].guest_phys_address <= addr) &&
- ((mem->regions[i].guest_phys_address + mem->regions[i].memory_size) > addr)) {
- return (void *) ((uword)vui->region_addr[i] + addr - (uword)mem->regions[i].guest_phys_address);
- }
- }
- DBG_SOCK("failed to map guest mem addr %lx", addr);
+ for (i = 0; i < mem->nregions; i++)
+ {
+ if ((mem->regions[i].guest_phys_address <= addr) &&
+ ((mem->regions[i].guest_phys_address +
+ mem->regions[i].memory_size) > addr))
+ {
+ return (void *) ((uword) vui->region_addr[i] + addr -
+ (uword) mem->regions[i].guest_phys_address);
+ }
+ }
+ DBG_SOCK ("failed to map guest mem addr %lx", addr);
return 0;
}
static clib_error_t *
-dpdk_create_vhost_user_if_internal (u32 * hw_if_index, u32 if_id, u8 *hwaddr)
+dpdk_create_vhost_user_if_internal (u32 * hw_if_index, u32 if_id, u8 * hwaddr)
{
- dpdk_main_t * dm = &dpdk_main;
- vlib_main_t * vm = vlib_get_main();
- vlib_thread_main_t * tm = vlib_get_thread_main();
- vnet_sw_interface_t * sw;
- clib_error_t * error;
- dpdk_device_and_queue_t * dq;
+ dpdk_main_t *dm = &dpdk_main;
+ vlib_main_t *vm = vlib_get_main ();
+ vlib_thread_main_t *tm = vlib_get_thread_main ();
+ vnet_sw_interface_t *sw;
+ clib_error_t *error;
+ dpdk_device_and_queue_t *dq;
int num_qpairs = 1;
dpdk_vu_intf_t *vui = NULL;
num_qpairs = dm->use_rss < 1 ? 1 : tm->n_vlib_mains;
- dpdk_device_t * xd = NULL;
+ dpdk_device_t *xd = NULL;
u8 addr[6];
int j;
vlib_worker_thread_barrier_sync (vm);
- int inactive_cnt = vec_len(dm->vu_inactive_interfaces_device_index);
+ int inactive_cnt = vec_len (dm->vu_inactive_interfaces_device_index);
// if there are any inactive ifaces
- if (inactive_cnt > 0) {
- // take last
- u32 vui_idx = dm->vu_inactive_interfaces_device_index[inactive_cnt - 1];
- if (vec_len(dm->devices) > vui_idx) {
- xd = vec_elt_at_index (dm->devices, vui_idx);
- if (xd->dev_type == VNET_DPDK_DEV_VHOST_USER) {
- DBG_SOCK("reusing inactive vhost-user interface sw_if_index %d", xd->vlib_sw_if_index);
- } else {
- clib_warning("error: inactive vhost-user interface sw_if_index %d not VHOST_USER type!",
- xd->vlib_sw_if_index);
- // reset so new interface is created
- xd = NULL;
- }
+ if (inactive_cnt > 0)
+ {
+ // take last
+ u32 vui_idx = dm->vu_inactive_interfaces_device_index[inactive_cnt - 1];
+ if (vec_len (dm->devices) > vui_idx)
+ {
+ xd = vec_elt_at_index (dm->devices, vui_idx);
+ if (xd->dev_type == VNET_DPDK_DEV_VHOST_USER)
+ {
+ DBG_SOCK
+ ("reusing inactive vhost-user interface sw_if_index %d",
+ xd->vlib_sw_if_index);
+ }
+ else
+ {
+ clib_warning
+ ("error: inactive vhost-user interface sw_if_index %d not VHOST_USER type!",
+ xd->vlib_sw_if_index);
+ // reset so new interface is created
+ xd = NULL;
+ }
+ }
+ // "remove" from inactive list
+ _vec_len (dm->vu_inactive_interfaces_device_index) -= 1;
}
- // "remove" from inactive list
- _vec_len(dm->vu_inactive_interfaces_device_index) -= 1;
- }
- if (xd) {
+ if (xd)
+ {
// existing interface used - do not overwrite if_id if not needed
- if (if_id != (u32)~0)
- xd->vu_if_id = if_id;
+ if (if_id != (u32) ~ 0)
+ xd->vu_if_id = if_id;
// reset virtqueues
vui = xd->vu_intf;
- for (j = 0; j < num_qpairs * VIRTIO_QNUM; j++) {
- memset(xd->vu_vhost_dev.virtqueue[j], 0, sizeof(struct vhost_virtqueue));
- xd->vu_vhost_dev.virtqueue[j]->kickfd = -1;
- xd->vu_vhost_dev.virtqueue[j]->callfd = -1;
- xd->vu_vhost_dev.virtqueue[j]->backend = -1;
- vui->vrings[j].packets = 0;
- vui->vrings[j].bytes = 0;
- }
+ for (j = 0; j < num_qpairs * VIRTIO_QNUM; j++)
+ {
+ memset (xd->vu_vhost_dev.virtqueue[j], 0,
+ sizeof (struct vhost_virtqueue));
+ xd->vu_vhost_dev.virtqueue[j]->kickfd = -1;
+ xd->vu_vhost_dev.virtqueue[j]->callfd = -1;
+ xd->vu_vhost_dev.virtqueue[j]->backend = -1;
+ vui->vrings[j].packets = 0;
+ vui->vrings[j].bytes = 0;
+ }
// reset lockp
- dpdk_device_lock_free(xd);
- dpdk_device_lock_init(xd);
+ dpdk_device_lock_free (xd);
+ dpdk_device_lock_init (xd);
// reset tx vectors
for (j = 0; j < tm->n_vlib_mains; j++)
- {
- vec_validate_ha (xd->tx_vectors[j], DPDK_TX_RING_SIZE,
- sizeof(tx_ring_hdr_t), CLIB_CACHE_LINE_BYTES);
- vec_reset_length (xd->tx_vectors[j]);
- }
+ {
+ vec_validate_ha (xd->tx_vectors[j], DPDK_TX_RING_SIZE,
+ sizeof (tx_ring_hdr_t), CLIB_CACHE_LINE_BYTES);
+ vec_reset_length (xd->tx_vectors[j]);
+ }
// reset rx vector
for (j = 0; j < xd->rx_q_used; j++)
- {
- vec_validate_aligned (xd->rx_vectors[j], VLIB_FRAME_SIZE-1,
- CLIB_CACHE_LINE_BYTES);
- vec_reset_length (xd->rx_vectors[j]);
- }
- } else {
+ {
+ vec_validate_aligned (xd->rx_vectors[j], VLIB_FRAME_SIZE - 1,
+ CLIB_CACHE_LINE_BYTES);
+ vec_reset_length (xd->rx_vectors[j]);
+ }
+ }
+ else
+ {
// vui was not retrieved from inactive ifaces - create new
vec_add2_aligned (dm->devices, xd, 1, CLIB_CACHE_LINE_BYTES);
xd->dev_type = VNET_DPDK_DEV_VHOST_USER;
@@ -266,86 +288,93 @@ dpdk_create_vhost_user_if_internal (u32 * hw_if_index, u32 if_id, u8 *hwaddr)
xd->tx_q_used = num_qpairs;
xd->vu_vhost_dev.virt_qp_nb = num_qpairs;
- vec_validate_aligned (xd->rx_vectors, xd->rx_q_used, CLIB_CACHE_LINE_BYTES);
+ vec_validate_aligned (xd->rx_vectors, xd->rx_q_used,
+ CLIB_CACHE_LINE_BYTES);
- if (if_id == (u32)~0)
- xd->vu_if_id = dm->next_vu_if_id++;
+ if (if_id == (u32) ~ 0)
+ xd->vu_if_id = dm->next_vu_if_id++;
else
- xd->vu_if_id = if_id;
+ xd->vu_if_id = if_id;
xd->device_index = xd - dm->devices;
xd->per_interface_next_index = ~0;
- xd->vu_intf = clib_mem_alloc (sizeof(*(xd->vu_intf)));
+ xd->vu_intf = clib_mem_alloc (sizeof (*(xd->vu_intf)));
- xd->vu_vhost_dev.mem = clib_mem_alloc (sizeof(struct virtio_memory) +
- VHOST_MEMORY_MAX_NREGIONS *
- sizeof(struct virtio_memory_regions));
+ xd->vu_vhost_dev.mem = clib_mem_alloc (sizeof (struct virtio_memory) +
+ VHOST_MEMORY_MAX_NREGIONS *
+ sizeof (struct
+ virtio_memory_regions));
/* Will be set when guest sends VHOST_USER_SET_MEM_TABLE cmd */
xd->vu_vhost_dev.mem->nregions = 0;
- /*
+ /*
* New virtqueue structure is an array of VHOST_MAX_QUEUE_PAIRS * 2
* We need to allocate numq pairs.
*/
vui = xd->vu_intf;
- for (j = 0; j < num_qpairs * VIRTIO_QNUM; j++) {
- xd->vu_vhost_dev.virtqueue[j] = clib_mem_alloc (sizeof(struct vhost_virtqueue));
- memset(xd->vu_vhost_dev.virtqueue[j], 0, sizeof(struct vhost_virtqueue));
- xd->vu_vhost_dev.virtqueue[j]->kickfd = -1;
- xd->vu_vhost_dev.virtqueue[j]->callfd = -1;
- xd->vu_vhost_dev.virtqueue[j]->backend = -1;
- vui->vrings[j].packets = 0;
- vui->vrings[j].bytes = 0;
- }
-
- dpdk_device_lock_init(xd);
-
- DBG_SOCK("tm->n_vlib_mains: %d. TX %d, RX: %d, num_qpairs: %d, Lock: %p",
- tm->n_vlib_mains, xd->tx_q_used, xd->rx_q_used, num_qpairs, xd->lockp);
+ for (j = 0; j < num_qpairs * VIRTIO_QNUM; j++)
+ {
+ xd->vu_vhost_dev.virtqueue[j] =
+ clib_mem_alloc (sizeof (struct vhost_virtqueue));
+ memset (xd->vu_vhost_dev.virtqueue[j], 0,
+ sizeof (struct vhost_virtqueue));
+ xd->vu_vhost_dev.virtqueue[j]->kickfd = -1;
+ xd->vu_vhost_dev.virtqueue[j]->callfd = -1;
+ xd->vu_vhost_dev.virtqueue[j]->backend = -1;
+ vui->vrings[j].packets = 0;
+ vui->vrings[j].bytes = 0;
+ }
+
+ dpdk_device_lock_init (xd);
+
+ DBG_SOCK
+ ("tm->n_vlib_mains: %d. TX %d, RX: %d, num_qpairs: %d, Lock: %p",
+ tm->n_vlib_mains, xd->tx_q_used, xd->rx_q_used, num_qpairs,
+ xd->lockp);
vec_validate_aligned (xd->tx_vectors, tm->n_vlib_mains,
- CLIB_CACHE_LINE_BYTES);
+ CLIB_CACHE_LINE_BYTES);
for (j = 0; j < tm->n_vlib_mains; j++)
- {
- vec_validate_ha (xd->tx_vectors[j], DPDK_TX_RING_SIZE,
- sizeof(tx_ring_hdr_t), CLIB_CACHE_LINE_BYTES);
- vec_reset_length (xd->tx_vectors[j]);
- }
+ {
+ vec_validate_ha (xd->tx_vectors[j], DPDK_TX_RING_SIZE,
+ sizeof (tx_ring_hdr_t), CLIB_CACHE_LINE_BYTES);
+ vec_reset_length (xd->tx_vectors[j]);
+ }
// reset rx vector
for (j = 0; j < xd->rx_q_used; j++)
- {
- vec_validate_aligned (xd->rx_vectors[j], VLIB_FRAME_SIZE-1,
- CLIB_CACHE_LINE_BYTES);
- vec_reset_length (xd->rx_vectors[j]);
- }
+ {
+ vec_validate_aligned (xd->rx_vectors[j], VLIB_FRAME_SIZE - 1,
+ CLIB_CACHE_LINE_BYTES);
+ vec_reset_length (xd->rx_vectors[j]);
+ }
- }
+ }
/*
* Generate random MAC address for the interface
*/
- if (hwaddr) {
- clib_memcpy(addr, hwaddr, sizeof(addr));
- } else {
- f64 now = vlib_time_now(vm);
- u32 rnd;
- rnd = (u32) (now * 1e6);
- rnd = random_u32 (&rnd);
-
- clib_memcpy (addr+2, &rnd, sizeof(rnd));
- addr[0] = 2;
- addr[1] = 0xfe;
- }
+ if (hwaddr)
+ {
+ clib_memcpy (addr, hwaddr, sizeof (addr));
+ }
+ else
+ {
+ f64 now = vlib_time_now (vm);
+ u32 rnd;
+ rnd = (u32) (now * 1e6);
+ rnd = random_u32 (&rnd);
+
+ clib_memcpy (addr + 2, &rnd, sizeof (rnd));
+ addr[0] = 2;
+ addr[1] = 0xfe;
+ }
error = ethernet_register_interface
- (dm->vnet_main,
- dpdk_device_class.index,
- xd->device_index,
+ (dm->vnet_main, dpdk_device_class.index, xd->device_index,
/* ethernet address */ addr,
- &xd->vlib_hw_if_index,
- 0);
+ &xd->vlib_hw_if_index, 0);
if (error)
return error;
@@ -355,61 +384,64 @@ dpdk_create_vhost_user_if_internal (u32 * hw_if_index, u32 if_id, u8 *hwaddr)
*hw_if_index = xd->vlib_hw_if_index;
- DBG_SOCK("xd->device_index: %d, dm->input_cpu_count: %d, "
- "dm->input_cpu_first_index: %d\n", xd->device_index,
- dm->input_cpu_count, dm->input_cpu_first_index);
+ DBG_SOCK ("xd->device_index: %d, dm->input_cpu_count: %d, "
+ "dm->input_cpu_first_index: %d\n", xd->device_index,
+ dm->input_cpu_count, dm->input_cpu_first_index);
int q, next_cpu = 0;
- for (q = 0; q < num_qpairs; q++) {
- int cpu = dm->input_cpu_first_index +
- (next_cpu % dm->input_cpu_count);
+ for (q = 0; q < num_qpairs; q++)
+ {
+ int cpu = dm->input_cpu_first_index + (next_cpu % dm->input_cpu_count);
unsigned lcore = vlib_worker_threads[cpu].dpdk_lcore_id;
- vec_validate(xd->cpu_socket_id_by_queue, q);
- xd->cpu_socket_id_by_queue[q] = rte_lcore_to_socket_id(lcore);
+ vec_validate (xd->cpu_socket_id_by_queue, q);
+ xd->cpu_socket_id_by_queue[q] = rte_lcore_to_socket_id (lcore);
- vec_add2(dm->devices_by_cpu[cpu], dq, 1);
+ vec_add2 (dm->devices_by_cpu[cpu], dq, 1);
dq->device = xd->device_index;
dq->queue_id = q;
- DBG_SOCK("CPU for %d = %d. QID: %d", *hw_if_index, cpu, dq->queue_id);
+ DBG_SOCK ("CPU for %d = %d. QID: %d", *hw_if_index, cpu, dq->queue_id);
// start polling if it was not started yet (because of no phys ifaces)
- if (tm->n_vlib_mains == 1 && dpdk_input_node.state != VLIB_NODE_STATE_POLLING)
- vlib_node_set_state (vm, dpdk_input_node.index, VLIB_NODE_STATE_POLLING);
+ if (tm->n_vlib_mains == 1
+ && dpdk_input_node.state != VLIB_NODE_STATE_POLLING)
+ vlib_node_set_state (vm, dpdk_input_node.index,
+ VLIB_NODE_STATE_POLLING);
if (tm->n_vlib_mains > 1)
- vlib_node_set_state (vlib_mains[cpu], dpdk_input_node.index,
- VLIB_NODE_STATE_POLLING);
+ vlib_node_set_state (vlib_mains[cpu], dpdk_input_node.index,
+ VLIB_NODE_STATE_POLLING);
next_cpu++;
- }
+ }
vlib_worker_thread_barrier_release (vm);
return 0;
}
#if RTE_VERSION >= RTE_VERSION_NUM(16, 4, 0, 0)
-static long get_huge_page_size(int fd)
+static long
+get_huge_page_size (int fd)
{
struct statfs s;
- fstatfs(fd, &s);
+ fstatfs (fd, &s);
return s.f_bsize;
}
#endif
static clib_error_t *
-dpdk_vhost_user_set_protocol_features(u32 hw_if_index, u64 prot_features)
+dpdk_vhost_user_set_protocol_features (u32 hw_if_index, u64 prot_features)
{
- dpdk_device_t * xd;
- xd = dpdk_vhost_user_device_from_hw_if_index(hw_if_index);
- assert(xd);
+ dpdk_device_t *xd;
+ xd = dpdk_vhost_user_device_from_hw_if_index (hw_if_index);
+ assert (xd);
xd->vu_vhost_dev.protocol_features = prot_features;
return 0;
}
static clib_error_t *
-dpdk_vhost_user_get_features(u32 hw_if_index, u64 * features)
+dpdk_vhost_user_get_features (u32 hw_if_index, u64 * features)
{
- *features = rte_vhost_feature_get();
+ *features = rte_vhost_feature_get ();
#if RTE_VERSION >= RTE_VERSION_NUM(16, 4, 0, 0)
#define OFFLOAD_FEATURES ((1ULL << VIRTIO_NET_F_HOST_TSO4) | \
@@ -426,33 +458,34 @@ dpdk_vhost_user_get_features(u32 hw_if_index, u64 * features)
*features &= (~OFFLOAD_FEATURES);
#endif
- DBG_SOCK("supported features: 0x%lx", *features);
+ DBG_SOCK ("supported features: 0x%lx", *features);
return 0;
}
static clib_error_t *
-dpdk_vhost_user_set_features(u32 hw_if_index, u64 features)
+dpdk_vhost_user_set_features (u32 hw_if_index, u64 features)
{
- dpdk_device_t * xd;
- u16 hdr_len = sizeof(struct virtio_net_hdr);
+ dpdk_device_t *xd;
+ u16 hdr_len = sizeof (struct virtio_net_hdr);
- if (!(xd = dpdk_vhost_user_device_from_hw_if_index(hw_if_index))) {
- clib_warning("not a vhost-user interface");
- return 0;
- }
+ if (!(xd = dpdk_vhost_user_device_from_hw_if_index (hw_if_index)))
+ {
+ clib_warning ("not a vhost-user interface");
+ return 0;
+ }
xd->vu_vhost_dev.features = features;
if (xd->vu_vhost_dev.features & (1 << VIRTIO_NET_F_MRG_RXBUF))
- hdr_len = sizeof(struct virtio_net_hdr_mrg_rxbuf);
+ hdr_len = sizeof (struct virtio_net_hdr_mrg_rxbuf);
int numqs = VIRTIO_QNUM;
u8 idx;
- int prot_feature = features &
- (1ULL << VHOST_USER_F_PROTOCOL_FEATURES);
+ int prot_feature = features & (1ULL << VHOST_USER_F_PROTOCOL_FEATURES);
numqs = xd->vu_vhost_dev.virt_qp_nb * VIRTIO_QNUM;
- for (idx = 0; idx < numqs; idx++) {
+ for (idx = 0; idx < numqs; idx++)
+ {
xd->vu_vhost_dev.virtqueue[idx]->vhost_hlen = hdr_len;
/*
* Spec says, if F_PROTOCOL_FEATURE is not set by the
@@ -460,122 +493,134 @@ dpdk_vhost_user_set_features(u32 hw_if_index, u64 features)
* enabled. If slave negotiates F_PROTOCOL_FEATURE, then
* slave is responsible to enable it.
*/
- if (! prot_feature)
- dpdk_vhost_user_set_vring_enable(hw_if_index, idx, 1);
- }
+ if (!prot_feature)
+ dpdk_vhost_user_set_vring_enable (hw_if_index, idx, 1);
+ }
return 0;
}
static clib_error_t *
-dpdk_vhost_user_set_mem_table(u32 hw_if_index, vhost_user_memory_t * vum, int fd[])
+dpdk_vhost_user_set_mem_table (u32 hw_if_index, vhost_user_memory_t * vum,
+ int fd[])
{
- struct virtio_memory * mem;
+ struct virtio_memory *mem;
int i;
- dpdk_device_t * xd;
- dpdk_vu_intf_t * vui;
+ dpdk_device_t *xd;
+ dpdk_vu_intf_t *vui;
- if (!(xd = dpdk_vhost_user_device_from_hw_if_index(hw_if_index))) {
- clib_warning("not a vhost-user interface");
- return 0;
- }
+ if (!(xd = dpdk_vhost_user_device_from_hw_if_index (hw_if_index)))
+ {
+ clib_warning ("not a vhost-user interface");
+ return 0;
+ }
vui = xd->vu_intf;
mem = xd->vu_vhost_dev.mem;
mem->nregions = vum->nregions;
- for (i=0; i < mem->nregions; i++) {
- u64 mapped_size, mapped_address;
-
- mem->regions[i].guest_phys_address = vum->regions[i].guest_phys_addr;
- mem->regions[i].guest_phys_address_end = vum->regions[i].guest_phys_addr +
- vum->regions[i].memory_size;
- mem->regions[i].memory_size = vum->regions[i].memory_size;
- mem->regions[i].userspace_address = vum->regions[i].userspace_addr;
-
- mapped_size = mem->regions[i].memory_size + vum->regions[i].mmap_offset;
- mapped_address = pointer_to_uword(mmap(NULL, mapped_size, PROT_READ | PROT_WRITE, MAP_SHARED, fd[i], 0));
-
- if (uword_to_pointer(mapped_address, void*) == MAP_FAILED)
+ for (i = 0; i < mem->nregions; i++)
{
- clib_warning("mmap error");
- return 0;
+ u64 mapped_size, mapped_address;
+
+ mem->regions[i].guest_phys_address = vum->regions[i].guest_phys_addr;
+ mem->regions[i].guest_phys_address_end =
+ vum->regions[i].guest_phys_addr + vum->regions[i].memory_size;
+ mem->regions[i].memory_size = vum->regions[i].memory_size;
+ mem->regions[i].userspace_address = vum->regions[i].userspace_addr;
+
+ mapped_size = mem->regions[i].memory_size + vum->regions[i].mmap_offset;
+ mapped_address =
+ pointer_to_uword (mmap
+ (NULL, mapped_size, PROT_READ | PROT_WRITE,
+ MAP_SHARED, fd[i], 0));
+
+ if (uword_to_pointer (mapped_address, void *) == MAP_FAILED)
+ {
+ clib_warning ("mmap error");
+ return 0;
+ }
+
+ mapped_address += vum->regions[i].mmap_offset;
+ vui->region_addr[i] = mapped_address;
+ vui->region_fd[i] = fd[i];
+ vui->region_offset[i] = vum->regions[i].mmap_offset;
+ mem->regions[i].address_offset =
+ mapped_address - mem->regions[i].guest_phys_address;
+
+ DBG_SOCK ("map memory region %d addr 0x%lx off 0x%lx len 0x%lx",
+ i, vui->region_addr[i], vui->region_offset[i], mapped_size);
+
+ if (vum->regions[i].guest_phys_addr == 0)
+ {
+ mem->base_address = vum->regions[i].userspace_addr;
+ mem->mapped_address = mem->regions[i].address_offset;
+ }
}
- mapped_address += vum->regions[i].mmap_offset;
- vui->region_addr[i] = mapped_address;
- vui->region_fd[i] = fd[i];
- vui->region_offset[i] = vum->regions[i].mmap_offset;
- mem->regions[i].address_offset = mapped_address - mem->regions[i].guest_phys_address;
-
- DBG_SOCK("map memory region %d addr 0x%lx off 0x%lx len 0x%lx",
- i, vui->region_addr[i], vui->region_offset[i], mapped_size);
-
- if (vum->regions[i].guest_phys_addr == 0) {
- mem->base_address = vum->regions[i].userspace_addr;
- mem->mapped_address = mem->regions[i].address_offset;
- }
- }
-
- disable_interface(xd);
+ disable_interface (xd);
return 0;
}
static clib_error_t *
-dpdk_vhost_user_set_vring_num(u32 hw_if_index, u8 idx, u32 num)
+dpdk_vhost_user_set_vring_num (u32 hw_if_index, u8 idx, u32 num)
{
- dpdk_device_t * xd;
+ dpdk_device_t *xd;
struct vhost_virtqueue *vq;
- DBG_SOCK("idx %u num %u", idx, num);
+ DBG_SOCK ("idx %u num %u", idx, num);
- if (!(xd = dpdk_vhost_user_device_from_hw_if_index(hw_if_index))) {
- clib_warning("not a vhost-user interface");
- return 0;
- }
+ if (!(xd = dpdk_vhost_user_device_from_hw_if_index (hw_if_index)))
+ {
+ clib_warning ("not a vhost-user interface");
+ return 0;
+ }
vq = xd->vu_vhost_dev.virtqueue[idx];
vq->size = num;
- stop_processing_packets(hw_if_index, idx);
+ stop_processing_packets (hw_if_index, idx);
return 0;
}
static clib_error_t *
-dpdk_vhost_user_set_vring_addr(u32 hw_if_index, u8 idx, uword desc, \
- uword used, uword avail, uword log)
+dpdk_vhost_user_set_vring_addr (u32 hw_if_index, u8 idx, uword desc,
+ uword used, uword avail, uword log)
{
- dpdk_device_t * xd;
+ dpdk_device_t *xd;
struct vhost_virtqueue *vq;
- DBG_SOCK("idx %u desc 0x%lx used 0x%lx avail 0x%lx log 0x%lx",
- idx, desc, used, avail, log);
+ DBG_SOCK ("idx %u desc 0x%lx used 0x%lx avail 0x%lx log 0x%lx",
+ idx, desc, used, avail, log);
- if (!(xd = dpdk_vhost_user_device_from_hw_if_index(hw_if_index))) {
- clib_warning("not a vhost-user interface");
- return 0;
- }
+ if (!(xd = dpdk_vhost_user_device_from_hw_if_index (hw_if_index)))
+ {
+ clib_warning ("not a vhost-user interface");
+ return 0;
+ }
vq = xd->vu_vhost_dev.virtqueue[idx];
- vq->desc = (struct vring_desc *) qva_to_vva(&xd->vu_vhost_dev, desc);
- vq->used = (struct vring_used *) qva_to_vva(&xd->vu_vhost_dev, used);
- vq->avail = (struct vring_avail *) qva_to_vva(&xd->vu_vhost_dev, avail);
+ vq->desc = (struct vring_desc *) qva_to_vva (&xd->vu_vhost_dev, desc);
+ vq->used = (struct vring_used *) qva_to_vva (&xd->vu_vhost_dev, used);
+ vq->avail = (struct vring_avail *) qva_to_vva (&xd->vu_vhost_dev, avail);
#if RTE_VERSION >= RTE_VERSION_NUM(16, 4, 0, 0)
vq->log_guest_addr = log;
#endif
- if (!(vq->desc && vq->used && vq->avail)) {
- clib_warning("falied to set vring addr");
- }
+ if (!(vq->desc && vq->used && vq->avail))
+ {
+ clib_warning ("falied to set vring addr");
+ }
- if (vq->last_used_idx != vq->used->idx) {
- clib_warning("last_used_idx (%u) and vq->used->idx (%u) mismatches; "
- "some packets maybe resent for Tx and dropped for Rx",
- vq->last_used_idx, vq->used->idx);
- vq->last_used_idx = vq->used->idx;
+ if (vq->last_used_idx != vq->used->idx)
+ {
+ clib_warning ("last_used_idx (%u) and vq->used->idx (%u) mismatches; "
+ "some packets maybe resent for Tx and dropped for Rx",
+ vq->last_used_idx, vq->used->idx);
+ vq->last_used_idx = vq->used->idx;
vq->last_used_idx_res = vq->used->idx;
- }
+ }
/*
* Inform the guest that there is no need to inform (kick) the
@@ -585,22 +630,23 @@ dpdk_vhost_user_set_vring_addr(u32 hw_if_index, u8 idx, uword desc, \
* The below function sets a flag in used table. Therefore,
* should be initialized after initializing vq->used.
*/
- rte_vhost_enable_guest_notification(&xd->vu_vhost_dev, idx, 0);
- stop_processing_packets(hw_if_index, idx);
+ rte_vhost_enable_guest_notification (&xd->vu_vhost_dev, idx, 0);
+ stop_processing_packets (hw_if_index, idx);
return 0;
}
static clib_error_t *
-dpdk_vhost_user_get_vring_base(u32 hw_if_index, u8 idx, u32 * num)
+dpdk_vhost_user_get_vring_base (u32 hw_if_index, u8 idx, u32 * num)
{
- dpdk_device_t * xd;
+ dpdk_device_t *xd;
struct vhost_virtqueue *vq;
- if (!(xd = dpdk_vhost_user_device_from_hw_if_index(hw_if_index))) {
- clib_warning("not a vhost-user interface");
- return 0;
- }
+ if (!(xd = dpdk_vhost_user_device_from_hw_if_index (hw_if_index)))
+ {
+ clib_warning ("not a vhost-user interface");
+ return 0;
+ }
vq = xd->vu_vhost_dev.virtqueue[idx];
*num = vq->last_used_idx;
@@ -612,10 +658,10 @@ dpdk_vhost_user_get_vring_base(u32 hw_if_index, u8 idx, u32 * num)
* on the descriptor specified by VHOST_USER_SET_VRING_KICK,
* and stop ring upon receiving VHOST_USER_GET_VRING_BASE.
*/
- DBG_SOCK("Stopping vring Q %u of device %d", idx, hw_if_index);
+ DBG_SOCK ("Stopping vring Q %u of device %d", idx, hw_if_index);
dpdk_vu_intf_t *vui = xd->vu_intf;
- vui->vrings[idx].enabled = 0; /* Reset local copy */
- vui->vrings[idx].callfd = -1; /* Reset FD */
+ vui->vrings[idx].enabled = 0; /* Reset local copy */
+ vui->vrings[idx].callfd = -1; /* Reset FD */
vq->enabled = 0;
vq->desc = NULL;
vq->used = NULL;
@@ -626,55 +672,59 @@ dpdk_vhost_user_get_vring_base(u32 hw_if_index, u8 idx, u32 * num)
/* Check if all Qs are disabled */
int numqs = xd->vu_vhost_dev.virt_qp_nb * VIRTIO_QNUM;
- for (idx = 0; idx < numqs; idx++) {
- if (xd->vu_vhost_dev.virtqueue[idx]->enabled)
- break;
- }
+ for (idx = 0; idx < numqs; idx++)
+ {
+ if (xd->vu_vhost_dev.virtqueue[idx]->enabled)
+ break;
+ }
/* If all vrings are disabed then disable device */
- if (idx == numqs) {
- DBG_SOCK("Device %d disabled", hw_if_index);
+ if (idx == numqs)
+ {
+ DBG_SOCK ("Device %d disabled", hw_if_index);
xd->vu_is_running = 0;
- }
+ }
return 0;
}
static clib_error_t *
-dpdk_vhost_user_set_vring_base(u32 hw_if_index, u8 idx, u32 num)
+dpdk_vhost_user_set_vring_base (u32 hw_if_index, u8 idx, u32 num)
{
- dpdk_device_t * xd;
+ dpdk_device_t *xd;
struct vhost_virtqueue *vq;
- DBG_SOCK("idx %u num %u", idx, num);
+ DBG_SOCK ("idx %u num %u", idx, num);
- if (!(xd = dpdk_vhost_user_device_from_hw_if_index(hw_if_index))) {
- clib_warning("not a vhost-user interface");
- return 0;
- }
+ if (!(xd = dpdk_vhost_user_device_from_hw_if_index (hw_if_index)))
+ {
+ clib_warning ("not a vhost-user interface");
+ return 0;
+ }
vq = xd->vu_vhost_dev.virtqueue[idx];
vq->last_used_idx = num;
vq->last_used_idx_res = num;
- stop_processing_packets(hw_if_index, idx);
+ stop_processing_packets (hw_if_index, idx);
return 0;
}
static clib_error_t *
-dpdk_vhost_user_set_vring_kick(u32 hw_if_index, u8 idx, int fd)
+dpdk_vhost_user_set_vring_kick (u32 hw_if_index, u8 idx, int fd)
{
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_t * xd;
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_t *xd;
dpdk_vu_vring *vring;
struct vhost_virtqueue *vq0, *vq1, *vq;
int index, vu_is_running = 0;
- if (!(xd = dpdk_vhost_user_device_from_hw_if_index(hw_if_index))) {
- clib_warning("not a vhost-user interface");
- return 0;
- }
+ if (!(xd = dpdk_vhost_user_device_from_hw_if_index (hw_if_index)))
+ {
+ clib_warning ("not a vhost-user interface");
+ return 0;
+ }
vq = xd->vu_vhost_dev.virtqueue[idx];
vq->kickfd = fd;
@@ -689,39 +739,43 @@ dpdk_vhost_user_set_vring_kick(u32 hw_if_index, u8 idx, int fd)
int numqs = VIRTIO_QNUM;
numqs = xd->vu_vhost_dev.virt_qp_nb * VIRTIO_QNUM;
- for (index = 0; index < numqs; index += 2) {
- vq0 = xd->vu_vhost_dev.virtqueue[index]; /* RX */
- vq1 = xd->vu_vhost_dev.virtqueue[index + 1]; /* TX */
- if (vq0->enabled && vq1->enabled)
+ for (index = 0; index < numqs; index += 2)
{
- vu_is_running = 1;
- break;
+ vq0 = xd->vu_vhost_dev.virtqueue[index]; /* RX */
+ vq1 = xd->vu_vhost_dev.virtqueue[index + 1]; /* TX */
+ if (vq0->enabled && vq1->enabled)
+ {
+ vu_is_running = 1;
+ break;
+ }
}
- }
- DBG_SOCK("SET_VRING_KICK - idx %d, running %d, fd: %d",
- idx, vu_is_running, fd);
+ DBG_SOCK ("SET_VRING_KICK - idx %d, running %d, fd: %d",
+ idx, vu_is_running, fd);
xd->vu_is_running = vu_is_running;
- if (xd->vu_is_running && xd->admin_up) {
- vnet_hw_interface_set_flags (dm->vnet_main,
- xd->vlib_hw_if_index, VNET_HW_INTERFACE_FLAG_LINK_UP |
- ETH_LINK_FULL_DUPLEX );
- }
+ if (xd->vu_is_running && xd->admin_up)
+ {
+ vnet_hw_interface_set_flags (dm->vnet_main,
+ xd->vlib_hw_if_index,
+ VNET_HW_INTERFACE_FLAG_LINK_UP |
+ ETH_LINK_FULL_DUPLEX);
+ }
return 0;
}
static int
-dpdk_vhost_user_set_vring_enable(u32 hw_if_index, u8 idx, int enable)
+dpdk_vhost_user_set_vring_enable (u32 hw_if_index, u8 idx, int enable)
{
- dpdk_device_t * xd;
+ dpdk_device_t *xd;
struct vhost_virtqueue *vq;
dpdk_vu_intf_t *vui;
- if (!(xd = dpdk_vhost_user_device_from_hw_if_index(hw_if_index))) {
- clib_warning("not a vhost-user interface");
- return 0;
- }
+ if (!(xd = dpdk_vhost_user_device_from_hw_if_index (hw_if_index)))
+ {
+ clib_warning ("not a vhost-user interface");
+ return 0;
+ }
vui = xd->vu_intf;
/*
@@ -731,15 +785,16 @@ dpdk_vhost_user_set_vring_enable(u32 hw_if_index, u8 idx, int enable)
* are set. If not, vq will be enabled when vring
* is kicked.
*/
- vui->vrings[idx].enabled = enable; /* Save local copy */
+ vui->vrings[idx].enabled = enable; /* Save local copy */
int numqs = xd->vu_vhost_dev.virt_qp_nb * VIRTIO_QNUM;
- while (numqs--) {
- if (! vui->vrings[numqs].enabled)
- break;
- }
+ while (numqs--)
+ {
+ if (!vui->vrings[numqs].enabled)
+ break;
+ }
- if (numqs == -1) /* All Qs are enabled */
+ if (numqs == -1) /* All Qs are enabled */
xd->need_txlock = 0;
else
xd->need_txlock = 1;
@@ -751,199 +806,215 @@ dpdk_vhost_user_set_vring_enable(u32 hw_if_index, u8 idx, int enable)
return 0;
}
-static clib_error_t * dpdk_vhost_user_callfd_read_ready (unix_file_t * uf)
+static clib_error_t *
+dpdk_vhost_user_callfd_read_ready (unix_file_t * uf)
{
- __attribute__((unused)) int n;
+ __attribute__ ((unused)) int n;
u8 buff[8];
- n = read(uf->file_descriptor, ((char*)&buff), 8);
+ n = read (uf->file_descriptor, ((char *) &buff), 8);
return 0;
}
static clib_error_t *
-dpdk_vhost_user_set_vring_call(u32 hw_if_index, u8 idx, int fd)
+dpdk_vhost_user_set_vring_call (u32 hw_if_index, u8 idx, int fd)
{
- dpdk_device_t * xd;
+ dpdk_device_t *xd;
struct vhost_virtqueue *vq;
- unix_file_t template = {0};
+ unix_file_t template = { 0 };
- DBG_SOCK("SET_VRING_CALL - idx %d, fd %d", idx, fd);
+ DBG_SOCK ("SET_VRING_CALL - idx %d, fd %d", idx, fd);
- if (!(xd = dpdk_vhost_user_device_from_hw_if_index(hw_if_index))) {
- clib_warning("not a vhost-user interface");
- return 0;
- }
+ if (!(xd = dpdk_vhost_user_device_from_hw_if_index (hw_if_index)))
+ {
+ clib_warning ("not a vhost-user interface");
+ return 0;
+ }
dpdk_vu_intf_t *vui = xd->vu_intf;
/* if there is old fd, delete it */
- if (vui->vrings[idx].callfd > 0) {
- unix_file_t * uf = pool_elt_at_index (unix_main.file_pool,
- vui->vrings[idx].callfd_idx);
- unix_file_del (&unix_main, uf);
- }
+ if (vui->vrings[idx].callfd > 0)
+ {
+ unix_file_t *uf = pool_elt_at_index (unix_main.file_pool,
+ vui->vrings[idx].callfd_idx);
+ unix_file_del (&unix_main, uf);
+ }
vui->vrings[idx].callfd = fd;
template.read_function = dpdk_vhost_user_callfd_read_ready;
template.file_descriptor = fd;
vui->vrings[idx].callfd_idx = unix_file_add (&unix_main, &template);
vq = xd->vu_vhost_dev.virtqueue[idx];
- vq->callfd = -1; /* We use locally saved vring->callfd; */
+ vq->callfd = -1; /* We use locally saved vring->callfd; */
return 0;
}
u8
-dpdk_vhost_user_want_interrupt(dpdk_device_t *xd, int idx)
+dpdk_vhost_user_want_interrupt (dpdk_device_t * xd, int idx)
{
- dpdk_vu_intf_t *vui = xd->vu_intf;
- ASSERT(vui != NULL);
+ dpdk_vu_intf_t *vui = xd->vu_intf;
+ ASSERT (vui != NULL);
- if (PREDICT_FALSE(vui->num_vrings <= 0))
- return 0;
+ if (PREDICT_FALSE (vui->num_vrings <= 0))
+ return 0;
- dpdk_vu_vring *vring = &(vui->vrings[idx]);
- struct vhost_virtqueue *vq = xd->vu_vhost_dev.virtqueue[idx];
+ dpdk_vu_vring *vring = &(vui->vrings[idx]);
+ struct vhost_virtqueue *vq = xd->vu_vhost_dev.virtqueue[idx];
- /* return if vm is interested in interrupts */
- return (vring->callfd > 0) && !(vq->avail->flags & VRING_AVAIL_F_NO_INTERRUPT);
+ /* return if vm is interested in interrupts */
+ return (vring->callfd > 0)
+ && !(vq->avail->flags & VRING_AVAIL_F_NO_INTERRUPT);
}
void
-dpdk_vhost_user_send_interrupt(vlib_main_t * vm, dpdk_device_t * xd, int idx)
+dpdk_vhost_user_send_interrupt (vlib_main_t * vm, dpdk_device_t * xd, int idx)
{
- dpdk_main_t * dm = &dpdk_main;
- dpdk_vu_intf_t *vui = xd->vu_intf;
- ASSERT(vui != NULL);
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_vu_intf_t *vui = xd->vu_intf;
+ ASSERT (vui != NULL);
- if (PREDICT_FALSE(vui->num_vrings <= 0))
- return;
+ if (PREDICT_FALSE (vui->num_vrings <= 0))
+ return;
- dpdk_vu_vring *vring = &(vui->vrings[idx]);
- struct vhost_virtqueue *vq = xd->vu_vhost_dev.virtqueue[idx];
+ dpdk_vu_vring *vring = &(vui->vrings[idx]);
+ struct vhost_virtqueue *vq = xd->vu_vhost_dev.virtqueue[idx];
- /* if vm is interested in interrupts */
- if((vring->callfd > 0) && !(vq->avail->flags & VRING_AVAIL_F_NO_INTERRUPT)) {
- eventfd_write(vring->callfd, (eventfd_t)1);
- vring->n_since_last_int = 0;
- vring->int_deadline = vlib_time_now(vm) + dm->conf->vhost_coalesce_time;
+ /* if vm is interested in interrupts */
+ if ((vring->callfd > 0) && !(vq->avail->flags & VRING_AVAIL_F_NO_INTERRUPT))
+ {
+ eventfd_write (vring->callfd, (eventfd_t) 1);
+ vring->n_since_last_int = 0;
+ vring->int_deadline =
+ vlib_time_now (vm) + dm->conf->vhost_coalesce_time;
}
}
/*
- * vhost-user interface management functions
+ * vhost-user interface management functions
*/
// initialize vui with specified attributes
-static void
-dpdk_vhost_user_vui_init(vnet_main_t * vnm,
- dpdk_device_t *xd, int sockfd,
- const char * sock_filename,
- u8 is_server, u64 feature_mask,
- u32 * sw_if_index)
+static void
+dpdk_vhost_user_vui_init (vnet_main_t * vnm,
+ dpdk_device_t * xd, int sockfd,
+ const char *sock_filename,
+ u8 is_server, u64 feature_mask, u32 * sw_if_index)
{
dpdk_vu_intf_t *vui = xd->vu_intf;
- memset(vui, 0, sizeof(*vui));
+ memset (vui, 0, sizeof (*vui));
vui->unix_fd = sockfd;
vui->num_vrings = xd->vu_vhost_dev.virt_qp_nb * VIRTIO_QNUM;
- DBG_SOCK("dpdk_vhost_user_vui_init VRINGS: %d", vui->num_vrings);
+ DBG_SOCK ("dpdk_vhost_user_vui_init VRINGS: %d", vui->num_vrings);
vui->sock_is_server = is_server;
- strncpy(vui->sock_filename, sock_filename, ARRAY_LEN(vui->sock_filename)-1);
+ strncpy (vui->sock_filename, sock_filename,
+ ARRAY_LEN (vui->sock_filename) - 1);
vui->sock_errno = 0;
vui->is_up = 0;
vui->feature_mask = feature_mask;
vui->active = 1;
vui->unix_file_index = ~0;
- vnet_hw_interface_set_flags (vnm, xd->vlib_hw_if_index, 0);
+ vnet_hw_interface_set_flags (vnm, xd->vlib_hw_if_index, 0);
if (sw_if_index)
- *sw_if_index = xd->vlib_sw_if_index;
+ *sw_if_index = xd->vlib_sw_if_index;
}
// register vui and start polling on it
-static void
-dpdk_vhost_user_vui_register(vlib_main_t * vm, dpdk_device_t *xd)
+static void
+dpdk_vhost_user_vui_register (vlib_main_t * vm, dpdk_device_t * xd)
{
- dpdk_main_t * dm = &dpdk_main;
+ dpdk_main_t *dm = &dpdk_main;
dpdk_vu_intf_t *vui = xd->vu_intf;
hash_set (dm->vu_sw_if_index_by_listener_fd, vui->unix_fd,
- xd->vlib_sw_if_index);
+ xd->vlib_sw_if_index);
}
-static void dpdk_unmap_all_mem_regions(dpdk_device_t * xd)
+static void
+dpdk_unmap_all_mem_regions (dpdk_device_t * xd)
{
int i, r;
dpdk_vu_intf_t *vui = xd->vu_intf;
- struct virtio_memory * mem = xd->vu_vhost_dev.mem;
+ struct virtio_memory *mem = xd->vu_vhost_dev.mem;
- for (i=0; i<mem->nregions; i++) {
- if (vui->region_addr[i] != -1) {
+ for (i = 0; i < mem->nregions; i++)
+ {
+ if (vui->region_addr[i] != -1)
+ {
- long page_sz = get_huge_page_size(vui->region_fd[i]);
+ long page_sz = get_huge_page_size (vui->region_fd[i]);
- ssize_t map_sz = RTE_ALIGN_CEIL(mem->regions[i].memory_size +
- vui->region_offset[i], page_sz);
+ ssize_t map_sz = RTE_ALIGN_CEIL (mem->regions[i].memory_size +
+ vui->region_offset[i], page_sz);
- r = munmap((void *)(vui->region_addr[i] - vui->region_offset[i]), map_sz);
+ r =
+ munmap ((void *) (vui->region_addr[i] - vui->region_offset[i]),
+ map_sz);
- DBG_SOCK("unmap memory region %d addr 0x%lx off 0x%lx len 0x%lx page_sz 0x%x",
- i, vui->region_addr[i], vui->region_offset[i], map_sz, page_sz);
+ DBG_SOCK
+ ("unmap memory region %d addr 0x%lx off 0x%lx len 0x%lx page_sz 0x%x",
+ i, vui->region_addr[i], vui->region_offset[i], map_sz, page_sz);
- vui->region_addr[i]= -1;
+ vui->region_addr[i] = -1;
- if (r == -1) {
- clib_unix_warning("failed to unmap memory region");
- }
- close(vui->region_fd[i]);
+ if (r == -1)
+ {
+ clib_unix_warning ("failed to unmap memory region");
+ }
+ close (vui->region_fd[i]);
+ }
}
- }
mem->nregions = 0;
}
static inline void
-dpdk_vhost_user_if_disconnect(dpdk_device_t * xd)
+dpdk_vhost_user_if_disconnect (dpdk_device_t * xd)
{
- dpdk_vu_intf_t *vui = xd->vu_intf;
- vnet_main_t * vnm = vnet_get_main();
- dpdk_main_t * dm = &dpdk_main;
- struct vhost_virtqueue *vq;
- int q;
-
- xd->admin_up = 0;
- vnet_hw_interface_set_flags (vnm, xd->vlib_hw_if_index, 0);
-
- if (vui->unix_file_index != ~0) {
- unix_file_del (&unix_main, unix_main.file_pool + vui->unix_file_index);
- vui->unix_file_index = ~0;
+ dpdk_vu_intf_t *vui = xd->vu_intf;
+ vnet_main_t *vnm = vnet_get_main ();
+ dpdk_main_t *dm = &dpdk_main;
+ struct vhost_virtqueue *vq;
+ int q;
+
+ xd->admin_up = 0;
+ vnet_hw_interface_set_flags (vnm, xd->vlib_hw_if_index, 0);
+
+ if (vui->unix_file_index != ~0)
+ {
+ unix_file_del (&unix_main, unix_main.file_pool + vui->unix_file_index);
+ vui->unix_file_index = ~0;
}
- hash_unset(dm->vu_sw_if_index_by_sock_fd, vui->unix_fd);
- hash_unset(dm->vu_sw_if_index_by_listener_fd, vui->unix_fd);
- close(vui->unix_fd);
- vui->unix_fd = -1;
- vui->is_up = 0;
-
- for (q = 0; q < vui->num_vrings; q++) {
- vq = xd->vu_vhost_dev.virtqueue[q];
- vui->vrings[q].enabled = 0; /* Reset local copy */
- vui->vrings[q].callfd = -1; /* Reset FD */
- vq->enabled = 0;
+ hash_unset (dm->vu_sw_if_index_by_sock_fd, vui->unix_fd);
+ hash_unset (dm->vu_sw_if_index_by_listener_fd, vui->unix_fd);
+ close (vui->unix_fd);
+ vui->unix_fd = -1;
+ vui->is_up = 0;
+
+ for (q = 0; q < vui->num_vrings; q++)
+ {
+ vq = xd->vu_vhost_dev.virtqueue[q];
+ vui->vrings[q].enabled = 0; /* Reset local copy */
+ vui->vrings[q].callfd = -1; /* Reset FD */
+ vq->enabled = 0;
#if RTE_VERSION >= RTE_VERSION_NUM(16, 4, 0, 0)
- vq->log_guest_addr = 0;
+ vq->log_guest_addr = 0;
#endif
- vq->desc = NULL;
- vq->used = NULL;
- vq->avail = NULL;
+ vq->desc = NULL;
+ vq->used = NULL;
+ vq->avail = NULL;
}
- xd->vu_is_running = 0;
+ xd->vu_is_running = 0;
- dpdk_unmap_all_mem_regions(xd);
- DBG_SOCK("interface ifindex %d disconnected", xd->vlib_sw_if_index);
+ dpdk_unmap_all_mem_regions (xd);
+ DBG_SOCK ("interface ifindex %d disconnected", xd->vlib_sw_if_index);
}
-static clib_error_t * dpdk_vhost_user_socket_read (unix_file_t * uf)
+static clib_error_t *
+dpdk_vhost_user_socket_read (unix_file_t * uf)
{
int n;
int fd, number_of_fds = 0;
@@ -951,30 +1022,30 @@ static clib_error_t * dpdk_vhost_user_socket_read (unix_file_t * uf)
vhost_user_msg_t msg;
struct msghdr mh;
struct iovec iov[1];
- dpdk_main_t * dm = &dpdk_main;
+ dpdk_main_t *dm = &dpdk_main;
dpdk_device_t *xd;
dpdk_vu_intf_t *vui;
struct cmsghdr *cmsg;
- uword * p;
+ uword *p;
u8 q;
- vnet_main_t * vnm = vnet_get_main();
+ vnet_main_t *vnm = vnet_get_main ();
p = hash_get (dm->vu_sw_if_index_by_sock_fd, uf->file_descriptor);
- if (p == 0) {
- DBG_SOCK ("FD %d doesn't belong to any interface",
- uf->file_descriptor);
+ if (p == 0)
+ {
+ DBG_SOCK ("FD %d doesn't belong to any interface", uf->file_descriptor);
return 0;
}
else
- xd = dpdk_vhost_user_device_from_sw_if_index(p[0]);
+ xd = dpdk_vhost_user_device_from_sw_if_index (p[0]);
- ASSERT(xd != NULL);
+ ASSERT (xd != NULL);
vui = xd->vu_intf;
- char control[CMSG_SPACE(VHOST_MEMORY_MAX_NREGIONS * sizeof(int))];
+ char control[CMSG_SPACE (VHOST_MEMORY_MAX_NREGIONS * sizeof (int))];
- memset(&mh, 0, sizeof(mh));
- memset(control, 0, sizeof(control));
+ memset (&mh, 0, sizeof (mh));
+ memset (control, 0, sizeof (control));
/* set the payload */
iov[0].iov_base = (void *) &msg;
@@ -983,127 +1054,134 @@ static clib_error_t * dpdk_vhost_user_socket_read (unix_file_t * uf)
mh.msg_iov = iov;
mh.msg_iovlen = 1;
mh.msg_control = control;
- mh.msg_controllen = sizeof(control);
+ mh.msg_controllen = sizeof (control);
- n = recvmsg(uf->file_descriptor, &mh, 0);
+ n = recvmsg (uf->file_descriptor, &mh, 0);
if (n != VHOST_USER_MSG_HDR_SZ)
goto close_socket;
- if (mh.msg_flags & MSG_CTRUNC) {
- goto close_socket;
- }
+ if (mh.msg_flags & MSG_CTRUNC)
+ {
+ goto close_socket;
+ }
- cmsg = CMSG_FIRSTHDR(&mh);
+ cmsg = CMSG_FIRSTHDR (&mh);
if (cmsg && (cmsg->cmsg_len > 0) && (cmsg->cmsg_level == SOL_SOCKET) &&
(cmsg->cmsg_type == SCM_RIGHTS) &&
- (cmsg->cmsg_len - CMSG_LEN(0) <= VHOST_MEMORY_MAX_NREGIONS * sizeof(int))) {
- number_of_fds = (cmsg->cmsg_len - CMSG_LEN(0)) / sizeof(int);
- clib_memcpy(fds, CMSG_DATA(cmsg), number_of_fds * sizeof(int));
- }
+ (cmsg->cmsg_len - CMSG_LEN (0) <=
+ VHOST_MEMORY_MAX_NREGIONS * sizeof (int)))
+ {
+ number_of_fds = (cmsg->cmsg_len - CMSG_LEN (0)) / sizeof (int);
+ clib_memcpy (fds, CMSG_DATA (cmsg), number_of_fds * sizeof (int));
+ }
- /* version 1, no reply bit set*/
- if ((msg.flags & 7) != 1) {
- DBG_SOCK("malformed message received. closing socket");
- goto close_socket;
- }
+ /* version 1, no reply bit set */
+ if ((msg.flags & 7) != 1)
+ {
+ DBG_SOCK ("malformed message received. closing socket");
+ goto close_socket;
+ }
{
- int rv __attribute__((unused));
- /* $$$$ pay attention to rv */
- rv = read(uf->file_descriptor, ((char*)&msg) + n, msg.size);
+ int rv __attribute__ ((unused));
+ /* $$$$ pay attention to rv */
+ rv = read (uf->file_descriptor, ((char *) &msg) + n, msg.size);
}
- DBG_SOCK("VPP VHOST message %s", vhost_message_str[msg.request]);
- switch (msg.request) {
+ DBG_SOCK ("VPP VHOST message %s", vhost_message_str[msg.request]);
+ switch (msg.request)
+ {
case VHOST_USER_GET_FEATURES:
- DBG_SOCK("if %d msg VHOST_USER_GET_FEATURES",
- xd->vlib_hw_if_index);
+ DBG_SOCK ("if %d msg VHOST_USER_GET_FEATURES", xd->vlib_hw_if_index);
msg.flags |= VHOST_USER_REPLY_MASK;
- dpdk_vhost_user_get_features(xd->vlib_hw_if_index, &msg.u64);
+ dpdk_vhost_user_get_features (xd->vlib_hw_if_index, &msg.u64);
msg.u64 &= vui->feature_mask;
- msg.size = sizeof(msg.u64);
+ msg.size = sizeof (msg.u64);
break;
case VHOST_USER_SET_FEATURES:
- DBG_SOCK("if %d msg VHOST_USER_SET_FEATURES features 0x%016lx",
- xd->vlib_hw_if_index, msg.u64);
+ DBG_SOCK ("if %d msg VHOST_USER_SET_FEATURES features 0x%016lx",
+ xd->vlib_hw_if_index, msg.u64);
- dpdk_vhost_user_set_features(xd->vlib_hw_if_index, msg.u64);
+ dpdk_vhost_user_set_features (xd->vlib_hw_if_index, msg.u64);
break;
case VHOST_USER_SET_MEM_TABLE:
- DBG_SOCK("if %d msg VHOST_USER_SET_MEM_TABLE nregions %d",
- xd->vlib_hw_if_index, msg.memory.nregions);
+ DBG_SOCK ("if %d msg VHOST_USER_SET_MEM_TABLE nregions %d",
+ xd->vlib_hw_if_index, msg.memory.nregions);
if ((msg.memory.nregions < 1) ||
- (msg.memory.nregions > VHOST_MEMORY_MAX_NREGIONS)) {
+ (msg.memory.nregions > VHOST_MEMORY_MAX_NREGIONS))
+ {
- DBG_SOCK("number of mem regions must be between 1 and %i",
- VHOST_MEMORY_MAX_NREGIONS);
+ DBG_SOCK ("number of mem regions must be between 1 and %i",
+ VHOST_MEMORY_MAX_NREGIONS);
- goto close_socket;
- }
+ goto close_socket;
+ }
- if (msg.memory.nregions != number_of_fds) {
- DBG_SOCK("each memory region must have FD");
- goto close_socket;
- }
+ if (msg.memory.nregions != number_of_fds)
+ {
+ DBG_SOCK ("each memory region must have FD");
+ goto close_socket;
+ }
- dpdk_vhost_user_set_mem_table(xd->vlib_hw_if_index, &msg.memory, fds);
+ dpdk_vhost_user_set_mem_table (xd->vlib_hw_if_index, &msg.memory, fds);
break;
case VHOST_USER_SET_VRING_NUM:
- DBG_SOCK("if %d msg VHOST_USER_SET_VRING_NUM idx %d num %d",
- xd->vlib_hw_if_index, msg.state.index, msg.state.num);
+ DBG_SOCK ("if %d msg VHOST_USER_SET_VRING_NUM idx %d num %d",
+ xd->vlib_hw_if_index, msg.state.index, msg.state.num);
- if ((msg.state.num > 32768) || /* maximum ring size is 32768 */
- (msg.state.num == 0) || /* it cannot be zero */
- (msg.state.num % 2)) /* must be power of 2 */
- goto close_socket;
+ if ((msg.state.num > 32768) || /* maximum ring size is 32768 */
+ (msg.state.num == 0) || /* it cannot be zero */
+ (msg.state.num % 2)) /* must be power of 2 */
+ goto close_socket;
- dpdk_vhost_user_set_vring_num(xd->vlib_hw_if_index, msg.state.index, msg.state.num);
+ dpdk_vhost_user_set_vring_num (xd->vlib_hw_if_index, msg.state.index,
+ msg.state.num);
break;
case VHOST_USER_SET_VRING_ADDR:
- DBG_SOCK("if %d msg VHOST_USER_SET_VRING_ADDR idx %d",
- xd->vlib_hw_if_index, msg.state.index);
-
- dpdk_vhost_user_set_vring_addr(xd->vlib_hw_if_index, msg.state.index,
- msg.addr.desc_user_addr,
- msg.addr.used_user_addr,
- msg.addr.avail_user_addr,
- msg.addr.log_guest_addr);
+ DBG_SOCK ("if %d msg VHOST_USER_SET_VRING_ADDR idx %d",
+ xd->vlib_hw_if_index, msg.state.index);
+
+ dpdk_vhost_user_set_vring_addr (xd->vlib_hw_if_index, msg.state.index,
+ msg.addr.desc_user_addr,
+ msg.addr.used_user_addr,
+ msg.addr.avail_user_addr,
+ msg.addr.log_guest_addr);
break;
case VHOST_USER_SET_OWNER:
- DBG_SOCK("if %d msg VHOST_USER_SET_OWNER",
- xd->vlib_hw_if_index);
+ DBG_SOCK ("if %d msg VHOST_USER_SET_OWNER", xd->vlib_hw_if_index);
break;
case VHOST_USER_RESET_OWNER:
- DBG_SOCK("if %d msg VHOST_USER_RESET_OWNER",
- xd->vlib_hw_if_index);
+ DBG_SOCK ("if %d msg VHOST_USER_RESET_OWNER", xd->vlib_hw_if_index);
break;
case VHOST_USER_SET_VRING_CALL:
q = (u8) (msg.u64 & 0xFF);
- DBG_SOCK("if %d msg VHOST_USER_SET_VRING_CALL u64 %lx, idx: %d",
- xd->vlib_hw_if_index, msg.u64, q);
+ DBG_SOCK ("if %d msg VHOST_USER_SET_VRING_CALL u64 %lx, idx: %d",
+ xd->vlib_hw_if_index, msg.u64, q);
if (!(msg.u64 & 0x100))
- {
- if (number_of_fds != 1)
- goto close_socket;
- fd = fds[0];
- } else {
- fd = -1;
- }
- dpdk_vhost_user_set_vring_call(xd->vlib_hw_if_index, q, fd);
+ {
+ if (number_of_fds != 1)
+ goto close_socket;
+ fd = fds[0];
+ }
+ else
+ {
+ fd = -1;
+ }
+ dpdk_vhost_user_set_vring_call (xd->vlib_hw_if_index, q, fd);
break;
@@ -1111,229 +1189,239 @@ static clib_error_t * dpdk_vhost_user_socket_read (unix_file_t * uf)
q = (u8) (msg.u64 & 0xFF);
- DBG_SOCK("if %d msg VHOST_USER_SET_VRING_KICK u64 %lx, idx: %d",
- xd->vlib_hw_if_index, msg.u64, q);
+ DBG_SOCK ("if %d msg VHOST_USER_SET_VRING_KICK u64 %lx, idx: %d",
+ xd->vlib_hw_if_index, msg.u64, q);
if (!(msg.u64 & 0x100))
- {
- if (number_of_fds != 1)
- goto close_socket;
+ {
+ if (number_of_fds != 1)
+ goto close_socket;
- vui->vrings[q].kickfd = fds[0];
- }
+ vui->vrings[q].kickfd = fds[0];
+ }
else
- vui->vrings[q].kickfd = -1;
+ vui->vrings[q].kickfd = -1;
- dpdk_vhost_user_set_vring_kick(xd->vlib_hw_if_index, q, vui->vrings[q].kickfd);
+ dpdk_vhost_user_set_vring_kick (xd->vlib_hw_if_index, q,
+ vui->vrings[q].kickfd);
break;
case VHOST_USER_SET_VRING_ERR:
q = (u8) (msg.u64 & 0xFF);
- DBG_SOCK("if %d msg VHOST_USER_SET_VRING_ERR u64 %lx, idx: %d",
- xd->vlib_hw_if_index, msg.u64, q);
+ DBG_SOCK ("if %d msg VHOST_USER_SET_VRING_ERR u64 %lx, idx: %d",
+ xd->vlib_hw_if_index, msg.u64, q);
if (!(msg.u64 & 0x100))
- {
- if (number_of_fds != 1)
- goto close_socket;
+ {
+ if (number_of_fds != 1)
+ goto close_socket;
- fd = fds[0];
- }
+ fd = fds[0];
+ }
else
- fd = -1;
+ fd = -1;
vui->vrings[q].errfd = fd;
break;
case VHOST_USER_SET_VRING_BASE:
- DBG_SOCK("if %d msg VHOST_USER_SET_VRING_BASE idx %d num %d",
- xd->vlib_hw_if_index, msg.state.index, msg.state.num);
+ DBG_SOCK ("if %d msg VHOST_USER_SET_VRING_BASE idx %d num %d",
+ xd->vlib_hw_if_index, msg.state.index, msg.state.num);
- dpdk_vhost_user_set_vring_base(xd->vlib_hw_if_index, msg.state.index, msg.state.num);
+ dpdk_vhost_user_set_vring_base (xd->vlib_hw_if_index, msg.state.index,
+ msg.state.num);
break;
case VHOST_USER_GET_VRING_BASE:
- DBG_SOCK("if %d msg VHOST_USER_GET_VRING_BASE idx %d num %d",
- xd->vlib_hw_if_index, msg.state.index, msg.state.num);
+ DBG_SOCK ("if %d msg VHOST_USER_GET_VRING_BASE idx %d num %d",
+ xd->vlib_hw_if_index, msg.state.index, msg.state.num);
msg.flags |= VHOST_USER_REPLY_MASK;
- msg.size = sizeof(msg.state);
+ msg.size = sizeof (msg.state);
- dpdk_vhost_user_get_vring_base(xd->vlib_hw_if_index, msg.state.index, &msg.state.num);
+ dpdk_vhost_user_get_vring_base (xd->vlib_hw_if_index, msg.state.index,
+ &msg.state.num);
break;
case VHOST_USER_NONE:
- DBG_SOCK("if %d msg VHOST_USER_NONE",
- xd->vlib_hw_if_index);
+ DBG_SOCK ("if %d msg VHOST_USER_NONE", xd->vlib_hw_if_index);
break;
case VHOST_USER_SET_LOG_BASE:
#if RTE_VERSION >= RTE_VERSION_NUM(16, 4, 0, 0)
- DBG_SOCK("if %d msg VHOST_USER_SET_LOG_BASE",
- xd->vlib_hw_if_index);
-
- if (msg.size != sizeof(msg.log)) {
- DBG_SOCK("invalid msg size for VHOST_USER_SET_LOG_BASE: %u instead of %lu",
- msg.size, sizeof(msg.log));
- goto close_socket;
- }
-
- if (!(xd->vu_vhost_dev.protocol_features & (1 << VHOST_USER_PROTOCOL_F_LOG_SHMFD))) {
- DBG_SOCK("VHOST_USER_PROTOCOL_F_LOG_SHMFD not set but VHOST_USER_SET_LOG_BASE received");
- goto close_socket;
- }
+ DBG_SOCK ("if %d msg VHOST_USER_SET_LOG_BASE", xd->vlib_hw_if_index);
+
+ if (msg.size != sizeof (msg.log))
+ {
+ DBG_SOCK
+ ("invalid msg size for VHOST_USER_SET_LOG_BASE: %u instead of %lu",
+ msg.size, sizeof (msg.log));
+ goto close_socket;
+ }
+
+ if (!
+ (xd->vu_vhost_dev.protocol_features & (1 <<
+ VHOST_USER_PROTOCOL_F_LOG_SHMFD)))
+ {
+ DBG_SOCK
+ ("VHOST_USER_PROTOCOL_F_LOG_SHMFD not set but VHOST_USER_SET_LOG_BASE received");
+ goto close_socket;
+ }
fd = fds[0];
/* align size to 2M page */
- long page_sz = get_huge_page_size(fd);
- ssize_t map_sz = RTE_ALIGN_CEIL(msg.log.size + msg.log.offset, page_sz);
+ long page_sz = get_huge_page_size (fd);
+ ssize_t map_sz =
+ RTE_ALIGN_CEIL (msg.log.size + msg.log.offset, page_sz);
- void *addr = mmap(0, map_sz, PROT_READ | PROT_WRITE,
- MAP_SHARED, fd, 0);
+ void *addr = mmap (0, map_sz, PROT_READ | PROT_WRITE,
+ MAP_SHARED, fd, 0);
- DBG_SOCK("map log region addr 0 len 0x%lx off 0x%lx fd %d mapped %p",
- map_sz, msg.log.offset, fd, addr);
+ DBG_SOCK ("map log region addr 0 len 0x%lx off 0x%lx fd %d mapped %p",
+ map_sz, msg.log.offset, fd, addr);
- if (addr == MAP_FAILED) {
- clib_warning("failed to map memory. errno is %d", errno);
- goto close_socket;
- }
+ if (addr == MAP_FAILED)
+ {
+ clib_warning ("failed to map memory. errno is %d", errno);
+ goto close_socket;
+ }
- xd->vu_vhost_dev.log_base += pointer_to_uword(addr) + msg.log.offset;
+ xd->vu_vhost_dev.log_base += pointer_to_uword (addr) + msg.log.offset;
xd->vu_vhost_dev.log_size = msg.log.size;
msg.flags |= VHOST_USER_REPLY_MASK;
- msg.size = sizeof(msg.u64);
+ msg.size = sizeof (msg.u64);
#else
- DBG_SOCK("if %d msg VHOST_USER_SET_LOG_BASE Not-Implemented",
- xd->vlib_hw_if_index);
+ DBG_SOCK ("if %d msg VHOST_USER_SET_LOG_BASE Not-Implemented",
+ xd->vlib_hw_if_index);
#endif
break;
case VHOST_USER_SET_LOG_FD:
- DBG_SOCK("if %d msg VHOST_USER_SET_LOG_FD",
- xd->vlib_hw_if_index);
+ DBG_SOCK ("if %d msg VHOST_USER_SET_LOG_FD", xd->vlib_hw_if_index);
break;
case VHOST_USER_GET_PROTOCOL_FEATURES:
- DBG_SOCK("if %d msg VHOST_USER_GET_PROTOCOL_FEATURES",
- xd->vlib_hw_if_index);
+ DBG_SOCK ("if %d msg VHOST_USER_GET_PROTOCOL_FEATURES",
+ xd->vlib_hw_if_index);
msg.flags |= VHOST_USER_REPLY_MASK;
msg.u64 = VHOST_USER_PROTOCOL_FEATURES;
- DBG_SOCK("VHOST_USER_PROTOCOL_FEATURES: %llx", VHOST_USER_PROTOCOL_FEATURES);
- msg.size = sizeof(msg.u64);
+ DBG_SOCK ("VHOST_USER_PROTOCOL_FEATURES: %llx",
+ VHOST_USER_PROTOCOL_FEATURES);
+ msg.size = sizeof (msg.u64);
break;
case VHOST_USER_SET_PROTOCOL_FEATURES:
- DBG_SOCK("if %d msg VHOST_USER_SET_PROTOCOL_FEATURES",
- xd->vlib_hw_if_index);
+ DBG_SOCK ("if %d msg VHOST_USER_SET_PROTOCOL_FEATURES",
+ xd->vlib_hw_if_index);
- DBG_SOCK("VHOST_USER_SET_PROTOCOL_FEATURES: 0x%lx",
- msg.u64);
- dpdk_vhost_user_set_protocol_features(xd->vlib_hw_if_index,
- msg.u64);
+ DBG_SOCK ("VHOST_USER_SET_PROTOCOL_FEATURES: 0x%lx", msg.u64);
+ dpdk_vhost_user_set_protocol_features (xd->vlib_hw_if_index, msg.u64);
break;
case VHOST_USER_SET_VRING_ENABLE:
- DBG_SOCK("%d VPP VHOST_USER_SET_VRING_ENABLE IDX: %d, Enable: %d",
- xd->vlib_hw_if_index, msg.state.index, msg.state.num);
+ DBG_SOCK ("%d VPP VHOST_USER_SET_VRING_ENABLE IDX: %d, Enable: %d",
+ xd->vlib_hw_if_index, msg.state.index, msg.state.num);
dpdk_vhost_user_set_vring_enable
- (xd->vlib_hw_if_index, msg.state.index, msg.state.num);
+ (xd->vlib_hw_if_index, msg.state.index, msg.state.num);
break;
case VHOST_USER_GET_QUEUE_NUM:
- DBG_SOCK("if %d msg VHOST_USER_GET_QUEUE_NUM:",
- xd->vlib_hw_if_index);
+ DBG_SOCK ("if %d msg VHOST_USER_GET_QUEUE_NUM:", xd->vlib_hw_if_index);
msg.flags |= VHOST_USER_REPLY_MASK;
msg.u64 = xd->vu_vhost_dev.virt_qp_nb;
- msg.size = sizeof(msg.u64);
+ msg.size = sizeof (msg.u64);
break;
default:
- DBG_SOCK("unknown vhost-user message %d received. closing socket",
- msg.request);
+ DBG_SOCK ("unknown vhost-user message %d received. closing socket",
+ msg.request);
goto close_socket;
- }
+ }
- /* if we have pointers to descriptor table, go up*/
+ /* if we have pointers to descriptor table, go up */
if (!vui->is_up &&
xd->vu_vhost_dev.virtqueue[VHOST_NET_VRING_IDX_TX]->desc &&
- xd->vu_vhost_dev.virtqueue[VHOST_NET_VRING_IDX_RX]->desc) {
+ xd->vu_vhost_dev.virtqueue[VHOST_NET_VRING_IDX_RX]->desc)
+ {
- DBG_SOCK("interface %d connected", xd->vlib_sw_if_index);
+ DBG_SOCK ("interface %d connected", xd->vlib_sw_if_index);
- vnet_hw_interface_set_flags (vnm, xd->vlib_hw_if_index, VNET_HW_INTERFACE_FLAG_LINK_UP);
+ vnet_hw_interface_set_flags (vnm, xd->vlib_hw_if_index,
+ VNET_HW_INTERFACE_FLAG_LINK_UP);
vui->is_up = 1;
xd->admin_up = 1;
- }
+ }
/* if we need to reply */
if (msg.flags & VHOST_USER_REPLY_MASK)
- {
- n = send(uf->file_descriptor, &msg, VHOST_USER_MSG_HDR_SZ + msg.size, 0);
+ {
+ n =
+ send (uf->file_descriptor, &msg, VHOST_USER_MSG_HDR_SZ + msg.size, 0);
if (n != (msg.size + VHOST_USER_MSG_HDR_SZ))
- goto close_socket;
- }
+ goto close_socket;
+ }
return 0;
close_socket:
- DBG_SOCK("error: close_socket");
- dpdk_vhost_user_if_disconnect(xd);
+ DBG_SOCK ("error: close_socket");
+ dpdk_vhost_user_if_disconnect (xd);
return 0;
}
-static clib_error_t * dpdk_vhost_user_socket_error (unix_file_t * uf)
+static clib_error_t *
+dpdk_vhost_user_socket_error (unix_file_t * uf)
{
- dpdk_main_t * dm = &dpdk_main;
+ dpdk_main_t *dm = &dpdk_main;
dpdk_device_t *xd;
- uword * p;
+ uword *p;
p = hash_get (dm->vu_sw_if_index_by_sock_fd, uf->file_descriptor);
- if (p == 0) {
- DBG_SOCK ("FD %d doesn't belong to any interface",
- uf->file_descriptor);
+ if (p == 0)
+ {
+ DBG_SOCK ("FD %d doesn't belong to any interface", uf->file_descriptor);
return 0;
}
else
- xd = dpdk_vhost_user_device_from_sw_if_index(p[0]);
+ xd = dpdk_vhost_user_device_from_sw_if_index (p[0]);
- dpdk_vhost_user_if_disconnect(xd);
+ dpdk_vhost_user_if_disconnect (xd);
return 0;
}
-static clib_error_t * dpdk_vhost_user_socksvr_accept_ready (unix_file_t * uf)
+static clib_error_t *
+dpdk_vhost_user_socksvr_accept_ready (unix_file_t * uf)
{
int client_fd, client_len;
struct sockaddr_un client;
- unix_file_t template = {0};
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_t * xd = NULL;
- dpdk_vu_intf_t * vui;
- uword * p;
-
- p = hash_get (dm->vu_sw_if_index_by_listener_fd,
- uf->file_descriptor);
- if (p == 0) {
- DBG_SOCK ("fd %d doesn't belong to any interface",
- uf->file_descriptor);
+ unix_file_t template = { 0 };
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_t *xd = NULL;
+ dpdk_vu_intf_t *vui;
+ uword *p;
+
+ p = hash_get (dm->vu_sw_if_index_by_listener_fd, uf->file_descriptor);
+ if (p == 0)
+ {
+ DBG_SOCK ("fd %d doesn't belong to any interface", uf->file_descriptor);
return 0;
}
- xd = dpdk_vhost_user_device_from_sw_if_index(p[0]);
- ASSERT(xd != NULL);
+ xd = dpdk_vhost_user_device_from_sw_if_index (p[0]);
+ ASSERT (xd != NULL);
vui = xd->vu_intf;
- client_len = sizeof(client);
+ client_len = sizeof (client);
client_fd = accept (uf->file_descriptor,
- (struct sockaddr *)&client,
- (socklen_t *)&client_len);
+ (struct sockaddr *) &client,
+ (socklen_t *) & client_len);
if (client_fd < 0)
- return clib_error_return_unix (0, "accept");
+ return clib_error_return_unix (0, "accept");
template.read_function = dpdk_vhost_user_socket_read;
template.error_function = dpdk_vhost_user_socket_error;
@@ -1342,41 +1430,45 @@ static clib_error_t * dpdk_vhost_user_socksvr_accept_ready (unix_file_t * uf)
vui->client_fd = client_fd;
hash_set (dm->vu_sw_if_index_by_sock_fd, vui->client_fd,
- xd->vlib_sw_if_index);
+ xd->vlib_sw_if_index);
return 0;
}
// init server socket on specified sock_filename
-static int dpdk_vhost_user_init_server_sock(const char * sock_filename, int *sockfd)
+static int
+dpdk_vhost_user_init_server_sock (const char *sock_filename, int *sockfd)
{
int rv = 0;
- struct sockaddr_un un = {};
+ struct sockaddr_un un = { };
int fd;
/* create listening socket */
- fd = socket(AF_UNIX, SOCK_STREAM, 0);
+ fd = socket (AF_UNIX, SOCK_STREAM, 0);
- if (fd < 0) {
- return VNET_API_ERROR_SYSCALL_ERROR_1;
- }
+ if (fd < 0)
+ {
+ return VNET_API_ERROR_SYSCALL_ERROR_1;
+ }
un.sun_family = AF_UNIX;
- strcpy((char *) un.sun_path, (char *) sock_filename);
+ strcpy ((char *) un.sun_path, (char *) sock_filename);
/* remove if exists */
- unlink( (char *) sock_filename);
+ unlink ((char *) sock_filename);
- if (bind(fd, (struct sockaddr *) &un, sizeof(un)) == -1) {
- rv = VNET_API_ERROR_SYSCALL_ERROR_2;
- goto error;
- }
+ if (bind (fd, (struct sockaddr *) &un, sizeof (un)) == -1)
+ {
+ rv = VNET_API_ERROR_SYSCALL_ERROR_2;
+ goto error;
+ }
- if (listen(fd, 1) == -1) {
- rv = VNET_API_ERROR_SYSCALL_ERROR_3;
- goto error;
- }
+ if (listen (fd, 1) == -1)
+ {
+ rv = VNET_API_ERROR_SYSCALL_ERROR_3;
+ goto error;
+ }
- unix_file_t template = {0};
+ unix_file_t template = { 0 };
template.read_function = dpdk_vhost_user_socksvr_accept_ready;
template.file_descriptor = fd;
unix_file_add (&unix_main, &template);
@@ -1384,7 +1476,7 @@ static int dpdk_vhost_user_init_server_sock(const char * sock_filename, int *soc
return rv;
error:
- close(fd);
+ close (fd);
return rv;
}
@@ -1392,73 +1484,84 @@ error:
* vhost-user interface control functions used from vpe api
*/
-int dpdk_vhost_user_create_if(vnet_main_t * vnm, vlib_main_t * vm,
- const char * sock_filename,
- u8 is_server,
- u32 * sw_if_index,
- u64 feature_mask,
- u8 renumber, u32 custom_dev_instance,
- u8 *hwaddr)
+int
+dpdk_vhost_user_create_if (vnet_main_t * vnm, vlib_main_t * vm,
+ const char *sock_filename,
+ u8 is_server,
+ u32 * sw_if_index,
+ u64 feature_mask,
+ u8 renumber, u32 custom_dev_instance, u8 * hwaddr)
{
- dpdk_main_t * dm = &dpdk_main;
+ dpdk_main_t *dm = &dpdk_main;
dpdk_device_t *xd;
u32 hw_if_idx = ~0;
int sockfd = -1;
int rv = 0;
// using virtio vhost user?
- if (dm->conf->use_virtio_vhost) {
- return vhost_user_create_if(vnm, vm, sock_filename, is_server,
- sw_if_index, feature_mask, renumber, custom_dev_instance, hwaddr);
- }
+ if (dm->conf->use_virtio_vhost)
+ {
+ return vhost_user_create_if (vnm, vm, sock_filename, is_server,
+ sw_if_index, feature_mask, renumber,
+ custom_dev_instance, hwaddr);
+ }
- if (is_server) {
- if ((rv = dpdk_vhost_user_init_server_sock (sock_filename, &sockfd)) != 0) {
- return rv;
+ if (is_server)
+ {
+ if ((rv =
+ dpdk_vhost_user_init_server_sock (sock_filename, &sockfd)) != 0)
+ {
+ return rv;
+ }
}
- }
- if (renumber) {
+ if (renumber)
+ {
// set next vhost-user if id if custom one is higher or equal
if (custom_dev_instance >= dm->next_vu_if_id)
- dm->next_vu_if_id = custom_dev_instance + 1;
+ dm->next_vu_if_id = custom_dev_instance + 1;
- dpdk_create_vhost_user_if_internal(&hw_if_idx, custom_dev_instance, hwaddr);
- } else
- dpdk_create_vhost_user_if_internal(&hw_if_idx, (u32)~0, hwaddr);
- DBG_SOCK("dpdk vhost-user interface created hw_if_index %d", hw_if_idx);
+ dpdk_create_vhost_user_if_internal (&hw_if_idx, custom_dev_instance,
+ hwaddr);
+ }
+ else
+ dpdk_create_vhost_user_if_internal (&hw_if_idx, (u32) ~ 0, hwaddr);
+ DBG_SOCK ("dpdk vhost-user interface created hw_if_index %d", hw_if_idx);
- xd = dpdk_vhost_user_device_from_hw_if_index(hw_if_idx);
- ASSERT(xd != NULL);
+ xd = dpdk_vhost_user_device_from_hw_if_index (hw_if_idx);
+ ASSERT (xd != NULL);
dpdk_vhost_user_vui_init (vnm, xd, sockfd, sock_filename, is_server,
- feature_mask, sw_if_index);
+ feature_mask, sw_if_index);
dpdk_vhost_user_vui_register (vm, xd);
return rv;
}
-int dpdk_vhost_user_modify_if(vnet_main_t * vnm, vlib_main_t * vm,
- const char * sock_filename,
- u8 is_server,
- u32 sw_if_index,
- u64 feature_mask,
- u8 renumber, u32 custom_dev_instance)
+int
+dpdk_vhost_user_modify_if (vnet_main_t * vnm, vlib_main_t * vm,
+ const char *sock_filename,
+ u8 is_server,
+ u32 sw_if_index,
+ u64 feature_mask,
+ u8 renumber, u32 custom_dev_instance)
{
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_t * xd;
- dpdk_vu_intf_t * vui = NULL;
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_t *xd;
+ dpdk_vu_intf_t *vui = NULL;
u32 sw_if_idx = ~0;
int sockfd = -1;
int rv = 0;
// using virtio vhost user?
- if (dm->conf->use_virtio_vhost) {
- return vhost_user_modify_if(vnm, vm, sock_filename, is_server,
- sw_if_index, feature_mask, renumber, custom_dev_instance);
- }
+ if (dm->conf->use_virtio_vhost)
+ {
+ return vhost_user_modify_if (vnm, vm, sock_filename, is_server,
+ sw_if_index, feature_mask, renumber,
+ custom_dev_instance);
+ }
- xd = dpdk_vhost_user_device_from_sw_if_index(sw_if_index);
+ xd = dpdk_vhost_user_device_from_sw_if_index (sw_if_index);
if (xd == NULL)
return VNET_API_ERROR_INVALID_SW_IF_INDEX;
@@ -1468,40 +1571,46 @@ int dpdk_vhost_user_modify_if(vnet_main_t * vnm, vlib_main_t * vm,
// interface is inactive
vui->active = 0;
// disconnect interface sockets
- dpdk_vhost_user_if_disconnect(xd);
+ dpdk_vhost_user_if_disconnect (xd);
- if (is_server) {
- if ((rv = dpdk_vhost_user_init_server_sock (sock_filename, &sockfd)) != 0) {
- return rv;
- }
- }
+ if (is_server)
+ {
+ if ((rv =
+ dpdk_vhost_user_init_server_sock (sock_filename, &sockfd)) != 0)
+ {
+ return rv;
+ }
+ }
dpdk_vhost_user_vui_init (vnm, xd, sockfd, sock_filename, is_server,
- feature_mask, &sw_if_idx);
+ feature_mask, &sw_if_idx);
- if (renumber) {
- vnet_interface_name_renumber (sw_if_idx, custom_dev_instance);
- }
+ if (renumber)
+ {
+ vnet_interface_name_renumber (sw_if_idx, custom_dev_instance);
+ }
dpdk_vhost_user_vui_register (vm, xd);
return rv;
}
-int dpdk_vhost_user_delete_if(vnet_main_t * vnm, vlib_main_t * vm,
- u32 sw_if_index)
+int
+dpdk_vhost_user_delete_if (vnet_main_t * vnm, vlib_main_t * vm,
+ u32 sw_if_index)
{
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_t * xd = NULL;
- dpdk_vu_intf_t * vui;
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_t *xd = NULL;
+ dpdk_vu_intf_t *vui;
int rv = 0;
// using virtio vhost user?
- if (dm->conf->use_virtio_vhost) {
- return vhost_user_delete_if(vnm, vm, sw_if_index);
- }
+ if (dm->conf->use_virtio_vhost)
+ {
+ return vhost_user_delete_if (vnm, vm, sw_if_index);
+ }
- xd = dpdk_vhost_user_device_from_sw_if_index(sw_if_index);
+ xd = dpdk_vhost_user_device_from_sw_if_index (sw_if_index);
if (xd == NULL)
return VNET_API_ERROR_INVALID_SW_IF_INDEX;
@@ -1511,147 +1620,172 @@ int dpdk_vhost_user_delete_if(vnet_main_t * vnm, vlib_main_t * vm,
// interface is inactive
vui->active = 0;
// disconnect interface sockets
- dpdk_vhost_user_if_disconnect(xd);
+ dpdk_vhost_user_if_disconnect (xd);
// add to inactive interface list
vec_add1 (dm->vu_inactive_interfaces_device_index, xd->device_index);
ethernet_delete_interface (vnm, xd->vlib_hw_if_index);
- DBG_SOCK ("deleted (deactivated) vhost-user interface sw_if_index %d", sw_if_index);
+ DBG_SOCK ("deleted (deactivated) vhost-user interface sw_if_index %d",
+ sw_if_index);
return rv;
}
-int dpdk_vhost_user_dump_ifs(vnet_main_t * vnm, vlib_main_t * vm, vhost_user_intf_details_t **out_vuids)
+int
+dpdk_vhost_user_dump_ifs (vnet_main_t * vnm, vlib_main_t * vm,
+ vhost_user_intf_details_t ** out_vuids)
{
- int rv = 0;
- dpdk_main_t * dm = &dpdk_main;
- dpdk_device_t * xd;
- dpdk_vu_intf_t * vui;
- struct virtio_net * vhost_dev;
- vhost_user_intf_details_t * r_vuids = NULL;
- vhost_user_intf_details_t * vuid = NULL;
- u32 * hw_if_indices = 0;
- vnet_hw_interface_t * hi;
- u8 *s = NULL;
- int i;
-
- if (!out_vuids)
- return -1;
-
- // using virtio vhost user?
- if (dm->conf->use_virtio_vhost) {
- return vhost_user_dump_ifs(vnm, vm, out_vuids);
- }
+ int rv = 0;
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_device_t *xd;
+ dpdk_vu_intf_t *vui;
+ struct virtio_net *vhost_dev;
+ vhost_user_intf_details_t *r_vuids = NULL;
+ vhost_user_intf_details_t *vuid = NULL;
+ u32 *hw_if_indices = 0;
+ vnet_hw_interface_t *hi;
+ u8 *s = NULL;
+ int i;
+
+ if (!out_vuids)
+ return -1;
- vec_foreach (xd, dm->devices) {
- if (xd->dev_type == VNET_DPDK_DEV_VHOST_USER &&
- xd->vu_intf->active)
- vec_add1(hw_if_indices, xd->vlib_hw_if_index);
+ // using virtio vhost user?
+ if (dm->conf->use_virtio_vhost)
+ {
+ return vhost_user_dump_ifs (vnm, vm, out_vuids);
}
- for (i = 0; i < vec_len (hw_if_indices); i++) {
+ vec_foreach (xd, dm->devices)
+ {
+ if (xd->dev_type == VNET_DPDK_DEV_VHOST_USER && xd->vu_intf->active)
+ vec_add1 (hw_if_indices, xd->vlib_hw_if_index);
+ }
+
+ for (i = 0; i < vec_len (hw_if_indices); i++)
+ {
hi = vnet_get_hw_interface (vnm, hw_if_indices[i]);
- xd = dpdk_vhost_user_device_from_hw_if_index(hw_if_indices[i]);
- if (!xd) {
- clib_warning("invalid vhost-user interface hw_if_index %d", hw_if_indices[i]);
- continue;
- }
+ xd = dpdk_vhost_user_device_from_hw_if_index (hw_if_indices[i]);
+ if (!xd)
+ {
+ clib_warning ("invalid vhost-user interface hw_if_index %d",
+ hw_if_indices[i]);
+ continue;
+ }
vui = xd->vu_intf;
- ASSERT(vui != NULL);
+ ASSERT (vui != NULL);
vhost_dev = &xd->vu_vhost_dev;
u32 virtio_net_hdr_sz = (vui->num_vrings > 0 ?
- vhost_dev->virtqueue[0]->vhost_hlen : 0);
+ vhost_dev->virtqueue[0]->vhost_hlen : 0);
- vec_add2(r_vuids, vuid, 1);
+ vec_add2 (r_vuids, vuid, 1);
vuid->sw_if_index = xd->vlib_sw_if_index;
vuid->virtio_net_hdr_sz = virtio_net_hdr_sz;
vuid->features = vhost_dev->features;
vuid->is_server = vui->sock_is_server;
- vuid->num_regions = (vhost_dev->mem != NULL ? vhost_dev->mem->nregions : 0);
+ vuid->num_regions =
+ (vhost_dev->mem != NULL ? vhost_dev->mem->nregions : 0);
vuid->sock_errno = vui->sock_errno;
- strncpy((char *)vuid->sock_filename, (char *)vui->sock_filename,
- ARRAY_LEN(vuid->sock_filename)-1);
+ strncpy ((char *) vuid->sock_filename, (char *) vui->sock_filename,
+ ARRAY_LEN (vuid->sock_filename) - 1);
s = format (s, "%v%c", hi->name, 0);
- strncpy((char *)vuid->if_name, (char *)s,
- ARRAY_LEN(vuid->if_name)-1);
- _vec_len(s) = 0;
+ strncpy ((char *) vuid->if_name, (char *) s,
+ ARRAY_LEN (vuid->if_name) - 1);
+ _vec_len (s) = 0;
}
- vec_free (s);
- vec_free (hw_if_indices);
+ vec_free (s);
+ vec_free (hw_if_indices);
- *out_vuids = r_vuids;
+ *out_vuids = r_vuids;
- return rv;
+ return rv;
}
/*
* Processing functions called from dpdk process fn
*/
-typedef struct {
- struct sockaddr_un sun;
- int sockfd;
- unix_file_t template;
- uword *event_data;
+typedef struct
+{
+ struct sockaddr_un sun;
+ int sockfd;
+ unix_file_t template;
+ uword *event_data;
} dpdk_vu_process_state;
-void dpdk_vhost_user_process_init (void **ctx)
+void
+dpdk_vhost_user_process_init (void **ctx)
{
- dpdk_vu_process_state *state = clib_mem_alloc (sizeof(dpdk_vu_process_state));
- memset(state, 0, sizeof(*state));
- state->sockfd = socket(AF_UNIX, SOCK_STREAM, 0);
- state->sun.sun_family = AF_UNIX;
- state->template.read_function = dpdk_vhost_user_socket_read;
- state->template.error_function = dpdk_vhost_user_socket_error;
- state->event_data = 0;
- *ctx = state;
+ dpdk_vu_process_state *state =
+ clib_mem_alloc (sizeof (dpdk_vu_process_state));
+ memset (state, 0, sizeof (*state));
+ state->sockfd = socket (AF_UNIX, SOCK_STREAM, 0);
+ state->sun.sun_family = AF_UNIX;
+ state->template.read_function = dpdk_vhost_user_socket_read;
+ state->template.error_function = dpdk_vhost_user_socket_error;
+ state->event_data = 0;
+ *ctx = state;
}
-void dpdk_vhost_user_process_cleanup (void *ctx)
+void
+dpdk_vhost_user_process_cleanup (void *ctx)
{
- clib_mem_free(ctx);
+ clib_mem_free (ctx);
}
-uword dpdk_vhost_user_process_if (vlib_main_t *vm, dpdk_device_t *xd, void *ctx)
+uword
+dpdk_vhost_user_process_if (vlib_main_t * vm, dpdk_device_t * xd, void *ctx)
{
- dpdk_main_t * dm = &dpdk_main;
- dpdk_vu_process_state *state = (dpdk_vu_process_state *)ctx;
- dpdk_vu_intf_t *vui = xd->vu_intf;
-
- if (vui->sock_is_server || !vui->active)
- return 0;
-
- if (vui->unix_fd == -1) {
- /* try to connect */
- strncpy(state->sun.sun_path, (char *) vui->sock_filename, sizeof(state->sun.sun_path) - 1);
-
- if (connect(state->sockfd, (struct sockaddr *) &(state->sun), sizeof(struct sockaddr_un)) == 0) {
- vui->sock_errno = 0;
- vui->unix_fd = state->sockfd;
- state->template.file_descriptor = state->sockfd;
- vui->unix_file_index = unix_file_add (&unix_main, &(state->template));
- hash_set (dm->vu_sw_if_index_by_sock_fd, state->sockfd, xd->vlib_sw_if_index);
-
- state->sockfd = socket(AF_UNIX, SOCK_STREAM, 0);
- if (state->sockfd < 0)
- return -1;
- } else {
- vui->sock_errno = errno;
- }
- } else {
- /* check if socket is alive */
- int error = 0;
- socklen_t len = sizeof (error);
- int retval = getsockopt(vui->unix_fd, SOL_SOCKET, SO_ERROR, &error, &len);
-
- if (retval)
- dpdk_vhost_user_if_disconnect(xd);
- }
+ dpdk_main_t *dm = &dpdk_main;
+ dpdk_vu_process_state *state = (dpdk_vu_process_state *) ctx;
+ dpdk_vu_intf_t *vui = xd->vu_intf;
+
+ if (vui->sock_is_server || !vui->active)
return 0;
+
+ if (vui->unix_fd == -1)
+ {
+ /* try to connect */
+ strncpy (state->sun.sun_path, (char *) vui->sock_filename,
+ sizeof (state->sun.sun_path) - 1);
+
+ if (connect
+ (state->sockfd, (struct sockaddr *) &(state->sun),
+ sizeof (struct sockaddr_un)) == 0)
+ {
+ vui->sock_errno = 0;
+ vui->unix_fd = state->sockfd;
+ state->template.file_descriptor = state->sockfd;
+ vui->unix_file_index =
+ unix_file_add (&unix_main, &(state->template));
+ hash_set (dm->vu_sw_if_index_by_sock_fd, state->sockfd,
+ xd->vlib_sw_if_index);
+
+ state->sockfd = socket (AF_UNIX, SOCK_STREAM, 0);
+ if (state->sockfd < 0)
+ return -1;
+ }
+ else
+ {
+ vui->sock_errno = errno;
+ }
+ }
+ else
+ {
+ /* check if socket is alive */
+ int error = 0;
+ socklen_t len = sizeof (error);
+ int retval =
+ getsockopt (vui->unix_fd, SOL_SOCKET, SO_ERROR, &error, &len);
+
+ if (retval)
+ dpdk_vhost_user_if_disconnect (xd);
+ }
+ return 0;
}
/*
@@ -1660,110 +1794,123 @@ uword dpdk_vhost_user_process_if (vlib_main_t *vm, dpdk_device_t *xd, void *ctx)
static clib_error_t *
dpdk_vhost_user_connect_command_fn (vlib_main_t * vm,
- unformat_input_t * input,
- vlib_cli_command_t * cmd)
+ unformat_input_t * input,
+ vlib_cli_command_t * cmd)
{
- dpdk_main_t * dm = &dpdk_main;
- unformat_input_t _line_input, * line_input = &_line_input;
- u8 * sock_filename = NULL;
+ dpdk_main_t *dm = &dpdk_main;
+ unformat_input_t _line_input, *line_input = &_line_input;
+ u8 *sock_filename = NULL;
u32 sw_if_index;
u8 is_server = 0;
- u64 feature_mask = (u64)~0;
+ u64 feature_mask = (u64) ~ 0;
u8 renumber = 0;
u32 custom_dev_instance = ~0;
u8 hwaddr[6];
u8 *hw = NULL;
- if (dm->conf->use_virtio_vhost) {
- return vhost_user_connect_command_fn(vm, input, cmd);
- }
+ if (dm->conf->use_virtio_vhost)
+ {
+ return vhost_user_connect_command_fn (vm, input, cmd);
+ }
/* Get a line of input. */
- if (! unformat_user (input, unformat_line_input, line_input))
+ if (!unformat_user (input, unformat_line_input, line_input))
return 0;
- while (unformat_check_input (line_input) != UNFORMAT_END_OF_INPUT) {
- if (unformat (line_input, "socket %s", &sock_filename))
- ;
- else if (unformat (line_input, "server"))
- is_server = 1;
- else if (unformat (line_input, "feature-mask 0x%llx", &feature_mask))
- ;
- else if (unformat (line_input, "hwaddr %U", unformat_ethernet_address, hwaddr))
- hw = hwaddr;
- else if (unformat (line_input, "renumber %d", &custom_dev_instance)) {
- renumber = 1;
+ while (unformat_check_input (line_input) != UNFORMAT_END_OF_INPUT)
+ {
+ if (unformat (line_input, "socket %s", &sock_filename))
+ ;
+ else if (unformat (line_input, "server"))
+ is_server = 1;
+ else if (unformat (line_input, "feature-mask 0x%llx", &feature_mask))
+ ;
+ else
+ if (unformat
+ (line_input, "hwaddr %U", unformat_ethernet_address, hwaddr))
+ hw = hwaddr;
+ else if (unformat (line_input, "renumber %d", &custom_dev_instance))
+ {
+ renumber = 1;
+ }
+ else
+ return clib_error_return (0, "unknown input `%U'",
+ format_unformat_error, input);
}
- else
- return clib_error_return (0, "unknown input `%U'",
- format_unformat_error, input);
- }
unformat_free (line_input);
- vnet_main_t *vnm = vnet_get_main();
+ vnet_main_t *vnm = vnet_get_main ();
if (sock_filename == NULL)
- return clib_error_return (0, "missing socket file");
+ return clib_error_return (0, "missing socket file");
- dpdk_vhost_user_create_if(vnm, vm, (char *)sock_filename,
- is_server, &sw_if_index, feature_mask,
- renumber, custom_dev_instance, hw);
+ dpdk_vhost_user_create_if (vnm, vm, (char *) sock_filename,
+ is_server, &sw_if_index, feature_mask,
+ renumber, custom_dev_instance, hw);
- vec_free(sock_filename);
- vlib_cli_output(vm, "%U\n", format_vnet_sw_if_index_name, vnet_get_main(), sw_if_index);
+ vec_free (sock_filename);
+ vlib_cli_output (vm, "%U\n", format_vnet_sw_if_index_name, vnet_get_main (),
+ sw_if_index);
return 0;
}
+/* *INDENT-OFF* */
VLIB_CLI_COMMAND (dpdk_vhost_user_connect_command, static) = {
.path = "create vhost-user",
.short_help = "create vhost-user socket <socket-filename> [server] [feature-mask <hex>] [renumber <dev_instance>]",
.function = dpdk_vhost_user_connect_command_fn,
};
+/* *INDENT-ON* */
static clib_error_t *
dpdk_vhost_user_delete_command_fn (vlib_main_t * vm,
- unformat_input_t * input,
- vlib_cli_command_t * cmd)
+ unformat_input_t * input,
+ vlib_cli_command_t * cmd)
{
- dpdk_main_t * dm = &dpdk_main;
- clib_error_t * error = 0;
- unformat_input_t _line_input, * line_input = &_line_input;
+ dpdk_main_t *dm = &dpdk_main;
+ clib_error_t *error = 0;
+ unformat_input_t _line_input, *line_input = &_line_input;
u32 sw_if_index = ~0;
- if (dm->conf->use_virtio_vhost) {
- return vhost_user_delete_command_fn(vm, input, cmd);
- }
+ if (dm->conf->use_virtio_vhost)
+ {
+ return vhost_user_delete_command_fn (vm, input, cmd);
+ }
/* Get a line of input. */
- if (! unformat_user (input, unformat_line_input, line_input))
+ if (!unformat_user (input, unformat_line_input, line_input))
return 0;
- while (unformat_check_input (line_input) != UNFORMAT_END_OF_INPUT) {
- if (unformat (line_input, "sw_if_index %d", &sw_if_index))
- ;
- else
- return clib_error_return (0, "unknown input `%U'",
- format_unformat_error, input);
- }
+ while (unformat_check_input (line_input) != UNFORMAT_END_OF_INPUT)
+ {
+ if (unformat (line_input, "sw_if_index %d", &sw_if_index))
+ ;
+ else
+ return clib_error_return (0, "unknown input `%U'",
+ format_unformat_error, input);
+ }
unformat_free (line_input);
- if (sw_if_index == ~0) {
+ if (sw_if_index == ~0)
+ {
error = clib_error_return (0, "invalid sw_if_index",
- format_unformat_error, input);
+ format_unformat_error, input);
return error;
- }
+ }
- vnet_main_t *vnm = vnet_get_main();
+ vnet_main_t *vnm = vnet_get_main ();
- dpdk_vhost_user_delete_if(vnm, vm, sw_if_index);
+ dpdk_vhost_user_delete_if (vnm, vm, sw_if_index);
return 0;
}
+/* *INDENT-OFF* */
VLIB_CLI_COMMAND (dpdk_vhost_user_delete_command, static) = {
.path = "delete vhost-user",
.short_help = "delete vhost-user sw_if_index <nn>",
.function = dpdk_vhost_user_delete_command_fn,
};
+/* *INDENT-ON* */
#define foreach_dpdk_vhost_feature \
_ (VIRTIO_NET_F_MRG_RXBUF) \
@@ -1772,146 +1919,188 @@ VLIB_CLI_COMMAND (dpdk_vhost_user_delete_command, static) = {
static clib_error_t *
show_dpdk_vhost_user_command_fn (vlib_main_t * vm,
- unformat_input_t * input,
- vlib_cli_command_t * cmd)
+ unformat_input_t * input,
+ vlib_cli_command_t * cmd)
{
- clib_error_t * error = 0;
- dpdk_main_t * dm = &dpdk_main;
- vnet_main_t * vnm = vnet_get_main();
- dpdk_device_t * xd;
- dpdk_vu_intf_t * vui;
- struct virtio_net * vhost_dev;
- u32 hw_if_index, * hw_if_indices = 0;
- vnet_hw_interface_t * hi;
+ clib_error_t *error = 0;
+ dpdk_main_t *dm = &dpdk_main;
+ vnet_main_t *vnm = vnet_get_main ();
+ dpdk_device_t *xd;
+ dpdk_vu_intf_t *vui;
+ struct virtio_net *vhost_dev;
+ u32 hw_if_index, *hw_if_indices = 0;
+ vnet_hw_interface_t *hi;
int i, j, q;
int show_descr = 0;
- struct virtio_memory * mem;
- struct feat_struct { u8 bit; char *str;};
+ struct virtio_memory *mem;
+ struct feat_struct
+ {
+ u8 bit;
+ char *str;
+ };
struct feat_struct *feat_entry;
static struct feat_struct feat_array[] = {
#define _(f) { .str = #f, .bit = f, },
- foreach_dpdk_vhost_feature
+ foreach_dpdk_vhost_feature
#undef _
- { .str = NULL }
+ {.str = NULL}
};
- if (dm->conf->use_virtio_vhost) {
- return show_vhost_user_command_fn(vm, input, cmd);
- }
-
- while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) {
- if (unformat (input, "%U", unformat_vnet_hw_interface, vnm, &hw_if_index)) {
- vec_add1 (hw_if_indices, hw_if_index);
- vlib_cli_output(vm, "add %d", hw_if_index);
+ if (dm->conf->use_virtio_vhost)
+ {
+ return show_vhost_user_command_fn (vm, input, cmd);
}
- else if (unformat (input, "descriptors") || unformat (input, "desc") )
- show_descr = 1;
- else {
- error = clib_error_return (0, "unknown input `%U'",
- format_unformat_error, input);
- goto done;
+
+ while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT)
+ {
+ if (unformat
+ (input, "%U", unformat_vnet_hw_interface, vnm, &hw_if_index))
+ {
+ vec_add1 (hw_if_indices, hw_if_index);
+ vlib_cli_output (vm, "add %d", hw_if_index);
+ }
+ else if (unformat (input, "descriptors") || unformat (input, "desc"))
+ show_descr = 1;
+ else
+ {
+ error = clib_error_return (0, "unknown input `%U'",
+ format_unformat_error, input);
+ goto done;
+ }
}
- }
- if (vec_len (hw_if_indices) == 0) {
- vec_foreach (xd, dm->devices) {
- if (xd->dev_type == VNET_DPDK_DEV_VHOST_USER && xd->vu_intf->active)
- vec_add1(hw_if_indices, xd->vlib_hw_if_index);
+ if (vec_len (hw_if_indices) == 0)
+ {
+ vec_foreach (xd, dm->devices)
+ {
+ if (xd->dev_type == VNET_DPDK_DEV_VHOST_USER && xd->vu_intf->active)
+ vec_add1 (hw_if_indices, xd->vlib_hw_if_index);
+ }
}
- }
vlib_cli_output (vm, "DPDK vhost-user interfaces");
vlib_cli_output (vm, "Global:\n coalesce frames %d time %e\n\n",
- dm->conf->vhost_coalesce_frames, dm->conf->vhost_coalesce_time);
-
- for (i = 0; i < vec_len (hw_if_indices); i++) {
- hi = vnet_get_hw_interface (vnm, hw_if_indices[i]);
+ dm->conf->vhost_coalesce_frames,
+ dm->conf->vhost_coalesce_time);
- if (!(xd = dpdk_vhost_user_device_from_hw_if_index(hw_if_indices[i]))) {
- error = clib_error_return (0, "not dpdk vhost-user interface: '%s'",
- hi->name);
- goto done;
- }
- vui = xd->vu_intf;
- vhost_dev = &xd->vu_vhost_dev;
- mem = vhost_dev->mem;
- u32 virtio_net_hdr_sz = (vui->num_vrings > 0 ?
- vhost_dev->virtqueue[0]->vhost_hlen : 0);
-
- vlib_cli_output (vm, "Interface: %v (ifindex %d)",
- hi->name, hw_if_indices[i]);
-
- vlib_cli_output (vm, "virtio_net_hdr_sz %d\n features (0x%llx): \n",
- virtio_net_hdr_sz, xd->vu_vhost_dev.features);
-
- feat_entry = (struct feat_struct *) &feat_array;
- while(feat_entry->str) {
- if (xd->vu_vhost_dev.features & (1 << feat_entry->bit))
- vlib_cli_output (vm, " %s (%d)", feat_entry->str, feat_entry->bit);
- feat_entry++;
- }
-
- vlib_cli_output (vm, "\n");
-
- vlib_cli_output (vm, " socket filename %s type %s errno \"%s\"\n\n",
- vui->sock_filename, vui->sock_is_server ? "server" : "client",
- strerror(vui->sock_errno));
-
- vlib_cli_output (vm, " Memory regions (total %d)\n", mem->nregions);
+ for (i = 0; i < vec_len (hw_if_indices); i++)
+ {
+ hi = vnet_get_hw_interface (vnm, hw_if_indices[i]);
- if (mem->nregions){
- vlib_cli_output(vm, " region fd guest_phys_addr memory_size userspace_addr mmap_offset mmap_addr\n");
- vlib_cli_output(vm, " ====== ===== ================== ================== ================== ================== ==================\n");
- }
- for (j = 0; j < mem->nregions; j++) {
- vlib_cli_output(vm, " %d %-5d 0x%016lx 0x%016lx 0x%016lx 0x%016lx 0x%016lx\n", j,
- vui->region_fd[j],
- mem->regions[j].guest_phys_address,
- mem->regions[j].memory_size,
- mem->regions[j].userspace_address,
- mem->regions[j].address_offset,
- vui->region_addr[j]);
- }
- for (q = 0; q < vui->num_vrings; q++) {
- struct vhost_virtqueue *vq = vhost_dev->virtqueue[q];
- const char *qtype = (q & 1) ? "TX" : "RX";
-
- vlib_cli_output(vm, "\n Virtqueue %d (%s)\n", q/2, qtype);
-
- vlib_cli_output(vm, " qsz %d last_used_idx %d last_used_idx_res %d\n",
- vq->size, vq->last_used_idx, vq->last_used_idx_res);
-
- if (vq->avail && vq->used)
- vlib_cli_output(vm, " avail.flags %x avail.idx %d used.flags %x used.idx %d\n",
- vq->avail->flags, vq->avail->idx, vq->used->flags, vq->used->idx);
-
- vlib_cli_output(vm, " kickfd %d callfd %d errfd %d enabled %d\n",
- vq->kickfd, vq->callfd, vui->vrings[q].errfd, vq->enabled);
-
- if (show_descr && vq->enabled) {
- vlib_cli_output(vm, "\n descriptor table:\n");
- vlib_cli_output(vm, " id addr len flags next user_addr\n");
- vlib_cli_output(vm, " ===== ================== ===== ====== ===== ==================\n");
- for(j = 0; j < vq->size; j++) {
- vlib_cli_output(vm, " %-5d 0x%016lx %-5d 0x%04x %-5d 0x%016lx\n",
- j,
- vq->desc[j].addr,
- vq->desc[j].len,
- vq->desc[j].flags,
- vq->desc[j].next,
- pointer_to_uword(map_guest_mem(xd, vq->desc[j].addr)));}
- }
+ if (!(xd = dpdk_vhost_user_device_from_hw_if_index (hw_if_indices[i])))
+ {
+ error = clib_error_return (0, "not dpdk vhost-user interface: '%s'",
+ hi->name);
+ goto done;
+ }
+ vui = xd->vu_intf;
+ vhost_dev = &xd->vu_vhost_dev;
+ mem = vhost_dev->mem;
+ u32 virtio_net_hdr_sz = (vui->num_vrings > 0 ?
+ vhost_dev->virtqueue[0]->vhost_hlen : 0);
+
+ vlib_cli_output (vm, "Interface: %v (ifindex %d)",
+ hi->name, hw_if_indices[i]);
+
+ vlib_cli_output (vm, "virtio_net_hdr_sz %d\n features (0x%llx): \n",
+ virtio_net_hdr_sz, xd->vu_vhost_dev.features);
+
+ feat_entry = (struct feat_struct *) &feat_array;
+ while (feat_entry->str)
+ {
+ if (xd->vu_vhost_dev.features & (1 << feat_entry->bit))
+ vlib_cli_output (vm, " %s (%d)", feat_entry->str,
+ feat_entry->bit);
+ feat_entry++;
+ }
+
+ vlib_cli_output (vm, "\n");
+
+ vlib_cli_output (vm, " socket filename %s type %s errno \"%s\"\n\n",
+ vui->sock_filename,
+ vui->sock_is_server ? "server" : "client",
+ strerror (vui->sock_errno));
+
+ vlib_cli_output (vm, " Memory regions (total %d)\n", mem->nregions);
+
+ if (mem->nregions)
+ {
+ vlib_cli_output (vm,
+ " region fd guest_phys_addr memory_size userspace_addr mmap_offset mmap_addr\n");
+ vlib_cli_output (vm,
+ " ====== ===== ================== ================== ================== ================== ==================\n");
+ }
+ for (j = 0; j < mem->nregions; j++)
+ {
+ vlib_cli_output (vm,
+ " %d %-5d 0x%016lx 0x%016lx 0x%016lx 0x%016lx 0x%016lx\n",
+ j, vui->region_fd[j],
+ mem->regions[j].guest_phys_address,
+ mem->regions[j].memory_size,
+ mem->regions[j].userspace_address,
+ mem->regions[j].address_offset,
+ vui->region_addr[j]);
+ }
+ for (q = 0; q < vui->num_vrings; q++)
+ {
+ struct vhost_virtqueue *vq = vhost_dev->virtqueue[q];
+ const char *qtype = (q & 1) ? "TX" : "RX";
+
+ vlib_cli_output (vm, "\n Virtqueue %d (%s)\n", q / 2, qtype);
+
+ vlib_cli_output (vm,
+ " qsz %d last_used_idx %d last_used_idx_res %d\n",
+ vq->size, vq->last_used_idx,
+ vq->last_used_idx_res);
+
+ if (vq->avail && vq->used)
+ vlib_cli_output (vm,
+ " avail.flags %x avail.idx %d used.flags %x used.idx %d\n",
+ vq->avail->flags, vq->avail->idx,
+ vq->used->flags, vq->used->idx);
+
+ vlib_cli_output (vm, " kickfd %d callfd %d errfd %d enabled %d\n",
+ vq->kickfd, vq->callfd, vui->vrings[q].errfd,
+ vq->enabled);
+
+ if (show_descr && vq->enabled)
+ {
+ vlib_cli_output (vm, "\n descriptor table:\n");
+ vlib_cli_output (vm,
+ " id addr len flags next user_addr\n");
+ vlib_cli_output (vm,
+ " ===== ================== ===== ====== ===== ==================\n");
+ for (j = 0; j < vq->size; j++)
+ {
+ vlib_cli_output (vm,
+ " %-5d 0x%016lx %-5d 0x%04x %-5d 0x%016lx\n",
+ j, vq->desc[j].addr, vq->desc[j].len,
+ vq->desc[j].flags, vq->desc[j].next,
+ pointer_to_uword (map_guest_mem
+ (xd, vq->desc[j].addr)));
+ }
+ }
+ }
+ vlib_cli_output (vm, "\n");
}
- vlib_cli_output (vm, "\n");
- }
done:
vec_free (hw_if_indices);
return error;
}
+/* *INDENT-OFF* */
VLIB_CLI_COMMAND (show_vhost_user_command, static) = {
.path = "show vhost-user",
.short_help = "show vhost-user interface",
.function = show_dpdk_vhost_user_command_fn,
};
+/* *INDENT-ON* */
#endif
+
+/*
+ * fd.io coding-style-patch-verification: ON
+ *
+ * Local Variables:
+ * eval: (c-set-style "gnu")
+ * End:
+ */