/* *------------------------------------------------------------------ * Copyright (c) 2017 Cisco and/or its affiliates. * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at: * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. *------------------------------------------------------------------ */ #include #include #include #include #ifdef __linux__ #include #elif __FreeBSD__ #include #endif /* __linux__ */ #include #include #include #include #include #include #include #include #include #include #include #include #include virtio_main_t virtio_main; #define _IOCTL(fd,a,...) \ if (ioctl (fd, a, __VA_ARGS__) < 0) \ { \ err = clib_error_return_unix (0, "ioctl(" #a ")"); \ goto error; \ } static clib_error_t * call_read_ready (clib_file_t * uf) { vnet_main_t *vnm = vnet_get_main (); u64 b; CLIB_UNUSED (ssize_t size) = read (uf->file_descriptor, &b, sizeof (b)); vnet_hw_if_rx_queue_set_int_pending (vnm, uf->private_data); return 0; } clib_error_t * virtio_vring_init (vlib_main_t * vm, virtio_if_t * vif, u16 idx, u16 sz) { vnet_virtio_vring_t *vring; int i; if (!is_pow2 (sz)) return clib_error_return (0, "ring size must be power of 2"); if (sz > 32768) return clib_error_return (0, "ring size must be 32768 or lower"); if (sz == 0) sz = 256; if (idx % 2) { vec_validate_aligned (vif->txq_vrings, TX_QUEUE_ACCESS (idx), CLIB_CACHE_LINE_BYTES); vring = vec_elt_at_index (vif->txq_vrings, TX_QUEUE_ACCESS (idx)); clib_spinlock_init (&vring->lockp); } else { vec_validate_aligned (vif->rxq_vrings, RX_QUEUE_ACCESS (idx), CLIB_CACHE_LINE_BYTES); vring = vec_elt_at_index (vif->rxq_vrings, RX_QUEUE_ACCESS (idx)); } i = sizeof (vnet_virtio_vring_desc_t) * sz; i = round_pow2 (i, CLIB_CACHE_LINE_BYTES); vring->desc = clib_mem_alloc_aligned (i, CLIB_CACHE_LINE_BYTES); clib_memset (vring->desc, 0, i); i = sizeof (vnet_virtio_vring_avail_t) + sz * sizeof (vring->avail->ring[0]); i = round_pow2 (i, CLIB_CACHE_LINE_BYTES); vring->avail = clib_mem_alloc_aligned (i, CLIB_CACHE_LINE_BYTES); clib_memset (vring->avail, 0, i); // tell kernel that we don't need interrupt vring->avail->flags = VRING_AVAIL_F_NO_INTERRUPT; i = sizeof (vnet_virtio_vring_used_t) + sz * sizeof (vnet_virtio_vring_used_elem_t); i = round_pow2 (i, CLIB_CACHE_LINE_BYTES); vring->used = clib_mem_alloc_aligned (i, CLIB_CACHE_LINE_BYTES); clib_memset (vring->used, 0, i); vring->queue_id = idx; ASSERT (vring->buffers == 0); vec_validate_aligned (vring->buffers, sz, CLIB_CACHE_LINE_BYTES); if (idx & 1) { clib_memset_u32 (vring->buffers, ~0, sz); // tx path: suppress the interrupts from kernel vring->call_fd = -1; } else vring->call_fd = eventfd (0, EFD_NONBLOCK | EFD_CLOEXEC); vring->total_packets = 0; vring->queue_size = sz; vring->kick_fd = eventfd (0, EFD_NONBLOCK | EFD_CLOEXEC); virtio_log_debug (vif, "vring %u size %u call_fd %d kick_fd %d", idx, vring->queue_size, vring->call_fd, vring->kick_fd); return 0; } inline void virtio_free_buffers (vlib_main_t *vm, vnet_virtio_vring_t *vring) { u16 used = vring->desc_in_use; u16 last = vring->last_used_idx; u16 mask = vring->queue_size - 1; while (used) { vlib_buffer_free (vm, &vring->buffers[last & mask], 1); last++; used--; } } clib_error_t * virtio_vring_free_rx (vlib_main_t * vm, virtio_if_t * vif, u32 idx) { vnet_virtio_vring_t *vring = vec_elt_at_index (vif->rxq_vrings, RX_QUEUE_ACCESS (idx)); clib_file_del_by_index (&file_main, vring->call_file_index); close (vring->kick_fd); close (vring->call_fd); if (vring->used) { virtio_free_buffers (vm, vring); clib_mem_free (vring->used); } if (vring->desc) clib_mem_free (vring->desc); if (vring->avail) clib_mem_free (vring->avail); vec_free (vring->buffers); return 0; } clib_error_t * virtio_vring_free_tx (vlib_main_t * vm, virtio_if_t * vif, u32 idx) { vnet_virtio_vring_t *vring = vec_elt_at_index (vif->txq_vrings, TX_QUEUE_ACCESS (idx)); close (vring->kick_fd); if (vring->used) { virtio_free_buffers (vm, vring); clib_mem_free (vring->used); } if (vring->desc) clib_mem_free (vring->desc); if (vring->avail) clib_mem_free (vring->avail); vec_free (vring->buffers); gro_flow_table_free (vring->flow_table); virtio_vring_buffering_free (vm, vring->buffering); clib_spinlock_free (&vring->lockp); return 0; } void virtio_set_packet_coalesce (virtio_if_t * vif) { vnet_main_t *vnm = vnet_get_main (); vnet_hw_interface_t *hw = vnet_get_hw_interface (vnm, vif->hw_if_index); vnet_virtio_vring_t *vring; vif->packet_coalesce = 1; vec_foreach (vring, vif->txq_vrings) { gro_flow_table_init (&vring->flow_table, vif->type & (VIRTIO_IF_TYPE_TAP | VIRTIO_IF_TYPE_PCI), hw->tx_node_index); } } clib_error_t * virtio_set_packet_buffering (virtio_if_t * vif, u16 buffering_size) { vnet_main_t *vnm = vnet_get_main (); vnet_hw_interface_t *hw = vnet_get_hw_interface (vnm, vif->hw_if_index); vnet_virtio_vring_t *vring; clib_error_t *error = 0; vec_foreach (vring, vif->txq_vrings) { if ((error = virtio_vring_buffering_init (&vring->buffering, hw->tx_node_index, buffering_size))) { break; } } return error; } static void virtio_vring_fill (vlib_main_t *vm, virtio_if_t *vif, vnet_virtio_vring_t *vring) { if (vif->is_packed) virtio_refill_vring_packed (vm, vif, vif->type, vring, vif->virtio_net_hdr_sz, virtio_input_node.index); else virtio_refill_vring_split (vm, vif, vif->type, vring, vif->virtio_net_hdr_sz, virtio_input_node.index); } void virtio_vring_set_rx_queues (vlib_main_t *vm, virtio_if_t *vif) { vnet_main_t *vnm = vnet_get_main (); vnet_virtio_vring_t *vring; u32 i = 0; vnet_hw_if_set_input_node (vnm, vif->hw_if_index, virtio_input_node.index); vec_foreach (vring, vif->rxq_vrings) { vring->queue_index = vnet_hw_if_register_rx_queue ( vnm, vif->hw_if_index, RX_QUEUE_ACCESS (vring->queue_id), VNET_HW_IF_RXQ_THREAD_ANY); vring->buffer_pool_index = vlib_buffer_pool_get_default_for_numa ( vm, vnet_hw_if_get_rx_queue_numa_node (vnm, vring->queue_index)); if (vif->type == VIRTIO_IF_TYPE_TAP || vif->type == VIRTIO_IF_TYPE_TUN) { clib_file_t f = { .read_function = call_read_ready, .flags = UNIX_FILE_EVENT_EDGE_TRIGGERED, .file_descriptor = vring->call_fd, .private_data = vring->queue_index, .description = format (0, "%U vring %u", format_virtio_device_name, vif->dev_instance, vring->queue_id), }; vring->call_file_index = clib_file_add (&file_main, &f); vnet_hw_if_set_rx_queue_file_index (vnm, vring->queue_index, vring->call_file_index); } else if ((vif->type == VIRTIO_IF_TYPE_PCI) && (vif->support_int_mode) && (vif->msix_enabled == VIRTIO_MSIX_ENABLED)) { u32 file_index; file_index = vlib_pci_get_msix_file_index (vm, vif->pci_dev_handle, i + 1); vnet_hw_if_set_rx_queue_file_index (vnm, vring->queue_index, file_index); i++; } vnet_hw_if_set_rx_queue_mode (vnm, vring->queue_index, VNET_HW_IF_RX_MODE_POLLING); vring->mode = VNET_HW_IF_RX_MODE_POLLING; virtio_vring_fill (vm, vif, vring); } vnet_hw_if_update_runtime_data (vnm, vif->hw_if_index); } void virtio_vring_set_tx_queues (vlib_main_t *vm, virtio_if_t *vif) { vnet_main_t *vnm = vnet_get_main (); vnet_virtio_vring_t *vring; uword n_threads = vlib_get_n_threads (); u8 consistent = vif->consistent_qp; vec_foreach (vring, vif->txq_vrings) { vring->queue_index = vnet_hw_if_register_tx_queue ( vnm, vif->hw_if_index, TX_QUEUE_ACCESS (vring->queue_id)); } if (vif->num_txqs == 0) { virtio_log_error (vif, "Interface %U has 0 txq", format_vnet_hw_if_index_name, vnm, vif->hw_if_index); return; } for (u32 j = 0; j < n_threads; j++) { u32 qi = vif->txq_vrings[j % vif->num_txqs].queue_index; vnet_hw_if_tx_queue_assign_thread (vnm, qi, (j + consistent) % n_threads); } vnet_hw_if_update_runtime_data (vnm, vif->hw_if_index); } inline void virtio_set_net_hdr_size (virtio_if_t * vif) { if (vif->features & VIRTIO_FEATURE (VIRTIO_NET_F_MRG_RXBUF) || vif->features & VIRTIO_FEATURE (VIRTIO_F_VERSION_1)) vif->virtio_net_hdr_sz = sizeof (vnet_virtio_net_hdr_v1_t); else vif->virtio_net_hdr_sz = sizeof (vnet_virtio_net_hdr_t); } inline void virtio_show (vlib_main_t *vm, u32 *hw_if_indices, u8 show_descr, virtio_if_type_t type) { u32 i, j, hw_if_index; virtio_if_t *vif; vnet_main_t *vnm = &vnet_main; virtio_main_t *mm = &virtio_main; vnet_virtio_vring_t *vring; struct feat_struct { u8 bit; char *str; }; struct feat_struct *feat_entry; static struct feat_struct feat_array[] = { #define _(s,b) { .str = #s, .bit = b, }, foreach_virtio_net_features #undef _ {.str = NULL} }; struct feat_struct *flag_entry; static struct feat_struct flags_array[] = { #define _(b,e,s) { .bit = b, .str = s, }, foreach_virtio_if_flag #undef _ {.str = NULL} }; if (!hw_if_indices) return; for (hw_if_index = 0; hw_if_index < vec_len (hw_if_indices); hw_if_index++) { vnet_hw_interface_t *hi = vnet_get_hw_interface (vnm, hw_if_indices[hw_if_index]); vif = pool_elt_at_index (mm->interfaces, hi->dev_instance); if (vif->type != type) continue; vlib_cli_output (vm, "Interface: %U (ifindex %d)", format_vnet_hw_if_index_name, vnm, hw_if_indices[hw_if_index], vif->hw_if_index); if (type == VIRTIO_IF_TYPE_PCI) { vlib_cli_output (vm, " PCI Address: %U", format_vlib_pci_addr, &vif->pci_addr); } if (type & (VIRTIO_IF_TYPE_TAP | VIRTIO_IF_TYPE_TUN)) { u8 *str = 0; if (vif->host_if_name) vlib_cli_output (vm, " name \"%s\"", vif->host_if_name); if (vif->net_ns) vlib_cli_output (vm, " host-ns \"%s\"", vif->net_ns); if (vif->host_mtu_size) vlib_cli_output (vm, " host-mtu-size \"%d\"", vif->host_mtu_size); if (type == VIRTIO_IF_TYPE_TAP) vlib_cli_output (vm, " host-mac-addr: %U", format_ethernet_address, vif->host_mac_addr); vlib_cli_output (vm, " host-carrier-up: %u", vif->host_carrier_up); vec_foreach_index (i, vif->vhost_fds) str = format (str, " %d", vif->vhost_fds[i]); vlib_cli_output (vm, " vhost-fds%v", str); vec_free (str); vec_foreach_index (i, vif->tap_fds) str = format (str, " %d", vif->tap_fds[i]); vlib_cli_output (vm, " tap-fds%v", str); vec_free (str); } vlib_cli_output (vm, " gso-enabled %d", vif->gso_enabled); vlib_cli_output (vm, " csum-enabled %d", vif->csum_offload_enabled); vlib_cli_output (vm, " packet-coalesce %d", vif->packet_coalesce); vlib_cli_output (vm, " packet-buffering %d", vif->packet_buffering); vlib_cli_output (vm, " rss-enabled %d", vif->rss_enabled); if (type & (VIRTIO_IF_TYPE_TAP | VIRTIO_IF_TYPE_PCI)) vlib_cli_output (vm, " Mac Address: %U", format_ethernet_address, vif->mac_addr); vlib_cli_output (vm, " Device instance: %u", vif->dev_instance); vlib_cli_output (vm, " flags 0x%x", vif->flags); flag_entry = (struct feat_struct *) &flags_array; while (flag_entry->str) { if (vif->flags & (1ULL << flag_entry->bit)) vlib_cli_output (vm, " %s (%d)", flag_entry->str, flag_entry->bit); flag_entry++; } if (type == VIRTIO_IF_TYPE_PCI) { device_status (vm, vif); } vlib_cli_output (vm, " features 0x%lx", vif->features); feat_entry = (struct feat_struct *) &feat_array; while (feat_entry->str) { if (vif->features & (1ULL << feat_entry->bit)) vlib_cli_output (vm, " %s (%d)", feat_entry->str, feat_entry->bit); feat_entry++; } vlib_cli_output (vm, " remote-features 0x%lx", vif->remote_features); feat_entry = (struct feat_struct *) &feat_array; while (feat_entry->str) { if (vif->remote_features & (1ULL << feat_entry->bit)) vlib_cli_output (vm, " %s (%d)", feat_entry->str, feat_entry->bit); feat_entry++; } vlib_cli_output (vm, " Number of RX Virtqueue %u", vif->num_rxqs); vlib_cli_output (vm, " Number of TX Virtqueue %u", vif->num_txqs); if (type == VIRTIO_IF_TYPE_PCI && vif->cxq_vring != NULL && vif->features & VIRTIO_FEATURE (VIRTIO_NET_F_CTRL_VQ)) vlib_cli_output (vm, " Number of CTRL Virtqueue 1"); vec_foreach_index (i, vif->rxq_vrings) { vring = vec_elt_at_index (vif->rxq_vrings, i); vlib_cli_output (vm, " Virtqueue (RX) %d", vring->queue_id); vlib_cli_output ( vm, " qsz %d, last_used_idx %d, desc_next %d, desc_in_use %d", vring->queue_size, vring->last_used_idx, vring->desc_next, vring->desc_in_use); if (vif->is_packed) { vlib_cli_output (vm, " driver_event.flags 0x%x driver_event.off_wrap %d device_event.flags 0x%x device_event.off_wrap %d", vring->driver_event->flags, vring->driver_event->off_wrap, vring->device_event->flags, vring->device_event->off_wrap); vlib_cli_output (vm, " avail wrap counter %d, used wrap counter %d", vring->avail_wrap_counter, vring->used_wrap_counter); } else vlib_cli_output (vm, " avail.flags 0x%x avail.idx %d used.flags 0x%x used.idx %d", vring->avail->flags, vring->avail->idx, vring->used->flags, vring->used->idx); if (type & (VIRTIO_IF_TYPE_TAP | VIRTIO_IF_TYPE_TUN)) { vlib_cli_output (vm, " kickfd %d, callfd %d", vring->kick_fd, vring->call_fd); } if (show_descr) { vlib_cli_output (vm, "\n descriptor table:\n"); vlib_cli_output (vm, " id addr len flags next/id user_addr\n"); vlib_cli_output (vm, " ===== ================== ===== ====== ======= ==================\n"); for (j = 0; j < vring->queue_size; j++) { if (vif->is_packed) { vnet_virtio_vring_packed_desc_t *desc = &vring->packed_desc[j]; vlib_cli_output (vm, " %-5d 0x%016lx %-5d 0x%04x %-8d 0x%016lx\n", j, desc->addr, desc->len, desc->flags, desc->id, desc->addr); } else { vnet_virtio_vring_desc_t *desc = &vring->desc[j]; vlib_cli_output (vm, " %-5d 0x%016lx %-5d 0x%04x %-8d 0x%016lx\n", j, desc->addr, desc->len, desc->flags, desc->next, desc->addr); } } } } vec_foreach_index (i, vif->txq_vrings) { vring = vec_elt_at_index (vif->txq_vrings, i); vlib_cli_output (vm, " Virtqueue (TX) %d", vring->queue_id); vlib_cli_output ( vm, " qsz %d, last_used_idx %d, desc_next %d, desc_in_use %d", vring->queue_size, vring->last_used_idx, vring->desc_next, vring->desc_in_use); if (vif->is_packed) { vlib_cli_output (vm, " driver_event.flags 0x%x driver_event.off_wrap %d device_event.flags 0x%x device_event.off_wrap %d", vring->driver_event->flags, vring->driver_event->off_wrap, vring->device_event->flags, vring->device_event->off_wrap); vlib_cli_output (vm, " avail wrap counter %d, used wrap counter %d", vring->avail_wrap_counter, vring->used_wrap_counter); } else vlib_cli_output (vm, " avail.flags 0x%x avail.idx %d used.flags 0x%x used.idx %d", vring->avail->flags, vring->avail->idx, vring->used->flags, vring->used->idx); if (type & (VIRTIO_IF_TYPE_TAP | VIRTIO_IF_TYPE_TUN)) { vlib_cli_output (vm, " kickfd %d, callfd %d", vring->kick_fd, vring->call_fd); } if (vring->flow_table) { vlib_cli_output (vm, " %U", gro_flow_table_format, vring->flow_table); } if (vif->packet_buffering) { vlib_cli_output (vm, " %U", virtio_vring_buffering_format, vring->buffering); } if (show_descr) { vlib_cli_output (vm, "\n descriptor table:\n"); vlib_cli_output (vm, " id addr len flags next/id user_addr\n"); vlib_cli_output (vm, " ===== ================== ===== ====== ======== ==================\n"); for (j = 0; j < vring->queue_size; j++) { if (vif->is_packed) { vnet_virtio_vring_packed_desc_t *desc = &vring->packed_desc[j]; vlib_cli_output (vm, " %-5d 0x%016lx %-5d 0x%04x %-8d 0x%016lx\n", j, desc->addr, desc->len, desc->flags, desc->id, desc->addr); } else { vnet_virtio_vring_desc_t *desc = &vring->desc[j]; vlib_cli_output (vm, " %-5d 0x%016lx %-5d 0x%04x %-8d 0x%016lx\n", j, desc->addr, desc->len, desc->flags, desc->next, desc->addr); } } } } if (type == VIRTIO_IF_TYPE_PCI && vif->cxq_vring != NULL && vif->features & VIRTIO_FEATURE (VIRTIO_NET_F_CTRL_VQ)) { vring = vif->cxq_vring; vlib_cli_output (vm, " Virtqueue (CTRL) %d", vring->queue_id); vlib_cli_output ( vm, " qsz %d, last_used_idx %d, desc_next %d, desc_in_use %d", vring->queue_size, vring->last_used_idx, vring->desc_next, vring->desc_in_use); if (vif->is_packed) { vlib_cli_output (vm, " driver_event.flags 0x%x driver_event.off_wrap %d device_event.flags 0x%x device_event.off_wrap %d", vring->driver_event->flags, vring->driver_event->off_wrap, vring->device_event->flags, vring->device_event->off_wrap); vlib_cli_output (vm, " avail wrap counter %d, used wrap counter %d", vring->avail_wrap_counter, vring->used_wrap_counter); } else { vlib_cli_output (vm, " avail.flags 0x%x avail.idx %d used.flags 0x%x used.idx %d", vring->avail->flags, vring->avail->idx, vring->used->flags, vring->used->idx); } if (show_descr) { vlib_cli_output (vm, "\n descriptor table:\n"); vlib_cli_output (vm, " id addr len flags next/id user_addr\n"); vlib_cli_output (vm, " ===== ================== ===== ====== ======== ==================\n"); for (j = 0; j < vring->queue_size; j++) { if (vif->is_packed) { vnet_virtio_vring_packed_desc_t *desc = &vring->packed_desc[j]; vlib_cli_output (vm, " %-5d 0x%016lx %-5d 0x%04x %-8d 0x%016lx\n", j, desc->addr, desc->len, desc->flags, desc->id, desc->addr); } else { vnet_virtio_vring_desc_t *desc = &vring->desc[j]; vlib_cli_output (vm, " %-5d 0x%016lx %-5d 0x%04x %-8d 0x%016lx\n", j, desc->addr, desc->len, desc->flags, desc->next, desc->addr); } } } } } } static clib_error_t * virtio_init (vlib_main_t * vm) { virtio_main_t *vim = &virtio_main; clib_error_t *error = 0; vim->log_default = vlib_log_register_class ("virtio", 0); vlib_log_debug (vim->log_default, "initialized"); return error; } VLIB_INIT_FUNCTION (virtio_init); /* * fd.io coding-style-patch-verification: ON * * Local Variables: * eval: (c-set-style "gnu") * End: */