[dpdk-dev] [PATCH v3 2/2] vhost: Add VHOST PMD
Yuanhan Liu
yuanhan.liu at linux.intel.com
Mon Nov 9 07:21:42 CET 2015
Hi Tetsuya,
Here I just got some minor nits after a very rough glimpse.
On Mon, Nov 09, 2015 at 02:17:01PM +0900, Tetsuya Mukawa wrote:
...
> +static uint16_t
> +eth_vhost_rx(void *q, struct rte_mbuf **bufs, uint16_t nb_bufs)
> +{
> + struct vhost_queue *r = q;
> + uint16_t nb_rx = 0;
> +
> + if (unlikely(rte_atomic32_read(&r->allow_queuing) == 0))
> + return 0;
> +
> + rte_atomic32_set(&r->while_queuing, 1);
> +
> + if (unlikely(rte_atomic32_read(&r->allow_queuing) == 0))
> + goto out;
> +
> + /* Dequeue packets from guest TX queue */
> + nb_rx = (uint16_t)rte_vhost_dequeue_burst(r->device,
> + r->virtqueue_id, r->mb_pool, bufs, nb_bufs);
Unnecessary cast, as rte_vhost_enqueue_burst is defined with uint16_t
return type.
> +
> + r->rx_pkts += nb_rx;
> +
> +out:
> + rte_atomic32_set(&r->while_queuing, 0);
> +
> + return nb_rx;
> +}
> +
> +static uint16_t
> +eth_vhost_tx(void *q, struct rte_mbuf **bufs, uint16_t nb_bufs)
> +{
> + struct vhost_queue *r = q;
> + uint16_t i, nb_tx = 0;
> +
> + if (unlikely(rte_atomic32_read(&r->allow_queuing) == 0))
> + return 0;
> +
> + rte_atomic32_set(&r->while_queuing, 1);
> +
> + if (unlikely(rte_atomic32_read(&r->allow_queuing) == 0))
> + goto out;
> +
> + /* Enqueue packets to guest RX queue */
> + nb_tx = (uint16_t)rte_vhost_enqueue_burst(r->device,
> + r->virtqueue_id, bufs, nb_bufs);
Ditto.
> +
> + r->tx_pkts += nb_tx;
> + r->err_pkts += nb_bufs - nb_tx;
> +
> + for (i = 0; likely(i < nb_tx); i++)
> + rte_pktmbuf_free(bufs[i]);
> +
> +out:
> + rte_atomic32_set(&r->while_queuing, 0);
> +
> + return nb_tx;
> +}
> +
> +static int
> +eth_dev_configure(struct rte_eth_dev *dev __rte_unused) { return 0; }
I personally would not prefer to saving few lines of code to sacrifice
the readability.
> +
> +static int
> +eth_rx_queue_setup(struct rte_eth_dev *dev, uint16_t rx_queue_id,
> + uint16_t nb_rx_desc __rte_unused,
> + unsigned int socket_id,
> + const struct rte_eth_rxconf *rx_conf __rte_unused,
> + struct rte_mempool *mb_pool)
> +{
> + struct pmd_internal *internal = dev->data->dev_private;
> + struct vhost_queue *vq;
> +
> + if (internal->rx_vhost_queues[rx_queue_id] != NULL)
> + rte_free(internal->rx_vhost_queues[rx_queue_id]);
Such NULL check is unnecessary; rte_free will handle it.
> +
> + vq = rte_zmalloc_socket(NULL, sizeof(struct vhost_queue),
> + RTE_CACHE_LINE_SIZE, socket_id);
> + if (vq == NULL) {
> + RTE_LOG(ERR, PMD, "Failed to allocate memory for rx queue\n");
> + return -ENOMEM;
> + }
> +
> + vq->mb_pool = mb_pool;
> + vq->virtqueue_id = rx_queue_id * VIRTIO_QNUM + VIRTIO_TXQ;
> + internal->rx_vhost_queues[rx_queue_id] = vq;
> + dev->data->rx_queues[rx_queue_id] = vq;
> + return 0;
> +}
> +
> +static int
> +eth_tx_queue_setup(struct rte_eth_dev *dev, uint16_t tx_queue_id,
> + uint16_t nb_tx_desc __rte_unused,
> + unsigned int socket_id,
> + const struct rte_eth_txconf *tx_conf __rte_unused)
> +{
> + struct pmd_internal *internal = dev->data->dev_private;
> + struct vhost_queue *vq;
> +
> + if (internal->tx_vhost_queues[tx_queue_id] != NULL)
> + rte_free(internal->tx_vhost_queues[tx_queue_id]);
Ditto.
> +
> + vq = rte_zmalloc_socket(NULL, sizeof(struct vhost_queue),
> + RTE_CACHE_LINE_SIZE, socket_id);
> + if (vq == NULL) {
> + RTE_LOG(ERR, PMD, "Failed to allocate memory for tx queue\n");
> + return -ENOMEM;
> + }
> +
> + vq->virtqueue_id = tx_queue_id * VIRTIO_QNUM + VIRTIO_RXQ;
> + internal->tx_vhost_queues[tx_queue_id] = vq;
> + dev->data->tx_queues[tx_queue_id] = vq;
> + return 0;
> +}
> +
> +
> +static void
> +eth_dev_info(struct rte_eth_dev *dev,
> + struct rte_eth_dev_info *dev_info)
> +{
> + struct pmd_internal *internal = dev->data->dev_private;
> +
> + dev_info->driver_name = drivername;
> + dev_info->max_mac_addrs = 1;
> + dev_info->max_rx_pktlen = (uint32_t)-1;
> + dev_info->max_rx_queues = (uint16_t)internal->nb_rx_queues;
> + dev_info->max_tx_queues = (uint16_t)internal->nb_tx_queues;
> + dev_info->min_rx_bufsize = 0;
> +}
> +
> +static void
> +eth_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *igb_stats)
> +{
> + unsigned i;
> + unsigned long rx_total = 0, tx_total = 0, tx_err_total = 0;
> + const struct pmd_internal *internal = dev->data->dev_private;
> +
> + for (i = 0; i < RTE_ETHDEV_QUEUE_STAT_CNTRS &&
> + i < internal->nb_rx_queues; i++) {
> + if (internal->rx_vhost_queues[i] == NULL)
> + continue;
> + igb_stats->q_ipackets[i] = internal->rx_vhost_queues[i]->rx_pkts;
> + rx_total += igb_stats->q_ipackets[i];
> + }
> +
> + for (i = 0; i < RTE_ETHDEV_QUEUE_STAT_CNTRS &&
> + i < internal->nb_tx_queues; i++) {
> + if (internal->tx_vhost_queues[i] == NULL)
> + continue;
> + igb_stats->q_opackets[i] = internal->tx_vhost_queues[i]->tx_pkts;
> + igb_stats->q_errors[i] = internal->tx_vhost_queues[i]->err_pkts;
> + tx_total += igb_stats->q_opackets[i];
> + tx_err_total += igb_stats->q_errors[i];
> + }
> +
> + igb_stats->ipackets = rx_total;
> + igb_stats->opackets = tx_total;
> + igb_stats->oerrors = tx_err_total;
> +}
> +
> +static void
> +eth_stats_reset(struct rte_eth_dev *dev)
> +{
> + unsigned i;
> + struct pmd_internal *internal = dev->data->dev_private;
> +
> + for (i = 0; i < internal->nb_rx_queues; i++) {
> + if (internal->rx_vhost_queues[i] == NULL)
> + continue;
> + internal->rx_vhost_queues[i]->rx_pkts = 0;
> + }
> + for (i = 0; i < internal->nb_tx_queues; i++) {
> + if (internal->tx_vhost_queues[i] == NULL)
> + continue;
> + internal->tx_vhost_queues[i]->tx_pkts = 0;
> + internal->tx_vhost_queues[i]->err_pkts = 0;
> + }
> +}
> +
> +static void
> +eth_queue_release(void *q __rte_unused) { ; }
> +static int
> +eth_link_update(struct rte_eth_dev *dev __rte_unused,
> + int wait_to_complete __rte_unused) { return 0; }
Ditto.
> +
> +static const struct eth_dev_ops ops = {
> + .dev_start = eth_dev_start,
> + .dev_stop = eth_dev_stop,
> + .dev_configure = eth_dev_configure,
> + .dev_infos_get = eth_dev_info,
> + .rx_queue_setup = eth_rx_queue_setup,
> + .tx_queue_setup = eth_tx_queue_setup,
> + .rx_queue_release = eth_queue_release,
> + .tx_queue_release = eth_queue_release,
> + .link_update = eth_link_update,
> + .stats_get = eth_stats_get,
> + .stats_reset = eth_stats_reset,
> +};
> +
> +static int
> +eth_dev_vhost_create(const char *name, int index,
> + char *iface_name,
> + int16_t queues,
> + const unsigned numa_node)
> +{
> + struct rte_eth_dev_data *data = NULL;
> + struct pmd_internal *internal = NULL;
> + struct rte_eth_dev *eth_dev = NULL;
> + struct ether_addr *eth_addr = NULL;
> +
> + RTE_LOG(INFO, PMD, "Creating VHOST-USER backend on numa socket %u\n",
> + numa_node);
> +
> + /* now do all data allocation - for eth_dev structure, dummy pci driver
> + * and internal (private) data
> + */
> + data = rte_zmalloc_socket(name, sizeof(*data), 0, numa_node);
> + if (data == NULL)
> + goto error;
> +
> + internal = rte_zmalloc_socket(name, sizeof(*internal), 0, numa_node);
> + if (internal == NULL)
> + goto error;
> +
> + eth_addr = rte_zmalloc_socket(name, sizeof(*eth_addr), 0, numa_node);
> + if (eth_addr == NULL)
> + goto error;
> + *eth_addr = base_eth_addr;
> + eth_addr->addr_bytes[5] = index;
> +
> + /* reserve an ethdev entry */
> + eth_dev = rte_eth_dev_allocate(name, RTE_ETH_DEV_VIRTUAL);
> + if (eth_dev == NULL)
> + goto error;
> +
> + /* now put it all together
> + * - store queue data in internal,
> + * - store numa_node info in ethdev data
> + * - point eth_dev_data to internals
> + * - and point eth_dev structure to new eth_dev_data structure
> + */
> + internal->nb_rx_queues = queues;
> + internal->nb_tx_queues = queues;
> + internal->dev_name = strdup(name);
> + if (internal->dev_name == NULL)
> + goto error;
> + internal->iface_name = strdup(iface_name);
> + if (internal->iface_name == NULL)
> + goto error;
If allocation failed here, you will find that internal->dev_name is not
freed.
> +
> + pthread_mutex_lock(&internal_list_lock);
> + TAILQ_INSERT_TAIL(&internals_list, internal, next);
> + pthread_mutex_unlock(&internal_list_lock);
> +
> + data->dev_private = internal;
> + data->port_id = eth_dev->data->port_id;
> + memmove(data->name, eth_dev->data->name, sizeof(data->name));
> + data->nb_rx_queues = queues;
> + data->nb_tx_queues = queues;
> + data->dev_link = pmd_link;
> + data->mac_addrs = eth_addr;
> +
> + /* We'll replace the 'data' originally allocated by eth_dev. So the
> + * vhost PMD resources won't be shared between multi processes.
> + */
> + eth_dev->data = data;
> + eth_dev->dev_ops = &ops;
> + eth_dev->driver = NULL;
> + eth_dev->data->dev_flags = RTE_ETH_DEV_DETACHABLE;
> + eth_dev->data->kdrv = RTE_KDRV_NONE;
> + eth_dev->data->drv_name = internal->dev_name;
> + eth_dev->data->numa_node = numa_node;
> +
> + /* finally assign rx and tx ops */
> + eth_dev->rx_pkt_burst = eth_vhost_rx;
> + eth_dev->tx_pkt_burst = eth_vhost_tx;
> +
> + return data->port_id;
> +
> +error:
> + rte_free(data);
> + rte_free(internal);
> + rte_free(eth_addr);
> +
> + return -1;
> +}
...
...
> +
> + if ((internal) && (internal->dev_name))
> + free(internal->dev_name);
> + if ((internal) && (internal->iface_name))
> + free(internal->iface_name);
> +
> + rte_free(eth_dev->data->mac_addrs);
> + rte_free(eth_dev->data);
> +
> + for (i = 0; i < internal->nb_rx_queues; i++) {
> + if (internal->rx_vhost_queues[i] != NULL)
> + rte_free(internal->rx_vhost_queues[i]);
> + }
> + for (i = 0; i < internal->nb_tx_queues; i++) {
> + if (internal->tx_vhost_queues[i] != NULL)
> + rte_free(internal->tx_vhost_queues[i]);
Ditto.
(Hopefully I could have a detailed review later, say next week).
--yliu
More information about the dev
mailing list