[dpdk-dev] [PATCH v2 06/10] net/virtio: fix queue setup consistency

Olivier Matz olivier.matz at 6wind.com
Wed Feb 7 23:01:03 CET 2018


Hi,

It's in my short plans, but unfortunately some other high priority tasks
were inserted before. Honnestly, I'm not sure I'll be able to make it
for the release, but I'll do my best.

Olivier



On Wed, Feb 07, 2018 at 08:31:07AM +0000, Xu, Qian Q wrote:
> Any update, Olivier? 
> We are near to release, and the bug-fix is important for the virtio vector path usage. Thanks. 
> 
> > -----Original Message-----
> > From: dev [mailto:dev-bounces at dpdk.org] On Behalf Of Olivier Matz
> > Sent: Thursday, February 1, 2018 4:28 PM
> > To: Yao, Lei A <lei.a.yao at intel.com>
> > Cc: dev at dpdk.org; yliu at fridaylinux.org; maxime.coquelin at redhat.com;
> > Thomas Monjalon <thomas at monjalon.net>; stable at dpdk.org
> > Subject: Re: [dpdk-dev] [PATCH v2 06/10] net/virtio: fix queue setup consistency
> > 
> > Hi Lei,
> > 
> > It's on my todo list, I'll check this as soon as possible.
> > 
> > Olivier
> > 
> > 
> > On Thu, Feb 01, 2018 at 03:14:15AM +0000, Yao, Lei A wrote:
> > > Hi, Olivier
> > >
> > > This is Lei from DPDK validation team in Intel. During our DPDK
> > > 18.02-rc1 test, I find the following patch will cause one serious issue with virtio
> > vector path:
> > > the traffic can't resume after stop/start the virtio device.
> > >
> > > The step like following:
> > > 1. Launch vhost-user port using testpmd at Host 2. Launch VM with
> > > virtio device, mergeable is off 3. Bind the virtio device to pmd
> > > driver, launch testpmd, let the tx/rx use vector path
> > >     virtio_xmit_pkts_simple
> > >     virtio_recv_pkts_vec
> > > 4. Send traffic to virtio device from vhost side, then stop the virtio
> > > device 5. Start the virtio device again After step 5, the traffic
> > > can't resume.
> > >
> > > Could you help check this and give a fix? This issue will impact the
> > > virtio pmd user experience heavily. By the way, this patch is already
> > > included into V17.11. Looks like we need give a patch to this LTS version.
> > Thanks a lot!
> > >
> > > BRs
> > > Lei
> > > > -----Original Message-----
> > > > From: dev [mailto:dev-bounces at dpdk.org] On Behalf Of Olivier Matz
> > > > Sent: Thursday, September 7, 2017 8:14 PM
> > > > To: dev at dpdk.org; yliu at fridaylinux.org; maxime.coquelin at redhat.com
> > > > Cc: stephen at networkplumber.org; stable at dpdk.org
> > > > Subject: [dpdk-dev] [PATCH v2 06/10] net/virtio: fix queue setup
> > > > consistency
> > > >
> > > > In rx/tx queue setup functions, some code is executed only if
> > > > use_simple_rxtx == 1. The value of this variable can change
> > > > depending on the offload flags or sse support. If Rx queue setup is
> > > > called before Tx queue setup, it can result in an invalid configuration:
> > > >
> > > > - dev_configure is called: use_simple_rxtx is initialized to 0
> > > > - rx queue setup is called: queues are initialized without simple path
> > > >   support
> > > > - tx queue setup is called: use_simple_rxtx switch to 1, and simple
> > > >   Rx/Tx handlers are selected
> > > >
> > > > Fix this by postponing a part of Rx/Tx queue initialization in
> > > > dev_start(), as it was the case in the initial implementation.
> > > >
> > > > Fixes: 48cec290a3d2 ("net/virtio: move queue configure code to
> > > > proper
> > > > place")
> > > > Cc: stable at dpdk.org
> > > >
> > > > Signed-off-by: Olivier Matz <olivier.matz at 6wind.com>
> > > > ---
> > > >  drivers/net/virtio/virtio_ethdev.c | 13 +++++++++++++
> > > > drivers/net/virtio/virtio_ethdev.h |  6 ++++++
> > > >  drivers/net/virtio/virtio_rxtx.c   | 40 ++++++++++++++++++++++++++++++-
> > > > -------
> > > >  3 files changed, 51 insertions(+), 8 deletions(-)
> > > >
> > > > diff --git a/drivers/net/virtio/virtio_ethdev.c
> > > > b/drivers/net/virtio/virtio_ethdev.c
> > > > index 8eee3ff80..c7888f103 100644
> > > > --- a/drivers/net/virtio/virtio_ethdev.c
> > > > +++ b/drivers/net/virtio/virtio_ethdev.c
> > > > @@ -1737,6 +1737,19 @@ virtio_dev_start(struct rte_eth_dev *dev)
> > > >  	struct virtnet_rx *rxvq;
> > > >  	struct virtnet_tx *txvq __rte_unused;
> > > >  	struct virtio_hw *hw = dev->data->dev_private;
> > > > +	int ret;
> > > > +
> > > > +	/* Finish the initialization of the queues */
> > > > +	for (i = 0; i < dev->data->nb_rx_queues; i++) {
> > > > +		ret = virtio_dev_rx_queue_setup_finish(dev, i);
> > > > +		if (ret < 0)
> > > > +			return ret;
> > > > +	}
> > > > +	for (i = 0; i < dev->data->nb_tx_queues; i++) {
> > > > +		ret = virtio_dev_tx_queue_setup_finish(dev, i);
> > > > +		if (ret < 0)
> > > > +			return ret;
> > > > +	}
> > > >
> > > >  	/* check if lsc interrupt feature is enabled */
> > > >  	if (dev->data->dev_conf.intr_conf.lsc) { diff --git
> > > > a/drivers/net/virtio/virtio_ethdev.h
> > > > b/drivers/net/virtio/virtio_ethdev.h
> > > > index c3413c6d9..2039bc547 100644
> > > > --- a/drivers/net/virtio/virtio_ethdev.h
> > > > +++ b/drivers/net/virtio/virtio_ethdev.h
> > > > @@ -92,10 +92,16 @@ int  virtio_dev_rx_queue_setup(struct
> > > > rte_eth_dev *dev, uint16_t rx_queue_id,
> > > >  		const struct rte_eth_rxconf *rx_conf,
> > > >  		struct rte_mempool *mb_pool);
> > > >
> > > > +int virtio_dev_rx_queue_setup_finish(struct rte_eth_dev *dev,
> > > > +				uint16_t rx_queue_id);
> > > > +
> > > >  int  virtio_dev_tx_queue_setup(struct rte_eth_dev *dev, uint16_t
> > > > tx_queue_id,
> > > >  		uint16_t nb_tx_desc, unsigned int socket_id,
> > > >  		const struct rte_eth_txconf *tx_conf);
> > > >
> > > > +int virtio_dev_tx_queue_setup_finish(struct rte_eth_dev *dev,
> > > > +				uint16_t tx_queue_id);
> > > > +
> > > >  uint16_t virtio_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
> > > >  		uint16_t nb_pkts);
> > > >
> > > > diff --git a/drivers/net/virtio/virtio_rxtx.c
> > > > b/drivers/net/virtio/virtio_rxtx.c
> > > > index e30377c51..a32e3229f 100644
> > > > --- a/drivers/net/virtio/virtio_rxtx.c
> > > > +++ b/drivers/net/virtio/virtio_rxtx.c
> > > > @@ -421,9 +421,6 @@ virtio_dev_rx_queue_setup(struct rte_eth_dev *dev,
> > > >  	struct virtio_hw *hw = dev->data->dev_private;
> > > >  	struct virtqueue *vq = hw->vqs[vtpci_queue_idx];
> > > >  	struct virtnet_rx *rxvq;
> > > > -	int error, nbufs;
> > > > -	struct rte_mbuf *m;
> > > > -	uint16_t desc_idx;
> > > >
> > > >  	PMD_INIT_FUNC_TRACE();
> > > >
> > > > @@ -440,10 +437,24 @@ virtio_dev_rx_queue_setup(struct rte_eth_dev
> > > > *dev,
> > > >  	}
> > > >  	dev->data->rx_queues[queue_idx] = rxvq;
> > > >
> > > > +	return 0;
> > > > +}
> > > > +
> > > > +int
> > > > +virtio_dev_rx_queue_setup_finish(struct rte_eth_dev *dev, uint16_t
> > > > queue_idx)
> > > > +{
> > > > +	uint16_t vtpci_queue_idx = 2 * queue_idx +
> > > > VTNET_SQ_RQ_QUEUE_IDX;
> > > > +	struct virtio_hw *hw = dev->data->dev_private;
> > > > +	struct virtqueue *vq = hw->vqs[vtpci_queue_idx];
> > > > +	struct virtnet_rx *rxvq = &vq->rxq;
> > > > +	struct rte_mbuf *m;
> > > > +	uint16_t desc_idx;
> > > > +	int error, nbufs;
> > > > +
> > > > +	PMD_INIT_FUNC_TRACE();
> > > >
> > > >  	/* Allocate blank mbufs for the each rx descriptor */
> > > >  	nbufs = 0;
> > > > -	error = ENOSPC;
> > > >
> > > >  	if (hw->use_simple_rxtx) {
> > > >  		for (desc_idx = 0; desc_idx < vq->vq_nentries; @@ -534,7
> > +545,6
> > > > @@ virtio_dev_tx_queue_setup(struct rte_eth_dev *dev,
> > > >  	struct virtqueue *vq = hw->vqs[vtpci_queue_idx];
> > > >  	struct virtnet_tx *txvq;
> > > >  	uint16_t tx_free_thresh;
> > > > -	uint16_t desc_idx;
> > > >
> > > >  	PMD_INIT_FUNC_TRACE();
> > > >
> > > > @@ -563,9 +573,24 @@ virtio_dev_tx_queue_setup(struct rte_eth_dev
> > > > *dev,
> > > >
> > > >  	vq->vq_free_thresh = tx_free_thresh;
> > > >
> > > > -	if (hw->use_simple_rxtx) {
> > > > -		uint16_t mid_idx  = vq->vq_nentries >> 1;
> > > > +	dev->data->tx_queues[queue_idx] = txvq;
> > > > +	return 0;
> > > > +}
> > > > +
> > > > +int
> > > > +virtio_dev_tx_queue_setup_finish(struct rte_eth_dev *dev,
> > > > +				uint16_t queue_idx)
> > > > +{
> > > > +	uint8_t vtpci_queue_idx = 2 * queue_idx +
> > > > VTNET_SQ_TQ_QUEUE_IDX;
> > > > +	struct virtio_hw *hw = dev->data->dev_private;
> > > > +	struct virtqueue *vq = hw->vqs[vtpci_queue_idx];
> > > > +	uint16_t mid_idx = vq->vq_nentries >> 1;
> > > > +	struct virtnet_tx *txvq = &vq->txq;
> > > > +	uint16_t desc_idx;
> > > >
> > > > +	PMD_INIT_FUNC_TRACE();
> > > > +
> > > > +	if (hw->use_simple_rxtx) {
> > > >  		for (desc_idx = 0; desc_idx < mid_idx; desc_idx++) {
> > > >  			vq->vq_ring.avail->ring[desc_idx] =
> > > >  				desc_idx + mid_idx;
> > > > @@ -587,7 +612,6 @@ virtio_dev_tx_queue_setup(struct rte_eth_dev
> > > > *dev,
> > > >
> > > >  	VIRTQUEUE_DUMP(vq);
> > > >
> > > > -	dev->data->tx_queues[queue_idx] = txvq;
> > > >  	return 0;
> > > >  }
> > > >
> > > > --
> > > > 2.11.0
> > >


More information about the dev mailing list