[dpdk-dev] [PATCH 1/2] pmd: add DCB for VF for ixgbe

Michal Jastrzebski michalx.k.jastrzebski at intel.com
Mon Jan 12 15:43:13 CET 2015


Date: Mon, 12 Jan 2015 15:39:40 +0100
Message-Id: <1421073581-6644-2-git-send-email-michalx.k.jastrzebski at intel.com>
X-Mailer: git-send-email 2.1.1
In-Reply-To: <1421073581-6644-1-git-send-email-michalx.k.jastrzebski at intel.com>
References: <1421073581-6644-1-git-send-email-michalx.k.jastrzebski at intel.com>

From: Pawel Wodkowski <pawelx.wodkowski at intel.com>


This patch add support for DCB in SRIOV mode. When no PFC

is enabled this feature might be used as multiple queues

(up to 8 or 4) for VF.



It incorporate following modifications:

 - Allow zero rx/tx queues to be passed to rte_eth_dev_configure().

   Rationale:

   in SRIOV mode PF use first free VF to RX/TX. If VF count

   is 16 or 32 all recources are assigned to VFs so PF can

   be used only for configuration.

 - split nb_q_per_pool to nb_rx_q_per_pool and nb_tx_q_per_pool

   Rationale:

   rx and tx number of queue might be different if RX and TX are

   configured in different mode. This allow to inform VF about

   proper number of queues.

 - extern mailbox API for DCB mode



Signed-off-by: Pawel Wodkowski <pawelx.wodkowski at intel.com>

---

 lib/librte_ether/rte_ethdev.c       |   84 +++++++++++++++++++++---------

 lib/librte_ether/rte_ethdev.h       |    5 +-

 lib/librte_pmd_e1000/igb_pf.c       |    3 +-

 lib/librte_pmd_ixgbe/ixgbe_ethdev.c |   10 ++--

 lib/librte_pmd_ixgbe/ixgbe_ethdev.h |    1 +

 lib/librte_pmd_ixgbe/ixgbe_pf.c     |   98 ++++++++++++++++++++++++++++++-----

 lib/librte_pmd_ixgbe/ixgbe_rxtx.c   |    7 ++-

 7 files changed, 159 insertions(+), 49 deletions(-)



diff --git a/lib/librte_ether/rte_ethdev.c b/lib/librte_ether/rte_ethdev.c

index 95f2ceb..4c1a494 100644

--- a/lib/librte_ether/rte_ethdev.c

+++ b/lib/librte_ether/rte_ethdev.c

@@ -333,7 +333,7 @@ rte_eth_dev_rx_queue_config(struct rte_eth_dev *dev, uint16_t nb_queues)

 		dev->data->rx_queues = rte_zmalloc("ethdev->rx_queues",

 				sizeof(dev->data->rx_queues[0]) * nb_queues,

 				RTE_CACHE_LINE_SIZE);

-		if (dev->data->rx_queues == NULL) {

+		if (dev->data->rx_queues == NULL && nb_queues > 0) {

 			dev->data->nb_rx_queues = 0;

 			return -(ENOMEM);

 		}

@@ -475,7 +475,7 @@ rte_eth_dev_tx_queue_config(struct rte_eth_dev *dev, uint16_t nb_queues)

 		dev->data->tx_queues = rte_zmalloc("ethdev->tx_queues",

 				sizeof(dev->data->tx_queues[0]) * nb_queues,

 				RTE_CACHE_LINE_SIZE);

-		if (dev->data->tx_queues == NULL) {

+		if (dev->data->tx_queues == NULL && nb_queues > 0) {

 			dev->data->nb_tx_queues = 0;

 			return -(ENOMEM);

 		}

@@ -507,6 +507,7 @@ rte_eth_dev_check_mq_mode(uint8_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,

 		      const struct rte_eth_conf *dev_conf)

 {

 	struct rte_eth_dev *dev = &rte_eth_devices[port_id];

+	struct rte_eth_dev_info dev_info;

 

 	if (RTE_ETH_DEV_SRIOV(dev).active != 0) {

 		/* check multi-queue mode */

@@ -524,11 +525,33 @@ rte_eth_dev_check_mq_mode(uint8_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,

 			return (-EINVAL);

 		}

 

+		if ((dev_conf->rxmode.mq_mode == ETH_MQ_RX_VMDQ_DCB) &&

+			(dev_conf->txmode.mq_mode == ETH_MQ_TX_VMDQ_DCB)) {

+			enum rte_eth_nb_pools rx_pools =

+						dev_conf->rx_adv_conf.vmdq_dcb_conf.nb_queue_pools;

+			enum rte_eth_nb_pools tx_pools =

+						dev_conf->tx_adv_conf.vmdq_dcb_tx_conf.nb_queue_pools;

+

+			if (rx_pools != tx_pools) {

+				/* Only equal number of pools is supported when

+				 * DCB+VMDq in SRIOV */

+				PMD_DEBUG_TRACE("ethdev port_id=%" PRIu8

+						" SRIOV active, DCB+VMDQ mode, "

+						"number of rx and tx pools is not eqaul\n",

+						port_id);

+				return (-EINVAL);

+			}

+		}

+

+		uint16_t nb_rx_q_per_pool = RTE_ETH_DEV_SRIOV(dev).nb_rx_q_per_pool;

+		uint16_t nb_tx_q_per_pool = RTE_ETH_DEV_SRIOV(dev).nb_tx_q_per_pool;

+

 		switch (dev_conf->rxmode.mq_mode) {

-		case ETH_MQ_RX_VMDQ_RSS:

 		case ETH_MQ_RX_VMDQ_DCB:

+			break;

+		case ETH_MQ_RX_VMDQ_RSS:

 		case ETH_MQ_RX_VMDQ_DCB_RSS:

-			/* DCB/RSS VMDQ in SRIOV mode, not implement yet */

+			/* RSS, DCB+RSS VMDQ in SRIOV mode, not implement yet */

 			PMD_DEBUG_TRACE("ethdev port_id=%" PRIu8

 					" SRIOV active, "

 					"unsupported VMDQ mq_mode rx %u\n",

@@ -537,37 +560,32 @@ rte_eth_dev_check_mq_mode(uint8_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,

 		default: /* ETH_MQ_RX_VMDQ_ONLY or ETH_MQ_RX_NONE */

 			/* if nothing mq mode configure, use default scheme */

 			dev->data->dev_conf.rxmode.mq_mode = ETH_MQ_RX_VMDQ_ONLY;

-			if (RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool > 1)

-				RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool = 1;

+			if (nb_rx_q_per_pool > 1)

+				nb_rx_q_per_pool = 1;

 			break;

 		}

 

 		switch (dev_conf->txmode.mq_mode) {

-		case ETH_MQ_TX_VMDQ_DCB:

-			/* DCB VMDQ in SRIOV mode, not implement yet */

-			PMD_DEBUG_TRACE("ethdev port_id=%" PRIu8

-					" SRIOV active, "

-					"unsupported VMDQ mq_mode tx %u\n",

-					port_id, dev_conf->txmode.mq_mode);

-			return (-EINVAL);

+		case ETH_MQ_TX_VMDQ_DCB: /* DCB VMDQ in SRIOV mode*/

+			break;

 		default: /* ETH_MQ_TX_VMDQ_ONLY or ETH_MQ_TX_NONE */

 			/* if nothing mq mode configure, use default scheme */

 			dev->data->dev_conf.txmode.mq_mode = ETH_MQ_TX_VMDQ_ONLY;

-			if (RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool > 1)

-				RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool = 1;

+			if (nb_tx_q_per_pool > 1)

+				nb_tx_q_per_pool = 1;

 			break;

 		}

 

 		/* check valid queue number */

-		if ((nb_rx_q > RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool) ||

-		    (nb_tx_q > RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool)) {

+		if (nb_rx_q > nb_rx_q_per_pool || nb_tx_q > nb_tx_q_per_pool) {

 			PMD_DEBUG_TRACE("ethdev port_id=%d SRIOV active, "

-				    "queue number must less equal to %d\n",

-					port_id, RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool);

+				    "rx/tx queue number must less equal to %d/%d\n",

+					port_id, RTE_ETH_DEV_SRIOV(dev).nb_rx_q_per_pool,

+					RTE_ETH_DEV_SRIOV(dev).nb_tx_q_per_pool);

 			return (-EINVAL);

 		}

 	} else {

-		/* For vmdb+dcb mode check our configuration before we go further */

+		/* For vmdq+dcb mode check our configuration before we go further */

 		if (dev_conf->rxmode.mq_mode == ETH_MQ_RX_VMDQ_DCB) {

 			const struct rte_eth_vmdq_dcb_conf *conf;

 

@@ -606,11 +624,20 @@ rte_eth_dev_check_mq_mode(uint8_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,

 			}

 		}

 

+		/* For DCB we need to obtain maximum number of queues dinamically,

+		 * as this depends on max VF exported in PF */

+		if ((dev_conf->rxmode.mq_mode == ETH_MQ_RX_DCB) ||

+			(dev_conf->txmode.mq_mode == ETH_MQ_TX_DCB)) {

+

+				FUNC_PTR_OR_ERR_RET(*dev->dev_ops->dev_infos_get, -ENOTSUP);

+				(*dev->dev_ops->dev_infos_get)(dev, &dev_info);

+		}

+

 		/* For DCB mode check our configuration before we go further */

 		if (dev_conf->rxmode.mq_mode == ETH_MQ_RX_DCB) {

 			const struct rte_eth_dcb_rx_conf *conf;

 

-			if (nb_rx_q != ETH_DCB_NUM_QUEUES) {

+			if (nb_rx_q != dev_info.max_rx_queues) {

 				PMD_DEBUG_TRACE("ethdev port_id=%d DCB, nb_rx_q "

 						"!= %d\n",

 						port_id, ETH_DCB_NUM_QUEUES);

@@ -630,7 +657,7 @@ rte_eth_dev_check_mq_mode(uint8_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,

 		if (dev_conf->txmode.mq_mode == ETH_MQ_TX_DCB) {

 			const struct rte_eth_dcb_tx_conf *conf;

 

-			if (nb_tx_q != ETH_DCB_NUM_QUEUES) {

+			if (nb_tx_q != dev_info.max_tx_queues) {

 				PMD_DEBUG_TRACE("ethdev port_id=%d DCB, nb_tx_q "

 						"!= %d\n",

 						port_id, ETH_DCB_NUM_QUEUES);

@@ -690,7 +717,10 @@ rte_eth_dev_configure(uint8_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,

 	}

 	if (nb_rx_q == 0) {

 		PMD_DEBUG_TRACE("ethdev port_id=%d nb_rx_q == 0\n", port_id);

-		return (-EINVAL);

+		/* In SRIOV there can be no free resource for PF. So permit use only

+		 * for configuration. */

+		if (RTE_ETH_DEV_SRIOV(dev).active == 0)

+			return (-EINVAL);

 	}

 

 	if (nb_tx_q > dev_info.max_tx_queues) {

@@ -698,9 +728,13 @@ rte_eth_dev_configure(uint8_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,

 				port_id, nb_tx_q, dev_info.max_tx_queues);

 		return (-EINVAL);

 	}

+

 	if (nb_tx_q == 0) {

 		PMD_DEBUG_TRACE("ethdev port_id=%d nb_tx_q == 0\n", port_id);

-		return (-EINVAL);

+		/* In SRIOV there can be no free resource for PF. So permit use only

+		 * for configuration. */

+		if (RTE_ETH_DEV_SRIOV(dev).active == 0)

+			return (-EINVAL);

 	}

 

 	/* Copy the dev_conf parameter into the dev structure */

@@ -750,7 +784,7 @@ rte_eth_dev_configure(uint8_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,

 							ETHER_MAX_LEN;

 	}

 

-	/* multipe queue mode checking */

+	/* multiple queue mode checking */

 	diag = rte_eth_dev_check_mq_mode(port_id, nb_rx_q, nb_tx_q, dev_conf);

 	if (diag != 0) {

 		PMD_DEBUG_TRACE("port%d rte_eth_dev_check_mq_mode = %d\n",

diff --git a/lib/librte_ether/rte_ethdev.h b/lib/librte_ether/rte_ethdev.h

index ce0528f..04fda83 100644

--- a/lib/librte_ether/rte_ethdev.h

+++ b/lib/librte_ether/rte_ethdev.h

@@ -299,7 +299,7 @@ enum rte_eth_rx_mq_mode {

 enum rte_eth_tx_mq_mode {

 	ETH_MQ_TX_NONE    = 0,  /**< It is in neither DCB nor VT mode. */

 	ETH_MQ_TX_DCB,          /**< For TX side,only DCB is on. */

-	ETH_MQ_TX_VMDQ_DCB,	/**< For TX side,both DCB and VT is on. */

+	ETH_MQ_TX_VMDQ_DCB,     /**< For TX side,both DCB and VT is on. */

 	ETH_MQ_TX_VMDQ_ONLY,    /**< Only VT on, no DCB */

 };

 

@@ -1569,7 +1569,8 @@ struct rte_eth_dev {

 

 struct rte_eth_dev_sriov {

 	uint8_t active;               /**< SRIOV is active with 16, 32 or 64 pools */

-	uint8_t nb_q_per_pool;        /**< rx queue number per pool */

+	uint8_t nb_rx_q_per_pool;        /**< rx queue number per pool */

+	uint8_t nb_tx_q_per_pool;        /**< tx queue number per pool */

 	uint16_t def_vmdq_idx;        /**< Default pool num used for PF */

 	uint16_t def_pool_q_idx;      /**< Default pool queue start reg index */

 };

diff --git a/lib/librte_pmd_e1000/igb_pf.c b/lib/librte_pmd_e1000/igb_pf.c

index bc3816a..9d2f858 100644

--- a/lib/librte_pmd_e1000/igb_pf.c

+++ b/lib/librte_pmd_e1000/igb_pf.c

@@ -115,7 +115,8 @@ void igb_pf_host_init(struct rte_eth_dev *eth_dev)

 		rte_panic("Cannot allocate memory for private VF data\n");

 

 	RTE_ETH_DEV_SRIOV(eth_dev).active = ETH_8_POOLS;

-	RTE_ETH_DEV_SRIOV(eth_dev).nb_q_per_pool = nb_queue;

+	RTE_ETH_DEV_SRIOV(eth_dev).nb_rx_q_per_pool = nb_queue;

+	RTE_ETH_DEV_SRIOV(eth_dev).nb_tx_q_per_pool = nb_queue;

 	RTE_ETH_DEV_SRIOV(eth_dev).def_vmdq_idx = vf_num;

 	RTE_ETH_DEV_SRIOV(eth_dev).def_pool_q_idx = (uint16_t)(vf_num * nb_queue);

 

diff --git a/lib/librte_pmd_ixgbe/ixgbe_ethdev.c b/lib/librte_pmd_ixgbe/ixgbe_ethdev.c

index 3fc3738..347f03c 100644

--- a/lib/librte_pmd_ixgbe/ixgbe_ethdev.c

+++ b/lib/librte_pmd_ixgbe/ixgbe_ethdev.c

@@ -3555,14 +3555,14 @@ static int ixgbe_set_vf_rate_limit(struct rte_eth_dev *dev, uint16_t vf,

 	struct ixgbe_hw *hw = IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);

 	struct ixgbe_vf_info *vfinfo =

 		*(IXGBE_DEV_PRIVATE_TO_P_VFDATA(dev->data->dev_private));

-	uint8_t  nb_q_per_pool = RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool;

+	uint8_t  nb_tx_q_per_pool = RTE_ETH_DEV_SRIOV(dev).nb_tx_q_per_pool;

 	uint32_t queue_stride =

 		IXGBE_MAX_RX_QUEUE_NUM / RTE_ETH_DEV_SRIOV(dev).active;

 	uint32_t queue_idx = vf * queue_stride, idx = 0, vf_idx;

-	uint32_t queue_end = queue_idx + nb_q_per_pool - 1;

+	uint32_t tx_queue_end = queue_idx + nb_tx_q_per_pool - 1;

 	uint16_t total_rate = 0;

 

-	if (queue_end >= hw->mac.max_tx_queues)

+	if (tx_queue_end >= hw->mac.max_tx_queues)

 		return -EINVAL;

 

 	if (vfinfo != NULL) {

@@ -3577,7 +3577,7 @@ static int ixgbe_set_vf_rate_limit(struct rte_eth_dev *dev, uint16_t vf,

 		return -EINVAL;

 

 	/* Store tx_rate for this vf. */

-	for (idx = 0; idx < nb_q_per_pool; idx++) {

+	for (idx = 0; idx < nb_tx_q_per_pool; idx++) {

 		if (((uint64_t)0x1 << idx) & q_msk) {

 			if (vfinfo[vf].tx_rate[idx] != tx_rate)

 				vfinfo[vf].tx_rate[idx] = tx_rate;

@@ -3595,7 +3595,7 @@ static int ixgbe_set_vf_rate_limit(struct rte_eth_dev *dev, uint16_t vf,

 	}

 

 	/* Set RTTBCNRC of each queue/pool for vf X  */

-	for (; queue_idx <= queue_end; queue_idx++) {

+	for (; queue_idx <= tx_queue_end; queue_idx++) {

 		if (0x1 & q_msk)

 			ixgbe_set_queue_rate_limit(dev, queue_idx, tx_rate);

 		q_msk = q_msk >> 1;

diff --git a/lib/librte_pmd_ixgbe/ixgbe_ethdev.h b/lib/librte_pmd_ixgbe/ixgbe_ethdev.h

index ca99170..ebf16e9 100644

--- a/lib/librte_pmd_ixgbe/ixgbe_ethdev.h

+++ b/lib/librte_pmd_ixgbe/ixgbe_ethdev.h

@@ -159,6 +159,7 @@ struct ixgbe_vf_info {

 	uint16_t tx_rate[IXGBE_MAX_QUEUE_NUM_PER_VF];

 	uint16_t vlan_count;

 	uint8_t spoofchk_enabled;

+	unsigned int vf_api;

 };

 

 /*

diff --git a/lib/librte_pmd_ixgbe/ixgbe_pf.c b/lib/librte_pmd_ixgbe/ixgbe_pf.c

index 51da1fd..4d30bcf 100644

--- a/lib/librte_pmd_ixgbe/ixgbe_pf.c

+++ b/lib/librte_pmd_ixgbe/ixgbe_pf.c

@@ -127,7 +127,8 @@ void ixgbe_pf_host_init(struct rte_eth_dev *eth_dev)

 		RTE_ETH_DEV_SRIOV(eth_dev).active = ETH_16_POOLS;

 	}

 

-	RTE_ETH_DEV_SRIOV(eth_dev).nb_q_per_pool = nb_queue;

+	RTE_ETH_DEV_SRIOV(eth_dev).nb_rx_q_per_pool = nb_queue;

+	RTE_ETH_DEV_SRIOV(eth_dev).nb_tx_q_per_pool = nb_queue;

 	RTE_ETH_DEV_SRIOV(eth_dev).def_vmdq_idx = vf_num;

 	RTE_ETH_DEV_SRIOV(eth_dev).def_pool_q_idx = (uint16_t)(vf_num * nb_queue);

 

@@ -189,7 +190,7 @@ int ixgbe_pf_host_configure(struct rte_eth_dev *eth_dev)

 	hw->mac.ops.set_vmdq(hw, 0, RTE_ETH_DEV_SRIOV(eth_dev).def_vmdq_idx);

 

 	/*

-	 * SW msut set GCR_EXT.VT_Mode the same as GPIE.VT_Mode

+	 * SW must set GCR_EXT.VT_Mode the same as GPIE.VT_Mode

 	 */

 	gcr_ext = IXGBE_READ_REG(hw, IXGBE_GCR_EXT);

 	gcr_ext &= ~IXGBE_GCR_EXT_VT_MODE_MASK;

@@ -214,19 +215,19 @@ int ixgbe_pf_host_configure(struct rte_eth_dev *eth_dev)

 	}

 

 	IXGBE_WRITE_REG(hw, IXGBE_GCR_EXT, gcr_ext);

-        IXGBE_WRITE_REG(hw, IXGBE_GPIE, gpie);

+	IXGBE_WRITE_REG(hw, IXGBE_GPIE, gpie);

 

-        /*

+	/*

 	 * enable vlan filtering and allow all vlan tags through

 	 */

-        vlanctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL);

-        vlanctrl |= IXGBE_VLNCTRL_VFE ; /* enable vlan filters */

-        IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlanctrl);

+	vlanctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL);

+	vlanctrl |= IXGBE_VLNCTRL_VFE ; /* enable vlan filters */

+	IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlanctrl);

 

-        /* VFTA - enable all vlan filters */

-        for (i = 0; i < IXGBE_MAX_VFTA; i++) {

-                IXGBE_WRITE_REG(hw, IXGBE_VFTA(i), 0xFFFFFFFF);

-        }

+	/* VFTA - enable all vlan filters */

+	for (i = 0; i < IXGBE_MAX_VFTA; i++) {

+		IXGBE_WRITE_REG(hw, IXGBE_VFTA(i), 0xFFFFFFFF);

+	}

 

 	/* Enable MAC Anti-Spoofing */

 	hw->mac.ops.set_mac_anti_spoofing(hw, FALSE, vf_num);

@@ -369,6 +370,73 @@ ixgbe_vf_reset(struct rte_eth_dev *dev, uint16_t vf, uint32_t *msgbuf)

 }

 

 static int

+ixgbe_negotiate_vf_api(struct rte_eth_dev *dev, uint32_t vf, uint32_t *msgbuf)

+{

+	struct ixgbe_vf_info *vfinfo =

+		*(IXGBE_DEV_PRIVATE_TO_P_VFDATA(dev->data->dev_private));

+	int api = msgbuf[1];

+

+	switch (api) {

+	case ixgbe_mbox_api_10:

+	case ixgbe_mbox_api_11:

+		vfinfo[vf].vf_api = api;

+		return 0;

+	default:

+		break;

+	}

+

+	RTE_LOG(DEBUG, PMD, "VF %d requested invalid api version %u\n", vf, api);

+	return -1;

+}

+

+static int

+ixgbe_get_vf_queues(struct rte_eth_dev *dev, uint32_t vf, uint32_t *msgbuf)

+{

+	struct ixgbe_vf_info *vfinfo =

+		*(IXGBE_DEV_PRIVATE_TO_P_VFDATA(dev->data->dev_private));

+	struct ixgbe_dcb_config *dcb_cfg =

+			IXGBE_DEV_PRIVATE_TO_DCB_CFG(dev->data->dev_private);

+

+	uint8_t num_tcs = dcb_cfg->num_tcs.pg_tcs;

+

+	/* verify the PF is supporting the correct APIs */

+	switch (vfinfo[vf].vf_api) {

+	case ixgbe_mbox_api_10:

+	case ixgbe_mbox_api_11:

+		break;

+	default:

+		return -1;

+	}

+

+	if (RTE_ETH_DEV_SRIOV(dev).active) {

+		if (dev->data->dev_conf.rxmode.mq_mode == ETH_MQ_RX_VMDQ_DCB)

+			msgbuf[IXGBE_VF_TX_QUEUES] = num_tcs;

+		else

+			msgbuf[IXGBE_VF_TX_QUEUES] = 1;

+

+		if (dev->data->dev_conf.txmode.mq_mode == ETH_MQ_TX_VMDQ_DCB)

+			msgbuf[IXGBE_VF_RX_QUEUES] = num_tcs;

+		else

+			msgbuf[IXGBE_VF_RX_QUEUES] = 1;

+	}	else {

+		/* only allow 1 Tx queue for bandwidth limiting */

+		msgbuf[IXGBE_VF_TX_QUEUES] = 1;

+		msgbuf[IXGBE_VF_RX_QUEUES] = 1;

+	}

+

+	/* notify VF of need for VLAN tag stripping, and correct queue */

+	if (num_tcs)

+		msgbuf[IXGBE_VF_TRANS_VLAN] = num_tcs;

+	else

+		msgbuf[IXGBE_VF_TRANS_VLAN] = 0;

+

+	/* notify VF of default queue */

+	msgbuf[IXGBE_VF_DEF_QUEUE] = 0;

+

+	return 0;

+}

+

+static int

 ixgbe_vf_set_mac_addr(struct rte_eth_dev *dev, uint32_t vf, uint32_t *msgbuf)

 {

 	struct ixgbe_hw *hw = IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);

@@ -512,6 +580,12 @@ ixgbe_rcv_msg_from_vf(struct rte_eth_dev *dev, uint16_t vf)

 	case IXGBE_VF_SET_VLAN:

 		retval = ixgbe_vf_set_vlan(dev, vf, msgbuf);

 		break;

+	case IXGBE_VF_API_NEGOTIATE:

+		retval = ixgbe_negotiate_vf_api(dev, vf, msgbuf);

+		break;

+	case IXGBE_VF_GET_QUEUES:

+		retval = ixgbe_get_vf_queues(dev, vf, msgbuf);

+		break;

 	default:

 		PMD_DRV_LOG(DEBUG, "Unhandled Msg %8.8x", (unsigned)msgbuf[0]);

 		retval = IXGBE_ERR_MBX;

@@ -526,7 +600,7 @@ ixgbe_rcv_msg_from_vf(struct rte_eth_dev *dev, uint16_t vf)

 

 	msgbuf[0] |= IXGBE_VT_MSGTYPE_CTS;

 

-	ixgbe_write_mbx(hw, msgbuf, 1, vf);

+	ixgbe_write_mbx(hw, msgbuf, mbx_size, vf);

 

 	return retval;

 }

diff --git a/lib/librte_pmd_ixgbe/ixgbe_rxtx.c b/lib/librte_pmd_ixgbe/ixgbe_rxtx.c

index e10d6a2..49b44fe 100644

--- a/lib/librte_pmd_ixgbe/ixgbe_rxtx.c

+++ b/lib/librte_pmd_ixgbe/ixgbe_rxtx.c

@@ -3166,10 +3166,9 @@ void ixgbe_configure_dcb(struct rte_eth_dev *dev)

 

 	/* check support mq_mode for DCB */

 	if ((dev_conf->rxmode.mq_mode != ETH_MQ_RX_VMDQ_DCB) &&

-	    (dev_conf->rxmode.mq_mode != ETH_MQ_RX_DCB))

-		return;

-

-	if (dev->data->nb_rx_queues != ETH_DCB_NUM_QUEUES)

+	    (dev_conf->rxmode.mq_mode != ETH_MQ_RX_DCB) &&

+	    (dev_conf->txmode.mq_mode != ETH_MQ_TX_VMDQ_DCB) &&

+	    (dev_conf->txmode.mq_mode != ETH_MQ_TX_DCB))

 		return;

 

 	/** Configure DCB hardware **/

-- 

1.7.9.5





More information about the dev mailing list