[dpdk-dev] [PATCH] net/liquidio: add support for queue re-configuration

Shijith Thotton shijith.thotton at caviumnetworks.com
Mon Nov 20 12:59:51 CET 2017


Support for re-configuration of number of queues per port and descriptor
size. Renamed variable representing number of descriptors as nb_desc
from max_count.

Signed-off-by: Shijith Thotton <shijith.thotton at caviumnetworks.com>
---
 drivers/net/liquidio/base/lio_23xx_vf.c |  54 +----------
 drivers/net/liquidio/base/lio_23xx_vf.h |   5 -
 drivers/net/liquidio/base/lio_hw_defs.h |   2 +
 drivers/net/liquidio/lio_ethdev.c       | 156 +++++++++++++++++++-------------
 drivers/net/liquidio/lio_rxtx.c         |  81 ++++++-----------
 drivers/net/liquidio/lio_struct.h       |   6 +-
 6 files changed, 132 insertions(+), 172 deletions(-)

diff --git a/drivers/net/liquidio/base/lio_23xx_vf.c b/drivers/net/liquidio/base/lio_23xx_vf.c
index e30c20d..7f7e98d 100644
--- a/drivers/net/liquidio/base/lio_23xx_vf.c
+++ b/drivers/net/liquidio/base/lio_23xx_vf.c
@@ -150,6 +150,8 @@
 
 		reg_val &= 0xEFFFFFFFFFFFFFFFL;
 
+		lio_write_csr(lio_dev, CN23XX_SLI_OQ_PKTS_SENT(q_no), reg_val);
+
 		reg_val =
 		    lio_read_csr(lio_dev, CN23XX_SLI_OQ_PKT_CONTROL(q_no));
 
@@ -211,7 +213,7 @@
 	/* Write the start of the input queue's ring and its size */
 	lio_write_csr64(lio_dev, CN23XX_SLI_IQ_BASE_ADDR64(iq_no),
 			iq->base_addr_dma);
-	lio_write_csr(lio_dev, CN23XX_SLI_IQ_SIZE(iq_no), iq->max_count);
+	lio_write_csr(lio_dev, CN23XX_SLI_IQ_SIZE(iq_no), iq->nb_desc);
 
 	/* Remember the doorbell & instruction count register addr
 	 * for this queue
@@ -243,7 +245,7 @@
 
 	lio_write_csr64(lio_dev, CN23XX_SLI_OQ_BASE_ADDR64(oq_no),
 			droq->desc_ring_dma);
-	lio_write_csr(lio_dev, CN23XX_SLI_OQ_SIZE(oq_no), droq->max_count);
+	lio_write_csr(lio_dev, CN23XX_SLI_OQ_SIZE(oq_no), droq->nb_desc);
 
 	lio_write_csr(lio_dev, CN23XX_SLI_OQ_BUFF_INFO_SIZE(oq_no),
 		      (droq->buffer_size | (OCTEON_RH_SIZE << 16)));
@@ -538,51 +540,3 @@
 	return 0;
 }
 
-int
-cn23xx_vf_set_io_queues_off(struct lio_device *lio_dev)
-{
-	uint32_t loop = CN23XX_VF_BUSY_READING_REG_LOOP_COUNT;
-	uint64_t q_no;
-
-	/* Disable the i/p and o/p queues for this Octeon.
-	 * IOQs will already be in reset.
-	 * If RST bit is set, wait for Quiet bit to be set
-	 * Once Quiet bit is set, clear the RST bit
-	 */
-	PMD_INIT_FUNC_TRACE();
-
-	for (q_no = 0; q_no < lio_dev->sriov_info.rings_per_vf; q_no++) {
-		volatile uint64_t reg_val;
-
-		reg_val = lio_read_csr64(lio_dev,
-					 CN23XX_SLI_IQ_PKT_CONTROL64(q_no));
-		while ((reg_val & CN23XX_PKT_INPUT_CTL_RST) && !(reg_val &
-					 CN23XX_PKT_INPUT_CTL_QUIET) && loop) {
-			reg_val = lio_read_csr64(
-					lio_dev,
-					CN23XX_SLI_IQ_PKT_CONTROL64(q_no));
-			loop = loop - 1;
-		}
-
-		if (loop == 0) {
-			lio_dev_err(lio_dev,
-				    "clearing the reset reg failed or setting the quiet reg failed for qno %lu\n",
-				    (unsigned long)q_no);
-			return -1;
-		}
-
-		reg_val = reg_val & ~CN23XX_PKT_INPUT_CTL_RST;
-		lio_write_csr64(lio_dev, CN23XX_SLI_IQ_PKT_CONTROL64(q_no),
-				reg_val);
-
-		reg_val = lio_read_csr64(lio_dev,
-					 CN23XX_SLI_IQ_PKT_CONTROL64(q_no));
-		if (reg_val & CN23XX_PKT_INPUT_CTL_RST) {
-			lio_dev_err(lio_dev, "unable to reset qno %lu\n",
-				    (unsigned long)q_no);
-			return -1;
-		}
-	}
-
-	return 0;
-}
diff --git a/drivers/net/liquidio/base/lio_23xx_vf.h b/drivers/net/liquidio/base/lio_23xx_vf.h
index ad8db0d..ea93524 100644
--- a/drivers/net/liquidio/base/lio_23xx_vf.h
+++ b/drivers/net/liquidio/base/lio_23xx_vf.h
@@ -80,11 +80,6 @@
 	return default_lio_conf;
 }
 
-/** Turns off the input and output queues for the device
- *  @param lio_dev which device io queues to disable
- */
-int cn23xx_vf_set_io_queues_off(struct lio_device *lio_dev);
-
 #define CN23XX_VF_BUSY_READING_REG_LOOP_COUNT	100000
 
 void cn23xx_vf_ask_pf_to_do_flr(struct lio_device *lio_dev);
diff --git a/drivers/net/liquidio/base/lio_hw_defs.h b/drivers/net/liquidio/base/lio_hw_defs.h
index fe5c3bb..5595075 100644
--- a/drivers/net/liquidio/base/lio_hw_defs.h
+++ b/drivers/net/liquidio/base/lio_hw_defs.h
@@ -113,6 +113,7 @@ enum lio_card_type {
 
 #define LIO_FW_VERSION_LENGTH		32
 
+#define LIO_Q_RECONF_MIN_VERSION	"1.7.0"
 #define LIO_VF_TRUST_MIN_VERSION	"1.7.1"
 
 /** Tag types used by Octeon cores in its work. */
@@ -156,6 +157,7 @@ enum octeon_tag_type {
 #define LIO_CMD_ADD_VLAN_FILTER		0x17
 #define LIO_CMD_DEL_VLAN_FILTER		0x18
 #define LIO_CMD_VXLAN_PORT_CONFIG	0x19
+#define LIO_CMD_QUEUE_COUNT_CTL		0x1f
 
 #define LIO_CMD_VXLAN_PORT_ADD		0x0
 #define LIO_CMD_VXLAN_PORT_DEL		0x1
diff --git a/drivers/net/liquidio/lio_ethdev.c b/drivers/net/liquidio/lio_ethdev.c
index 84b8a32..6b5d52e 100644
--- a/drivers/net/liquidio/lio_ethdev.c
+++ b/drivers/net/liquidio/lio_ethdev.c
@@ -1228,12 +1228,10 @@ struct rte_lio_xstats_name_off {
 
 	fw_mapped_oq = lio_dev->linfo.rxpciq[q_no].s.q_no;
 
-	if ((lio_dev->droq[fw_mapped_oq]) &&
-	    (num_rx_descs != lio_dev->droq[fw_mapped_oq]->max_count)) {
-		lio_dev_err(lio_dev,
-			    "Reconfiguring Rx descs not supported. Configure descs to same value %u or restart application\n",
-			    lio_dev->droq[fw_mapped_oq]->max_count);
-		return -ENOTSUP;
+	/* Free previous allocation if any */
+	if (eth_dev->data->rx_queues[q_no] != NULL) {
+		lio_dev_rx_queue_release(eth_dev->data->rx_queues[q_no]);
+		eth_dev->data->rx_queues[q_no] = NULL;
 	}
 
 	mbp_priv = rte_mempool_get_priv(mp);
@@ -1267,10 +1265,6 @@ struct rte_lio_xstats_name_off {
 	int oq_no;
 
 	if (droq) {
-		/* Run time queue deletion not supported */
-		if (droq->lio_dev->port_configured)
-			return;
-
 		oq_no = droq->q_no;
 		lio_delete_droq_queue(droq->lio_dev, oq_no);
 	}
@@ -1314,12 +1308,10 @@ struct rte_lio_xstats_name_off {
 
 	lio_dev_dbg(lio_dev, "setting up tx queue %u\n", q_no);
 
-	if ((lio_dev->instr_queue[fw_mapped_iq] != NULL) &&
-	    (num_tx_descs != lio_dev->instr_queue[fw_mapped_iq]->max_count)) {
-		lio_dev_err(lio_dev,
-			    "Reconfiguring Tx descs not supported. Configure descs to same value %u or restart application\n",
-			    lio_dev->instr_queue[fw_mapped_iq]->max_count);
-		return -ENOTSUP;
+	/* Free previous allocation if any */
+	if (eth_dev->data->tx_queues[q_no] != NULL) {
+		lio_dev_tx_queue_release(eth_dev->data->tx_queues[q_no]);
+		eth_dev->data->tx_queues[q_no] = NULL;
 	}
 
 	retval = lio_setup_iq(lio_dev, q_no, lio_dev->linfo.txpciq[q_no],
@@ -1331,7 +1323,7 @@ struct rte_lio_xstats_name_off {
 	}
 
 	retval = lio_setup_sglists(lio_dev, q_no, fw_mapped_iq,
-				lio_dev->instr_queue[fw_mapped_iq]->max_count,
+				lio_dev->instr_queue[fw_mapped_iq]->nb_desc,
 				socket_id);
 
 	if (retval) {
@@ -1362,10 +1354,6 @@ struct rte_lio_xstats_name_off {
 
 
 	if (tq) {
-		/* Run time queue deletion not supported */
-		if (tq->lio_dev->port_configured)
-			return;
-
 		/* Free sg_list */
 		lio_delete_sglist(tq);
 
@@ -1534,6 +1522,8 @@ struct rte_lio_xstats_name_off {
 
 	lio_send_rx_ctrl_cmd(eth_dev, 0);
 
+	lio_wait_for_instr_fetch(lio_dev);
+
 	/* Clear recorded link status */
 	lio_dev->linfo.link.link_status64 = 0;
 }
@@ -1607,34 +1597,14 @@ struct rte_lio_xstats_name_off {
 lio_dev_close(struct rte_eth_dev *eth_dev)
 {
 	struct lio_device *lio_dev = LIO_DEV(eth_dev);
-	uint32_t i;
 
 	lio_dev_info(lio_dev, "closing port %d\n", eth_dev->data->port_id);
 
 	if (lio_dev->intf_open)
 		lio_dev_stop(eth_dev);
 
-	lio_wait_for_instr_fetch(lio_dev);
-
-	lio_dev->fn_list.disable_io_queues(lio_dev);
-
-	cn23xx_vf_set_io_queues_off(lio_dev);
-
-	/* Reset iq regs (IQ_DBELL).
-	 * Clear sli_pktx_cnts (OQ_PKTS_SENT).
-	 */
-	for (i = 0; i < lio_dev->nb_rx_queues; i++) {
-		struct lio_droq *droq = lio_dev->droq[i];
-
-		if (droq == NULL)
-			break;
-
-		uint32_t pkt_count = rte_read32(droq->pkts_sent_reg);
-
-		lio_dev_dbg(lio_dev,
-			    "pending oq count %u\n", pkt_count);
-		rte_write32(pkt_count, droq->pkts_sent_reg);
-	}
+	/* Reset ioq regs */
+	lio_dev->fn_list.setup_device_regs(lio_dev);
 
 	if (lio_dev->pci_dev->kdrv == RTE_KDRV_IGB_UIO) {
 		cn23xx_vf_ask_pf_to_do_flr(lio_dev);
@@ -1724,7 +1694,76 @@ struct rte_lio_xstats_name_off {
 		lio_dev_err(lio_dev, "TNL_TX_CSUM command timed out\n");
 }
 
-static int lio_dev_configure(struct rte_eth_dev *eth_dev)
+static int
+lio_send_queue_count_update(struct rte_eth_dev *eth_dev, int num_txq,
+			    int num_rxq)
+{
+	struct lio_device *lio_dev = LIO_DEV(eth_dev);
+	struct lio_dev_ctrl_cmd ctrl_cmd;
+	struct lio_ctrl_pkt ctrl_pkt;
+
+	if (strcmp(lio_dev->firmware_version, LIO_Q_RECONF_MIN_VERSION) < 0) {
+		lio_dev_err(lio_dev, "Require firmware version >= %s\n",
+			    LIO_Q_RECONF_MIN_VERSION);
+		return -ENOTSUP;
+	}
+
+	/* flush added to prevent cmd failure
+	 * incase the queue is full
+	 */
+	lio_flush_iq(lio_dev, lio_dev->instr_queue[0]);
+
+	memset(&ctrl_pkt, 0, sizeof(struct lio_ctrl_pkt));
+	memset(&ctrl_cmd, 0, sizeof(struct lio_dev_ctrl_cmd));
+
+	ctrl_cmd.eth_dev = eth_dev;
+	ctrl_cmd.cond = 0;
+
+	ctrl_pkt.ncmd.s.cmd = LIO_CMD_QUEUE_COUNT_CTL;
+	ctrl_pkt.ncmd.s.param1 = num_txq;
+	ctrl_pkt.ncmd.s.param2 = num_rxq;
+	ctrl_pkt.ctrl_cmd = &ctrl_cmd;
+
+	if (lio_send_ctrl_pkt(lio_dev, &ctrl_pkt)) {
+		lio_dev_err(lio_dev, "Failed to send queue count control command\n");
+		return -1;
+	}
+
+	if (lio_wait_for_ctrl_cmd(lio_dev, &ctrl_cmd)) {
+		lio_dev_err(lio_dev, "Queue count control command timed out\n");
+		return -1;
+	}
+
+	return 0;
+}
+
+static int
+lio_reconf_queues(struct rte_eth_dev *eth_dev, int num_txq, int num_rxq)
+{
+	struct lio_device *lio_dev = LIO_DEV(eth_dev);
+
+	if (lio_dev->nb_rx_queues != num_rxq ||
+	    lio_dev->nb_tx_queues != num_txq) {
+		if (lio_send_queue_count_update(eth_dev, num_txq, num_rxq))
+			return -1;
+		lio_dev->nb_rx_queues = num_rxq;
+		lio_dev->nb_tx_queues = num_txq;
+	}
+
+	if (lio_dev->intf_open)
+		lio_dev_stop(eth_dev);
+
+	/* Reset ioq registers */
+	if (lio_dev->fn_list.setup_device_regs(lio_dev)) {
+		lio_dev_err(lio_dev, "Failed to configure device registers\n");
+		return -1;
+	}
+
+	return 0;
+}
+
+static int
+lio_dev_configure(struct rte_eth_dev *eth_dev)
 {
 	struct lio_device *lio_dev = LIO_DEV(eth_dev);
 	uint16_t timeout = LIO_MAX_CMD_TIMEOUT;
@@ -1737,22 +1776,21 @@ static int lio_dev_configure(struct rte_eth_dev *eth_dev)
 
 	PMD_INIT_FUNC_TRACE();
 
-	/* Re-configuring firmware not supported.
-	 * Can't change tx/rx queues per port from initial value.
+	/* Inform firmware about change in number of queues to use.
+	 * Disable IO queues and reset registers for re-configuration.
 	 */
-	if (lio_dev->port_configured) {
-		if ((lio_dev->nb_rx_queues != eth_dev->data->nb_rx_queues) ||
-		    (lio_dev->nb_tx_queues != eth_dev->data->nb_tx_queues)) {
-			lio_dev_err(lio_dev,
-				    "rxq/txq re-conf not supported. Restart application with new value.\n");
-			return -ENOTSUP;
-		}
-		return 0;
-	}
+	if (lio_dev->port_configured)
+		return lio_reconf_queues(eth_dev,
+					 eth_dev->data->nb_tx_queues,
+					 eth_dev->data->nb_rx_queues);
 
 	lio_dev->nb_rx_queues = eth_dev->data->nb_rx_queues;
 	lio_dev->nb_tx_queues = eth_dev->data->nb_tx_queues;
 
+	/* Set max number of queues which can be re-configured. */
+	lio_dev->max_rx_queues = eth_dev->data->nb_rx_queues;
+	lio_dev->max_tx_queues = eth_dev->data->nb_tx_queues;
+
 	resp_size = sizeof(struct lio_if_cfg_resp);
 	sc = lio_alloc_soft_command(lio_dev, 0, resp_size, 0);
 	if (sc == NULL)
@@ -1879,9 +1917,6 @@ static int lio_dev_configure(struct rte_eth_dev *eth_dev)
 
 	lio_free_soft_command(sc);
 
-	/* Disable iq_0 for reconf */
-	lio_dev->fn_list.disable_io_queues(lio_dev);
-
 	/* Reset ioq regs */
 	lio_dev->fn_list.setup_device_regs(lio_dev);
 
@@ -2021,11 +2056,6 @@ static int lio_dev_configure(struct rte_eth_dev *eth_dev)
 		rte_delay_ms(LIO_PCI_FLR_WAIT * 2);
 	}
 
-	if (cn23xx_vf_set_io_queues_off(lio_dev)) {
-		lio_dev_err(lio_dev, "Setting io queues off failed\n");
-		goto error;
-	}
-
 	if (lio_dev->fn_list.setup_device_regs(lio_dev)) {
 		lio_dev_err(lio_dev, "Failed to configure device registers\n");
 		goto error;
diff --git a/drivers/net/liquidio/lio_rxtx.c b/drivers/net/liquidio/lio_rxtx.c
index 376893a..759556c 100644
--- a/drivers/net/liquidio/lio_rxtx.c
+++ b/drivers/net/liquidio/lio_rxtx.c
@@ -42,7 +42,7 @@
 
 #define LIO_MAX_SG 12
 /* Flush iq if available tx_desc fall below LIO_FLUSH_WM */
-#define LIO_FLUSH_WM(_iq) ((_iq)->max_count / 2)
+#define LIO_FLUSH_WM(_iq) ((_iq)->nb_desc / 2)
 #define LIO_PKT_IN_DONE_CNT_MASK 0x00000000FFFFFFFFULL
 
 static void
@@ -70,7 +70,7 @@
 {
 	uint32_t i;
 
-	for (i = 0; i < droq->max_count; i++) {
+	for (i = 0; i < droq->nb_desc; i++) {
 		if (droq->recv_buf_list[i].buffer) {
 			rte_pktmbuf_free((struct rte_mbuf *)
 					 droq->recv_buf_list[i].buffer);
@@ -89,7 +89,7 @@
 	uint32_t i;
 	void *buf;
 
-	for (i = 0; i < droq->max_count; i++) {
+	for (i = 0; i < droq->nb_desc; i++) {
 		buf = rte_pktmbuf_alloc(droq->mpool);
 		if (buf == NULL) {
 			lio_dev_err(lio_dev, "buffer alloc failed\n");
@@ -164,7 +164,7 @@
 {
 	droq->info_mz = rte_eth_dma_zone_reserve(lio_dev->eth_dev,
 						 "info_list", droq->q_no,
-						 (droq->max_count *
+						 (droq->nb_desc *
 							LIO_DROQ_INFO_SIZE),
 						 RTE_CACHE_LINE_SIZE,
 						 socket_id);
@@ -206,10 +206,10 @@
 
 	c_refill_threshold = LIO_OQ_REFILL_THRESHOLD_CFG(lio_dev);
 
-	droq->max_count = num_descs;
+	droq->nb_desc = num_descs;
 	droq->buffer_size = desc_size;
 
-	desc_ring_size = droq->max_count * LIO_DROQ_DESC_SIZE;
+	desc_ring_size = droq->nb_desc * LIO_DROQ_DESC_SIZE;
 	droq->desc_ring_mz = rte_eth_dma_zone_reserve(lio_dev->eth_dev,
 						      "droq", q_no,
 						      desc_ring_size,
@@ -228,7 +228,7 @@
 	lio_dev_dbg(lio_dev, "droq[%d]: desc_ring: virt: 0x%p, dma: %lx\n",
 		    q_no, droq->desc_ring, (unsigned long)droq->desc_ring_dma);
 	lio_dev_dbg(lio_dev, "droq[%d]: num_desc: %d\n", q_no,
-		    droq->max_count);
+		    droq->nb_desc);
 
 	droq->info_list = lio_alloc_info_buffer(lio_dev, droq, socket_id);
 	if (droq->info_list == NULL) {
@@ -237,7 +237,7 @@
 	}
 
 	droq->recv_buf_list = rte_zmalloc_socket("recv_buf_list",
-						 (droq->max_count *
+						 (droq->nb_desc *
 							LIO_DROQ_RECVBUF_SIZE),
 						 RTE_CACHE_LINE_SIZE,
 						 socket_id);
@@ -274,11 +274,6 @@
 
 	PMD_INIT_FUNC_TRACE();
 
-	if (lio_dev->droq[oq_no]) {
-		lio_dev_dbg(lio_dev, "Droq %d in use\n", oq_no);
-		return 0;
-	}
-
 	/* Allocate the DS for the new droq. */
 	droq = rte_zmalloc_socket("ethdev RX queue", sizeof(*droq),
 				  RTE_CACHE_LINE_SIZE, socket_id);
@@ -303,7 +298,7 @@
 	/* Send credit for octeon output queues. credits are always
 	 * sent after the output queue is enabled.
 	 */
-	rte_write32(lio_dev->droq[oq_no]->max_count,
+	rte_write32(lio_dev->droq[oq_no]->nb_desc,
 		    lio_dev->droq[oq_no]->pkts_credit_reg);
 	rte_wmb();
 
@@ -342,13 +337,13 @@
 			do {
 				droq->refill_idx = lio_incr_index(
 							droq->refill_idx, 1,
-							droq->max_count);
+							droq->nb_desc);
 				desc_refilled++;
 				droq->refill_count--;
 			} while (droq->recv_buf_list[droq->refill_idx].buffer);
 		}
 		refill_index = lio_incr_index(refill_index, 1,
-					      droq->max_count);
+					      droq->nb_desc);
 	}	/* while */
 
 	return desc_refilled;
@@ -379,7 +374,7 @@
 
 	desc_ring = droq->desc_ring;
 
-	while (droq->refill_count && (desc_refilled < droq->max_count)) {
+	while (droq->refill_count && (desc_refilled < droq->nb_desc)) {
 		/* If a valid buffer exists (happens if there is no dispatch),
 		 * reuse the buffer, else allocate.
 		 */
@@ -402,7 +397,7 @@
 		droq->info_list[droq->refill_idx].length = 0;
 
 		droq->refill_idx = lio_incr_index(droq->refill_idx, 1,
-						  droq->max_count);
+						  droq->nb_desc);
 		desc_refilled++;
 		droq->refill_count--;
 	}
@@ -449,7 +444,7 @@
 		buf_cnt = lio_droq_get_bufcount(droq->buffer_size,
 						(uint32_t)info->length);
 		droq->read_idx = lio_incr_index(droq->read_idx, buf_cnt,
-						droq->max_count);
+						droq->nb_desc);
 		droq->refill_count += buf_cnt;
 	} else {
 		if (info->length <= droq->buffer_size) {
@@ -462,7 +457,7 @@
 			droq->recv_buf_list[droq->read_idx].buffer = NULL;
 			droq->read_idx = lio_incr_index(
 						droq->read_idx, 1,
-						droq->max_count);
+						droq->nb_desc);
 			droq->refill_count++;
 
 			if (likely(nicbuf != NULL)) {
@@ -556,7 +551,7 @@
 				pkt_len += cpy_len;
 				droq->read_idx = lio_incr_index(
 							droq->read_idx,
-							1, droq->max_count);
+							1, droq->nb_desc);
 				droq->refill_count++;
 
 				/* Prefetch buffer pointers when on a
@@ -737,7 +732,7 @@
 	iq->base_addr_dma = iq->iq_mz->iova;
 	iq->base_addr = (uint8_t *)iq->iq_mz->addr;
 
-	iq->max_count = num_descs;
+	iq->nb_desc = num_descs;
 
 	/* Initialize a list to holds requests that have been posted to Octeon
 	 * but has yet to be fetched by octeon
@@ -756,7 +751,7 @@
 
 	lio_dev_dbg(lio_dev, "IQ[%d]: base: %p basedma: %lx count: %d\n",
 		    iq_no, iq->base_addr, (unsigned long)iq->base_addr_dma,
-		    iq->max_count);
+		    iq->nb_desc);
 
 	iq->lio_dev = lio_dev;
 	iq->txpciq.txpciq64 = txpciq.txpciq64;
@@ -853,14 +848,6 @@
 {
 	uint32_t iq_no = (uint32_t)txpciq.s.q_no;
 
-	if (lio_dev->instr_queue[iq_no]) {
-		lio_dev_dbg(lio_dev, "IQ is in use. Cannot create the IQ: %d again\n",
-			    iq_no);
-		lio_dev->instr_queue[iq_no]->txpciq.txpciq64 = txpciq.txpciq64;
-		lio_dev->instr_queue[iq_no]->app_ctx = app_ctx;
-		return 0;
-	}
-
 	lio_dev->instr_queue[iq_no] = rte_zmalloc_socket("ethdev TX queue",
 						sizeof(struct lio_instr_queue),
 						RTE_CACHE_LINE_SIZE, socket_id);
@@ -870,23 +857,15 @@
 	lio_dev->instr_queue[iq_no]->q_index = q_index;
 	lio_dev->instr_queue[iq_no]->app_ctx = app_ctx;
 
-	if (lio_init_instr_queue(lio_dev, txpciq, num_descs, socket_id))
-		goto release_lio_iq;
+	if (lio_init_instr_queue(lio_dev, txpciq, num_descs, socket_id)) {
+		rte_free(lio_dev->instr_queue[iq_no]);
+		lio_dev->instr_queue[iq_no] = NULL;
+		return -1;
+	}
 
 	lio_dev->num_iqs++;
-	if (lio_dev->fn_list.enable_io_queues(lio_dev))
-		goto delete_lio_iq;
 
 	return 0;
-
-delete_lio_iq:
-	lio_delete_instr_queue(lio_dev, iq_no);
-	lio_dev->num_iqs--;
-release_lio_iq:
-	rte_free(lio_dev->instr_queue[iq_no]);
-	lio_dev->instr_queue[iq_no] = NULL;
-
-	return -1;
 }
 
 int
@@ -957,14 +936,14 @@
 	 * position if queue gets full before Octeon could fetch any instr.
 	 */
 	if (rte_atomic64_read(&iq->instr_pending) >=
-			(int32_t)(iq->max_count - 1)) {
+			(int32_t)(iq->nb_desc - 1)) {
 		st.status = LIO_IQ_SEND_FAILED;
 		st.index = -1;
 		return st;
 	}
 
 	if (rte_atomic64_read(&iq->instr_pending) >=
-			(int32_t)(iq->max_count - 2))
+			(int32_t)(iq->nb_desc - 2))
 		st.status = LIO_IQ_SEND_STOP;
 
 	copy_cmd_into_iq(iq, cmd);
@@ -972,7 +951,7 @@
 	/* "index" is returned, host_write_index is modified. */
 	st.index = iq->host_write_index;
 	iq->host_write_index = lio_incr_index(iq->host_write_index, 1,
-					      iq->max_count);
+					      iq->nb_desc);
 	iq->fill_cnt++;
 
 	/* Flush the command into memory. We need to be sure the data is in
@@ -1074,7 +1053,7 @@
 
 skip_this:
 		inst_count++;
-		old = lio_incr_index(old, 1, iq->max_count);
+		old = lio_incr_index(old, 1, iq->nb_desc);
 	}
 
 	iq->flush_index = old;
@@ -1094,7 +1073,7 @@
 	/* Add last_done and modulo with the IQ size to get new index */
 	iq->lio_read_index = (iq->lio_read_index +
 			(uint32_t)(last_done & LIO_PKT_IN_DONE_CNT_MASK)) %
-			iq->max_count;
+			iq->nb_desc;
 }
 
 int
@@ -1552,7 +1531,7 @@ struct lio_soft_command *
 static inline uint32_t
 lio_iq_get_available(struct lio_device *lio_dev, uint32_t q_no)
 {
-	return ((lio_dev->instr_queue[q_no]->max_count - 1) -
+	return ((lio_dev->instr_queue[q_no]->nb_desc - 1) -
 		(uint32_t)rte_atomic64_read(
 				&lio_dev->instr_queue[q_no]->instr_pending));
 }
@@ -1562,7 +1541,7 @@ struct lio_soft_command *
 {
 	return ((uint32_t)rte_atomic64_read(
 				&lio_dev->instr_queue[q_no]->instr_pending) >=
-				(lio_dev->instr_queue[q_no]->max_count - 2));
+				(lio_dev->instr_queue[q_no]->nb_desc - 2));
 }
 
 static int
diff --git a/drivers/net/liquidio/lio_struct.h b/drivers/net/liquidio/lio_struct.h
index 10e3976..4b4670f 100644
--- a/drivers/net/liquidio/lio_struct.h
+++ b/drivers/net/liquidio/lio_struct.h
@@ -131,7 +131,7 @@ struct lio_droq {
 	rte_atomic64_t pkts_pending;
 
 	/** Number of  descriptors in this ring. */
-	uint32_t max_count;
+	uint32_t nb_desc;
 
 	/** The number of descriptors pending refill. */
 	uint32_t refill_count;
@@ -298,8 +298,8 @@ struct lio_instr_queue {
 
 	uint32_t status:8;
 
-	/** Maximum no. of instructions in this queue. */
-	uint32_t max_count;
+	/** Number of  descriptors in this ring. */
+	uint32_t nb_desc;
 
 	/** Index in input ring where the driver should write the next packet */
 	uint32_t host_write_index;
-- 
1.8.3.1



More information about the dev mailing list