From: Pawel Wodkowski <pawelx.wodkow...@intel.com>

This patch add support for DCB in SRIOV mode. When no PFC
is enabled this feature might be used as multiple queues
(up to 8 or 4) for VF.

It incorporate following modifications:
 - Allow zero rx/tx queues to be passed to rte_eth_dev_configure().
   Rationale:
   in SRIOV mode PF use first free VF to RX/TX. If VF count
   is 16 or 32 all recources are assigned to VFs so PF can
   be used only for configuration.
 - split nb_q_per_pool to nb_rx_q_per_pool and nb_tx_q_per_pool
   Rationale:
   rx and tx number of queue might be different if RX and TX are
   configured in different mode. This allow to inform VF about
   proper number of queues.
 - extern mailbox API for DCB mode

Signed-off-by: Pawel Wodkowski <pawelx.wodkowski at intel.com>
---
 lib/librte_ether/rte_ethdev.c       |   84 +++++++++++++++++++++---------
 lib/librte_ether/rte_ethdev.h       |    5 +-
 lib/librte_pmd_e1000/igb_pf.c       |    3 +-
 lib/librte_pmd_ixgbe/ixgbe_ethdev.c |   10 ++--
 lib/librte_pmd_ixgbe/ixgbe_ethdev.h |    1 +
 lib/librte_pmd_ixgbe/ixgbe_pf.c     |   98 ++++++++++++++++++++++++++++++-----
 lib/librte_pmd_ixgbe/ixgbe_rxtx.c   |    7 ++-
 7 files changed, 159 insertions(+), 49 deletions(-)

diff --git a/lib/librte_ether/rte_ethdev.c b/lib/librte_ether/rte_ethdev.c
index 95f2ceb..4c1a494 100644
--- a/lib/librte_ether/rte_ethdev.c
+++ b/lib/librte_ether/rte_ethdev.c
@@ -333,7 +333,7 @@ rte_eth_dev_rx_queue_config(struct rte_eth_dev *dev, 
uint16_t nb_queues)
                dev->data->rx_queues = rte_zmalloc("ethdev->rx_queues",
                                sizeof(dev->data->rx_queues[0]) * nb_queues,
                                RTE_CACHE_LINE_SIZE);
-               if (dev->data->rx_queues == NULL) {
+               if (dev->data->rx_queues == NULL && nb_queues > 0) {
                        dev->data->nb_rx_queues = 0;
                        return -(ENOMEM);
                }
@@ -475,7 +475,7 @@ rte_eth_dev_tx_queue_config(struct rte_eth_dev *dev, 
uint16_t nb_queues)
                dev->data->tx_queues = rte_zmalloc("ethdev->tx_queues",
                                sizeof(dev->data->tx_queues[0]) * nb_queues,
                                RTE_CACHE_LINE_SIZE);
-               if (dev->data->tx_queues == NULL) {
+               if (dev->data->tx_queues == NULL && nb_queues > 0) {
                        dev->data->nb_tx_queues = 0;
                        return -(ENOMEM);
                }
@@ -507,6 +507,7 @@ rte_eth_dev_check_mq_mode(uint8_t port_id, uint16_t 
nb_rx_q, uint16_t nb_tx_q,
                      const struct rte_eth_conf *dev_conf)
 {
        struct rte_eth_dev *dev = &rte_eth_devices[port_id];
+       struct rte_eth_dev_info dev_info;

        if (RTE_ETH_DEV_SRIOV(dev).active != 0) {
                /* check multi-queue mode */
@@ -524,11 +525,33 @@ rte_eth_dev_check_mq_mode(uint8_t port_id, uint16_t 
nb_rx_q, uint16_t nb_tx_q,
                        return (-EINVAL);
                }

+               if ((dev_conf->rxmode.mq_mode == ETH_MQ_RX_VMDQ_DCB) &&
+                       (dev_conf->txmode.mq_mode == ETH_MQ_TX_VMDQ_DCB)) {
+                       enum rte_eth_nb_pools rx_pools =
+                                               
dev_conf->rx_adv_conf.vmdq_dcb_conf.nb_queue_pools;
+                       enum rte_eth_nb_pools tx_pools =
+                                               
dev_conf->tx_adv_conf.vmdq_dcb_tx_conf.nb_queue_pools;
+
+                       if (rx_pools != tx_pools) {
+                               /* Only equal number of pools is supported when
+                                * DCB+VMDq in SRIOV */
+                               PMD_DEBUG_TRACE("ethdev port_id=%" PRIu8
+                                               " SRIOV active, DCB+VMDQ mode, "
+                                               "number of rx and tx pools is 
not eqaul\n",
+                                               port_id);
+                               return (-EINVAL);
+                       }
+               }
+
+               uint16_t nb_rx_q_per_pool = 
RTE_ETH_DEV_SRIOV(dev).nb_rx_q_per_pool;
+               uint16_t nb_tx_q_per_pool = 
RTE_ETH_DEV_SRIOV(dev).nb_tx_q_per_pool;
+
                switch (dev_conf->rxmode.mq_mode) {
-               case ETH_MQ_RX_VMDQ_RSS:
                case ETH_MQ_RX_VMDQ_DCB:
+                       break;
+               case ETH_MQ_RX_VMDQ_RSS:
                case ETH_MQ_RX_VMDQ_DCB_RSS:
-                       /* DCB/RSS VMDQ in SRIOV mode, not implement yet */
+                       /* RSS, DCB+RSS VMDQ in SRIOV mode, not implement yet */
                        PMD_DEBUG_TRACE("ethdev port_id=%" PRIu8
                                        " SRIOV active, "
                                        "unsupported VMDQ mq_mode rx %u\n",
@@ -537,37 +560,32 @@ rte_eth_dev_check_mq_mode(uint8_t port_id, uint16_t 
nb_rx_q, uint16_t nb_tx_q,
                default: /* ETH_MQ_RX_VMDQ_ONLY or ETH_MQ_RX_NONE */
                        /* if nothing mq mode configure, use default scheme */
                        dev->data->dev_conf.rxmode.mq_mode = 
ETH_MQ_RX_VMDQ_ONLY;
-                       if (RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool > 1)
-                               RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool = 1;
+                       if (nb_rx_q_per_pool > 1)
+                               nb_rx_q_per_pool = 1;
                        break;
                }

                switch (dev_conf->txmode.mq_mode) {
-               case ETH_MQ_TX_VMDQ_DCB:
-                       /* DCB VMDQ in SRIOV mode, not implement yet */
-                       PMD_DEBUG_TRACE("ethdev port_id=%" PRIu8
-                                       " SRIOV active, "
-                                       "unsupported VMDQ mq_mode tx %u\n",
-                                       port_id, dev_conf->txmode.mq_mode);
-                       return (-EINVAL);
+               case ETH_MQ_TX_VMDQ_DCB: /* DCB VMDQ in SRIOV mode*/
+                       break;
                default: /* ETH_MQ_TX_VMDQ_ONLY or ETH_MQ_TX_NONE */
                        /* if nothing mq mode configure, use default scheme */
                        dev->data->dev_conf.txmode.mq_mode = 
ETH_MQ_TX_VMDQ_ONLY;
-                       if (RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool > 1)
-                               RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool = 1;
+                       if (nb_tx_q_per_pool > 1)
+                               nb_tx_q_per_pool = 1;
                        break;
                }

                /* check valid queue number */
-               if ((nb_rx_q > RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool) ||
-                   (nb_tx_q > RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool)) {
+               if (nb_rx_q > nb_rx_q_per_pool || nb_tx_q > nb_tx_q_per_pool) {
                        PMD_DEBUG_TRACE("ethdev port_id=%d SRIOV active, "
-                                   "queue number must less equal to %d\n",
-                                       port_id, 
RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool);
+                                   "rx/tx queue number must less equal to 
%d/%d\n",
+                                       port_id, 
RTE_ETH_DEV_SRIOV(dev).nb_rx_q_per_pool,
+                                       
RTE_ETH_DEV_SRIOV(dev).nb_tx_q_per_pool);
                        return (-EINVAL);
                }
        } else {
-               /* For vmdb+dcb mode check our configuration before we go 
further */
+               /* For vmdq+dcb mode check our configuration before we go 
further */
                if (dev_conf->rxmode.mq_mode == ETH_MQ_RX_VMDQ_DCB) {
                        const struct rte_eth_vmdq_dcb_conf *conf;

@@ -606,11 +624,20 @@ rte_eth_dev_check_mq_mode(uint8_t port_id, uint16_t 
nb_rx_q, uint16_t nb_tx_q,
                        }
                }

+               /* For DCB we need to obtain maximum number of queues 
dinamically,
+                * as this depends on max VF exported in PF */
+               if ((dev_conf->rxmode.mq_mode == ETH_MQ_RX_DCB) ||
+                       (dev_conf->txmode.mq_mode == ETH_MQ_TX_DCB)) {
+
+                               
FUNC_PTR_OR_ERR_RET(*dev->dev_ops->dev_infos_get, -ENOTSUP);
+                               (*dev->dev_ops->dev_infos_get)(dev, &dev_info);
+               }
+
                /* For DCB mode check our configuration before we go further */
                if (dev_conf->rxmode.mq_mode == ETH_MQ_RX_DCB) {
                        const struct rte_eth_dcb_rx_conf *conf;

-                       if (nb_rx_q != ETH_DCB_NUM_QUEUES) {
+                       if (nb_rx_q != dev_info.max_rx_queues) {
                                PMD_DEBUG_TRACE("ethdev port_id=%d DCB, nb_rx_q 
"
                                                "!= %d\n",
                                                port_id, ETH_DCB_NUM_QUEUES);
@@ -630,7 +657,7 @@ rte_eth_dev_check_mq_mode(uint8_t port_id, uint16_t 
nb_rx_q, uint16_t nb_tx_q,
                if (dev_conf->txmode.mq_mode == ETH_MQ_TX_DCB) {
                        const struct rte_eth_dcb_tx_conf *conf;

-                       if (nb_tx_q != ETH_DCB_NUM_QUEUES) {
+                       if (nb_tx_q != dev_info.max_tx_queues) {
                                PMD_DEBUG_TRACE("ethdev port_id=%d DCB, nb_tx_q 
"
                                                "!= %d\n",
                                                port_id, ETH_DCB_NUM_QUEUES);
@@ -690,7 +717,10 @@ rte_eth_dev_configure(uint8_t port_id, uint16_t nb_rx_q, 
uint16_t nb_tx_q,
        }
        if (nb_rx_q == 0) {
                PMD_DEBUG_TRACE("ethdev port_id=%d nb_rx_q == 0\n", port_id);
-               return (-EINVAL);
+               /* In SRIOV there can be no free resource for PF. So permit use 
only
+                * for configuration. */
+               if (RTE_ETH_DEV_SRIOV(dev).active == 0)
+                       return (-EINVAL);
        }

        if (nb_tx_q > dev_info.max_tx_queues) {
@@ -698,9 +728,13 @@ rte_eth_dev_configure(uint8_t port_id, uint16_t nb_rx_q, 
uint16_t nb_tx_q,
                                port_id, nb_tx_q, dev_info.max_tx_queues);
                return (-EINVAL);
        }
+
        if (nb_tx_q == 0) {
                PMD_DEBUG_TRACE("ethdev port_id=%d nb_tx_q == 0\n", port_id);
-               return (-EINVAL);
+               /* In SRIOV there can be no free resource for PF. So permit use 
only
+                * for configuration. */
+               if (RTE_ETH_DEV_SRIOV(dev).active == 0)
+                       return (-EINVAL);
        }

        /* Copy the dev_conf parameter into the dev structure */
@@ -750,7 +784,7 @@ rte_eth_dev_configure(uint8_t port_id, uint16_t nb_rx_q, 
uint16_t nb_tx_q,
                                                        ETHER_MAX_LEN;
        }

-       /* multipe queue mode checking */
+       /* multiple queue mode checking */
        diag = rte_eth_dev_check_mq_mode(port_id, nb_rx_q, nb_tx_q, dev_conf);
        if (diag != 0) {
                PMD_DEBUG_TRACE("port%d rte_eth_dev_check_mq_mode = %d\n",
diff --git a/lib/librte_ether/rte_ethdev.h b/lib/librte_ether/rte_ethdev.h
index ce0528f..04fda83 100644
--- a/lib/librte_ether/rte_ethdev.h
+++ b/lib/librte_ether/rte_ethdev.h
@@ -299,7 +299,7 @@ enum rte_eth_rx_mq_mode {
 enum rte_eth_tx_mq_mode {
        ETH_MQ_TX_NONE    = 0,  /**< It is in neither DCB nor VT mode. */
        ETH_MQ_TX_DCB,          /**< For TX side,only DCB is on. */
-       ETH_MQ_TX_VMDQ_DCB,     /**< For TX side,both DCB and VT is on. */
+       ETH_MQ_TX_VMDQ_DCB,     /**< For TX side,both DCB and VT is on. */
        ETH_MQ_TX_VMDQ_ONLY,    /**< Only VT on, no DCB */
 };

@@ -1569,7 +1569,8 @@ struct rte_eth_dev {

 struct rte_eth_dev_sriov {
        uint8_t active;               /**< SRIOV is active with 16, 32 or 64 
pools */
-       uint8_t nb_q_per_pool;        /**< rx queue number per pool */
+       uint8_t nb_rx_q_per_pool;        /**< rx queue number per pool */
+       uint8_t nb_tx_q_per_pool;        /**< tx queue number per pool */
        uint16_t def_vmdq_idx;        /**< Default pool num used for PF */
        uint16_t def_pool_q_idx;      /**< Default pool queue start reg index */
 };
diff --git a/lib/librte_pmd_e1000/igb_pf.c b/lib/librte_pmd_e1000/igb_pf.c
index bc3816a..9d2f858 100644
--- a/lib/librte_pmd_e1000/igb_pf.c
+++ b/lib/librte_pmd_e1000/igb_pf.c
@@ -115,7 +115,8 @@ void igb_pf_host_init(struct rte_eth_dev *eth_dev)
                rte_panic("Cannot allocate memory for private VF data\n");

        RTE_ETH_DEV_SRIOV(eth_dev).active = ETH_8_POOLS;
-       RTE_ETH_DEV_SRIOV(eth_dev).nb_q_per_pool = nb_queue;
+       RTE_ETH_DEV_SRIOV(eth_dev).nb_rx_q_per_pool = nb_queue;
+       RTE_ETH_DEV_SRIOV(eth_dev).nb_tx_q_per_pool = nb_queue;
        RTE_ETH_DEV_SRIOV(eth_dev).def_vmdq_idx = vf_num;
        RTE_ETH_DEV_SRIOV(eth_dev).def_pool_q_idx = (uint16_t)(vf_num * 
nb_queue);

diff --git a/lib/librte_pmd_ixgbe/ixgbe_ethdev.c 
b/lib/librte_pmd_ixgbe/ixgbe_ethdev.c
index 3fc3738..347f03c 100644
--- a/lib/librte_pmd_ixgbe/ixgbe_ethdev.c
+++ b/lib/librte_pmd_ixgbe/ixgbe_ethdev.c
@@ -3555,14 +3555,14 @@ static int ixgbe_set_vf_rate_limit(struct rte_eth_dev 
*dev, uint16_t vf,
        struct ixgbe_hw *hw = IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);
        struct ixgbe_vf_info *vfinfo =
                *(IXGBE_DEV_PRIVATE_TO_P_VFDATA(dev->data->dev_private));
-       uint8_t  nb_q_per_pool = RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool;
+       uint8_t  nb_tx_q_per_pool = RTE_ETH_DEV_SRIOV(dev).nb_tx_q_per_pool;
        uint32_t queue_stride =
                IXGBE_MAX_RX_QUEUE_NUM / RTE_ETH_DEV_SRIOV(dev).active;
        uint32_t queue_idx = vf * queue_stride, idx = 0, vf_idx;
-       uint32_t queue_end = queue_idx + nb_q_per_pool - 1;
+       uint32_t tx_queue_end = queue_idx + nb_tx_q_per_pool - 1;
        uint16_t total_rate = 0;

-       if (queue_end >= hw->mac.max_tx_queues)
+       if (tx_queue_end >= hw->mac.max_tx_queues)
                return -EINVAL;

        if (vfinfo != NULL) {
@@ -3577,7 +3577,7 @@ static int ixgbe_set_vf_rate_limit(struct rte_eth_dev 
*dev, uint16_t vf,
                return -EINVAL;

        /* Store tx_rate for this vf. */
-       for (idx = 0; idx < nb_q_per_pool; idx++) {
+       for (idx = 0; idx < nb_tx_q_per_pool; idx++) {
                if (((uint64_t)0x1 << idx) & q_msk) {
                        if (vfinfo[vf].tx_rate[idx] != tx_rate)
                                vfinfo[vf].tx_rate[idx] = tx_rate;
@@ -3595,7 +3595,7 @@ static int ixgbe_set_vf_rate_limit(struct rte_eth_dev 
*dev, uint16_t vf,
        }

        /* Set RTTBCNRC of each queue/pool for vf X  */
-       for (; queue_idx <= queue_end; queue_idx++) {
+       for (; queue_idx <= tx_queue_end; queue_idx++) {
                if (0x1 & q_msk)
                        ixgbe_set_queue_rate_limit(dev, queue_idx, tx_rate);
                q_msk = q_msk >> 1;
diff --git a/lib/librte_pmd_ixgbe/ixgbe_ethdev.h 
b/lib/librte_pmd_ixgbe/ixgbe_ethdev.h
index ca99170..ebf16e9 100644
--- a/lib/librte_pmd_ixgbe/ixgbe_ethdev.h
+++ b/lib/librte_pmd_ixgbe/ixgbe_ethdev.h
@@ -159,6 +159,7 @@ struct ixgbe_vf_info {
        uint16_t tx_rate[IXGBE_MAX_QUEUE_NUM_PER_VF];
        uint16_t vlan_count;
        uint8_t spoofchk_enabled;
+       unsigned int vf_api;
 };

 /*
diff --git a/lib/librte_pmd_ixgbe/ixgbe_pf.c b/lib/librte_pmd_ixgbe/ixgbe_pf.c
index 51da1fd..4d30bcf 100644
--- a/lib/librte_pmd_ixgbe/ixgbe_pf.c
+++ b/lib/librte_pmd_ixgbe/ixgbe_pf.c
@@ -127,7 +127,8 @@ void ixgbe_pf_host_init(struct rte_eth_dev *eth_dev)
                RTE_ETH_DEV_SRIOV(eth_dev).active = ETH_16_POOLS;
        }

-       RTE_ETH_DEV_SRIOV(eth_dev).nb_q_per_pool = nb_queue;
+       RTE_ETH_DEV_SRIOV(eth_dev).nb_rx_q_per_pool = nb_queue;
+       RTE_ETH_DEV_SRIOV(eth_dev).nb_tx_q_per_pool = nb_queue;
        RTE_ETH_DEV_SRIOV(eth_dev).def_vmdq_idx = vf_num;
        RTE_ETH_DEV_SRIOV(eth_dev).def_pool_q_idx = (uint16_t)(vf_num * 
nb_queue);

@@ -189,7 +190,7 @@ int ixgbe_pf_host_configure(struct rte_eth_dev *eth_dev)
        hw->mac.ops.set_vmdq(hw, 0, RTE_ETH_DEV_SRIOV(eth_dev).def_vmdq_idx);

        /*
-        * SW msut set GCR_EXT.VT_Mode the same as GPIE.VT_Mode
+        * SW must set GCR_EXT.VT_Mode the same as GPIE.VT_Mode
         */
        gcr_ext = IXGBE_READ_REG(hw, IXGBE_GCR_EXT);
        gcr_ext &= ~IXGBE_GCR_EXT_VT_MODE_MASK;
@@ -214,19 +215,19 @@ int ixgbe_pf_host_configure(struct rte_eth_dev *eth_dev)
        }

        IXGBE_WRITE_REG(hw, IXGBE_GCR_EXT, gcr_ext);
-        IXGBE_WRITE_REG(hw, IXGBE_GPIE, gpie);
+       IXGBE_WRITE_REG(hw, IXGBE_GPIE, gpie);

-        /*
+       /*
         * enable vlan filtering and allow all vlan tags through
         */
-        vlanctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL);
-        vlanctrl |= IXGBE_VLNCTRL_VFE ; /* enable vlan filters */
-        IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlanctrl);
+       vlanctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL);
+       vlanctrl |= IXGBE_VLNCTRL_VFE ; /* enable vlan filters */
+       IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlanctrl);

-        /* VFTA - enable all vlan filters */
-        for (i = 0; i < IXGBE_MAX_VFTA; i++) {
-                IXGBE_WRITE_REG(hw, IXGBE_VFTA(i), 0xFFFFFFFF);
-        }
+       /* VFTA - enable all vlan filters */
+       for (i = 0; i < IXGBE_MAX_VFTA; i++) {
+               IXGBE_WRITE_REG(hw, IXGBE_VFTA(i), 0xFFFFFFFF);
+       }

        /* Enable MAC Anti-Spoofing */
        hw->mac.ops.set_mac_anti_spoofing(hw, FALSE, vf_num);
@@ -369,6 +370,73 @@ ixgbe_vf_reset(struct rte_eth_dev *dev, uint16_t vf, 
uint32_t *msgbuf)
 }

 static int
+ixgbe_negotiate_vf_api(struct rte_eth_dev *dev, uint32_t vf, uint32_t *msgbuf)
+{
+       struct ixgbe_vf_info *vfinfo =
+               *(IXGBE_DEV_PRIVATE_TO_P_VFDATA(dev->data->dev_private));
+       int api = msgbuf[1];
+
+       switch (api) {
+       case ixgbe_mbox_api_10:
+       case ixgbe_mbox_api_11:
+               vfinfo[vf].vf_api = api;
+               return 0;
+       default:
+               break;
+       }
+
+       RTE_LOG(DEBUG, PMD, "VF %d requested invalid api version %u\n", vf, 
api);
+       return -1;
+}
+
+static int
+ixgbe_get_vf_queues(struct rte_eth_dev *dev, uint32_t vf, uint32_t *msgbuf)
+{
+       struct ixgbe_vf_info *vfinfo =
+               *(IXGBE_DEV_PRIVATE_TO_P_VFDATA(dev->data->dev_private));
+       struct ixgbe_dcb_config *dcb_cfg =
+                       IXGBE_DEV_PRIVATE_TO_DCB_CFG(dev->data->dev_private);
+
+       uint8_t num_tcs = dcb_cfg->num_tcs.pg_tcs;
+
+       /* verify the PF is supporting the correct APIs */
+       switch (vfinfo[vf].vf_api) {
+       case ixgbe_mbox_api_10:
+       case ixgbe_mbox_api_11:
+               break;
+       default:
+               return -1;
+       }
+
+       if (RTE_ETH_DEV_SRIOV(dev).active) {
+               if (dev->data->dev_conf.rxmode.mq_mode == ETH_MQ_RX_VMDQ_DCB)
+                       msgbuf[IXGBE_VF_TX_QUEUES] = num_tcs;
+               else
+                       msgbuf[IXGBE_VF_TX_QUEUES] = 1;
+
+               if (dev->data->dev_conf.txmode.mq_mode == ETH_MQ_TX_VMDQ_DCB)
+                       msgbuf[IXGBE_VF_RX_QUEUES] = num_tcs;
+               else
+                       msgbuf[IXGBE_VF_RX_QUEUES] = 1;
+       }       else {
+               /* only allow 1 Tx queue for bandwidth limiting */
+               msgbuf[IXGBE_VF_TX_QUEUES] = 1;
+               msgbuf[IXGBE_VF_RX_QUEUES] = 1;
+       }
+
+       /* notify VF of need for VLAN tag stripping, and correct queue */
+       if (num_tcs)
+               msgbuf[IXGBE_VF_TRANS_VLAN] = num_tcs;
+       else
+               msgbuf[IXGBE_VF_TRANS_VLAN] = 0;
+
+       /* notify VF of default queue */
+       msgbuf[IXGBE_VF_DEF_QUEUE] = 0;
+
+       return 0;
+}
+
+static int
 ixgbe_vf_set_mac_addr(struct rte_eth_dev *dev, uint32_t vf, uint32_t *msgbuf)
 {
        struct ixgbe_hw *hw = IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);
@@ -512,6 +580,12 @@ ixgbe_rcv_msg_from_vf(struct rte_eth_dev *dev, uint16_t vf)
        case IXGBE_VF_SET_VLAN:
                retval = ixgbe_vf_set_vlan(dev, vf, msgbuf);
                break;
+       case IXGBE_VF_API_NEGOTIATE:
+               retval = ixgbe_negotiate_vf_api(dev, vf, msgbuf);
+               break;
+       case IXGBE_VF_GET_QUEUES:
+               retval = ixgbe_get_vf_queues(dev, vf, msgbuf);
+               break;
        default:
                PMD_DRV_LOG(DEBUG, "Unhandled Msg %8.8x", (unsigned)msgbuf[0]);
                retval = IXGBE_ERR_MBX;
@@ -526,7 +600,7 @@ ixgbe_rcv_msg_from_vf(struct rte_eth_dev *dev, uint16_t vf)

        msgbuf[0] |= IXGBE_VT_MSGTYPE_CTS;

-       ixgbe_write_mbx(hw, msgbuf, 1, vf);
+       ixgbe_write_mbx(hw, msgbuf, mbx_size, vf);

        return retval;
 }
diff --git a/lib/librte_pmd_ixgbe/ixgbe_rxtx.c 
b/lib/librte_pmd_ixgbe/ixgbe_rxtx.c
index e10d6a2..49b44fe 100644
--- a/lib/librte_pmd_ixgbe/ixgbe_rxtx.c
+++ b/lib/librte_pmd_ixgbe/ixgbe_rxtx.c
@@ -3166,10 +3166,9 @@ void ixgbe_configure_dcb(struct rte_eth_dev *dev)

        /* check support mq_mode for DCB */
        if ((dev_conf->rxmode.mq_mode != ETH_MQ_RX_VMDQ_DCB) &&
-           (dev_conf->rxmode.mq_mode != ETH_MQ_RX_DCB))
-               return;
-
-       if (dev->data->nb_rx_queues != ETH_DCB_NUM_QUEUES)
+           (dev_conf->rxmode.mq_mode != ETH_MQ_RX_DCB) &&
+           (dev_conf->txmode.mq_mode != ETH_MQ_TX_VMDQ_DCB) &&
+           (dev_conf->txmode.mq_mode != ETH_MQ_TX_DCB))
                return;

        /** Configure DCB hardware **/
-- 
1.7.9.5

Reply via email to