> On 9/8/2022 9:44 AM, Chaoyong He wrote:
> > Adds the vNIC initialization logic for the flower PF vNIC. The flower
> > firmware exposes this vNIC for the purposes of fallback traffic in the
> > switchdev use-case.
> >
> > Adds minimal dev_ops for this PF device. Because the device is being
> > exposed externally to DPDK it should also be configured using DPDK
> > helpers like rte_eth_configure(). For these helpers to work the flower
> > logic needs to implements a minimal set of dev_ops.
> >
> > Signed-off-by: Chaoyong He <chaoyong...@corigine.com>
> > Reviewed-by: Niklas Söderlund <niklas.soderl...@corigine.com>
> > ---
> >   drivers/net/nfp/flower/nfp_flower.c            | 398
> ++++++++++++++++++++++++-
> >   drivers/net/nfp/flower/nfp_flower.h            |   6 +
> >   drivers/net/nfp/flower/nfp_flower_ovs_compat.h |  37 +++
> 
> Can you please detail why OVS specific header is required? Having application
> specific code in PMD can be sign of some design issue, that is why can you
> please explain more what it does?
> 

Basically, there exist two layers polling mode to move a pkt from firmware to 
OVS.

When our card using flower application firmware receive pkt and find the pkt 
can't be offloaded, 
it will record the input port in a place of the pkt, we call it metadata.

There exist a rte_ring for each representor port.

We use the pf device as a multiplexer, which keeps polling pkts from the 
firmware. 
Based on the metadata, it will enqueue the pkt into the rte_ring of the 
corresponding representor port.

On the OVS side, it will keeps try to dequeue the pkt from the rte_ring of the 
representor port.
Once it gets the pkt, the OVS will go its logic and treat the pkt as `struct 
dp_packet`.

So we copy the definition of `struct dp_packet` from OVS to prevent the 
coredump caused by memory read/write out of range.

Another possible way is defining a big enough mbuf_priv_len using macro to 
prevent this structure definition from OVS.
Is this the right way? 

> <...>
> 
> > +static int
> > +nfp_flower_init_pf_vnic(struct nfp_net_hw *hw) {
> > +   int ret;
> > +   uint16_t i;
> > +   uint16_t n_txq;
> > +   uint16_t n_rxq;
> > +   uint16_t port_id;
> > +   unsigned int numa_node;
> > +   struct rte_mempool *mp;
> > +   struct nfp_pf_dev *pf_dev;
> > +   struct rte_eth_dev *eth_dev;
> > +   struct nfp_app_fw_flower *app_fw_flower;
> > +
> > +   static const struct rte_eth_conf port_conf = {
> > +           .rxmode = {
> > +                   .mq_mode  = RTE_ETH_MQ_RX_RSS,
> > +                   .offloads = RTE_ETH_RX_OFFLOAD_CHECKSUM,
> > +           },
> > +           .txmode = {
> > +                   .mq_mode = RTE_ETH_MQ_TX_NONE,
> > +           },
> > +   };
> > +
> > +   /* Set up some pointers here for ease of use */
> > +   pf_dev = hw->pf_dev;
> > +   app_fw_flower = NFP_PRIV_TO_APP_FW_FLOWER(pf_dev-
> >app_fw_priv);
> > +
> > +   /*
> > +    * Perform the "common" part of setting up a flower vNIC.
> > +    * Mostly reading configuration from hardware.
> > +    */
> > +   ret = nfp_flower_init_vnic_common(hw, "pf_vnic");
> > +   if (ret != 0)
> > +           goto done;
> > +
> > +   hw->eth_dev = rte_eth_dev_allocate("nfp_pf_vnic");
> > +   if (hw->eth_dev == NULL) {
> > +           ret = -ENOMEM;
> > +           goto done;
> > +   }
> > +
> > +   /* Grab the pointer to the newly created rte_eth_dev here */
> > +   eth_dev = hw->eth_dev;
> > +
> > +   numa_node = rte_socket_id();
> > +
> > +   /* Fill in some of the eth_dev fields */
> > +   eth_dev->device = &pf_dev->pci_dev->device;
> > +   eth_dev->data->dev_private = hw;
> > +
> > +   /* Create a mbuf pool for the PF */
> > +   app_fw_flower->pf_pktmbuf_pool = nfp_flower_pf_mp_create();
> > +   if (app_fw_flower->pf_pktmbuf_pool == NULL) {
> > +           ret = -ENOMEM;
> > +           goto port_release;
> > +   }
> > +
> > +   mp = app_fw_flower->pf_pktmbuf_pool;
> > +
> > +   /* Add Rx/Tx functions */
> > +   eth_dev->dev_ops = &nfp_flower_pf_vnic_ops;
> > +
> > +   /* PF vNIC gets a random MAC */
> > +   eth_dev->data->mac_addrs = rte_zmalloc("mac_addr",
> RTE_ETHER_ADDR_LEN, 0);
> > +   if (eth_dev->data->mac_addrs == NULL) {
> > +           ret = -ENOMEM;
> > +           goto mempool_cleanup;
> > +   }
> > +
> > +   rte_eth_random_addr(eth_dev->data->mac_addrs->addr_bytes);
> > +   rte_eth_dev_probing_finish(eth_dev);
> > +
> > +   /* Configure the PF device now */
> > +   n_rxq = hw->max_rx_queues;
> > +   n_txq = hw->max_tx_queues;
> > +   port_id = hw->eth_dev->data->port_id;
> > +
> > +   ret = rte_eth_dev_configure(port_id, n_rxq, n_txq, &port_conf);
> 
> Still not sure about PMD calling 'rte_eth_dev_configure()', can you please
> give more details on what specific configuration is expected with that call?

The main configuration we need is the number of rx/tx queue.
So we should use the internal api `eth_dev_rx/tx_queue_config` to instead?

Reply via email to