Hi Thomas,

On 01/10/2021 11:11, Thomas Monjalon wrote:
01/10/2021 08:47, Andrew Rybchenko:
On 9/30/21 10:30 PM, Ivan Malov wrote:
Hi Thomas,

On 30/09/2021 19:18, Thomas Monjalon wrote:
23/09/2021 13:20, Ivan Malov:
In 2019, commit [1] announced changes in DEV_RX_OFFLOAD namespace
intending to add new flags, RSS_HASH and FLOW_MARK. Since then,
only the former has been added. The problem hasn't been solved.
Applications still assume that no efforts are needed to enable
flow mark and similar meta data delivery.

The team behind net/sfc driver has to take over the efforts since
the problem has started impacting us. Riverhead, a cutting edge
Xilinx smart NIC family, has two Rx prefix types. Rx meta data
is available only from long Rx prefix. Switching between the
prefix formats can't happen in started state. Hence, we run
into the same problem which [1] was aiming to solve.

Sorry I don't understand what is Rx prefix?

A small chunk of per-packet metadata in Rx packet buffer preceding the
actual packet data. In terms of mbuf, this could be something lying
before m->data_off.

I've never seen the word "Rx prefix".
In general we talk about mbuf headroom and mbuf metadata,
the rest being the mbuf payload and mbuf tailroom.
I guess you mean mbuf metadata in the space of the struct rte_mbuf?

In this paragraph I describe the two ways how the NIC itself can provide metadata buffers of different sizes. Hence the term "Rx prefix". As you understand, the NIC HW is unaware of DPDK, mbufs and whatever else SW concepts. To NIC, this is "Rx prefix", that is, a chunk of per-packet metadata *preceding* the actual packet data. It's responsibility of the PMD to treat this the right way, care about headroom, payload and tailroom. I describe the two Rx prefix formats in NIC terminology just to provide the gist of the problem.


Rx meta data (mark, flag, tunnel ID) delivery is not an offload
on its own since the corresponding flows must be active to set
the data in the first place. Hence, adding offload flags
similar to RSS_HASH is not a good idea.

What means "active" here?

Active = inserted and functional. What this paragraph is trying to say
is that when you enable, say, RSS_HASH, that implies both computation of
the hash and the driver's ability to extract in from packets
("delivery"). But when it comes to MARK, it's just "delivery". No
"offload" here: the NIC won't set any mark in packets unless you create
a flow rule to make it do so. That's the gist of it.

OK
Yes I agree RTE_FLOW_ACTION_TYPE_MARK doesn't need any offload flag.
Same for RTE_FLOW_ACTION_TYPE_SET_META.

Patch [1/5] of this series adds a generic API to let applications
negotiate delivery of Rx meta data during initialisation period.

What is a metadata?
Do you mean RTE_FLOW_ITEM_TYPE_META and RTE_FLOW_ITEM_TYPE_MARK?
Metadata word could cover any field in the mbuf struct so it is vague.

Metadata here is *any* additional information provided by the NIC for each received packet. For example, Rx flag, Rx mark, RSS hash, packet classification info, you name it. I'd like to stress out that the suggested API comes with flags each of which is crystal clear on what concrete kind of metadata it covers, eg. Rx mark.


This way, an application knows right from the start which parts
of Rx meta data won't be delivered. Hence, no necessity to try
inserting flows requesting such data and handle the failures.

Sorry I don't understand the problem you want to solve.
And sorry for not noticing earlier.

No worries. *Some* PMDs do not enable delivery of, say, Rx mark with the
packets by default (for performance reasons). If the application tries
to insert a flow with action MARK, the PMD may not be able to enable
delivery of Rx mark without the need to re-start Rx sub-system. And
that's fraught with traffic disruption and similar bad consequences. In
order to address it, we need to let the application express its interest
in receiving mark with packets as early as possible. This way, the PMD
can enable Rx mark delivery in advance. And, as an additional benefit,
the application can learn *from the very beginning* whether it will be
possible to use the feature or not. If this API tells the application
that no mark delivery will be enabled, then the application can just
skip many unnecessary attempts to insert wittingly unsupported flows
during runtime.

I'm puzzled, because we could have the same reasoning for any offload.

We're not discussing *offloads*. An offload is when NIC *computes something* and *delivers* it. We are discussing precisely *delivery*.

I don't understand why we are focusing on mark only
We are not focusing on mark on purpose. It's just how our discussion goes. I chose mark (could've chosen flag or anything else) just to show you an example.

I would prefer we find a generic solution using the rte_flow API. > Can we make 
rte_flow_validate() working before port start?
If validating a fake rule doesn't make sense,
why not having a new function accepting a single action as parameter?

A noble idea, but if we feed the entire flow rule to the driver for validation, then the driver must not look specifically for actions FLAG or MARK in it (to enable or disable metadata delivery). This way, the driver is obliged to also validate match criteria, attributes, etc. And, if something is unsupported (say, some specific item), the driver will have to reject the rule as a whole thus leaving the application to join the dots itself.

Say, you ask the driver to validate the following rule:
pattern blah-blah-1 / blah-blah-2 / end action flag / end
intending to check support for FLAG delivery. Suppose, the driver doesn't support pattern item "blah-blah-1". It will throw an error right after seeing this unsupported item and won't even go further to see the action FLAG. How can application know whether its request for FLAG was heard or not?

And I'd not bind delivery of metadata to flow API. Consider the following example. We have a DPDK application sitting at the *host* and we have a *guest* with its *own* DPDK instance. The guest DPDK has asked the NIC (by virtue of flow API) to mark all outgoing packets. This packets reach the *host* DPDK. Say, the host application just wants to see the marked packets from the guest. Its own, (the host's) use of flow API is a don't care here. The host doesn't want to mark packets itself, it wants to see packets marked by the guest.


Thomas, if I'm not mistaken, net/mlx5 dv_xmeta_en driver option
is vendor-specific way to address the same problem.

Not exactly, it is configuring the capabilities:
   +------+-----------+-----------+-------------+-------------+
   | Mode | ``MARK``  | ``META``  | ``META`` Tx | FDB/Through |
   +======+===========+===========+=============+=============+
   | 0    | 24 bits   | 32 bits   | 32 bits     | no          |
   +------+-----------+-----------+-------------+-------------+
   | 1    | 24 bits   | vary 0-32 | 32 bits     | yes         |
   +------+-----------+-----------+-------------+-------------+
   | 2    | vary 0-24 | 32 bits   | 32 bits     | yes         |
   +------+-----------+-----------+-------------+-------------+



--
Ivan M

Reply via email to