> -----Original Message-----
> From: Jerin Jacob <jer...@marvell.com>
> Sent: Thursday, January 30, 2025 10:18 PM
> To: Naga Harish K, S V <s.v.naga.haris...@intel.com>; Shijith Thotton
> <sthot...@marvell.com>; dev@dpdk.org
> Cc: Pavan Nikhilesh Bhagavatula <pbhagavat...@marvell.com>; Pathak,
> Pravin <pravin.pat...@intel.com>; Hemant Agrawal
> <hemant.agra...@nxp.com>; Sachin Saxena <sachin.sax...@nxp.com>;
> Mattias R_nnblom <mattias.ronnb...@ericsson.com>; Liang Ma
> <lian...@liangbit.com>; Mccarthy, Peter <peter.mccar...@intel.com>; Van
> Haaren, Harry <harry.van.haa...@intel.com>; Carrillo, Erik G
> <erik.g.carri...@intel.com>; Gujjar, Abhinandan S
> <abhinandan.guj...@intel.com>; Amit Prakash Shukla
> <amitpraka...@marvell.com>; Burakov, Anatoly
> <anatoly.bura...@intel.com>
> Subject: RE: [RFC PATCH] eventdev: adapter API to configure multiple Rx
> queues
> 
> 
> > -----Original Message-----
> > From: Naga Harish K, S V <s.v.naga.haris...@intel.com>
> > Sent: Thursday, January 30, 2025 9:01 PM
> > To: Jerin Jacob <jer...@marvell.com>; Shijith Thotton
> > <sthot...@marvell.com>; dev@dpdk.org
> > Cc: Pavan Nikhilesh Bhagavatula <pbhagavat...@marvell.com>; Pathak,
> > Pravin <pravin.pat...@intel.com>; Hemant Agrawal
> > <hemant.agra...@nxp.com>; Sachin Saxena <sachin.sax...@nxp.com>;
> > Mattias R_nnblom <mattias.ronnb...@ericsson.com>; Liang Ma
> > <lian...@liangbit.com>; Mccarthy, Peter <peter.mccar...@intel.com>;
> > Van Haaren, Harry <harry.van.haa...@intel.com>; Carrillo, Erik G
> > <erik.g.carri...@intel.com>; Gujjar, Abhinandan S
> > <abhinandan.guj...@intel.com>; Amit Prakash Shukla
> > <amitpraka...@marvell.com>; Burakov, Anatoly
> > <anatoly.bura...@intel.com>
> > Subject: [EXTERNAL] RE: [RFC PATCH] eventdev: adapter API to configure
> > multiple Rx queues
> >
> > > -----Original Message----- > From: Jerin Jacob <jerinj@ marvell. 
> > > com>
> > > > Sent: Wednesday, January 29, 2025 1: 13 PM > To: Naga Harish K, S
> > > > V
> > > <s. v. naga. harish. k@ intel. com>; Shijith Thotton > <sthotton@
> > > marvell. com>;
> >
> >
> >
> > > -----Original Message-----
> > > From: Jerin Jacob <jer...@marvell.com>
> > > Sent: Wednesday, January 29, 2025 1:13 PM
> > > To: Naga Harish K, S V <s.v.naga.haris...@intel.com>; Shijith
> > > Thotton <sthot...@marvell.com>; dev@dpdk.org
> > > Cc: Pavan Nikhilesh Bhagavatula <pbhagavat...@marvell.com>; Pathak,
> > > Pravin <pravin.pat...@intel.com>; Hemant Agrawal
> > > <hemant.agra...@nxp.com>; Sachin Saxena <sachin.sax...@nxp.com>;
> > > Mattias R_nnblom <mattias.ronnb...@ericsson.com>; Liang Ma
> > > <lian...@liangbit.com>; Mccarthy, Peter <peter.mccar...@intel.com>;
> > > Van Haaren, Harry <harry.van.haa...@intel.com>; Carrillo, Erik G
> > > <erik.g.carri...@intel.com>; Gujjar, Abhinandan S
> > > <abhinandan.guj...@intel.com>; Amit Prakash Shukla
> > > <amitpraka...@marvell.com>; Burakov, Anatoly
> > > <anatoly.bura...@intel.com>
> > > Subject: RE: [RFC PATCH] eventdev: adapter API to configure multiple
> > > Rx queues
> > >
> > >
> > >
> > > > -----Original Message-----
> > > > From: Naga Harish K, S V <s.v.naga.haris...@intel.com>
> > > > Sent: Wednesday, January 29, 2025 10:35 AM
> > > > To: Shijith Thotton <sthot...@marvell.com>; dev@dpdk.org
> > > > Cc: Pavan Nikhilesh Bhagavatula <pbhagavat...@marvell.com>;
> > > > Pathak, Pravin <pravin.pat...@intel.com>; Hemant Agrawal
> > > > <hemant.agra...@nxp.com>; Sachin Saxena <sachin.sax...@nxp.com>;
> > > > Mattias R_nnblom <mattias.ronnb...@ericsson.com>; Jerin Jacob
> > > > <jer...@marvell.com>; Liang Ma <lian...@liangbit.com>; Mccarthy,
> > > > Peter <peter.mccar...@intel.com>; Van Haaren, Harry
> > > > <harry.van.haa...@intel.com>; Carrillo, Erik G
> > > > <erik.g.carri...@intel.com>; Gujjar, Abhinandan S
> > > > <abhinandan.guj...@intel.com>; Amit Prakash Shukla
> > > > <amitpraka...@marvell.com>; Burakov, Anatoly
> > > > <anatoly.bura...@intel.com>
> > > > Subject: [EXTERNAL] RE: [RFC PATCH] eventdev: adapter API to
> > > > configure multiple Rx queues
> > > > > >
> > > > > >This requires a change to the
> > > > > >rte_event_eth_rx_adapter_queue_add()
> > > > > >stable API parameters.
> > > > > >This is an ABI breakage and may not be possible now.
> > > > > >It requires changes to many current applications that are using
> > > > > >the
> > > > > >rte_event_eth_rx_adapter_queue_add() stable API.
> > > > > >
> > > > >
> > > > > What I meant by mapping was to retain the stable API parameters
> > > > > as they
> > > are.
> > > > > Internally, the API can use the proposed eventdev PMD operation
> > > > > (eth_rx_adapter_queues_add) without causing an ABI break, as
> > > > > shown
> > > below.
> > > > >
> > > > > int rte_event_eth_rx_adapter_queue_add(uint8_t id, uint16_t
> eth_dev_id,
> > > > >                 int32_t rx_queue_id,
> > > > >                 const struct rte_event_eth_rx_adapter_queue_conf 
> > > > > *conf) {
> > > > >         if (rx_queue_id == -1)
> > > > >                 dev->dev_ops->eth_rx_adapter_queues_add)(
> > > > >                         dev, &rte_eth_devices[eth_dev_id], 0,
> > > > >                         conf, 0);
> > > > >         else
> > > > >                 dev->dev_ops->eth_rx_adapter_queues_add)(
> > > > >                         dev, &rte_eth_devices[eth_dev_id], 
> > > > > &rx_queue_id,
> > > > >                         conf, 1); }
> > > > >
> > > > > With above change, old op (eth_rx_adapter_queue_add) can be
> > > > > removed as both API (stable and proposed) will be using
> > > eth_rx_adapter_queues_add.
> > >
> > >
> > > Since this thread is not converging and looks like it is due to confusion.
> > > I am trying to summarize my understanding to define the next
> > > steps(like if needed, we need to reach tech board if there are no
> > > consensus)
> > >
> > >
> > > Problem statement:
> > > ==================
> > > 1) Implementation of rte_event_eth_rx_adapter_queue_add() in HW
> > > typically uses an administrative function to enable it. Typically,
> > > it translated to sending a mailbox to PF driver etc.
> > > So, this function takes "time" to complete in HW implementations.
> > > 2) For SW implementations, this won't take time as there is no other
> > > actors involved.
> > > 3) There are customer use cases, they add 300+
> > > rte_event_eth_rx_adapter_queue_add() on application bootup, that is
> > > introducing significant boot time for the application.
> > > Number of queues are function of number of ethdev ports, number  of
> > > ethdev Rx queues per port and number of event queues.
> > >
> > >
> > > Expected outcome of problem statement:
> > > ======================================
> > > 1) The cases where application knows queue mapping(typically at boot
> > > time case), application can call burst variant of
> > > rte_event_eth_rx_adapter_queue_add()
> > > function
> > > to amortize the cost. Similar scheme used DPDK in control path API
> > > where latency is critical, like rte_acl_add_rules() or rte_flow via
> > > template scheme.
> > > 2) Solution should not break ABI or any impact to SW drivers.
> > > 3) Avoid duplicating the code as much as possible
> > >
> > >
> > > Proposed solution:
> > > ==================
> > > 1) Update eventdev_eth_rx_adapter_queue_add_t() PMD (Internal ABI)
> > > API to take burst parameters
> > > 2) Add new rte_event_eth_rx_adapter_queue*s*_add() function and wire
> > > to use updated PMD API
> > > 3) Use rte_event_eth_rx_adapter_queue_add() as
> > > rte_event_eth_rx_adapter_queue*s*_add(...., 1)
> > >
> > > If so, I am not sure what is the cons of this approach, it will let
> > > to have optimized applications when
> > > a) Application knows the queue mapping at priorly (typically in boot
> > > time)
> > > b) Allow HW drivers to optimize without breaking anything for SW
> > > drivers
> > > c) Provide applications to decide burst vs non burst selection based
> > > on the needed and performance requirements
> >
> > The proposed API benefits only some hardware platforms that have
> > optimized the "queue_add" eventdev PMD implementation for burst mode.
> > It may not benefit SW drivers/other HW platforms.
> 
> The sprint is to have ONE API for all drivers(SW or HW). If one driver is not 
> able
> to leverage feature is OK as long it is NOT breaking anything. We been
> accommodating ton of capabilities(like
> RTE_EVENT_DEV_CAP_DISTRIBUTED_SCHED)
> And SW driver specific public API(like
> rte_event_eth_rx_adapter_service_id_get()) to have Common API. As long as
> it does not break each other and application has clarity on the usage (when to
> use the API) I don’t see any issue. Do you see any issue with that forward
> progress approach?
> 

This approach is fine, as long as it is not breaking the other platforms.

> 
> > There will not be much difference in calling the existing API
> > (rte_event_eth_rx_adapter_queue_add()) in a loop vs using the new API
> > for the above cases.
> 
> That is just A implementation view. Right? I have explained in the problem
> statement which is the not case for some drivers.(Even SW driver can leverage
> such burst function using SIMD etc, if one driver wants to)
> 

Not Just from the implementation point of view, but from the latency 
improvement also.
Anyway, I am fine with the new API approach.

> >
> > If the new proposed API benefits all platforms, then it is useful.
> 
> See above.
> 
> 
> > This is the point I am making from the beginning, it is not captured
> > in the summary.

Reply via email to