> Currently, we expect that only one callback can be active at any given
> moment, for a particular queue configuration, which is relatively easy
> to implement in a thread-safe way. However, we're about to add support
> for multiple queues per lcore, which will greatly increase the
> possibility of various race conditions.
> 
> We could have used something like an RCU for this use case, but absent
> of a pressing need for thread safety we'll go the easy way and just
> mandate that the API's are to be called when all affected ports are
> stopped, and document this limitation. This greatly simplifies the
> `rte_power_monitor`-related code.

I think you need to update RN too with that.
Another thing - do you really need the whole port stopped?
>From what I understand - you work on queues, so it is enough for you
that related RX queue is stopped.
So, to make things a bit more robust, in pmgmt_queue_enable/disable 
you can call rte_eth_rx_queue_info_get() and check queue state.
 
> Signed-off-by: Anatoly Burakov <anatoly.bura...@intel.com>
> ---
>  lib/power/meson.build          |   3 +
>  lib/power/rte_power_pmd_mgmt.c | 106 ++++++++-------------------------
>  lib/power/rte_power_pmd_mgmt.h |   6 ++
>  3 files changed, 35 insertions(+), 80 deletions(-)
> 
> diff --git a/lib/power/meson.build b/lib/power/meson.build
> index c1097d32f1..4f6a242364 100644
> --- a/lib/power/meson.build
> +++ b/lib/power/meson.build
> @@ -21,4 +21,7 @@ headers = files(
>          'rte_power_pmd_mgmt.h',
>          'rte_power_guest_channel.h',
>  )
> +if cc.has_argument('-Wno-cast-qual')
> +    cflags += '-Wno-cast-qual'
> +endif
>  deps += ['timer', 'ethdev']
> diff --git a/lib/power/rte_power_pmd_mgmt.c b/lib/power/rte_power_pmd_mgmt.c
> index db03cbf420..0707c60a4f 100644
> --- a/lib/power/rte_power_pmd_mgmt.c
> +++ b/lib/power/rte_power_pmd_mgmt.c
> @@ -40,8 +40,6 @@ struct pmd_queue_cfg {
>       /**< Callback mode for this queue */
>       const struct rte_eth_rxtx_callback *cur_cb;
>       /**< Callback instance */
> -     volatile bool umwait_in_progress;
> -     /**< are we currently sleeping? */
>       uint64_t empty_poll_stats;
>       /**< Number of empty polls */
>  } __rte_cache_aligned;
> @@ -92,30 +90,11 @@ clb_umwait(uint16_t port_id, uint16_t qidx, struct 
> rte_mbuf **pkts __rte_unused,
>                       struct rte_power_monitor_cond pmc;
>                       uint16_t ret;
> 
> -                     /*
> -                      * we might get a cancellation request while being
> -                      * inside the callback, in which case the wakeup
> -                      * wouldn't work because it would've arrived too early.
> -                      *
> -                      * to get around this, we notify the other thread that
> -                      * we're sleeping, so that it can spin until we're done.
> -                      * unsolicited wakeups are perfectly safe.
> -                      */
> -                     q_conf->umwait_in_progress = true;
> -
> -                     rte_atomic_thread_fence(__ATOMIC_SEQ_CST);
> -
> -                     /* check if we need to cancel sleep */
> -                     if (q_conf->pwr_mgmt_state == PMD_MGMT_ENABLED) {
> -                             /* use monitoring condition to sleep */
> -                             ret = rte_eth_get_monitor_addr(port_id, qidx,
> -                                             &pmc);
> -                             if (ret == 0)
> -                                     rte_power_monitor(&pmc, UINT64_MAX);
> -                     }
> -                     q_conf->umwait_in_progress = false;
> -
> -                     rte_atomic_thread_fence(__ATOMIC_SEQ_CST);
> +                     /* use monitoring condition to sleep */
> +                     ret = rte_eth_get_monitor_addr(port_id, qidx,
> +                                     &pmc);
> +                     if (ret == 0)
> +                             rte_power_monitor(&pmc, UINT64_MAX);
>               }
>       } else
>               q_conf->empty_poll_stats = 0;
> @@ -183,6 +162,7 @@ rte_power_ethdev_pmgmt_queue_enable(unsigned int 
> lcore_id, uint16_t port_id,
>  {
>       struct pmd_queue_cfg *queue_cfg;
>       struct rte_eth_dev_info info;
> +     rte_rx_callback_fn clb;
>       int ret;
> 
>       RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -EINVAL);
> @@ -232,17 +212,7 @@ rte_power_ethdev_pmgmt_queue_enable(unsigned int 
> lcore_id, uint16_t port_id,
>                       ret = -ENOTSUP;
>                       goto end;
>               }
> -             /* initialize data before enabling the callback */
> -             queue_cfg->empty_poll_stats = 0;
> -             queue_cfg->cb_mode = mode;
> -             queue_cfg->umwait_in_progress = false;
> -             queue_cfg->pwr_mgmt_state = PMD_MGMT_ENABLED;
> -
> -             /* ensure we update our state before callback starts */
> -             rte_atomic_thread_fence(__ATOMIC_SEQ_CST);
> -
> -             queue_cfg->cur_cb = rte_eth_add_rx_callback(port_id, queue_id,
> -                             clb_umwait, NULL);
> +             clb = clb_umwait;
>               break;
>       }
>       case RTE_POWER_MGMT_TYPE_SCALE:
> @@ -269,16 +239,7 @@ rte_power_ethdev_pmgmt_queue_enable(unsigned int 
> lcore_id, uint16_t port_id,
>                       ret = -ENOTSUP;
>                       goto end;
>               }
> -             /* initialize data before enabling the callback */
> -             queue_cfg->empty_poll_stats = 0;
> -             queue_cfg->cb_mode = mode;
> -             queue_cfg->pwr_mgmt_state = PMD_MGMT_ENABLED;
> -
> -             /* this is not necessary here, but do it anyway */
> -             rte_atomic_thread_fence(__ATOMIC_SEQ_CST);
> -
> -             queue_cfg->cur_cb = rte_eth_add_rx_callback(port_id,
> -                             queue_id, clb_scale_freq, NULL);
> +             clb = clb_scale_freq;
>               break;
>       }
>       case RTE_POWER_MGMT_TYPE_PAUSE:
> @@ -286,18 +247,21 @@ rte_power_ethdev_pmgmt_queue_enable(unsigned int 
> lcore_id, uint16_t port_id,
>               if (global_data.tsc_per_us == 0)
>                       calc_tsc();
> 
> -             /* initialize data before enabling the callback */
> -             queue_cfg->empty_poll_stats = 0;
> -             queue_cfg->cb_mode = mode;
> -             queue_cfg->pwr_mgmt_state = PMD_MGMT_ENABLED;
> -
> -             /* this is not necessary here, but do it anyway */
> -             rte_atomic_thread_fence(__ATOMIC_SEQ_CST);
> -
> -             queue_cfg->cur_cb = rte_eth_add_rx_callback(port_id, queue_id,
> -                             clb_pause, NULL);
> +             clb = clb_pause;
>               break;
> +     default:
> +             RTE_LOG(DEBUG, POWER, "Invalid power management type\n");
> +             ret = -EINVAL;
> +             goto end;
>       }
> +
> +     /* initialize data before enabling the callback */
> +     queue_cfg->empty_poll_stats = 0;
> +     queue_cfg->cb_mode = mode;
> +     queue_cfg->pwr_mgmt_state = PMD_MGMT_ENABLED;
> +     queue_cfg->cur_cb = rte_eth_add_rx_callback(port_id, queue_id,
> +                     clb, NULL);
> +
>       ret = 0;
>  end:
>       return ret;
> @@ -323,27 +287,8 @@ rte_power_ethdev_pmgmt_queue_disable(unsigned int 
> lcore_id,
>       /* stop any callbacks from progressing */
>       queue_cfg->pwr_mgmt_state = PMD_MGMT_DISABLED;
> 
> -     /* ensure we update our state before continuing */
> -     rte_atomic_thread_fence(__ATOMIC_SEQ_CST);
> -
>       switch (queue_cfg->cb_mode) {
> -     case RTE_POWER_MGMT_TYPE_MONITOR:
> -     {
> -             bool exit = false;
> -             do {
> -                     /*
> -                      * we may request cancellation while the other thread
> -                      * has just entered the callback but hasn't started
> -                      * sleeping yet, so keep waking it up until we know it's
> -                      * done sleeping.
> -                      */
> -                     if (queue_cfg->umwait_in_progress)
> -                             rte_power_monitor_wakeup(lcore_id);
> -                     else
> -                             exit = true;
> -             } while (!exit);
> -     }
> -     /* fall-through */
> +     case RTE_POWER_MGMT_TYPE_MONITOR: /* fall-through */
>       case RTE_POWER_MGMT_TYPE_PAUSE:
>               rte_eth_remove_rx_callback(port_id, queue_id,
>                               queue_cfg->cur_cb);
> @@ -356,10 +301,11 @@ rte_power_ethdev_pmgmt_queue_disable(unsigned int 
> lcore_id,
>               break;
>       }
>       /*
> -      * we don't free the RX callback here because it is unsafe to do so
> -      * unless we know for a fact that all data plane threads have stopped.
> +      * the API doc mandates that the user stops all processing on affected
> +      * ports before calling any of these API's, so we can assume that the
> +      * callbacks can be freed. we're intentionally casting away const-ness.
>        */
> -     queue_cfg->cur_cb = NULL;
> +     rte_free((void *)queue_cfg->cur_cb);
> 
>       return 0;
>  }
> diff --git a/lib/power/rte_power_pmd_mgmt.h b/lib/power/rte_power_pmd_mgmt.h
> index 7a0ac24625..7557f5d7e1 100644
> --- a/lib/power/rte_power_pmd_mgmt.h
> +++ b/lib/power/rte_power_pmd_mgmt.h
> @@ -43,6 +43,9 @@ enum rte_power_pmd_mgmt_type {
>   *
>   * @note This function is not thread-safe.
>   *
> + * @warning This function must be called when all affected Ethernet ports are
> + *   stopped and no Rx/Tx is in progress!
> + *
>   * @param lcore_id
>   *   The lcore the Rx queue will be polled from.
>   * @param port_id
> @@ -69,6 +72,9 @@ rte_power_ethdev_pmgmt_queue_enable(unsigned int lcore_id,
>   *
>   * @note This function is not thread-safe.
>   *
> + * @warning This function must be called when all affected Ethernet ports are
> + *   stopped and no Rx/Tx is in progress!
> + *
>   * @param lcore_id
>   *   The lcore the Rx queue is polled from.
>   * @param port_id
> --
> 2.25.1

Reply via email to