Funny, everyone is trying to get more performance out of vhost-user and you're are adding qos... :-)
I am still looking into it, so don't consider this a throughout review. On Fri, Oct 16, 2015 at 02:28:10PM +0100, Ian Stokes wrote: > This patch provides the modifications required in netdev-dpdk.c and > vswitch.xml to allow for a DPDK user space QoS algorithm. > > This patch adds a QoS configuration structure for netdev-dpdk and > expected QoS operations 'dpdk_qos_ops'. Various helper functions > are also supplied. > > Also included are the modifications required for vswitch.xml to allow a > new QoS implementation for netdev-dpdk devices. This includes a new QoS type > `egress-policer` as well as its expected QoS table entries. > > The QoS functionality implemented for DPDK devices is `egress-policer`. > This can be used to drop egress packets at a configurable rate. > > The INSTALL.DPDK.md guide has also been modified to provide an example > configuration of `egress-policer` QoS. > > Signed-off-by: Ian Stokes <ian.sto...@intel.com> > --- > INSTALL.DPDK.md | 20 +++ > lib/netdev-dpdk.c | 416 > +++++++++++++++++++++++++++++++++++++++++++++++++- > vswitchd/vswitch.xml | 52 +++++++ > 3 files changed, 481 insertions(+), 7 deletions(-) > > diff --git a/INSTALL.DPDK.md b/INSTALL.DPDK.md > index 7bf110c..c64b8ce 100644 > --- a/INSTALL.DPDK.md > +++ b/INSTALL.DPDK.md > @@ -207,6 +207,26 @@ Using the DPDK with ovs-vswitchd: > ./ovs-ofctl add-flow br0 in_port=2,action=output:1 > ``` > > +8. QoS usage example > + > + Assuming you have a vhost-user port transmitting traffic consisting of > + packets of size 64 bytes, the following command would limit the egress > + transmission rate of the port to ~1,000,000 packets per second: > + > + `ovs-vsctl set port vhost-user0 qos=@newqos -- --id=@newqos create qos > + type=egress-policer other-config:cir=46000000 other-config:cbs=2048` > + > + To examine the QoS configuration of the port: > + > + `ovs-appctl -t ovs-vswitchd qos/show vhost-user0` > + > + To clear the QoS configuration from the port and ovsdb use the following: > + > + `ovs-vsctl -- destroy QoS vhost-user0 -- clear Port vhost-user0 qos` > + > + For more details regarding egress-policer parameters please refer to the > + vswitch.xml. > + > Performance Tuning: > ------------------- > > diff --git a/lib/netdev-dpdk.c b/lib/netdev-dpdk.c > index 93b0589..cd4c532 100644 > --- a/lib/netdev-dpdk.c > +++ b/lib/netdev-dpdk.c > @@ -44,6 +44,7 @@ > #include "ovs-rcu.h" > #include "packets.h" > #include "shash.h" > +#include "smap.h" > #include "sset.h" > #include "unaligned.h" > #include "timeval.h" > @@ -52,12 +53,14 @@ > > #include "rte_config.h" > #include "rte_mbuf.h" > +#include "rte_meter.h" > #include "rte_virtio_net.h" > > VLOG_DEFINE_THIS_MODULE(dpdk); > static struct vlog_rate_limit rl = VLOG_RATE_LIMIT_INIT(5, 20); > > #define DPDK_PORT_WATCHDOG_INTERVAL 5 > +#define DPDK_MAX_QOS_NAME_SIZE 10 > > #define OVS_CACHE_LINE_SIZE CACHE_LINE_SIZE > #define OVS_VPORT_DPDK "ovs_dpdk" > @@ -142,6 +145,107 @@ static int rte_eal_init_ret = ENODEV; > > static struct ovs_mutex dpdk_mutex = OVS_MUTEX_INITIALIZER; > > +/* Quality of Service */ > + > +/* An instance of a QoS configuration. Always associated with a particular > + * network device. > + * > + * Each QoS implementation subclasses this with whatever additional data it > + * needs. > + */ > +struct qos_conf { > + const struct dpdk_qos_ops *ops; > +}; > + > +/* A particular implementation of dpdk QoS operations. > + * > + * The functions below return 0 if successful or a positive errno value on > + * failure, except where otherwise noted. All of them must be provided, > except > + * where otherwise noted. > + */ > +struct dpdk_qos_ops { > + > + /* Name of the QoS type */ > + const char *qos_name; > + > + /* Called to construct the QoS implementation on 'netdev'. The > + * implementation should make the appropriate calls to configure QoS > + * according to 'details'. The implementation may assume that any current > + * QoS configuration already installed should be destroyed before > + * constructing the new configuration. > + * > + * The contents of 'details' should be documented as valid for 'ovs_name' > + * in the "other_config" column in the "QoS" table in > vswitchd/vswitch.xml > + * (which is built as ovs-vswitchd.conf.db(8)). > + * > + * This function must return 0 if and only if it sets 'netdev->qos_conf' > + * to an initialized 'struct qos_conf'. > + * > + * For all QoS implementations it should always be non-null. > + */ > + int (*qos_construct)(struct netdev *netdev, const struct smap *details); > + > + /* Destroys the data structures allocated by the implementation as part > of > + * 'qos_conf. > + * > + * For all QoS implementations it should always be non-null. > + */ > + void (*qos_destruct)(struct netdev *netdev, struct qos_conf *conf); > + > + /* Retrieves details of 'netdev->qos_conf' configuration into 'details'. > + * > + * The contents of 'details' should be documented as valid for 'ovs_name' > + * in the "other_config" column in the "QoS" table in > vswitchd/vswitch.xml > + * (which is built as ovs-vswitchd.conf.db(8)). > + */ > + int (*qos_get)(const struct netdev *netdev, struct smap *details); > + > + /* Reconfigures 'netdev->qos_conf' according to 'details', performing any > + * required calls to complete the reconfiguration. > + * > + * The contents of 'details' should be documented as valid for 'ovs_name' > + * in the "other_config" column in the "QoS" table in > vswitchd/vswitch.xml > + * (which is built as ovs-vswitchd.conf.db(8)). > + * > + * This function may be null if 'qos_conf' is not configurable. > + */ > + int (*qos_set)(struct netdev *netdev, const struct smap *details); > + > + /* Modify an array of rte_mbufs. The modification is specific to > + * each qos implementation. > + * > + * The function should take and array of mbufs and an int representing > + * the current number of mbufs present in the array. > + * > + * After the function has performed a qos modification to the array of > + * mbufs it returns an int representing the number of mbufs now present > in > + * the array. This value is can then be passed to the port send function > + * along with the modified array for transmission. > + * > + * For all QoS implementations it should always be non-null. > + */ > + int (*qos_alg_process)(struct netdev *netdev, struct rte_mbuf **pkts, > + int pkt_cnt); > +}; > + > +/* dpdk_qos_ops for each type of user space QoS implementation */ > +static const struct dpdk_qos_ops egress_policer_ops; > + > +/* > + * Array of dpdk_qos_ops, contains pointer to all supported QoS > + * operations. > + */ > +static const struct dpdk_qos_ops *const qos_confs[] = { > + &egress_policer_ops, > + NULL > +}; > + > +/* Action that can be set for a packet for rte_meter */ > +enum egress_policer_action { > + GREEN = e_RTE_METER_GREEN, > + DROP = 1, > +}; > + > /* Contains all 'struct dpdk_dev's. */ > static struct ovs_list dpdk_list OVS_GUARDED_BY(dpdk_mutex) > = OVS_LIST_INITIALIZER(&dpdk_list); > @@ -235,6 +339,11 @@ struct netdev_dpdk { > > /* In dpdk_list. */ > struct ovs_list list_node OVS_GUARDED_BY(dpdk_mutex); > + > + /* QoS configuration and lock for the device */ > + struct qos_conf *qos_conf; > + rte_spinlock_t qos_lock; > + > }; > > struct netdev_rxq_dpdk { > @@ -614,6 +723,10 @@ netdev_dpdk_init(struct netdev *netdev_, unsigned int > port_no, > goto unlock; > } > > + /* Initialise QoS configuration to NULL and qos lock to unlocked */ > + netdev->qos_conf = NULL; > + rte_spinlock_init(&netdev->qos_lock); > + > netdev_->n_txq = NR_QUEUE; > netdev_->n_rxq = NR_QUEUE; > netdev->real_n_txq = NR_QUEUE; > @@ -1038,6 +1151,21 @@ netdev_dpdk_rxq_recv(struct netdev_rxq *rxq_, struct > dp_packet **packets, > return 0; > } > > +static inline int > +netdev_dpdk_qos_process__(struct netdev_dpdk *dev, struct rte_mbuf **pkts, > + int cnt) > +{ > + struct netdev *netdev = &dev->up; > + People are working really hard to get the data path as fast as possible and this seems to be too expensive: > + rte_spinlock_lock(&dev->qos_lock); > + if (dev->qos_conf != NULL) { > + cnt = dev->qos_conf->ops->qos_alg_process(netdev, pkts, cnt); > + } > + rte_spinlock_unlock(&dev->qos_lock); Maybe we could do something like this: if (dev->qos_conf != NULL) { rte_spinlock_lock(&dev->qos_lock); if (dev->qos_conf != NULL) { cnt = dev->qos_conf->ops->qos_alg_process(netdev, pkts, cnt); } rte_spinlock_unlock(&dev->qos_lock); } Though I still have to look at qos_alg_process() to see if the locking could be done in another way. > + > + return cnt; > +} > + > static inline void > netdev_dpdk_vhost_update_tx_counters(struct netdev_stats *stats, > struct dp_packet **packets, > @@ -1072,6 +1200,9 @@ __netdev_dpdk_vhost_send(struct netdev *netdev, struct > dp_packet **pkts, > goto out; > } > > + /* Check has QoS has been configured for the netdev */ > + cnt = netdev_dpdk_qos_process__(vhost_dev, cur_pkts, cnt); > + > /* There is vHost TX single queue, So we need to lock it for TX. */ > rte_spinlock_lock(&vhost_dev->vhost_tx_lock); > > @@ -1216,6 +1347,10 @@ dpdk_do_tx_copy(struct netdev *netdev, int qid, struct > dp_packet **pkts, > if (dev->type == DPDK_DEV_VHOST) { > __netdev_dpdk_vhost_send(netdev, (struct dp_packet **) mbufs, > newcnt, true); > } else { > + > + /* Check if QoS has been configured for this netdev. */ > + newcnt = netdev_dpdk_qos_process__(dev, mbufs, newcnt); > + > dpdk_queue_pkts(dev, qid, mbufs, newcnt); > dpdk_queue_flush(dev, qid); > } > @@ -1275,9 +1410,15 @@ netdev_dpdk_send__(struct netdev_dpdk *dev, int qid, > > if (OVS_UNLIKELY(size > dev->max_packet_len)) { > if (next_tx_idx != i) { > + cnt = i - next_tx_idx; > + > + cnt = netdev_dpdk_qos_process__(dev, > + (struct rte_mbuf**)pkts, > + cnt); I have another concern at this point because for each batch it will grab a lock pretty much serializing multiple queues. > + > dpdk_queue_pkts(dev, qid, > (struct rte_mbuf **)&pkts[next_tx_idx], > - i-next_tx_idx); > + cnt); > } > > VLOG_WARN_RL(&rl, "Too big size %d max_packet_len %d", > @@ -1289,9 +1430,12 @@ netdev_dpdk_send__(struct netdev_dpdk *dev, int qid, > } > } > if (next_tx_idx != cnt) { > - dpdk_queue_pkts(dev, qid, > - (struct rte_mbuf **)&pkts[next_tx_idx], > - cnt-next_tx_idx); > + cnt -= next_tx_idx; > + > + cnt = netdev_dpdk_qos_process__(dev, (struct rte_mbuf**)pkts, > cnt); Same here. and it seems it should update tx_dropped too. Thanks, fbl > + > + dpdk_queue_pkts(dev, qid, (struct rte_mbuf **)&pkts[next_tx_idx], > + cnt); > } > > if (OVS_UNLIKELY(dropped)) { > @@ -2032,6 +2176,264 @@ unlock_dpdk: > return err; > } > > +/* QoS Functions */ > + > +/* > + * Initialize QoS configuration operations. > + */ > +static void > +qos_conf_init(struct qos_conf *conf, const struct dpdk_qos_ops *ops) > +{ > + conf->ops = ops; > +} > + > +/* > + * Search existing QoS operations in qos_ops and compare each set of > + * operations qos_name to name. Return a dpdk_qos_ops pointer to a match, > + * else return NULL > + */ > +static const struct dpdk_qos_ops * > +qos_lookup_name(const char *name) > +{ > + const struct dpdk_qos_ops *const *opsp; > + > + for (opsp = qos_confs; *opsp != NULL; opsp++) { > + const struct dpdk_qos_ops *ops = *opsp; > + if (!strncmp(name, ops->qos_name,DPDK_MAX_QOS_NAME_SIZE)) { > + return ops; > + } > + } > + return NULL; > +} > + > +/* > + * Call qos_destruct to clean up items associated with the netdevs > + * qos_conf. Set netdevs qos_conf to NULL. > + */ > +static void > +qos_delete_conf(struct netdev *netdev_) > +{ > + struct netdev_dpdk *netdev = netdev_dpdk_cast(netdev_); > + > + rte_spinlock_lock(&netdev->qos_lock); > + if (netdev->qos_conf) { > + if (netdev->qos_conf->ops->qos_destruct) { > + netdev->qos_conf->ops->qos_destruct(netdev_, netdev->qos_conf); > + } > + netdev->qos_conf = NULL; > + } > + rte_spinlock_unlock(&netdev->qos_lock); > +} > + > +static int > +netdev_dpdk_get_qos_types(const struct netdev *netdev OVS_UNUSED, > + struct sset *types) > +{ > + const struct dpdk_qos_ops *const *opsp; > + > + for (opsp = qos_confs; *opsp != NULL; opsp++) { > + const struct dpdk_qos_ops *ops = *opsp; > + if (ops->qos_construct && ops->qos_name[0] != '\0') { > + sset_add(types, ops->qos_name); > + } > + } > + return 0; > +} > + > +static int > +netdev_dpdk_get_qos(const struct netdev *netdev_, > + const char **typep, struct smap *details) > +{ > + struct netdev_dpdk *netdev = netdev_dpdk_cast(netdev_); > + int error = 0; > + > + ovs_mutex_lock(&netdev->mutex); > + if(netdev->qos_conf) { > + *typep = netdev->qos_conf->ops->qos_name; > + error = (netdev->qos_conf->ops->qos_get > + ? netdev->qos_conf->ops->qos_get(netdev_, details): 0); > + } > + ovs_mutex_unlock(&netdev->mutex); > + > + return error; > +} > + > +static int > +netdev_dpdk_set_qos(struct netdev *netdev_, > + const char *type, const struct smap *details) > +{ > + struct netdev_dpdk *netdev = netdev_dpdk_cast(netdev_); > + const struct dpdk_qos_ops *new_ops = NULL; > + int error = 0; > + > + /* If type is empty the current QoS configuration for the dpdk-netdev can > + * be destroyed */ > + if(type[0] == '\0') { > + qos_delete_conf(netdev_); > + } > + > + new_ops = qos_lookup_name(type); > + if (!new_ops || !new_ops->qos_construct) { > + return EOPNOTSUPP; > + } > + > + ovs_mutex_lock(&netdev->mutex); > + if (netdev->qos_conf) { > + if (new_ops == netdev->qos_conf->ops) { > + error = new_ops->qos_set ? new_ops->qos_set(netdev_, details) : > 0; > + } > + else { > + /* Delete existing QoS configuration. */ > + qos_delete_conf(netdev_); > + ovs_assert(netdev->qos_conf == NULL); > + > + /* Install new QoS configuration. */ > + error = new_ops->qos_construct(netdev_, details); > + ovs_assert((error == 0) == (netdev->qos_conf != NULL)); > + } > + } else { > + error = new_ops->qos_construct(netdev_, details); > + ovs_assert((error == 0) == (netdev->qos_conf != NULL)); > + } > + > + ovs_mutex_unlock(&netdev->mutex); > + return error; > +} > + > +/* egress-policer details */ > + > +struct egress_policer { > + struct qos_conf qos_conf; > + struct rte_meter_srtcm_params app_srtcm_params; > + struct rte_meter_srtcm test_meter; > +}; > + > +static struct egress_policer * > +egress_policer_get__(const struct netdev *netdev_) > +{ > + struct netdev_dpdk *netdev = netdev_dpdk_cast(netdev_); > + return CONTAINER_OF(netdev->qos_conf, struct egress_policer, qos_conf); > +} > + > +static int > +egress_policer_qos_construct(struct netdev *netdev_, > + const struct smap *details) > +{ > + struct netdev_dpdk *netdev = netdev_dpdk_cast(netdev_); > + struct egress_policer *policer; > + const char *cir_s; > + const char *cbs_s; > + int err = 0; > + > + rte_spinlock_lock(&netdev->qos_lock); > + policer = xmalloc(sizeof *policer); > + qos_conf_init(&policer->qos_conf, &egress_policer_ops); > + netdev->qos_conf = &policer->qos_conf; > + cir_s = smap_get(details, "cir"); > + cbs_s = smap_get(details, "cbs"); > + policer->app_srtcm_params.cir = cir_s ? strtoull(cir_s, NULL, 10) : 0; > + policer->app_srtcm_params.cbs = cbs_s ? strtoull(cbs_s, NULL, 10) : 0; > + policer->app_srtcm_params.ebs = 0; > + err = rte_meter_srtcm_config(&policer->test_meter, > + &policer->app_srtcm_params); > + rte_spinlock_unlock(&netdev->qos_lock); > + > + return err; > +} > + > +static void > +egress_policer_qos_destruct(struct netdev *netdev_ OVS_UNUSED, > + struct qos_conf *conf) > +{ > + struct egress_policer *policer = CONTAINER_OF(conf, struct > egress_policer, > + qos_conf); > + free(policer); > +} > + > +static int > +egress_policer_qos_get(const struct netdev *netdev, struct smap *details) > +{ > + struct egress_policer *policer = egress_policer_get__(netdev); > + smap_add_format(details, "cir", "%llu", > + 1ULL * policer->app_srtcm_params.cir); > + smap_add_format(details, "cbs", "%llu", > + 1ULL * policer->app_srtcm_params.cbs); > + return 0; > +} > + > +static int > +egress_policer_qos_set(struct netdev *netdev_, const struct smap *details) > +{ > + struct egress_policer *policer; > + const char *cir_s; > + const char *cbs_s; > + int err = 0; > + > + policer = egress_policer_get__(netdev_); > + cir_s = smap_get(details, "cir"); > + cbs_s = smap_get(details, "cbs"); > + policer->app_srtcm_params.cir = cir_s ? strtoull(cir_s, NULL, 10) : 0; > + policer->app_srtcm_params.cbs = cbs_s ? strtoull(cbs_s, NULL, 10) : 0; > + policer->app_srtcm_params.ebs = 0; > + err = rte_meter_srtcm_config(&policer->test_meter, > + &policer->app_srtcm_params); > + > + return err; > +} > + > +static inline int > +egress_policer_pkt_handle__(struct rte_meter_srtcm *meter, > + struct rte_mbuf *pkt, uint64_t time) > +{ > + uint8_t output_color; > + uint32_t pkt_len = rte_pktmbuf_pkt_len(pkt) - sizeof(struct ether_hdr); > + enum egress_policer_action action = GREEN; > + > + /* color input is not used for blind modes */ > + output_color = (uint8_t) rte_meter_srtcm_color_blind_check(meter, time, > + pkt_len); > + > + /* Check output color, 0 corresponds to GREEN i.e. packet can be > + * transmitted. If greater than 0 then action is set to DROP. */ > + if (output_color) { > + action = DROP; > + } > + > + return action; > +} > + > +static int > +egress_policer_alg_process(struct netdev *netdev_, struct rte_mbuf **pkts, > + int pkt_cnt) > +{ > + int i = 0; > + int cnt = pkt_cnt; > + struct egress_policer *policer = egress_policer_get__(netdev_); > + struct rte_mbuf *pkt = NULL; > + uint64_t current_time = rte_rdtsc(); > + > + for(i = 0; i < pkt_cnt; i++) { > + pkt = pkts[i]; > + /* Handle current packet */ > + if (egress_policer_pkt_handle__(&policer->test_meter, pkt, > + current_time) == DROP) { > + rte_pktmbuf_free(pkt); > + cnt = cnt - 1; > + } > + } > + > + return cnt; > +} > + > +static const struct dpdk_qos_ops egress_policer_ops = { > + "egress-policer", /* qos_name */ > + egress_policer_qos_construct, > + egress_policer_qos_destruct, > + egress_policer_qos_get, > + egress_policer_qos_set, > + egress_policer_alg_process > +}; > + > #define NETDEV_DPDK_CLASS(NAME, INIT, CONSTRUCT, DESTRUCT, MULTIQ, SEND, \ > GET_CARRIER, GET_STATS, GET_FEATURES, GET_STATUS, RXQ_RECV) \ > { \ > @@ -2069,10 +2471,10 @@ unlock_dpdk: > NULL, /* set_advertisements */ \ > \ > NULL, /* set_policing */ \ > - NULL, /* get_qos_types */ \ > + netdev_dpdk_get_qos_types, \ > NULL, /* get_qos_capabilities */ \ > - NULL, /* get_qos */ \ > - NULL, /* set_qos */ \ > + netdev_dpdk_get_qos, \ > + netdev_dpdk_set_qos, \ > NULL, /* get_queue */ \ > NULL, /* set_queue */ \ > NULL, /* delete_queue */ \ > diff --git a/vswitchd/vswitch.xml b/vswitchd/vswitch.xml > index 4dbd9e3..5be8eb6 100644 > --- a/vswitchd/vswitch.xml > +++ b/vswitchd/vswitch.xml > @@ -3298,6 +3298,26 @@ > for information on how this classifier works. > </dd> > </dl> > + <dl> > + <dt><code>egress-policer</code></dt> > + <dd> > + An egress policer algorithm. This implementation uses the DPDK > + rte_meter library. The rte_meter library provides an implementation > + of srTCM (RFC2697) which allows the metering and policing of > traffic. > + The implementation in OVS creates a color blind srTCM meter with a > + single token bucket used to police traffic. It should be noted that > + when the rte_meter is configured as part of QoS there will be a > + performance overhead as the rte_mter itself will consume CPU cycles > + in order to police traffic. These CPU cycles ordinarily are used > for > + packet proccessing. As such the drop in performance will be noticed > + in terms of overall aggregate traffic throughput. > + > + For more information regarding the usage of the DPDK rte_meter see > + > <code>http://dpdk.org/doc/guides/sample_app_ug/qos_metering.html</code> > + For more information regarding srTCM see > + <code>https://tools.ietf.org/html/rfc2697</code> > + </dd> > + </dl> > </column> > > <column name="queues"> > @@ -3334,6 +3354,38 @@ > </column> > </group> > > + <group title="Configuration for egress-policer QoS"> > + <p> > + <ref table="QoS"/> <ref table="QoS" column="type"/> > + <code>egress-policer</code> provides egress policing for userspace > + port types with DPDK. > + > + It has the following key-value pairs defined. > + </p> > + > + <column name="other_config" key="cir" type='{"type": "integer"}'> > + The Committed Information Rate (CIR) is measured in bytes of IP > + packets per second, i.e. it includes the IP header, but not link > + specific (e.g. Ethernet) headers. This represents the bytes per > second > + rate at which the token bucket will be updated. The cir value is > + calculated by (pps x packet data size). For example assuming a user > + wishes to limit a stream consisting of 64 byte packets to 1 million > + packets per second the CIR would be set to to to 46000000. This value > + can be broken into '1,000,000 x 46'. Where 1,000,000 is the policing > + rate for the number of packets per second and 46 represents the size > + of the packet data for a 64 byte ip packet. > + </column> > + <column name="other_config" key="cbs" type='{"type": "integer"}'> > + The Committed Burst Size (CBS) is measured in bytes and represents a > + token bucket. At a minimum this value should be be set to the > expected > + largest size packet in the traffic stream. In practice larger values > + may be used to increase the size of the token bucket. If a packet can > + be transmitted then the cbs will be decremented by the number of > + bytes/tokens of the packet. If there are not enough tokens in the cbs > + bucket the packet will be dropped. > + </column> > + </group> > + > <group title="Common Columns"> > The overall purpose of these columns is described under <code>Common > Columns</code> at the beginning of this document. > -- > 1.7.4.1 > > _______________________________________________ > dev mailing list > dev@openvswitch.org > http://openvswitch.org/mailman/listinfo/dev _______________________________________________ dev mailing list dev@openvswitch.org http://openvswitch.org/mailman/listinfo/dev