Re: [dpdk-dev] [PATCH] ethdev: fix flow expansion matching types

2018-07-11 Thread Adrien Mazarguil
On Wed, Jul 11, 2018 at 08:49:35AM +0200, Nelio Laranjeiro wrote:
> Node RSS types are generally covering more RSS kind than the user is
> requesting, it should accept to expand even if only a single bit is
> remains after masking.  Setting the correct RSS kind for the rule
> remains the driver job.
> 
> Fixes: 959823ce4c47 ("ethdev: add flow API to expand RSS flows")
> 
> Signed-off-by: Nelio Laranjeiro 

Acked-by: Adrien Mazarguil 

-- 
Adrien Mazarguil
6WIND


[dpdk-dev] [PATCH] net/thunderx: Block sq door writes on zero pkts

2018-07-11 Thread Kiran Kumar
With current code, we are performing sq door writes even with 0 pkts.
this will create pressure on register bus. This patch will block these
writes.

Fixes: 1c421f18e0 ("net/thunderx: add single and multi-segment Tx")

Signed-off-by: Kiran Kumar 
---
 drivers/net/thunderx/nicvf_rxtx.c | 12 +++-
 1 file changed, 7 insertions(+), 5 deletions(-)

diff --git a/drivers/net/thunderx/nicvf_rxtx.c 
b/drivers/net/thunderx/nicvf_rxtx.c
index 72305d9..8075a8e 100644
--- a/drivers/net/thunderx/nicvf_rxtx.c
+++ b/drivers/net/thunderx/nicvf_rxtx.c
@@ -162,12 +162,14 @@ nicvf_xmit_pkts(void *tx_queue, struct rte_mbuf 
**tx_pkts, uint16_t nb_pkts)
free_desc -= TX_DESC_PER_PKT;
}
 
-   sq->tail = tail;
-   sq->xmit_bufs += i;
-   rte_wmb();
+   if (likely(i)) {
+   sq->tail = tail;
+   sq->xmit_bufs += i;
+   rte_wmb();
 
-   /* Inform HW to xmit the packets */
-   nicvf_addr_write(sq->sq_door, i * TX_DESC_PER_PKT);
+   /* Inform HW to xmit the packets */
+   nicvf_addr_write(sq->sq_door, i * TX_DESC_PER_PKT);
+   }
return i;
 }
 
-- 
2.7.4



[dpdk-dev] [PATCH v3 00/21] net/mlx5: flow rework

2018-07-11 Thread Nelio Laranjeiro
Re-work flow engine to support port redirection actions through TC.

This first series depends on [1] which is implemented in commit 
"net/mlx5: support inner RSS computation" and on [2].
Next series will bring the port redirection as announced[3].

[1] https://mails.dpdk.org/archives/dev/2018-July/107378.html
[2] https://mails.dpdk.org/archives/dev/2018-June/104192.html
[3] https://mails.dpdk.org/archives/dev/2018-May/103043.html

Changes in v3:

- remove redundant parameters in drop queues internal API.
- simplify the RSS expansion by only adding missing items in the pattern.
- document all functions.

Nelio Laranjeiro (21):
  net/mlx5: remove flow support
  net/mlx5: handle drop queues as regular queues
  net/mlx5: replace verbs priorities by flow
  net/mlx5: support flow Ethernet item along with drop action
  net/mlx5: add flow queue action
  net/mlx5: add flow stop/start
  net/mlx5: add flow VLAN item
  net/mlx5: add flow IPv4 item
  net/mlx5: add flow IPv6 item
  net/mlx5: add flow UDP item
  net/mlx5: add flow TCP item
  net/mlx5: add mark/flag flow action
  net/mlx5: use a macro for the RSS key size
  net/mlx5: add RSS flow action
  net/mlx5: remove useless arguments in hrxq API
  net/mlx5: support inner RSS computation
  net/mlx5: add flow VXLAN item
  net/mlx5: add flow VXLAN-GPE item
  net/mlx5: add flow GRE item
  net/mlx5: add flow MPLS item
  net/mlx5: add count flow action

 drivers/net/mlx5/mlx5.c|   22 +-
 drivers/net/mlx5/mlx5.h|   18 +-
 drivers/net/mlx5/mlx5_ethdev.c |   14 +-
 drivers/net/mlx5/mlx5_flow.c   | 4815 
 drivers/net/mlx5/mlx5_prm.h|3 +
 drivers/net/mlx5/mlx5_rss.c|7 +-
 drivers/net/mlx5/mlx5_rxq.c|  281 +-
 drivers/net/mlx5/mlx5_rxtx.h   |   21 +-
 8 files changed, 2632 insertions(+), 2549 deletions(-)

-- 
2.18.0



[dpdk-dev] [PATCH v3 02/21] net/mlx5: handle drop queues as regular queues

2018-07-11 Thread Nelio Laranjeiro
Drop queues are essentially used in flows due to Verbs API, the
information if the fate of the flow is a drop or not is already present
in the flow.  Due to this, drop queues can be fully mapped on regular
queues.

Signed-off-by: Nelio Laranjeiro 
---
 drivers/net/mlx5/mlx5.c  |  24 ++--
 drivers/net/mlx5/mlx5.h  |  14 ++-
 drivers/net/mlx5/mlx5_flow.c |  94 +++---
 drivers/net/mlx5/mlx5_rxq.c  | 232 +++
 drivers/net/mlx5/mlx5_rxtx.h |   6 +
 5 files changed, 308 insertions(+), 62 deletions(-)

diff --git a/drivers/net/mlx5/mlx5.c b/drivers/net/mlx5/mlx5.c
index df7f39844..e9780ac8f 100644
--- a/drivers/net/mlx5/mlx5.c
+++ b/drivers/net/mlx5/mlx5.c
@@ -261,7 +261,6 @@ mlx5_dev_close(struct rte_eth_dev *dev)
priv->txqs_n = 0;
priv->txqs = NULL;
}
-   mlx5_flow_delete_drop_queue(dev);
mlx5_mprq_free_mp(dev);
mlx5_mr_release(dev);
if (priv->pd != NULL) {
@@ -1139,22 +1138,15 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
mlx5_link_update(eth_dev, 0);
/* Store device configuration on private structure. */
priv->config = config;
-   /* Create drop queue. */
-   err = mlx5_flow_create_drop_queue(eth_dev);
-   if (err) {
-   DRV_LOG(ERR, "port %u drop queue allocation failed: %s",
-   eth_dev->data->port_id, strerror(rte_errno));
-   err = rte_errno;
-   goto error;
-   }
/* Supported Verbs flow priority number detection. */
-   if (verb_priorities == 0)
-   verb_priorities = mlx5_get_max_verbs_prio(eth_dev);
-   if (verb_priorities < MLX5_VERBS_FLOW_PRIO_8) {
-   DRV_LOG(ERR, "port %u wrong Verbs flow priorities: %u",
-   eth_dev->data->port_id, verb_priorities);
-   err = ENOTSUP;
-   goto error;
+   if (verb_priorities == 0) {
+   err = mlx5_verbs_max_prio(eth_dev);
+   if (err < 0) {
+   DRV_LOG(ERR, "port %u wrong Verbs flow priorities",
+   eth_dev->data->port_id);
+   goto error;
+   }
+   verb_priorities = err;
}
priv->config.max_verbs_prio = verb_priorities;
/*
diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h
index cc01310e0..227429848 100644
--- a/drivers/net/mlx5/mlx5.h
+++ b/drivers/net/mlx5/mlx5.h
@@ -139,9 +139,6 @@ enum mlx5_verbs_alloc_type {
MLX5_VERBS_ALLOC_TYPE_RX_QUEUE,
 };
 
-/* 8 Verbs priorities. */
-#define MLX5_VERBS_FLOW_PRIO_8 8
-
 /**
  * Verbs allocator needs a context to know in the callback which kind of
  * resources it is allocating.
@@ -153,6 +150,12 @@ struct mlx5_verbs_alloc_ctx {
 
 LIST_HEAD(mlx5_mr_list, mlx5_mr);
 
+/* Flow drop context necessary due to Verbs API. */
+struct mlx5_drop {
+   struct mlx5_hrxq *hrxq; /* Hash Rx queue queue. */
+   struct mlx5_rxq_ibv *rxq; /* Verbs Rx queue. */
+};
+
 struct priv {
LIST_ENTRY(priv) mem_event_cb; /* Called by memory event callback. */
struct rte_eth_dev_data *dev_data;  /* Pointer to device data. */
@@ -182,7 +185,7 @@ struct priv {
struct rte_intr_handle intr_handle; /* Interrupt handler. */
unsigned int (*reta_idx)[]; /* RETA index table. */
unsigned int reta_idx_n; /* RETA index size. */
-   struct mlx5_hrxq_drop *flow_drop_queue; /* Flow drop queue. */
+   struct mlx5_drop drop_queue; /* Flow drop queues. */
struct mlx5_flows flows; /* RTE Flow rules. */
struct mlx5_flows ctrl_flows; /* Control flow rules. */
struct {
@@ -314,7 +317,8 @@ int mlx5_traffic_restart(struct rte_eth_dev *dev);
 
 /* mlx5_flow.c */
 
-unsigned int mlx5_get_max_verbs_prio(struct rte_eth_dev *dev);
+int mlx5_verbs_max_prio(struct rte_eth_dev *dev);
+void mlx5_flow_print(struct rte_flow *flow);
 int mlx5_flow_validate(struct rte_eth_dev *dev,
   const struct rte_flow_attr *attr,
   const struct rte_flow_item items[],
diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index a45cb06e1..5e325be37 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -75,6 +75,58 @@ struct ibv_spec_header {
uint16_t size;
 };
 
+ /**
+  * Get the maximum number of priority available.
+  *
+  * @param[in] dev
+  *   Pointer to Ethernet device.
+  *
+  * @return
+  *   number of supported Verbs flow priority on success, a negative errno
+  *   value otherwise and rte_errno is set.
+  */
+int
+mlx5_verbs_max_prio(struct rte_eth_dev *dev)
+{
+   struct {
+   struct ibv_flow_attr attr;
+   struct ibv_flow_spec_eth eth;
+   struct ibv_flow_spec_action_drop drop;
+   } flow_attr = {
+   .attr = {
+   .num_of_specs = 2,
+   },
+   .eth = {
+  

[dpdk-dev] [PATCH v3 03/21] net/mlx5: replace verbs priorities by flow

2018-07-11 Thread Nelio Laranjeiro
Previous work introduce verbs priorities, whereas the PMD is making
translation between Flow priority into Verbs.  Rename this to make more
sense on what the PMD has to translate.

Signed-off-by: Nelio Laranjeiro 
Acked-by: Yongseok Koh 
---
 drivers/net/mlx5/mlx5.c  | 15 ---
 drivers/net/mlx5/mlx5.h  |  4 ++--
 drivers/net/mlx5/mlx5_flow.c | 24 
 3 files changed, 18 insertions(+), 25 deletions(-)

diff --git a/drivers/net/mlx5/mlx5.c b/drivers/net/mlx5/mlx5.c
index e9780ac8f..74248f098 100644
--- a/drivers/net/mlx5/mlx5.c
+++ b/drivers/net/mlx5/mlx5.c
@@ -717,7 +717,6 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
unsigned int tunnel_en = 0;
unsigned int mpls_en = 0;
unsigned int swp = 0;
-   unsigned int verb_priorities = 0;
unsigned int mprq = 0;
unsigned int mprq_min_stride_size_n = 0;
unsigned int mprq_max_stride_size_n = 0;
@@ -1139,16 +1138,10 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
/* Store device configuration on private structure. */
priv->config = config;
/* Supported Verbs flow priority number detection. */
-   if (verb_priorities == 0) {
-   err = mlx5_verbs_max_prio(eth_dev);
-   if (err < 0) {
-   DRV_LOG(ERR, "port %u wrong Verbs flow priorities",
-   eth_dev->data->port_id);
-   goto error;
-   }
-   verb_priorities = err;
-   }
-   priv->config.max_verbs_prio = verb_priorities;
+   err = mlx5_flow_discover_priorities(eth_dev);
+   if (err < 0)
+   goto error;
+   priv->config.flow_prio = err;
/*
 * Once the device is added to the list of memory event
 * callback, its global MR cache table cannot be expanded
diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h
index 227429848..9949cd3fa 100644
--- a/drivers/net/mlx5/mlx5.h
+++ b/drivers/net/mlx5/mlx5.h
@@ -122,7 +122,7 @@ struct mlx5_dev_config {
unsigned int min_rxqs_num;
/* Rx queue count threshold to enable MPRQ. */
} mprq; /* Configurations for Multi-Packet RQ. */
-   unsigned int max_verbs_prio; /* Number of Verb flow priorities. */
+   unsigned int flow_prio; /* Number of flow priorities. */
unsigned int tso_max_payload_sz; /* Maximum TCP payload for TSO. */
unsigned int ind_table_max_size; /* Maximum indirection table size. */
int txq_inline; /* Maximum packet size for inlining. */
@@ -317,7 +317,7 @@ int mlx5_traffic_restart(struct rte_eth_dev *dev);
 
 /* mlx5_flow.c */
 
-int mlx5_verbs_max_prio(struct rte_eth_dev *dev);
+int mlx5_flow_discover_priorities(struct rte_eth_dev *dev);
 void mlx5_flow_print(struct rte_flow *flow);
 int mlx5_flow_validate(struct rte_eth_dev *dev,
   const struct rte_flow_attr *attr,
diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index 5e325be37..8fdc6d7bb 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -76,17 +76,17 @@ struct ibv_spec_header {
 };
 
  /**
-  * Get the maximum number of priority available.
+  * Discover the maximum number of priority available.
   *
   * @param[in] dev
   *   Pointer to Ethernet device.
   *
   * @return
-  *   number of supported Verbs flow priority on success, a negative errno
-  *   value otherwise and rte_errno is set.
+  *   number of supported flow priority on success, a negative errno value
+  *   otherwise and rte_errno is set.
   */
 int
-mlx5_verbs_max_prio(struct rte_eth_dev *dev)
+mlx5_flow_discover_priorities(struct rte_eth_dev *dev)
 {
struct {
struct ibv_flow_attr attr;
@@ -106,25 +106,25 @@ mlx5_verbs_max_prio(struct rte_eth_dev *dev)
},
};
struct ibv_flow *flow;
-   uint32_t verb_priorities;
struct mlx5_hrxq *drop = mlx5_hrxq_drop_new(dev);
+   uint16_t vprio[] = { 8, 16 };
+   int i;
 
if (!drop) {
rte_errno = ENOTSUP;
return -rte_errno;
}
-   for (verb_priorities = 0; 1; verb_priorities++) {
-   flow_attr.attr.priority = verb_priorities;
-   flow = mlx5_glue->create_flow(drop->qp,
- &flow_attr.attr);
+   for (i = 0; i != RTE_DIM(vprio); i++) {
+   flow_attr.attr.priority = vprio[i] - 1;
+   flow = mlx5_glue->create_flow(drop->qp, &flow_attr.attr);
if (!flow)
break;
claim_zero(mlx5_glue->destroy_flow(flow));
}
mlx5_hrxq_drop_release(dev);
DRV_LOG(INFO, "port %u flow maximum priority: %d",
-   dev->data->port_id, verb_priorities);
-   return verb_priorities;
+   dev->data->port_id, vprio[i - 1]);
+   return vprio[i - 1];
 }
 
 /**
@@ -318,7 +318,7 @@ mlx5_ctrl_flow_vlan

[dpdk-dev] [PATCH v3 01/21] net/mlx5: remove flow support

2018-07-11 Thread Nelio Laranjeiro
This start a series to re-work the flow engine in mlx5 to easily support
flow conversion to Verbs or TC.  This is necessary to handle both regular
flows and representors flows.

As the full file needs to be clean-up to re-write all items/actions
processing, this patch starts to disable the regular code and only let the
PMD to start in isolated mode.

After this patch flow API will not be usable.

Signed-off-by: Nelio Laranjeiro 
Acked-by: Yongseok Koh 
---
 drivers/net/mlx5/mlx5_flow.c | 3095 +-
 drivers/net/mlx5/mlx5_rxtx.h |1 -
 2 files changed, 80 insertions(+), 3016 deletions(-)

diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index 45207d70e..a45cb06e1 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -31,2406 +31,49 @@
 #include "mlx5_prm.h"
 #include "mlx5_glue.h"
 
-/* Flow priority for control plane flows. */
-#define MLX5_CTRL_FLOW_PRIORITY 1
-
-/* Internet Protocol versions. */
-#define MLX5_IPV4 4
-#define MLX5_IPV6 6
-#define MLX5_GRE 47
-
-#ifndef HAVE_IBV_DEVICE_COUNTERS_SET_SUPPORT
-struct ibv_flow_spec_counter_action {
-   int dummy;
-};
-#endif
-
-/* Dev ops structure defined in mlx5.c */
-extern const struct eth_dev_ops mlx5_dev_ops;
-extern const struct eth_dev_ops mlx5_dev_ops_isolate;
-
-/** Structure give to the conversion functions. */
-struct mlx5_flow_data {
-   struct rte_eth_dev *dev; /** Ethernet device. */
-   struct mlx5_flow_parse *parser; /** Parser context. */
-   struct rte_flow_error *error; /** Error context. */
-};
-
-static int
-mlx5_flow_create_eth(const struct rte_flow_item *item,
-const void *default_mask,
-struct mlx5_flow_data *data);
-
-static int
-mlx5_flow_create_vlan(const struct rte_flow_item *item,
- const void *default_mask,
- struct mlx5_flow_data *data);
-
-static int
-mlx5_flow_create_ipv4(const struct rte_flow_item *item,
- const void *default_mask,
- struct mlx5_flow_data *data);
-
-static int
-mlx5_flow_create_ipv6(const struct rte_flow_item *item,
- const void *default_mask,
- struct mlx5_flow_data *data);
-
-static int
-mlx5_flow_create_udp(const struct rte_flow_item *item,
-const void *default_mask,
-struct mlx5_flow_data *data);
-
-static int
-mlx5_flow_create_tcp(const struct rte_flow_item *item,
-const void *default_mask,
-struct mlx5_flow_data *data);
-
-static int
-mlx5_flow_create_vxlan(const struct rte_flow_item *item,
-  const void *default_mask,
-  struct mlx5_flow_data *data);
-
-static int
-mlx5_flow_create_vxlan_gpe(const struct rte_flow_item *item,
-  const void *default_mask,
-  struct mlx5_flow_data *data);
-
-static int
-mlx5_flow_create_gre(const struct rte_flow_item *item,
-const void *default_mask,
-struct mlx5_flow_data *data);
-
-static int
-mlx5_flow_create_mpls(const struct rte_flow_item *item,
- const void *default_mask,
- struct mlx5_flow_data *data);
-
-struct mlx5_flow_parse;
-
-static void
-mlx5_flow_create_copy(struct mlx5_flow_parse *parser, void *src,
- unsigned int size);
-
-static int
-mlx5_flow_create_flag_mark(struct mlx5_flow_parse *parser, uint32_t mark_id);
-
-static int
-mlx5_flow_create_count(struct rte_eth_dev *dev, struct mlx5_flow_parse 
*parser);
-
-/* Hash RX queue types. */
-enum hash_rxq_type {
-   HASH_RXQ_TCPV4,
-   HASH_RXQ_UDPV4,
-   HASH_RXQ_IPV4,
-   HASH_RXQ_TCPV6,
-   HASH_RXQ_UDPV6,
-   HASH_RXQ_IPV6,
-   HASH_RXQ_ETH,
-   HASH_RXQ_TUNNEL,
-};
-
-/* Initialization data for hash RX queue. */
-struct hash_rxq_init {
-   uint64_t hash_fields; /* Fields that participate in the hash. */
-   uint64_t dpdk_rss_hf; /* Matching DPDK RSS hash fields. */
-   unsigned int flow_priority; /* Flow priority to use. */
-   unsigned int ip_version; /* Internet protocol. */
-};
-
-/* Initialization data for hash RX queues. */
-const struct hash_rxq_init hash_rxq_init[] = {
-   [HASH_RXQ_TCPV4] = {
-   .hash_fields = (IBV_RX_HASH_SRC_IPV4 |
-   IBV_RX_HASH_DST_IPV4 |
-   IBV_RX_HASH_SRC_PORT_TCP |
-   IBV_RX_HASH_DST_PORT_TCP),
-   .dpdk_rss_hf = ETH_RSS_NONFRAG_IPV4_TCP,
-   .flow_priority = 0,
-   .ip_version = MLX5_IPV4,
-   },
-   [HASH_RXQ_UDPV4] = {
-   .hash_fields = (IBV_RX_HASH_SRC_IPV4 |
-   IBV_RX_HASH_DST_IPV4 |
-   IBV_RX_HASH_SRC_PORT_UDP |
-   IBV_RX_HASH_DST_PORT_UDP),
-   .dpd

[dpdk-dev] [PATCH v3 04/21] net/mlx5: support flow Ethernet item along with drop action

2018-07-11 Thread Nelio Laranjeiro
Signed-off-by: Nelio Laranjeiro 
---
 drivers/net/mlx5/mlx5.c  |   1 +
 drivers/net/mlx5/mlx5_flow.c | 664 +--
 2 files changed, 627 insertions(+), 38 deletions(-)

diff --git a/drivers/net/mlx5/mlx5.c b/drivers/net/mlx5/mlx5.c
index 74248f098..6d3421fae 100644
--- a/drivers/net/mlx5/mlx5.c
+++ b/drivers/net/mlx5/mlx5.c
@@ -242,6 +242,7 @@ mlx5_dev_close(struct rte_eth_dev *dev)
/* In case mlx5_dev_stop() has not been called. */
mlx5_dev_interrupt_handler_uninstall(dev);
mlx5_traffic_disable(dev);
+   mlx5_flow_flush(dev, NULL);
/* Prevent crashes when queues are still in use. */
dev->rx_pkt_burst = removed_rx_burst;
dev->tx_pkt_burst = removed_tx_burst;
diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index 8fdc6d7bb..036a8d440 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -35,11 +35,50 @@
 extern const struct eth_dev_ops mlx5_dev_ops;
 extern const struct eth_dev_ops mlx5_dev_ops_isolate;
 
+/* Pattern Layer bits. */
+#define MLX5_FLOW_LAYER_OUTER_L2 (1u << 0)
+#define MLX5_FLOW_LAYER_OUTER_L3_IPV4 (1u << 1)
+#define MLX5_FLOW_LAYER_OUTER_L3_IPV6 (1u << 2)
+#define MLX5_FLOW_LAYER_OUTER_L4_UDP (1u << 3)
+#define MLX5_FLOW_LAYER_OUTER_L4_TCP (1u << 4)
+#define MLX5_FLOW_LAYER_OUTER_VLAN (1u << 5)
+/* Masks. */
+#define MLX5_FLOW_LAYER_OUTER_L3 \
+   (MLX5_FLOW_LAYER_OUTER_L3_IPV4 | MLX5_FLOW_LAYER_OUTER_L3_IPV6)
+#define MLX5_FLOW_LAYER_OUTER_L4 \
+   (MLX5_FLOW_LAYER_OUTER_L4_UDP | MLX5_FLOW_LAYER_OUTER_L4_TCP)
+
+/* Actions that modify the fate of matching traffic. */
+#define MLX5_FLOW_FATE_DROP (1u << 0)
+
+/** Handles information leading to a drop fate. */
+struct mlx5_flow_verbs {
+   unsigned int size; /**< Size of the attribute. */
+   struct {
+   struct ibv_flow_attr *attr;
+   /**< Pointer to the Specification buffer. */
+   uint8_t *specs; /**< Pointer to the specifications. */
+   };
+   struct ibv_flow *flow; /**< Verbs flow pointer. */
+   struct mlx5_hrxq *hrxq; /**< Hash Rx queue object. */
+};
+
+/* Flow structure. */
 struct rte_flow {
TAILQ_ENTRY(rte_flow) next; /**< Pointer to the next flow structure. */
+   struct rte_flow_attr attributes; /**< User flow attribute. */
+   uint32_t layers;
+   /**< Bit-fields of present layers see MLX5_FLOW_LAYER_*. */
+   uint32_t fate;
+   /**< Bit-fields of present fate see MLX5_FLOW_FATE_*. */
+   struct mlx5_flow_verbs verbs; /* Verbs drop flow. */
 };
 
 static const struct rte_flow_ops mlx5_flow_ops = {
+   .validate = mlx5_flow_validate,
+   .create = mlx5_flow_create,
+   .destroy = mlx5_flow_destroy,
+   .flush = mlx5_flow_flush,
.isolate = mlx5_flow_isolate,
 };
 
@@ -128,13 +167,415 @@ mlx5_flow_discover_priorities(struct rte_eth_dev *dev)
 }
 
 /**
- * Convert a flow.
+ * Verify the @p attributes will be correctly understood by the NIC and store
+ * them in the @p flow if everything is correct.
  *
- * @param dev
+ * @param[in] dev
  *   Pointer to Ethernet device.
- * @param list
- *   Pointer to a TAILQ flow list.
- * @param[in] attr
+ * @param[in] attributes
+ *   Pointer to flow attributes
+ * @param[in, out] flow
+ *   Pointer to the rte_flow structure.
+ * @param[out] error
+ *   Pointer to error structure.
+ *
+ * @return
+ *   0 on success, a negative errno value otherwise and rte_errno is set.
+ */
+static int
+mlx5_flow_attributes(struct rte_eth_dev *dev,
+const struct rte_flow_attr *attributes,
+struct rte_flow *flow,
+struct rte_flow_error *error)
+{
+   uint32_t priority_max =
+   ((struct priv *)dev->data->dev_private)->config.flow_prio;
+
+   if (attributes->group)
+   return rte_flow_error_set(error, ENOTSUP,
+ RTE_FLOW_ERROR_TYPE_ATTR_GROUP,
+ NULL,
+ "groups is not supported");
+   if (attributes->priority >= priority_max)
+   return rte_flow_error_set(error, ENOTSUP,
+ RTE_FLOW_ERROR_TYPE_ATTR_PRIORITY,
+ NULL,
+ "priority out of range");
+   if (attributes->egress)
+   return rte_flow_error_set(error, ENOTSUP,
+ RTE_FLOW_ERROR_TYPE_ATTR_EGRESS,
+ NULL,
+ "egress is not supported");
+   if (attributes->transfer)
+   return rte_flow_error_set(error, ENOTSUP,
+ RTE_FLOW_ERROR_TYPE_ATTR_TRANSFER,
+ NULL,
+ "transfer is not supported");
+   if (!attri

[dpdk-dev] [PATCH v3 07/21] net/mlx5: add flow VLAN item

2018-07-11 Thread Nelio Laranjeiro
Signed-off-by: Nelio Laranjeiro 
---
 drivers/net/mlx5/mlx5_flow.c | 127 +++
 1 file changed, 127 insertions(+)

diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index 77f1bd5cc..659979283 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -382,6 +382,130 @@ mlx5_flow_item_eth(const struct rte_flow_item *item, 
struct rte_flow *flow,
return size;
 }
 
+/**
+ * Update the VLAN tag in the Verbs Ethernet specification.
+ *
+ * @param[in, out] attr
+ *   Pointer to Verbs attributes structure.
+ * @param[in] eth
+ *   Verbs structure containing the VLAN information to copy.
+ */
+static void
+mlx5_flow_item_vlan_update(struct ibv_flow_attr *attr,
+  struct ibv_flow_spec_eth *eth)
+{
+   unsigned int i;
+   enum ibv_flow_spec_type search = IBV_FLOW_SPEC_ETH;
+   struct ibv_spec_header *hdr = (struct ibv_spec_header *)
+   ((uint8_t *)attr + sizeof(struct ibv_flow_attr));
+
+   for (i = 0; i != attr->num_of_specs; ++i) {
+   if (hdr->type == search) {
+   struct ibv_flow_spec_eth *e =
+   (struct ibv_flow_spec_eth *)hdr;
+
+   e->val.vlan_tag = eth->val.vlan_tag;
+   e->mask.vlan_tag = eth->mask.vlan_tag;
+   e->val.ether_type = eth->val.ether_type;
+   e->mask.ether_type = eth->mask.ether_type;
+   break;
+   }
+   hdr = (struct ibv_spec_header *)((uint8_t *)hdr + hdr->size);
+   }
+}
+
+/**
+ * Convert the @p item into @p flow (or by updating the already present
+ * Ethernet Verbs) specification after ensuring the NIC will understand and
+ * process it correctly.
+ * If the necessary size for the conversion is greater than the @p flow_size,
+ * nothing is written in @p flow, the validation is still performed.
+ *
+ * @param[in] item
+ *   Item specification.
+ * @param[in, out] flow
+ *   Pointer to flow structure.
+ * @param[in] flow_size
+ *   Size in bytes of the available space in @p flow, if too small, nothing is
+ *   written.
+ * @param[out] error
+ *   Pointer to error structure.
+ *
+ * @return
+ *   On success the number of bytes consumed/necessary, if the returned value
+ *   is lesser or equal to @p flow_size, the @p item has fully been converted,
+ *   otherwise another call with this returned memory size should be done.
+ *   On error, a negative errno value is returned and rte_errno is set.
+ */
+static int
+mlx5_flow_item_vlan(const struct rte_flow_item *item, struct rte_flow *flow,
+   const size_t flow_size, struct rte_flow_error *error)
+{
+   const struct rte_flow_item_vlan *spec = item->spec;
+   const struct rte_flow_item_vlan *mask = item->mask;
+   const struct rte_flow_item_vlan nic_mask = {
+   .tci = RTE_BE16(0x0fff),
+   .inner_type = RTE_BE16(0x),
+   };
+   unsigned int size = sizeof(struct ibv_flow_spec_eth);
+   struct ibv_flow_spec_eth eth = {
+   .type = IBV_FLOW_SPEC_ETH,
+   .size = size,
+   };
+   int ret;
+   const uint32_t l34m = MLX5_FLOW_LAYER_OUTER_L3 |
+   MLX5_FLOW_LAYER_OUTER_L4;
+   const uint32_t vlanm = MLX5_FLOW_LAYER_OUTER_VLAN;
+   const uint32_t l2m = MLX5_FLOW_LAYER_OUTER_L2;
+
+   if (flow->layers & vlanm)
+   return rte_flow_error_set(error, ENOTSUP,
+ RTE_FLOW_ERROR_TYPE_ITEM,
+ item,
+ "VLAN layer already configured");
+   else if ((flow->layers & l34m) != 0)
+   return rte_flow_error_set(error, ENOTSUP,
+ RTE_FLOW_ERROR_TYPE_ITEM,
+ item,
+ "L2 layer cannot follow L3/L4 layer");
+   if (!mask)
+   mask = &rte_flow_item_vlan_mask;
+   ret = mlx5_flow_item_acceptable
+   (item, (const uint8_t *)mask,
+(const uint8_t *)&nic_mask,
+sizeof(struct rte_flow_item_vlan), error);
+   if (ret)
+   return ret;
+   if (spec) {
+   eth.val.vlan_tag = spec->tci;
+   eth.mask.vlan_tag = mask->tci;
+   eth.val.vlan_tag &= eth.mask.vlan_tag;
+   eth.val.ether_type = spec->inner_type;
+   eth.mask.ether_type = mask->inner_type;
+   eth.val.ether_type &= eth.mask.ether_type;
+   }
+   /*
+* From verbs perspective an empty VLAN is equivalent
+* to a packet without VLAN layer.
+*/
+   if (!eth.mask.vlan_tag)
+   return rte_flow_error_set(error, EINVAL,
+ RTE_FLOW_ERROR_TYPE_ITEM_SPEC,
+   

[dpdk-dev] [PATCH v3 05/21] net/mlx5: add flow queue action

2018-07-11 Thread Nelio Laranjeiro
Signed-off-by: Nelio Laranjeiro 
Acked-by: Yongseok Koh 
---
 drivers/net/mlx5/mlx5_flow.c | 97 
 1 file changed, 86 insertions(+), 11 deletions(-)

diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index 036a8d440..6041a4573 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -50,6 +50,7 @@ extern const struct eth_dev_ops mlx5_dev_ops_isolate;
 
 /* Actions that modify the fate of matching traffic. */
 #define MLX5_FLOW_FATE_DROP (1u << 0)
+#define MLX5_FLOW_FATE_QUEUE (1u << 1)
 
 /** Handles information leading to a drop fate. */
 struct mlx5_flow_verbs {
@@ -71,7 +72,8 @@ struct rte_flow {
/**< Bit-fields of present layers see MLX5_FLOW_LAYER_*. */
uint32_t fate;
/**< Bit-fields of present fate see MLX5_FLOW_FATE_*. */
-   struct mlx5_flow_verbs verbs; /* Verbs drop flow. */
+   struct mlx5_flow_verbs verbs; /* Verbs flow. */
+   uint16_t queue; /**< Destination queue to redirect traffic to. */
 };
 
 static const struct rte_flow_ops mlx5_flow_ops = {
@@ -492,6 +494,52 @@ mlx5_flow_action_drop(const struct rte_flow_action *action,
return size;
 }
 
+/**
+ * Convert the @p action into @p flow after ensuring the NIC will understand
+ * and process it correctly.
+ *
+ * @param[in] dev
+ *   Pointer to Ethernet device structure.
+ * @param[in] action
+ *   Action configuration.
+ * @param[in, out] flow
+ *   Pointer to flow structure.
+ * @param[out] error
+ *   Pointer to error structure.
+ *
+ * @return
+ *   0 on success, a negative errno value otherwise and rte_errno is set.
+ */
+static int
+mlx5_flow_action_queue(struct rte_eth_dev *dev,
+  const struct rte_flow_action *action,
+  struct rte_flow *flow,
+  struct rte_flow_error *error)
+{
+   struct priv *priv = dev->data->dev_private;
+   const struct rte_flow_action_queue *queue = action->conf;
+
+   if (flow->fate)
+   return rte_flow_error_set(error, ENOTSUP,
+ RTE_FLOW_ERROR_TYPE_ACTION,
+ action,
+ "multiple fate actions are not"
+ " supported");
+   if (queue->index >= priv->rxqs_n)
+   return rte_flow_error_set(error, EINVAL,
+ RTE_FLOW_ERROR_TYPE_ACTION_CONF,
+ &queue->index,
+ "queue index out of range");
+   if (!(*priv->rxqs)[queue->index])
+   return rte_flow_error_set(error, EINVAL,
+ RTE_FLOW_ERROR_TYPE_ACTION_CONF,
+ &queue->index,
+ "queue is not configured");
+   flow->queue = queue->index;
+   flow->fate |= MLX5_FLOW_FATE_QUEUE;
+   return 0;
+}
+
 /**
  * Convert the @p action into @p flow after ensuring the NIC will understand
  * and process it correctly.
@@ -520,7 +568,7 @@ mlx5_flow_action_drop(const struct rte_flow_action *action,
  *   On error, a negative errno value is returned and rte_errno is set.
  */
 static int
-mlx5_flow_actions(struct rte_eth_dev *dev __rte_unused,
+mlx5_flow_actions(struct rte_eth_dev *dev,
  const struct rte_flow_action actions[],
  struct rte_flow *flow, const size_t flow_size,
  struct rte_flow_error *error)
@@ -537,6 +585,9 @@ mlx5_flow_actions(struct rte_eth_dev *dev __rte_unused,
ret = mlx5_flow_action_drop(actions, flow, remain,
error);
break;
+   case RTE_FLOW_ACTION_TYPE_QUEUE:
+   ret = mlx5_flow_action_queue(dev, actions, flow, error);
+   break;
default:
return rte_flow_error_set(error, ENOTSUP,
  RTE_FLOW_ERROR_TYPE_ACTION,
@@ -661,7 +712,10 @@ mlx5_flow_remove(struct rte_eth_dev *dev, struct rte_flow 
*flow)
}
}
if (flow->verbs.hrxq) {
-   mlx5_hrxq_drop_release(dev);
+   if (flow->fate & MLX5_FLOW_FATE_DROP)
+   mlx5_hrxq_drop_release(dev);
+   else if (flow->fate & MLX5_FLOW_FATE_QUEUE)
+   mlx5_hrxq_release(dev, flow->verbs.hrxq);
flow->verbs.hrxq = NULL;
}
 }
@@ -683,17 +737,38 @@ static int
 mlx5_flow_apply(struct rte_eth_dev *dev, struct rte_flow *flow,
struct rte_flow_error *error)
 {
-   flow->verbs.hrxq = mlx5_hrxq_drop_new(dev);
-   if (!flow->verbs.hrxq)
-   return rte_flow_error_set
-   (error, errno,
-RTE_FLOW_ERROR_TYPE_UNSPECIFIED,
-

[dpdk-dev] [PATCH v3 06/21] net/mlx5: add flow stop/start

2018-07-11 Thread Nelio Laranjeiro
Signed-off-by: Nelio Laranjeiro 
Acked-by: Yongseok Koh 
---
 drivers/net/mlx5/mlx5_flow.c | 24 
 1 file changed, 20 insertions(+), 4 deletions(-)

diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index 6041a4573..77f1bd5cc 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -908,9 +908,12 @@ mlx5_flow_list_flush(struct rte_eth_dev *dev, struct 
mlx5_flows *list)
  *   Pointer to a TAILQ flow list.
  */
 void
-mlx5_flow_stop(struct rte_eth_dev *dev __rte_unused,
-  struct mlx5_flows *list __rte_unused)
+mlx5_flow_stop(struct rte_eth_dev *dev, struct mlx5_flows *list)
 {
+   struct rte_flow *flow;
+
+   TAILQ_FOREACH_REVERSE(flow, list, mlx5_flows, next)
+   mlx5_flow_remove(dev, flow);
 }
 
 /**
@@ -925,10 +928,23 @@ mlx5_flow_stop(struct rte_eth_dev *dev __rte_unused,
  *   0 on success, a negative errno value otherwise and rte_errno is set.
  */
 int
-mlx5_flow_start(struct rte_eth_dev *dev __rte_unused,
-   struct mlx5_flows *list __rte_unused)
+mlx5_flow_start(struct rte_eth_dev *dev, struct mlx5_flows *list)
 {
+   struct rte_flow *flow;
+   struct rte_flow_error error;
+   int ret = 0;
+
+   TAILQ_FOREACH(flow, list, next) {
+   ret = mlx5_flow_apply(dev, flow, &error);
+   if (ret < 0)
+   goto error;
+   }
return 0;
+error:
+   ret = rte_errno; /* Save rte_errno before cleanup. */
+   mlx5_flow_stop(dev, list);
+   rte_errno = ret; /* Restore rte_errno. */
+   return -rte_errno;
 }
 
 /**
-- 
2.18.0



[dpdk-dev] [PATCH v3 10/21] net/mlx5: add flow UDP item

2018-07-11 Thread Nelio Laranjeiro
Signed-off-by: Nelio Laranjeiro 
Acked-by: Yongseok Koh 
---
 drivers/net/mlx5/mlx5_flow.c | 97 +---
 1 file changed, 91 insertions(+), 6 deletions(-)

diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index 513f70d40..0096ed8a2 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -52,6 +52,9 @@ extern const struct eth_dev_ops mlx5_dev_ops_isolate;
 #define MLX5_FLOW_FATE_DROP (1u << 0)
 #define MLX5_FLOW_FATE_QUEUE (1u << 1)
 
+/* possible L3 layers protocols filtering. */
+#define MLX5_IP_PROTOCOL_UDP 17
+
 /** Handles information leading to a drop fate. */
 struct mlx5_flow_verbs {
unsigned int size; /**< Size of the attribute. */
@@ -68,10 +71,12 @@ struct mlx5_flow_verbs {
 struct rte_flow {
TAILQ_ENTRY(rte_flow) next; /**< Pointer to the next flow structure. */
struct rte_flow_attr attributes; /**< User flow attribute. */
+   uint32_t l3_protocol_en:1; /**< Protocol filtering requested. */
uint32_t layers;
/**< Bit-fields of present layers see MLX5_FLOW_LAYER_*. */
uint32_t fate;
/**< Bit-fields of present fate see MLX5_FLOW_FATE_*. */
+   uint8_t l3_protocol; /**< valid when l3_protocol_en is set. */
struct mlx5_flow_verbs verbs; /* Verbs flow. */
uint16_t queue; /**< Destination queue to redirect traffic to. */
 };
@@ -568,8 +573,6 @@ mlx5_flow_item_ipv4(const struct rte_flow_item *item, 
struct rte_flow *flow,
if (ret < 0)
return ret;
flow->layers |= MLX5_FLOW_LAYER_OUTER_L3_IPV4;
-   if (size > flow_size)
-   return size;
if (spec) {
ipv4.val = (struct ibv_flow_ipv4_ext_filter){
.src_ip = spec->hdr.src_addr,
@@ -589,7 +592,10 @@ mlx5_flow_item_ipv4(const struct rte_flow_item *item, 
struct rte_flow *flow,
ipv4.val.proto &= ipv4.mask.proto;
ipv4.val.tos &= ipv4.mask.tos;
}
-   mlx5_flow_spec_verbs_add(flow, &ipv4, size);
+   flow->l3_protocol_en = !!ipv4.mask.proto;
+   flow->l3_protocol = ipv4.val.proto;
+   if (size <= flow_size)
+   mlx5_flow_spec_verbs_add(flow, &ipv4, size);
return size;
 }
 
@@ -660,8 +666,6 @@ mlx5_flow_item_ipv6(const struct rte_flow_item *item, 
struct rte_flow *flow,
if (ret < 0)
return ret;
flow->layers |= MLX5_FLOW_LAYER_OUTER_L3_IPV6;
-   if (size > flow_size)
-   return size;
if (spec) {
unsigned int i;
uint32_t vtc_flow_val;
@@ -701,7 +705,85 @@ mlx5_flow_item_ipv6(const struct rte_flow_item *item, 
struct rte_flow *flow,
ipv6.val.next_hdr &= ipv6.mask.next_hdr;
ipv6.val.hop_limit &= ipv6.mask.hop_limit;
}
-   mlx5_flow_spec_verbs_add(flow, &ipv6, size);
+   flow->l3_protocol_en = !!ipv6.mask.next_hdr;
+   flow->l3_protocol = ipv6.val.next_hdr;
+   if (size <= flow_size)
+   mlx5_flow_spec_verbs_add(flow, &ipv6, size);
+   return size;
+}
+
+/**
+ * Convert the @p item into a Verbs specification after ensuring the NIC
+ * will understand and process it correctly.
+ * If the necessary size for the conversion is greater than the @p flow_size,
+ * nothing is written in @p flow, the validation is still performed.
+ *
+ * @param[in] item
+ *   Item specification.
+ * @param[in, out] flow
+ *   Pointer to flow structure.
+ * @param[in] flow_size
+ *   Size in bytes of the available space in @p flow, if too small, nothing is
+ *   written.
+ * @param[out] error
+ *   Pointer to error structure.
+ *
+ * @return
+ *   On success the number of bytes consumed/necessary, if the returned value
+ *   is lesser or equal to @p flow_size, the @p item has fully been converted,
+ *   otherwise another call with this returned memory size should be done.
+ *   On error, a negative errno value is returned and rte_errno is set.
+ */
+static int
+mlx5_flow_item_udp(const struct rte_flow_item *item, struct rte_flow *flow,
+  const size_t flow_size, struct rte_flow_error *error)
+{
+   const struct rte_flow_item_udp *spec = item->spec;
+   const struct rte_flow_item_udp *mask = item->mask;
+   unsigned int size = sizeof(struct ibv_flow_spec_tcp_udp);
+   struct ibv_flow_spec_tcp_udp udp = {
+   .type = IBV_FLOW_SPEC_UDP,
+   .size = size,
+   };
+   int ret;
+
+   if (!(flow->layers & MLX5_FLOW_LAYER_OUTER_L3))
+   return rte_flow_error_set(error, ENOTSUP,
+ RTE_FLOW_ERROR_TYPE_ITEM,
+ item,
+ "L3 is mandatory to filter on L4");
+   if (flow->layers & MLX5_FLOW_LAYER_OUTER_L4)
+   return rte_flow_error_set(error, ENOTSUP,
+ RTE_FLOW_ERROR_TYPE_ITEM,
+ 

[dpdk-dev] [PATCH v3 08/21] net/mlx5: add flow IPv4 item

2018-07-11 Thread Nelio Laranjeiro
Signed-off-by: Nelio Laranjeiro 
---
 drivers/net/mlx5/mlx5_flow.c | 90 
 1 file changed, 90 insertions(+)

diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index 659979283..c05b8498d 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -506,6 +506,93 @@ mlx5_flow_item_vlan(const struct rte_flow_item *item, 
struct rte_flow *flow,
return size;
 }
 
+/**
+ * Convert the @p item into a Verbs specification after ensuring the NIC
+ * will understand and process it correctly.
+ * If the necessary size for the conversion is greater than the @p flow_size,
+ * nothing is written in @p flow, the validation is still performed.
+ *
+ * @param[in] item
+ *   Item specification.
+ * @param[in, out] flow
+ *   Pointer to flow structure.
+ * @param[in] flow_size
+ *   Size in bytes of the available space in @p flow, if too small, nothing is
+ *   written.
+ * @param[out] error
+ *   Pointer to error structure.
+ *
+ * @return
+ *   On success the number of bytes consumed/necessary, if the returned value
+ *   is lesser or equal to @p flow_size, the @p item has fully been converted,
+ *   otherwise another call with this returned memory size should be done.
+ *   On error, a negative errno value is returned and rte_errno is set.
+ */
+static int
+mlx5_flow_item_ipv4(const struct rte_flow_item *item, struct rte_flow *flow,
+   const size_t flow_size, struct rte_flow_error *error)
+{
+   const struct rte_flow_item_ipv4 *spec = item->spec;
+   const struct rte_flow_item_ipv4 *mask = item->mask;
+   const struct rte_flow_item_ipv4 nic_mask = {
+   .hdr = {
+   .src_addr = RTE_BE32(0x),
+   .dst_addr = RTE_BE32(0x),
+   .type_of_service = 0xff,
+   .next_proto_id = 0xff,
+   },
+   };
+   unsigned int size = sizeof(struct ibv_flow_spec_ipv4_ext);
+   struct ibv_flow_spec_ipv4_ext ipv4 = {
+   .type = IBV_FLOW_SPEC_IPV4_EXT,
+   .size = size,
+   };
+   int ret;
+
+   if (flow->layers & MLX5_FLOW_LAYER_OUTER_L3)
+   return rte_flow_error_set(error, ENOTSUP,
+ RTE_FLOW_ERROR_TYPE_ITEM,
+ item,
+ "multiple L3 layers not supported");
+   else if (flow->layers & MLX5_FLOW_LAYER_OUTER_L4)
+   return rte_flow_error_set(error, ENOTSUP,
+ RTE_FLOW_ERROR_TYPE_ITEM,
+ item,
+ "L3 cannot follow an L4 layer.");
+   if (!mask)
+   mask = &rte_flow_item_ipv4_mask;
+   ret = mlx5_flow_item_acceptable
+   (item, (const uint8_t *)mask,
+(const uint8_t *)&nic_mask,
+sizeof(struct rte_flow_item_ipv4), error);
+   if (ret < 0)
+   return ret;
+   flow->layers |= MLX5_FLOW_LAYER_OUTER_L3_IPV4;
+   if (size > flow_size)
+   return size;
+   if (spec) {
+   ipv4.val = (struct ibv_flow_ipv4_ext_filter){
+   .src_ip = spec->hdr.src_addr,
+   .dst_ip = spec->hdr.dst_addr,
+   .proto = spec->hdr.next_proto_id,
+   .tos = spec->hdr.type_of_service,
+   };
+   ipv4.mask = (struct ibv_flow_ipv4_ext_filter){
+   .src_ip = mask->hdr.src_addr,
+   .dst_ip = mask->hdr.dst_addr,
+   .proto = mask->hdr.next_proto_id,
+   .tos = mask->hdr.type_of_service,
+   };
+   /* Remove unwanted bits from values. */
+   ipv4.val.src_ip &= ipv4.mask.src_ip;
+   ipv4.val.dst_ip &= ipv4.mask.dst_ip;
+   ipv4.val.proto &= ipv4.mask.proto;
+   ipv4.val.tos &= ipv4.mask.tos;
+   }
+   mlx5_flow_spec_verbs_add(flow, &ipv4, size);
+   return size;
+}
+
 /**
  * Convert the @p pattern into a Verbs specifications after ensuring the NIC
  * will understand and process it correctly.
@@ -551,6 +638,9 @@ mlx5_flow_items(const struct rte_flow_item pattern[],
case RTE_FLOW_ITEM_TYPE_VLAN:
ret = mlx5_flow_item_vlan(pattern, flow, remain, error);
break;
+   case RTE_FLOW_ITEM_TYPE_IPV4:
+   ret = mlx5_flow_item_ipv4(pattern, flow, remain, error);
+   break;
default:
return rte_flow_error_set(error, ENOTSUP,
  RTE_FLOW_ERROR_TYPE_ITEM,
-- 
2.18.0



[dpdk-dev] [PATCH v3 09/21] net/mlx5: add flow IPv6 item

2018-07-11 Thread Nelio Laranjeiro
Signed-off-by: Nelio Laranjeiro 
Acked-by: Yongseok Koh 
---
 drivers/net/mlx5/mlx5_flow.c | 115 +++
 1 file changed, 115 insertions(+)

diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index c05b8498d..513f70d40 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -593,6 +593,118 @@ mlx5_flow_item_ipv4(const struct rte_flow_item *item, 
struct rte_flow *flow,
return size;
 }
 
+/**
+ * Convert the @p item into a Verbs specification after ensuring the NIC
+ * will understand and process it correctly.
+ * If the necessary size for the conversion is greater than the @p flow_size,
+ * nothing is written in @p flow, the validation is still performed.
+ *
+ * @param[in] item
+ *   Item specification.
+ * @param[in, out] flow
+ *   Pointer to flow structure.
+ * @param[in] flow_size
+ *   Size in bytes of the available space in @p flow, if too small, nothing is
+ *   written.
+ * @param[out] error
+ *   Pointer to error structure.
+ *
+ * @return
+ *   On success the number of bytes consumed/necessary, if the returned value
+ *   is lesser or equal to @p flow_size, the @p item has fully been converted,
+ *   otherwise another call with this returned memory size should be done.
+ *   On error, a negative errno value is returned and rte_errno is set.
+ */
+static int
+mlx5_flow_item_ipv6(const struct rte_flow_item *item, struct rte_flow *flow,
+   const size_t flow_size, struct rte_flow_error *error)
+{
+   const struct rte_flow_item_ipv6 *spec = item->spec;
+   const struct rte_flow_item_ipv6 *mask = item->mask;
+   const struct rte_flow_item_ipv6 nic_mask = {
+   .hdr = {
+   .src_addr =
+   "\xff\xff\xff\xff\xff\xff\xff\xff"
+   "\xff\xff\xff\xff\xff\xff\xff\xff",
+   .dst_addr =
+   "\xff\xff\xff\xff\xff\xff\xff\xff"
+   "\xff\xff\xff\xff\xff\xff\xff\xff",
+   .vtc_flow = RTE_BE32(0x),
+   .proto = 0xff,
+   .hop_limits = 0xff,
+   },
+   };
+   unsigned int size = sizeof(struct ibv_flow_spec_ipv6);
+   struct ibv_flow_spec_ipv6 ipv6 = {
+   .type = IBV_FLOW_SPEC_IPV6,
+   .size = size,
+   };
+   int ret;
+
+   if (flow->layers & MLX5_FLOW_LAYER_OUTER_L3)
+   return rte_flow_error_set(error, ENOTSUP,
+ RTE_FLOW_ERROR_TYPE_ITEM,
+ item,
+ "multiple L3 layers not supported");
+   else if (flow->layers & MLX5_FLOW_LAYER_OUTER_L4)
+   return rte_flow_error_set(error, ENOTSUP,
+ RTE_FLOW_ERROR_TYPE_ITEM,
+ item,
+ "L3 cannot follow an L4 layer.");
+   if (!mask)
+   mask = &rte_flow_item_ipv6_mask;
+   ret = mlx5_flow_item_acceptable
+   (item, (const uint8_t *)mask,
+(const uint8_t *)&nic_mask,
+sizeof(struct rte_flow_item_ipv6), error);
+   if (ret < 0)
+   return ret;
+   flow->layers |= MLX5_FLOW_LAYER_OUTER_L3_IPV6;
+   if (size > flow_size)
+   return size;
+   if (spec) {
+   unsigned int i;
+   uint32_t vtc_flow_val;
+   uint32_t vtc_flow_mask;
+
+   memcpy(&ipv6.val.src_ip, spec->hdr.src_addr,
+  RTE_DIM(ipv6.val.src_ip));
+   memcpy(&ipv6.val.dst_ip, spec->hdr.dst_addr,
+  RTE_DIM(ipv6.val.dst_ip));
+   memcpy(&ipv6.mask.src_ip, mask->hdr.src_addr,
+  RTE_DIM(ipv6.mask.src_ip));
+   memcpy(&ipv6.mask.dst_ip, mask->hdr.dst_addr,
+  RTE_DIM(ipv6.mask.dst_ip));
+   vtc_flow_val = rte_be_to_cpu_32(spec->hdr.vtc_flow);
+   vtc_flow_mask = rte_be_to_cpu_32(mask->hdr.vtc_flow);
+   ipv6.val.flow_label =
+   rte_cpu_to_be_32((vtc_flow_val & IPV6_HDR_FL_MASK) >>
+IPV6_HDR_FL_SHIFT);
+   ipv6.val.traffic_class = (vtc_flow_val & IPV6_HDR_TC_MASK) >>
+IPV6_HDR_TC_SHIFT;
+   ipv6.val.next_hdr = spec->hdr.proto;
+   ipv6.val.hop_limit = spec->hdr.hop_limits;
+   ipv6.mask.flow_label =
+   rte_cpu_to_be_32((vtc_flow_mask & IPV6_HDR_FL_MASK) >>
+IPV6_HDR_FL_SHIFT);
+   ipv6.mask.traffic_class = (vtc_flow_mask & IPV6_HDR_TC_MASK) >>
+ IPV6_HDR_TC_SHIFT;
+   ipv6.mask.next_hdr = mask->hdr.pr

[dpdk-dev] [PATCH v3 12/21] net/mlx5: add mark/flag flow action

2018-07-11 Thread Nelio Laranjeiro
Signed-off-by: Nelio Laranjeiro 
---
 drivers/net/mlx5/mlx5_flow.c | 252 +++
 drivers/net/mlx5/mlx5_rxtx.h |   1 +
 2 files changed, 253 insertions(+)

diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index f646eee01..1280db486 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -52,6 +52,10 @@ extern const struct eth_dev_ops mlx5_dev_ops_isolate;
 #define MLX5_FLOW_FATE_DROP (1u << 0)
 #define MLX5_FLOW_FATE_QUEUE (1u << 1)
 
+/* Modify a packet. */
+#define MLX5_FLOW_MOD_FLAG (1u << 0)
+#define MLX5_FLOW_MOD_MARK (1u << 1)
+
 /* possible L3 layers protocols filtering. */
 #define MLX5_IP_PROTOCOL_TCP 6
 #define MLX5_IP_PROTOCOL_UDP 17
@@ -75,6 +79,8 @@ struct rte_flow {
uint32_t l3_protocol_en:1; /**< Protocol filtering requested. */
uint32_t layers;
/**< Bit-fields of present layers see MLX5_FLOW_LAYER_*. */
+   uint32_t modifier;
+   /**< Bit-fields of present modifier see MLX5_FLOW_MOD_*. */
uint32_t fate;
/**< Bit-fields of present fate see MLX5_FLOW_FATE_*. */
uint8_t l3_protocol; /**< valid when l3_protocol_en is set. */
@@ -984,6 +990,12 @@ mlx5_flow_action_drop(const struct rte_flow_action *action,
  action,
  "multiple fate actions are not"
  " supported");
+   if (flow->modifier & (MLX5_FLOW_MOD_FLAG | MLX5_FLOW_MOD_MARK))
+   return rte_flow_error_set(error, ENOTSUP,
+ RTE_FLOW_ERROR_TYPE_ACTION,
+ action,
+ "drop is not compatible with"
+ " flag/mark action");
if (size < flow_size)
mlx5_flow_spec_verbs_add(flow, &drop, size);
flow->fate |= MLX5_FLOW_FATE_DROP;
@@ -1036,6 +1048,161 @@ mlx5_flow_action_queue(struct rte_eth_dev *dev,
return 0;
 }
 
+/**
+ * Convert the @p action into a Verbs specification after ensuring the NIC
+ * will understand and process it correctly.
+ * If the necessary size for the conversion is greater than the @p flow_size,
+ * nothing is written in @p flow, the validation is still performed.
+ *
+ * @param[in] action
+ *   Action configuration.
+ * @param[in, out] flow
+ *   Pointer to flow structure.
+ * @param[in] flow_size
+ *   Size in bytes of the available space in @p flow, if too small, nothing is
+ *   written.
+ * @param[out] error
+ *   Pointer to error structure.
+ *
+ * @return
+ *   On success the number of bytes consumed/necessary, if the returned value
+ *   is lesser or equal to @p flow_size, the @p action has fully been
+ *   converted, otherwise another call with this returned memory size should
+ *   be done.
+ *   On error, a negative errno value is returned and rte_errno is set.
+ */
+static int
+mlx5_flow_action_flag(const struct rte_flow_action *action,
+ struct rte_flow *flow, const size_t flow_size,
+ struct rte_flow_error *error)
+{
+   unsigned int size = sizeof(struct ibv_flow_spec_action_tag);
+   struct ibv_flow_spec_action_tag tag = {
+   .type = IBV_FLOW_SPEC_ACTION_TAG,
+   .size = size,
+   .tag_id = mlx5_flow_mark_set(MLX5_FLOW_MARK_DEFAULT),
+   };
+
+   if (flow->modifier & MLX5_FLOW_MOD_FLAG)
+   return rte_flow_error_set(error, ENOTSUP,
+ RTE_FLOW_ERROR_TYPE_ACTION,
+ action,
+ "flag action already present");
+   if (flow->fate & MLX5_FLOW_FATE_DROP)
+   return rte_flow_error_set(error, ENOTSUP,
+ RTE_FLOW_ERROR_TYPE_ACTION,
+ action,
+ "flag is not compatible with drop"
+ " action");
+   if (flow->modifier & MLX5_FLOW_MOD_MARK)
+   return 0;
+   flow->modifier |= MLX5_FLOW_MOD_FLAG;
+   if (size <= flow_size)
+   mlx5_flow_spec_verbs_add(flow, &tag, size);
+   return size;
+}
+
+/**
+ * Update verbs specification to modify the flag to mark.
+ *
+ * @param[in, out] flow
+ *   Pointer to the rte_flow structure.
+ * @param[in] mark_id
+ *   Mark identifier to replace the flag.
+ */
+static void
+mlx5_flow_verbs_mark_update(struct rte_flow *flow, uint32_t mark_id)
+{
+   struct ibv_spec_header *hdr;
+   int i;
+
+   /* Update Verbs specification. */
+   hdr = (struct ibv_spec_header *)flow->verbs.specs;
+   if (!hdr)
+   return;
+   for (i = 0; i != flow->verbs.attr->num_of_specs; ++i) {
+   if (hdr->type == IBV_FLOW_SPEC_ACTION_TAG) {
+   struct ibv_flow_spec_action_

[dpdk-dev] [PATCH v3 13/21] net/mlx5: use a macro for the RSS key size

2018-07-11 Thread Nelio Laranjeiro
ConnectX 4-5 support only 40 bytes of RSS key, using a compiled size
hash key is not necessary.

Signed-off-by: Nelio Laranjeiro 
---
 drivers/net/mlx5/mlx5_ethdev.c | 14 +++---
 drivers/net/mlx5/mlx5_flow.c   |  4 ++--
 drivers/net/mlx5/mlx5_prm.h|  3 +++
 drivers/net/mlx5/mlx5_rss.c|  7 ---
 drivers/net/mlx5/mlx5_rxq.c| 12 +++-
 drivers/net/mlx5/mlx5_rxtx.h   |  1 -
 6 files changed, 23 insertions(+), 18 deletions(-)

diff --git a/drivers/net/mlx5/mlx5_ethdev.c b/drivers/net/mlx5/mlx5_ethdev.c
index 05f66f7b6..6e44d5ff0 100644
--- a/drivers/net/mlx5/mlx5_ethdev.c
+++ b/drivers/net/mlx5/mlx5_ethdev.c
@@ -377,15 +377,15 @@ mlx5_dev_configure(struct rte_eth_dev *dev)
 
if (use_app_rss_key &&
(dev->data->dev_conf.rx_adv_conf.rss_conf.rss_key_len !=
-rss_hash_default_key_len)) {
-   DRV_LOG(ERR, "port %u RSS key len must be %zu Bytes long",
-   dev->data->port_id, rss_hash_default_key_len);
+MLX5_RSS_HASH_KEY_LEN)) {
+   DRV_LOG(ERR, "port %u RSS key len must be %s Bytes long",
+   dev->data->port_id, RTE_STR(MLX5_RSS_HASH_KEY_LEN));
rte_errno = EINVAL;
return -rte_errno;
}
priv->rss_conf.rss_key =
rte_realloc(priv->rss_conf.rss_key,
-   rss_hash_default_key_len, 0);
+   MLX5_RSS_HASH_KEY_LEN, 0);
if (!priv->rss_conf.rss_key) {
DRV_LOG(ERR, "port %u cannot allocate RSS hash key memory (%u)",
dev->data->port_id, rxqs_n);
@@ -396,8 +396,8 @@ mlx5_dev_configure(struct rte_eth_dev *dev)
   use_app_rss_key ?
   dev->data->dev_conf.rx_adv_conf.rss_conf.rss_key :
   rss_hash_default_key,
-  rss_hash_default_key_len);
-   priv->rss_conf.rss_key_len = rss_hash_default_key_len;
+  MLX5_RSS_HASH_KEY_LEN);
+   priv->rss_conf.rss_key_len = MLX5_RSS_HASH_KEY_LEN;
priv->rss_conf.rss_hf = dev->data->dev_conf.rx_adv_conf.rss_conf.rss_hf;
priv->rxqs = (void *)dev->data->rx_queues;
priv->txqs = (void *)dev->data->tx_queues;
@@ -515,7 +515,7 @@ mlx5_dev_infos_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *info)
info->if_index = if_nametoindex(ifname);
info->reta_size = priv->reta_idx_n ?
priv->reta_idx_n : config->ind_table_max_size;
-   info->hash_key_size = rss_hash_default_key_len;
+   info->hash_key_size = MLX5_RSS_HASH_KEY_LEN;
info->speed_capa = priv->link_speed_capa;
info->flow_type_rss_offloads = ~MLX5_RSS_HF_MASK;
mlx5_set_default_params(dev, info);
diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index 1280db486..77483bd1f 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -1493,11 +1493,11 @@ mlx5_flow_apply(struct rte_eth_dev *dev, struct 
rte_flow *flow,
struct mlx5_hrxq *hrxq;
 
hrxq = mlx5_hrxq_get(dev, rss_hash_default_key,
-rss_hash_default_key_len, 0,
+MLX5_RSS_HASH_KEY_LEN, 0,
 &flow->queue, 1, 0, 0);
if (!hrxq)
hrxq = mlx5_hrxq_new(dev, rss_hash_default_key,
-rss_hash_default_key_len, 0,
+MLX5_RSS_HASH_KEY_LEN, 0,
 &flow->queue, 1, 0, 0);
if (!hrxq)
return rte_flow_error_set(error, rte_errno,
diff --git a/drivers/net/mlx5/mlx5_prm.h b/drivers/net/mlx5/mlx5_prm.h
index f9fae1e50..0870d32fd 100644
--- a/drivers/net/mlx5/mlx5_prm.h
+++ b/drivers/net/mlx5/mlx5_prm.h
@@ -21,6 +21,9 @@
 #include 
 #include "mlx5_autoconf.h"
 
+/* RSS hash key size. */
+#define MLX5_RSS_HASH_KEY_LEN 40
+
 /* Get CQE owner bit. */
 #define MLX5_CQE_OWNER(op_own) ((op_own) & MLX5_CQE_OWNER_MASK)
 
diff --git a/drivers/net/mlx5/mlx5_rss.c b/drivers/net/mlx5/mlx5_rss.c
index d69b4c09e..b95778a8c 100644
--- a/drivers/net/mlx5/mlx5_rss.c
+++ b/drivers/net/mlx5/mlx5_rss.c
@@ -50,10 +50,11 @@ mlx5_rss_hash_update(struct rte_eth_dev *dev,
return -rte_errno;
}
if (rss_conf->rss_key && rss_conf->rss_key_len) {
-   if (rss_conf->rss_key_len != rss_hash_default_key_len) {
+   if (rss_conf->rss_key_len != MLX5_RSS_HASH_KEY_LEN) {
DRV_LOG(ERR,
-   "port %u RSS key len must be %zu Bytes long",
-   dev->data->port_id, rss_hash_default_key_len);
+   "port %u RSS key len must be %s Bytes long",
+   dev->data->port_id,
+   RTE_STR(MLX5_RSS_HASH_KEY_LEN));
 

[dpdk-dev] [PATCH v3 11/21] net/mlx5: add flow TCP item

2018-07-11 Thread Nelio Laranjeiro
Signed-off-by: Nelio Laranjeiro 
Acked-by: Yongseok Koh 
---
 drivers/net/mlx5/mlx5_flow.c | 79 
 1 file changed, 79 insertions(+)

diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index 0096ed8a2..f646eee01 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -53,6 +53,7 @@ extern const struct eth_dev_ops mlx5_dev_ops_isolate;
 #define MLX5_FLOW_FATE_QUEUE (1u << 1)
 
 /* possible L3 layers protocols filtering. */
+#define MLX5_IP_PROTOCOL_TCP 6
 #define MLX5_IP_PROTOCOL_UDP 17
 
 /** Handles information leading to a drop fate. */
@@ -787,6 +788,81 @@ mlx5_flow_item_udp(const struct rte_flow_item *item, 
struct rte_flow *flow,
return size;
 }
 
+/**
+ * Convert the @p item into a Verbs specification after ensuring the NIC
+ * will understand and process it correctly.
+ * If the necessary size for the conversion is greater than the @p flow_size,
+ * nothing is written in @p flow, the validation is still performed.
+ *
+ * @param[in] item
+ *   Item specification.
+ * @param[in, out] flow
+ *   Pointer to flow structure.
+ * @param[in] flow_size
+ *   Size in bytes of the available space in @p flow, if too small, nothing is
+ *   written.
+ * @param[out] error
+ *   Pointer to error structure.
+ *
+ * @return
+ *   On success the number of bytes consumed/necessary, if the returned value
+ *   is lesser or equal to @p flow_size, the @p item has fully been converted,
+ *   otherwise another call with this returned memory size should be done.
+ *   On error, a negative errno value is returned and rte_errno is set.
+ */
+static int
+mlx5_flow_item_tcp(const struct rte_flow_item *item, struct rte_flow *flow,
+  const size_t flow_size, struct rte_flow_error *error)
+{
+   const struct rte_flow_item_tcp *spec = item->spec;
+   const struct rte_flow_item_tcp *mask = item->mask;
+   unsigned int size = sizeof(struct ibv_flow_spec_tcp_udp);
+   struct ibv_flow_spec_tcp_udp tcp = {
+   .type = IBV_FLOW_SPEC_TCP,
+   .size = size,
+   };
+   int ret;
+
+   if (flow->l3_protocol_en && flow->l3_protocol != MLX5_IP_PROTOCOL_TCP)
+   return rte_flow_error_set(error, ENOTSUP,
+ RTE_FLOW_ERROR_TYPE_ITEM,
+ item,
+ "protocol filtering not compatible"
+ " with TCP layer");
+   if (!(flow->layers & MLX5_FLOW_LAYER_OUTER_L3))
+   return rte_flow_error_set(error, ENOTSUP,
+ RTE_FLOW_ERROR_TYPE_ITEM,
+ item,
+ "L3 is mandatory to filter on L4");
+   if (flow->layers & MLX5_FLOW_LAYER_OUTER_L4)
+   return rte_flow_error_set(error, ENOTSUP,
+ RTE_FLOW_ERROR_TYPE_ITEM,
+ item,
+ "L4 layer is already present");
+   if (!mask)
+   mask = &rte_flow_item_tcp_mask;
+   ret = mlx5_flow_item_acceptable
+   (item, (const uint8_t *)mask,
+(const uint8_t *)&rte_flow_item_tcp_mask,
+sizeof(struct rte_flow_item_tcp), error);
+   if (ret < 0)
+   return ret;
+   flow->layers |= MLX5_FLOW_LAYER_OUTER_L4_TCP;
+   if (size > flow_size)
+   return size;
+   if (spec) {
+   tcp.val.dst_port = spec->hdr.dst_port;
+   tcp.val.src_port = spec->hdr.src_port;
+   tcp.mask.dst_port = mask->hdr.dst_port;
+   tcp.mask.src_port = mask->hdr.src_port;
+   /* Remove unwanted bits from values. */
+   tcp.val.src_port &= tcp.mask.src_port;
+   tcp.val.dst_port &= tcp.mask.dst_port;
+   }
+   mlx5_flow_spec_verbs_add(flow, &tcp, size);
+   return size;
+}
+
 /**
  * Convert the @p pattern into a Verbs specifications after ensuring the NIC
  * will understand and process it correctly.
@@ -841,6 +917,9 @@ mlx5_flow_items(const struct rte_flow_item pattern[],
case RTE_FLOW_ITEM_TYPE_UDP:
ret = mlx5_flow_item_udp(pattern, flow, remain, error);
break;
+   case RTE_FLOW_ITEM_TYPE_TCP:
+   ret = mlx5_flow_item_tcp(pattern, flow, remain, error);
+   break;
default:
return rte_flow_error_set(error, ENOTSUP,
  RTE_FLOW_ERROR_TYPE_ITEM,
-- 
2.18.0



[dpdk-dev] [PATCH v3 15/21] net/mlx5: remove useless arguments in hrxq API

2018-07-11 Thread Nelio Laranjeiro
RSS level is necessary to had a bit in the hash_fields which is already
provided in this API, for the tunnel, it is necessary to request such
queue to compute the checksum on the inner most, this last one should
always be activated.

Signed-off-by: Nelio Laranjeiro 
Acked-by: Yongseok Koh 
---
 drivers/net/mlx5/mlx5_flow.c |  4 ++--
 drivers/net/mlx5/mlx5_rxq.c  | 39 +---
 drivers/net/mlx5/mlx5_rxtx.h |  8 ++--
 3 files changed, 13 insertions(+), 38 deletions(-)

diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index f098e727a..f2acac1f5 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -1871,13 +1871,13 @@ mlx5_flow_apply(struct rte_eth_dev *dev, struct 
rte_flow *flow,
 MLX5_RSS_HASH_KEY_LEN,
 verbs->hash_fields,
 (*flow->queue),
-flow->rss.queue_num, 0, 0);
+flow->rss.queue_num);
if (!hrxq)
hrxq = mlx5_hrxq_new(dev, flow->key,
 MLX5_RSS_HASH_KEY_LEN,
 verbs->hash_fields,
 (*flow->queue),
-flow->rss.queue_num, 0, 0);
+flow->rss.queue_num);
if (!hrxq) {
rte_flow_error_set
(error, rte_errno,
diff --git a/drivers/net/mlx5/mlx5_rxq.c b/drivers/net/mlx5/mlx5_rxq.c
index d50b82c69..071740b6d 100644
--- a/drivers/net/mlx5/mlx5_rxq.c
+++ b/drivers/net/mlx5/mlx5_rxq.c
@@ -1740,10 +1740,6 @@ mlx5_ind_table_ibv_verify(struct rte_eth_dev *dev)
  *   first queue index will be taken for the indirection table.
  * @param queues_n
  *   Number of queues.
- * @param tunnel
- *   Tunnel type, implies tunnel offloading like inner checksum if available.
- * @param rss_level
- *   RSS hash on tunnel level.
  *
  * @return
  *   The Verbs object initialised, NULL otherwise and rte_errno is set.
@@ -1752,17 +1748,13 @@ struct mlx5_hrxq *
 mlx5_hrxq_new(struct rte_eth_dev *dev,
  const uint8_t *rss_key, uint32_t rss_key_len,
  uint64_t hash_fields,
- const uint16_t *queues, uint32_t queues_n,
- uint32_t tunnel, uint32_t rss_level)
+ const uint16_t *queues, uint32_t queues_n)
 {
struct priv *priv = dev->data->dev_private;
struct mlx5_hrxq *hrxq;
struct mlx5_ind_table_ibv *ind_tbl;
struct ibv_qp *qp;
int err;
-#ifdef HAVE_IBV_DEVICE_TUNNEL_SUPPORT
-   struct mlx5dv_qp_init_attr qp_init_attr = {0};
-#endif
 
queues_n = hash_fields ? queues_n : 1;
ind_tbl = mlx5_ind_table_ibv_get(dev, queues, queues_n);
@@ -1777,11 +1769,6 @@ mlx5_hrxq_new(struct rte_eth_dev *dev,
rss_key = rss_hash_default_key;
}
 #ifdef HAVE_IBV_DEVICE_TUNNEL_SUPPORT
-   if (tunnel) {
-   qp_init_attr.comp_mask =
-   MLX5DV_QP_INIT_ATTR_MASK_QP_CREATE_FLAGS;
-   qp_init_attr.create_flags = MLX5DV_QP_CREATE_TUNNEL_OFFLOADS;
-   }
qp = mlx5_glue->dv_create_qp
(priv->ctx,
 &(struct ibv_qp_init_attr_ex){
@@ -1797,14 +1784,17 @@ mlx5_hrxq_new(struct rte_eth_dev *dev,
.rx_hash_key = rss_key ?
   (void *)(uintptr_t)rss_key :
   rss_hash_default_key,
-   .rx_hash_fields_mask = hash_fields |
-   (tunnel && rss_level > 1 ?
-   (uint32_t)IBV_RX_HASH_INNER : 0),
+   .rx_hash_fields_mask = hash_fields,
},
.rwq_ind_tbl = ind_tbl->ind_table,
.pd = priv->pd,
 },
-&qp_init_attr);
+&(struct mlx5dv_qp_init_attr){
+   .comp_mask = (hash_fields & IBV_RX_HASH_INNER) ?
+MLX5DV_QP_INIT_ATTR_MASK_QP_CREATE_FLAGS :
+0,
+   .create_flags = MLX5DV_QP_CREATE_TUNNEL_OFFLOADS,
+});
 #else
qp = mlx5_glue->create_qp_ex
(priv->ctx,
@@ -1838,8 +1828,6 @@ mlx5_hrxq_new(struct rte_eth_dev *dev,
hrxq->qp = qp;
hrxq->rss_key_len = rss_key_len;
hrxq->hash_fields = hash_fields;
-   hrxq->tunnel = tunnel;
-   hrxq->rss_level = rss_level;
memcpy(hrxq->rss_key, rss_key, rss_key_len);
rte_atomic32_inc(&hrxq->re

[dpdk-dev] [PATCH v3 14/21] net/mlx5: add RSS flow action

2018-07-11 Thread Nelio Laranjeiro
Signed-off-by: Nelio Laranjeiro 
---
 drivers/net/mlx5/mlx5_flow.c | 682 +++
 1 file changed, 537 insertions(+), 145 deletions(-)

diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index 77483bd1f..f098e727a 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -51,6 +51,7 @@ extern const struct eth_dev_ops mlx5_dev_ops_isolate;
 /* Actions that modify the fate of matching traffic. */
 #define MLX5_FLOW_FATE_DROP (1u << 0)
 #define MLX5_FLOW_FATE_QUEUE (1u << 1)
+#define MLX5_FLOW_FATE_RSS (1u << 2)
 
 /* Modify a packet. */
 #define MLX5_FLOW_MOD_FLAG (1u << 0)
@@ -60,8 +61,68 @@ extern const struct eth_dev_ops mlx5_dev_ops_isolate;
 #define MLX5_IP_PROTOCOL_TCP 6
 #define MLX5_IP_PROTOCOL_UDP 17
 
+/* Priority reserved for default flows. */
+#define MLX5_FLOW_PRIO_RSVD ((uint32_t)-1)
+
+enum mlx5_expansion {
+   MLX5_EXPANSION_ROOT,
+   MLX5_EXPANSION_ETH,
+   MLX5_EXPANSION_IPV4,
+   MLX5_EXPANSION_IPV4_UDP,
+   MLX5_EXPANSION_IPV4_TCP,
+   MLX5_EXPANSION_IPV6,
+   MLX5_EXPANSION_IPV6_UDP,
+   MLX5_EXPANSION_IPV6_TCP,
+};
+
+/** Supported expansion of items. */
+static const struct rte_flow_expand_node mlx5_support_expansion[] = {
+   [MLX5_EXPANSION_ROOT] = {
+   .next = RTE_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_ETH,
+MLX5_EXPANSION_IPV4,
+MLX5_EXPANSION_IPV6),
+   .type = RTE_FLOW_ITEM_TYPE_END,
+   },
+   [MLX5_EXPANSION_ETH] = {
+   .next = RTE_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_IPV4,
+MLX5_EXPANSION_IPV6),
+   .type = RTE_FLOW_ITEM_TYPE_ETH,
+   },
+   [MLX5_EXPANSION_IPV4] = {
+   .next = RTE_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_IPV4_UDP,
+MLX5_EXPANSION_IPV4_TCP),
+   .type = RTE_FLOW_ITEM_TYPE_IPV4,
+   .rss_types = ETH_RSS_IPV4 | ETH_RSS_FRAG_IPV4 |
+   ETH_RSS_NONFRAG_IPV4_OTHER,
+   },
+   [MLX5_EXPANSION_IPV4_UDP] = {
+   .type = RTE_FLOW_ITEM_TYPE_UDP,
+   .rss_types = ETH_RSS_NONFRAG_IPV4_UDP,
+   },
+   [MLX5_EXPANSION_IPV4_TCP] = {
+   .type = RTE_FLOW_ITEM_TYPE_TCP,
+   .rss_types = ETH_RSS_NONFRAG_IPV4_TCP,
+   },
+   [MLX5_EXPANSION_IPV6] = {
+   .next = RTE_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_IPV6_UDP,
+MLX5_EXPANSION_IPV6_TCP),
+   .type = RTE_FLOW_ITEM_TYPE_IPV6,
+   .rss_types = ETH_RSS_IPV6 | ETH_RSS_FRAG_IPV6 |
+   ETH_RSS_NONFRAG_IPV6_OTHER,
+   },
+   [MLX5_EXPANSION_IPV6_UDP] = {
+   .type = RTE_FLOW_ITEM_TYPE_UDP,
+   .rss_types = ETH_RSS_NONFRAG_IPV6_UDP,
+   },
+   [MLX5_EXPANSION_IPV6_TCP] = {
+   .type = RTE_FLOW_ITEM_TYPE_TCP,
+   .rss_types = ETH_RSS_NONFRAG_IPV6_TCP,
+   },
+};
+
 /** Handles information leading to a drop fate. */
 struct mlx5_flow_verbs {
+   LIST_ENTRY(mlx5_flow_verbs) next;
unsigned int size; /**< Size of the attribute. */
struct {
struct ibv_flow_attr *attr;
@@ -70,6 +131,7 @@ struct mlx5_flow_verbs {
};
struct ibv_flow *flow; /**< Verbs flow pointer. */
struct mlx5_hrxq *hrxq; /**< Hash Rx queue object. */
+   uint64_t hash_fields; /**< Verbs hash Rx queue hash fields. */
 };
 
 /* Flow structure. */
@@ -84,8 +146,12 @@ struct rte_flow {
uint32_t fate;
/**< Bit-fields of present fate see MLX5_FLOW_FATE_*. */
uint8_t l3_protocol; /**< valid when l3_protocol_en is set. */
-   struct mlx5_flow_verbs verbs; /* Verbs flow. */
-   uint16_t queue; /**< Destination queue to redirect traffic to. */
+   LIST_HEAD(verbs, mlx5_flow_verbs) verbs; /**< Verbs flows list. */
+   struct mlx5_flow_verbs *cur_verbs;
+   /**< Current Verbs flow structure being filled. */
+   struct rte_flow_action_rss rss;/**< RSS context. */
+   uint8_t key[MLX5_RSS_HASH_KEY_LEN]; /**< RSS hash key. */
+   uint16_t (*queue)[]; /**< Destination queues to redirect traffic to. */
 };
 
 static const struct rte_flow_ops mlx5_flow_ops = {
@@ -128,16 +194,38 @@ struct ibv_spec_header {
uint16_t size;
 };
 
- /**
-  * Discover the maximum number of priority available.
-  *
-  * @param[in] dev
-  *   Pointer to Ethernet device.
-  *
-  * @return
-  *   number of supported flow priority on success, a negative errno value
-  *   otherwise and rte_errno is set.
-  */
+/*
+ * Number of sub priorities.
+ * For each kind of pattern matching i.e. L2, L3, L4 to have a correct
+ * matching on the NIC (firmware dependent) L4 most have the higher priority
+ * followed by L3 and ending with L2.
+ */
+#define MLX5_PRIO

[dpdk-dev] [PATCH v3 16/21] net/mlx5: support inner RSS computation

2018-07-11 Thread Nelio Laranjeiro
Signed-off-by: Nelio Laranjeiro 
---
 drivers/net/mlx5/mlx5_flow.c | 241 ++-
 1 file changed, 180 insertions(+), 61 deletions(-)

diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index f2acac1f5..edceb17ed 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -35,18 +35,42 @@
 extern const struct eth_dev_ops mlx5_dev_ops;
 extern const struct eth_dev_ops mlx5_dev_ops_isolate;
 
-/* Pattern Layer bits. */
+/* Pattern outer Layer bits. */
 #define MLX5_FLOW_LAYER_OUTER_L2 (1u << 0)
 #define MLX5_FLOW_LAYER_OUTER_L3_IPV4 (1u << 1)
 #define MLX5_FLOW_LAYER_OUTER_L3_IPV6 (1u << 2)
 #define MLX5_FLOW_LAYER_OUTER_L4_UDP (1u << 3)
 #define MLX5_FLOW_LAYER_OUTER_L4_TCP (1u << 4)
 #define MLX5_FLOW_LAYER_OUTER_VLAN (1u << 5)
-/* Masks. */
+
+/* Pattern inner Layer bits. */
+#define MLX5_FLOW_LAYER_INNER_L2 (1u << 6)
+#define MLX5_FLOW_LAYER_INNER_L3_IPV4 (1u << 7)
+#define MLX5_FLOW_LAYER_INNER_L3_IPV6 (1u << 8)
+#define MLX5_FLOW_LAYER_INNER_L4_UDP (1u << 9)
+#define MLX5_FLOW_LAYER_INNER_L4_TCP (1u << 10)
+#define MLX5_FLOW_LAYER_INNER_VLAN (1u << 11)
+
+/* Outer Masks. */
 #define MLX5_FLOW_LAYER_OUTER_L3 \
(MLX5_FLOW_LAYER_OUTER_L3_IPV4 | MLX5_FLOW_LAYER_OUTER_L3_IPV6)
 #define MLX5_FLOW_LAYER_OUTER_L4 \
(MLX5_FLOW_LAYER_OUTER_L4_UDP | MLX5_FLOW_LAYER_OUTER_L4_TCP)
+#define MLX5_FLOW_LAYER_OUTER \
+   (MLX5_FLOW_LAYER_OUTER_L2 | MLX5_FLOW_LAYER_OUTER_L3 | \
+MLX5_FLOW_LAYER_OUTER_L4)
+
+/* Tunnel Masks. */
+#define MLX5_FLOW_LAYER_TUNNEL 0
+
+/* Inner Masks. */
+#define MLX5_FLOW_LAYER_INNER_L3 \
+   (MLX5_FLOW_LAYER_INNER_L3_IPV4 | MLX5_FLOW_LAYER_INNER_L3_IPV6)
+#define MLX5_FLOW_LAYER_INNER_L4 \
+   (MLX5_FLOW_LAYER_INNER_L4_UDP | MLX5_FLOW_LAYER_INNER_L4_TCP)
+#define MLX5_FLOW_LAYER_INNER \
+   (MLX5_FLOW_LAYER_INNER_L2 | MLX5_FLOW_LAYER_INNER_L3 | \
+MLX5_FLOW_LAYER_INNER_L4)
 
 /* Actions that modify the fate of matching traffic. */
 #define MLX5_FLOW_FATE_DROP (1u << 0)
@@ -66,6 +90,14 @@ extern const struct eth_dev_ops mlx5_dev_ops_isolate;
 
 enum mlx5_expansion {
MLX5_EXPANSION_ROOT,
+   MLX5_EXPANSION_ROOT_OUTER,
+   MLX5_EXPANSION_OUTER_ETH,
+   MLX5_EXPANSION_OUTER_IPV4,
+   MLX5_EXPANSION_OUTER_IPV4_UDP,
+   MLX5_EXPANSION_OUTER_IPV4_TCP,
+   MLX5_EXPANSION_OUTER_IPV6,
+   MLX5_EXPANSION_OUTER_IPV6_UDP,
+   MLX5_EXPANSION_OUTER_IPV6_TCP,
MLX5_EXPANSION_ETH,
MLX5_EXPANSION_IPV4,
MLX5_EXPANSION_IPV4_UDP,
@@ -83,6 +115,50 @@ static const struct rte_flow_expand_node 
mlx5_support_expansion[] = {
 MLX5_EXPANSION_IPV6),
.type = RTE_FLOW_ITEM_TYPE_END,
},
+   [MLX5_EXPANSION_ROOT_OUTER] = {
+   .next = RTE_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_OUTER_ETH,
+MLX5_EXPANSION_OUTER_IPV4,
+MLX5_EXPANSION_OUTER_IPV6),
+   .type = RTE_FLOW_ITEM_TYPE_END,
+   },
+   [MLX5_EXPANSION_OUTER_ETH] = {
+   .next = RTE_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_OUTER_IPV4,
+MLX5_EXPANSION_OUTER_IPV6),
+   .type = RTE_FLOW_ITEM_TYPE_ETH,
+   .rss_types = 0,
+   },
+   [MLX5_EXPANSION_OUTER_IPV4] = {
+   .next = RTE_FLOW_EXPAND_RSS_NEXT
+   (MLX5_EXPANSION_OUTER_IPV4_UDP,
+MLX5_EXPANSION_OUTER_IPV4_TCP),
+   .type = RTE_FLOW_ITEM_TYPE_IPV4,
+   .rss_types = ETH_RSS_IPV4 | ETH_RSS_FRAG_IPV4 |
+   ETH_RSS_NONFRAG_IPV4_OTHER,
+   },
+   [MLX5_EXPANSION_OUTER_IPV4_UDP] = {
+   .type = RTE_FLOW_ITEM_TYPE_UDP,
+   .rss_types = ETH_RSS_NONFRAG_IPV4_UDP,
+   },
+   [MLX5_EXPANSION_OUTER_IPV4_TCP] = {
+   .type = RTE_FLOW_ITEM_TYPE_TCP,
+   .rss_types = ETH_RSS_NONFRAG_IPV4_TCP,
+   },
+   [MLX5_EXPANSION_OUTER_IPV6] = {
+   .next = RTE_FLOW_EXPAND_RSS_NEXT
+   (MLX5_EXPANSION_OUTER_IPV6_UDP,
+MLX5_EXPANSION_OUTER_IPV6_TCP),
+   .type = RTE_FLOW_ITEM_TYPE_IPV6,
+   .rss_types = ETH_RSS_IPV6 | ETH_RSS_FRAG_IPV6 |
+   ETH_RSS_NONFRAG_IPV6_OTHER,
+   },
+   [MLX5_EXPANSION_OUTER_IPV6_UDP] = {
+   .type = RTE_FLOW_ITEM_TYPE_UDP,
+   .rss_types = ETH_RSS_NONFRAG_IPV6_UDP,
+   },
+   [MLX5_EXPANSION_OUTER_IPV6_TCP] = {
+   .type = RTE_FLOW_ITEM_TYPE_TCP,
+   .rss_types = ETH_RSS_NONFRAG_IPV6_TCP,
+   },
[MLX5_EXPANSION_ETH] = {
.next = RTE_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_IPV4,
 MLX5_EXPANSION_IPV6),
@@ -453,6 +529,32 @@ mlx5_flow_spec_verbs_add(struct rte_fl

[dpdk-dev] [PATCH v3 17/21] net/mlx5: add flow VXLAN item

2018-07-11 Thread Nelio Laranjeiro
Signed-off-by: Nelio Laranjeiro 
Acked-by: Yongseok Koh 
---
 drivers/net/mlx5/mlx5_flow.c | 190 +--
 drivers/net/mlx5/mlx5_rxtx.h |   1 +
 2 files changed, 163 insertions(+), 28 deletions(-)

diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index edceb17ed..95507059e 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -51,6 +51,9 @@ extern const struct eth_dev_ops mlx5_dev_ops_isolate;
 #define MLX5_FLOW_LAYER_INNER_L4_TCP (1u << 10)
 #define MLX5_FLOW_LAYER_INNER_VLAN (1u << 11)
 
+/* Pattern tunnel Layer bits. */
+#define MLX5_FLOW_LAYER_VXLAN (1u << 12)
+
 /* Outer Masks. */
 #define MLX5_FLOW_LAYER_OUTER_L3 \
(MLX5_FLOW_LAYER_OUTER_L3_IPV4 | MLX5_FLOW_LAYER_OUTER_L3_IPV6)
@@ -61,7 +64,7 @@ extern const struct eth_dev_ops mlx5_dev_ops_isolate;
 MLX5_FLOW_LAYER_OUTER_L4)
 
 /* Tunnel Masks. */
-#define MLX5_FLOW_LAYER_TUNNEL 0
+#define MLX5_FLOW_LAYER_TUNNEL MLX5_FLOW_LAYER_VXLAN
 
 /* Inner Masks. */
 #define MLX5_FLOW_LAYER_INNER_L3 \
@@ -98,6 +101,7 @@ enum mlx5_expansion {
MLX5_EXPANSION_OUTER_IPV6,
MLX5_EXPANSION_OUTER_IPV6_UDP,
MLX5_EXPANSION_OUTER_IPV6_TCP,
+   MLX5_EXPANSION_VXLAN,
MLX5_EXPANSION_ETH,
MLX5_EXPANSION_IPV4,
MLX5_EXPANSION_IPV4_UDP,
@@ -136,6 +140,7 @@ static const struct rte_flow_expand_node 
mlx5_support_expansion[] = {
ETH_RSS_NONFRAG_IPV4_OTHER,
},
[MLX5_EXPANSION_OUTER_IPV4_UDP] = {
+   .next = RTE_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_VXLAN),
.type = RTE_FLOW_ITEM_TYPE_UDP,
.rss_types = ETH_RSS_NONFRAG_IPV4_UDP,
},
@@ -152,6 +157,7 @@ static const struct rte_flow_expand_node 
mlx5_support_expansion[] = {
ETH_RSS_NONFRAG_IPV6_OTHER,
},
[MLX5_EXPANSION_OUTER_IPV6_UDP] = {
+   .next = RTE_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_VXLAN),
.type = RTE_FLOW_ITEM_TYPE_UDP,
.rss_types = ETH_RSS_NONFRAG_IPV6_UDP,
},
@@ -159,6 +165,10 @@ static const struct rte_flow_expand_node 
mlx5_support_expansion[] = {
.type = RTE_FLOW_ITEM_TYPE_TCP,
.rss_types = ETH_RSS_NONFRAG_IPV6_TCP,
},
+   [MLX5_EXPANSION_VXLAN] = {
+   .next = RTE_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_ETH),
+   .type = RTE_FLOW_ITEM_TYPE_VXLAN,
+   },
[MLX5_EXPANSION_ETH] = {
.next = RTE_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_IPV4,
 MLX5_EXPANSION_IPV6),
@@ -226,6 +236,8 @@ struct rte_flow {
struct mlx5_flow_verbs *cur_verbs;
/**< Current Verbs flow structure being filled. */
struct rte_flow_action_rss rss;/**< RSS context. */
+   uint32_t tunnel_ptype;
+   /**< Store tunnel packet type data to store in Rx queue. */
uint8_t key[MLX5_RSS_HASH_KEY_LEN]; /**< RSS hash key. */
uint16_t (*queue)[]; /**< Destination queues to redirect traffic to. */
 };
@@ -1158,6 +1170,103 @@ mlx5_flow_item_tcp(const struct rte_flow_item *item, 
struct rte_flow *flow,
return size;
 }
 
+/**
+ * Convert the @p item into a Verbs specification after ensuring the NIC
+ * will understand and process it correctly.
+ * If the necessary size for the conversion is greater than the @p flow_size,
+ * nothing is written in @p flow, the validation is still performed.
+ *
+ * @param[in] item
+ *   Item specification.
+ * @param[in, out] flow
+ *   Pointer to flow structure.
+ * @param[in] flow_size
+ *   Size in bytes of the available space in @p flow, if too small, nothing is
+ *   written.
+ * @param[out] error
+ *   Pointer to error structure.
+ *
+ * @return
+ *   On success the number of bytes consumed/necessary, if the returned value
+ *   is lesser or equal to @p flow_size, the @p item has fully been converted,
+ *   otherwise another call with this returned memory size should be done.
+ *   On error, a negative errno value is returned and rte_errno is set.
+ */
+static int
+mlx5_flow_item_vxlan(const struct rte_flow_item *item, struct rte_flow *flow,
+const size_t flow_size, struct rte_flow_error *error)
+{
+   const struct rte_flow_item_vxlan *spec = item->spec;
+   const struct rte_flow_item_vxlan *mask = item->mask;
+   unsigned int size = sizeof(struct ibv_flow_spec_tunnel);
+   struct ibv_flow_spec_tunnel vxlan = {
+   .type = IBV_FLOW_SPEC_VXLAN_TUNNEL,
+   .size = size,
+   };
+   int ret;
+   union vni {
+   uint32_t vlan_id;
+   uint8_t vni[4];
+   } id = { .vlan_id = 0, };
+
+   if (flow->layers & MLX5_FLOW_LAYER_TUNNEL)
+   return rte_flow_error_set(error, ENOTSUP,
+ RTE_FLOW_ERROR_TYPE_ITEM,
+ item,
+  

[dpdk-dev] [PATCH v3 19/21] net/mlx5: add flow GRE item

2018-07-11 Thread Nelio Laranjeiro
Signed-off-by: Nelio Laranjeiro 
---
 drivers/net/mlx5/mlx5_flow.c | 193 ++-
 drivers/net/mlx5/mlx5_rxtx.h |   2 +-
 2 files changed, 192 insertions(+), 3 deletions(-)

diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index e06df0eb5..3412c856a 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -54,6 +54,7 @@ extern const struct eth_dev_ops mlx5_dev_ops_isolate;
 /* Pattern tunnel Layer bits. */
 #define MLX5_FLOW_LAYER_VXLAN (1u << 12)
 #define MLX5_FLOW_LAYER_VXLAN_GPE (1u << 13)
+#define MLX5_FLOW_LAYER_GRE (1u << 14)
 
 /* Outer Masks. */
 #define MLX5_FLOW_LAYER_OUTER_L3 \
@@ -66,7 +67,8 @@ extern const struct eth_dev_ops mlx5_dev_ops_isolate;
 
 /* Tunnel Masks. */
 #define MLX5_FLOW_LAYER_TUNNEL \
-   (MLX5_FLOW_LAYER_VXLAN | MLX5_FLOW_LAYER_VXLAN_GPE)
+   (MLX5_FLOW_LAYER_VXLAN | MLX5_FLOW_LAYER_VXLAN_GPE | \
+MLX5_FLOW_LAYER_GRE)
 
 /* Inner Masks. */
 #define MLX5_FLOW_LAYER_INNER_L3 \
@@ -89,6 +91,7 @@ extern const struct eth_dev_ops mlx5_dev_ops_isolate;
 /* possible L3 layers protocols filtering. */
 #define MLX5_IP_PROTOCOL_TCP 6
 #define MLX5_IP_PROTOCOL_UDP 17
+#define MLX5_IP_PROTOCOL_GRE 47
 
 /* Priority reserved for default flows. */
 #define MLX5_FLOW_PRIO_RSVD ((uint32_t)-1)
@@ -105,6 +108,7 @@ enum mlx5_expansion {
MLX5_EXPANSION_OUTER_IPV6_TCP,
MLX5_EXPANSION_VXLAN,
MLX5_EXPANSION_VXLAN_GPE,
+   MLX5_EXPANSION_GRE,
MLX5_EXPANSION_ETH,
MLX5_EXPANSION_IPV4,
MLX5_EXPANSION_IPV4_UDP,
@@ -137,7 +141,8 @@ static const struct rte_flow_expand_node 
mlx5_support_expansion[] = {
[MLX5_EXPANSION_OUTER_IPV4] = {
.next = RTE_FLOW_EXPAND_RSS_NEXT
(MLX5_EXPANSION_OUTER_IPV4_UDP,
-MLX5_EXPANSION_OUTER_IPV4_TCP),
+MLX5_EXPANSION_OUTER_IPV4_TCP,
+MLX5_EXPANSION_GRE),
.type = RTE_FLOW_ITEM_TYPE_IPV4,
.rss_types = ETH_RSS_IPV4 | ETH_RSS_FRAG_IPV4 |
ETH_RSS_NONFRAG_IPV4_OTHER,
@@ -180,6 +185,10 @@ static const struct rte_flow_expand_node 
mlx5_support_expansion[] = {
 MLX5_EXPANSION_IPV6),
.type = RTE_FLOW_ITEM_TYPE_VXLAN_GPE,
},
+   [MLX5_EXPANSION_GRE] = {
+   .next = RTE_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_IPV4),
+   .type = RTE_FLOW_ITEM_TYPE_GRE,
+   },
[MLX5_EXPANSION_ETH] = {
.next = RTE_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_IPV4,
 MLX5_EXPANSION_IPV6),
@@ -328,6 +337,10 @@ static struct mlx5_flow_tunnel_info tunnels_info[] = {
.tunnel = MLX5_FLOW_LAYER_VXLAN_GPE,
.ptype = RTE_PTYPE_TUNNEL_VXLAN_GPE | RTE_PTYPE_L4_UDP,
},
+   {
+   .tunnel = MLX5_FLOW_LAYER_GRE,
+   .ptype = RTE_PTYPE_TUNNEL_GRE,
+   },
 };
 
 /**
@@ -966,6 +979,18 @@ mlx5_flow_item_ipv6(const struct rte_flow_item *item, 
struct rte_flow *flow,
  RTE_FLOW_ERROR_TYPE_ITEM,
  item,
  "L3 cannot follow an L4 layer.");
+   /*
+* IPv6 is not recognised by the NIC inside a GRE tunnel.
+* Such support has to be disabled as the rule will be
+* accepted.  Issue reproduced with Mellanox OFED 4.3-3.0.2.1 and
+* Mellanox OFED 4.4-1.0.0.0.
+*/
+   if (tunnel && flow->layers & MLX5_FLOW_LAYER_GRE)
+   return rte_flow_error_set(error, ENOTSUP,
+ RTE_FLOW_ERROR_TYPE_ITEM,
+ item,
+ "IPv6 inside a GRE tunnel is"
+ " not recognised.");
if (!mask)
mask = &rte_flow_item_ipv6_mask;
ret = mlx5_flow_item_acceptable
@@ -1405,6 +1430,167 @@ mlx5_flow_item_vxlan_gpe(struct rte_eth_dev *dev,
return size;
 }
 
+/**
+ * Update the protocol in Verbs IPv4/IPv6 spec.
+ *
+ * @param[in, out] attr
+ *   Pointer to Verbs attributes structure.
+ * @param[in] search
+ *   Specification type to search in order to update the IP protocol.
+ * @param[in] protocol
+ *   Protocol value to set if none is present in the specification.
+ */
+static void
+mlx5_flow_item_gre_ip_protocol_update(struct ibv_flow_attr *attr,
+ enum ibv_flow_spec_type search,
+ uint8_t protocol)
+{
+   unsigned int i;
+   struct ibv_spec_header *hdr = (struct ibv_spec_header *)
+   ((uint8_t *)attr + sizeof(struct ibv_flow_attr));
+
+   if (!attr)
+   return;
+   for (i = 0; i != attr->num_of_specs; ++i) {
+   if (hdr->type == search) {
+  

[dpdk-dev] [PATCH v3 18/21] net/mlx5: add flow VXLAN-GPE item

2018-07-11 Thread Nelio Laranjeiro
Signed-off-by: Nelio Laranjeiro 
---
 drivers/net/mlx5/mlx5_flow.c | 219 ---
 drivers/net/mlx5/mlx5_rxtx.h |   5 +-
 2 files changed, 209 insertions(+), 15 deletions(-)

diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index 95507059e..e06df0eb5 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -53,6 +53,7 @@ extern const struct eth_dev_ops mlx5_dev_ops_isolate;
 
 /* Pattern tunnel Layer bits. */
 #define MLX5_FLOW_LAYER_VXLAN (1u << 12)
+#define MLX5_FLOW_LAYER_VXLAN_GPE (1u << 13)
 
 /* Outer Masks. */
 #define MLX5_FLOW_LAYER_OUTER_L3 \
@@ -64,7 +65,8 @@ extern const struct eth_dev_ops mlx5_dev_ops_isolate;
 MLX5_FLOW_LAYER_OUTER_L4)
 
 /* Tunnel Masks. */
-#define MLX5_FLOW_LAYER_TUNNEL MLX5_FLOW_LAYER_VXLAN
+#define MLX5_FLOW_LAYER_TUNNEL \
+   (MLX5_FLOW_LAYER_VXLAN | MLX5_FLOW_LAYER_VXLAN_GPE)
 
 /* Inner Masks. */
 #define MLX5_FLOW_LAYER_INNER_L3 \
@@ -102,6 +104,7 @@ enum mlx5_expansion {
MLX5_EXPANSION_OUTER_IPV6_UDP,
MLX5_EXPANSION_OUTER_IPV6_TCP,
MLX5_EXPANSION_VXLAN,
+   MLX5_EXPANSION_VXLAN_GPE,
MLX5_EXPANSION_ETH,
MLX5_EXPANSION_IPV4,
MLX5_EXPANSION_IPV4_UDP,
@@ -140,7 +143,8 @@ static const struct rte_flow_expand_node 
mlx5_support_expansion[] = {
ETH_RSS_NONFRAG_IPV4_OTHER,
},
[MLX5_EXPANSION_OUTER_IPV4_UDP] = {
-   .next = RTE_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_VXLAN),
+   .next = RTE_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_VXLAN,
+MLX5_EXPANSION_VXLAN_GPE),
.type = RTE_FLOW_ITEM_TYPE_UDP,
.rss_types = ETH_RSS_NONFRAG_IPV4_UDP,
},
@@ -157,7 +161,8 @@ static const struct rte_flow_expand_node 
mlx5_support_expansion[] = {
ETH_RSS_NONFRAG_IPV6_OTHER,
},
[MLX5_EXPANSION_OUTER_IPV6_UDP] = {
-   .next = RTE_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_VXLAN),
+   .next = RTE_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_VXLAN,
+MLX5_EXPANSION_VXLAN_GPE),
.type = RTE_FLOW_ITEM_TYPE_UDP,
.rss_types = ETH_RSS_NONFRAG_IPV6_UDP,
},
@@ -169,6 +174,12 @@ static const struct rte_flow_expand_node 
mlx5_support_expansion[] = {
.next = RTE_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_ETH),
.type = RTE_FLOW_ITEM_TYPE_VXLAN,
},
+   [MLX5_EXPANSION_VXLAN_GPE] = {
+   .next = RTE_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_ETH,
+MLX5_EXPANSION_IPV4,
+MLX5_EXPANSION_IPV6),
+   .type = RTE_FLOW_ITEM_TYPE_VXLAN_GPE,
+   },
[MLX5_EXPANSION_ETH] = {
.next = RTE_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_IPV4,
 MLX5_EXPANSION_IPV6),
@@ -236,8 +247,6 @@ struct rte_flow {
struct mlx5_flow_verbs *cur_verbs;
/**< Current Verbs flow structure being filled. */
struct rte_flow_action_rss rss;/**< RSS context. */
-   uint32_t tunnel_ptype;
-   /**< Store tunnel packet type data to store in Rx queue. */
uint8_t key[MLX5_RSS_HASH_KEY_LEN]; /**< RSS hash key. */
uint16_t (*queue)[]; /**< Destination queues to redirect traffic to. */
 };
@@ -304,6 +313,23 @@ static const uint32_t 
priority_map_5[][MLX5_PRIORITY_MAP_MAX] = {
{ 9, 10, 11 }, { 12, 13, 14 },
 };
 
+/* Tunnel information. */
+struct mlx5_flow_tunnel_info {
+   uint32_t tunnel; /**< Tunnel bit (see MLX5_FLOW_*). */
+   uint32_t ptype; /**< Tunnel Ptype (see RTE_PTYPE_*). */
+};
+
+static struct mlx5_flow_tunnel_info tunnels_info[] = {
+   {
+   .tunnel = MLX5_FLOW_LAYER_VXLAN,
+   .ptype = RTE_PTYPE_TUNNEL_VXLAN | RTE_PTYPE_L4_UDP,
+   },
+   {
+   .tunnel = MLX5_FLOW_LAYER_VXLAN_GPE,
+   .ptype = RTE_PTYPE_TUNNEL_VXLAN_GPE | RTE_PTYPE_L4_UDP,
+   },
+};
+
 /**
  * Discover the maximum number of priority available.
  *
@@ -1263,7 +1289,119 @@ mlx5_flow_item_vxlan(const struct rte_flow_item *item, 
struct rte_flow *flow,
flow->cur_verbs->attr->priority = MLX5_PRIORITY_MAP_L2;
}
flow->layers |= MLX5_FLOW_LAYER_VXLAN;
-   flow->tunnel_ptype = RTE_PTYPE_TUNNEL_VXLAN | RTE_PTYPE_L4_UDP;
+   return size;
+}
+
+/**
+ * Convert the @p item into a Verbs specification after ensuring the NIC
+ * will understand and process it correctly.
+ * If the necessary size for the conversion is greater than the @p flow_size,
+ * nothing is written in @p flow, the validation is still performed.
+ *
+ * @param dev
+ *   Pointer to Ethernet device.
+ * @param[in] item
+ *   Item specification.
+ * @param[in, out] flow
+ *   Pointer to flow structure.
+ * @param[in] flow_size
+ *   Size in

[dpdk-dev] [PATCH v3 20/21] net/mlx5: add flow MPLS item

2018-07-11 Thread Nelio Laranjeiro
Signed-off-by: Nelio Laranjeiro 
---
 drivers/net/mlx5/mlx5_flow.c | 102 ++-
 drivers/net/mlx5/mlx5_rxtx.h |   2 +-
 2 files changed, 101 insertions(+), 3 deletions(-)

diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index 3412c856a..4ad581ed9 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -55,6 +55,7 @@ extern const struct eth_dev_ops mlx5_dev_ops_isolate;
 #define MLX5_FLOW_LAYER_VXLAN (1u << 12)
 #define MLX5_FLOW_LAYER_VXLAN_GPE (1u << 13)
 #define MLX5_FLOW_LAYER_GRE (1u << 14)
+#define MLX5_FLOW_LAYER_MPLS (1u << 15)
 
 /* Outer Masks. */
 #define MLX5_FLOW_LAYER_OUTER_L3 \
@@ -68,7 +69,7 @@ extern const struct eth_dev_ops mlx5_dev_ops_isolate;
 /* Tunnel Masks. */
 #define MLX5_FLOW_LAYER_TUNNEL \
(MLX5_FLOW_LAYER_VXLAN | MLX5_FLOW_LAYER_VXLAN_GPE | \
-MLX5_FLOW_LAYER_GRE)
+MLX5_FLOW_LAYER_GRE | MLX5_FLOW_LAYER_MPLS)
 
 /* Inner Masks. */
 #define MLX5_FLOW_LAYER_INNER_L3 \
@@ -92,6 +93,7 @@ extern const struct eth_dev_ops mlx5_dev_ops_isolate;
 #define MLX5_IP_PROTOCOL_TCP 6
 #define MLX5_IP_PROTOCOL_UDP 17
 #define MLX5_IP_PROTOCOL_GRE 47
+#define MLX5_IP_PROTOCOL_MPLS 147
 
 /* Priority reserved for default flows. */
 #define MLX5_FLOW_PRIO_RSVD ((uint32_t)-1)
@@ -109,6 +111,7 @@ enum mlx5_expansion {
MLX5_EXPANSION_VXLAN,
MLX5_EXPANSION_VXLAN_GPE,
MLX5_EXPANSION_GRE,
+   MLX5_EXPANSION_MPLS,
MLX5_EXPANSION_ETH,
MLX5_EXPANSION_IPV4,
MLX5_EXPANSION_IPV4_UDP,
@@ -134,7 +137,8 @@ static const struct rte_flow_expand_node 
mlx5_support_expansion[] = {
},
[MLX5_EXPANSION_OUTER_ETH] = {
.next = RTE_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_OUTER_IPV4,
-MLX5_EXPANSION_OUTER_IPV6),
+MLX5_EXPANSION_OUTER_IPV6,
+MLX5_EXPANSION_MPLS),
.type = RTE_FLOW_ITEM_TYPE_ETH,
.rss_types = 0,
},
@@ -189,6 +193,11 @@ static const struct rte_flow_expand_node 
mlx5_support_expansion[] = {
.next = RTE_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_IPV4),
.type = RTE_FLOW_ITEM_TYPE_GRE,
},
+   [MLX5_EXPANSION_MPLS] = {
+   .next = RTE_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_IPV4,
+MLX5_EXPANSION_IPV6),
+   .type = RTE_FLOW_ITEM_TYPE_MPLS,
+   },
[MLX5_EXPANSION_ETH] = {
.next = RTE_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_IPV4,
 MLX5_EXPANSION_IPV6),
@@ -341,6 +350,14 @@ static struct mlx5_flow_tunnel_info tunnels_info[] = {
.tunnel = MLX5_FLOW_LAYER_GRE,
.ptype = RTE_PTYPE_TUNNEL_GRE,
},
+   {
+   .tunnel = MLX5_FLOW_LAYER_MPLS | MLX5_FLOW_LAYER_OUTER_L4_UDP,
+   .ptype = RTE_PTYPE_TUNNEL_MPLS_IN_GRE | RTE_PTYPE_L4_UDP,
+   },
+   {
+   .tunnel = MLX5_FLOW_LAYER_MPLS,
+   .ptype = RTE_PTYPE_TUNNEL_MPLS_IN_GRE,
+   },
 };
 
 /**
@@ -1591,6 +1608,84 @@ mlx5_flow_item_gre(const struct rte_flow_item *item,
return size;
 }
 
+/**
+ * Convert the @p item into a Verbs specification after ensuring the NIC
+ * will understand and process it correctly.
+ * If the necessary size for the conversion is greater than the @p flow_size,
+ * nothing is written in @p flow, the validation is still performed.
+ *
+ * @param[in] item
+ *   Item specification.
+ * @param[in, out] flow
+ *   Pointer to flow structure.
+ * @param[in] flow_size
+ *   Size in bytes of the available space in @p flow, if too small, nothing is
+ *   written.
+ * @param[out] error
+ *   Pointer to error structure.
+ *
+ * @return
+ *   On success the number of bytes consumed/necessary, if the returned value
+ *   is lesser or equal to @p flow_size, the @p item has fully been converted,
+ *   otherwise another call with this returned memory size should be done.
+ *   On error, a negative errno value is returned and rte_errno is set.
+ */
+static int
+mlx5_flow_item_mpls(const struct rte_flow_item *item __rte_unused,
+   struct rte_flow *flow __rte_unused,
+   const size_t flow_size __rte_unused,
+   struct rte_flow_error *error)
+{
+#ifdef HAVE_IBV_DEVICE_MPLS_SUPPORT
+   const struct rte_flow_item_mpls *spec = item->spec;
+   const struct rte_flow_item_mpls *mask = item->mask;
+   unsigned int size = sizeof(struct ibv_flow_spec_mpls);
+   struct ibv_flow_spec_mpls mpls = {
+   .type = IBV_FLOW_SPEC_MPLS,
+   .size = size,
+   };
+   int ret;
+
+   if (flow->l3_protocol_en && flow->l3_protocol != MLX5_IP_PROTOCOL_MPLS)
+   return rte_flow_error_set(error, ENOTSUP,
+ RTE_FLOW

[dpdk-dev] [PATCH v3 21/21] net/mlx5: add count flow action

2018-07-11 Thread Nelio Laranjeiro
This is only supported by Mellanox OFED.

Signed-off-by: Nelio Laranjeiro 
Acked-by: Yongseok Koh 
---
 drivers/net/mlx5/mlx5.h  |   2 +
 drivers/net/mlx5/mlx5_flow.c | 242 +++
 2 files changed, 244 insertions(+)

diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h
index 9949cd3fa..131be334c 100644
--- a/drivers/net/mlx5/mlx5.h
+++ b/drivers/net/mlx5/mlx5.h
@@ -188,6 +188,8 @@ struct priv {
struct mlx5_drop drop_queue; /* Flow drop queues. */
struct mlx5_flows flows; /* RTE Flow rules. */
struct mlx5_flows ctrl_flows; /* Control flow rules. */
+   LIST_HEAD(counters, mlx5_flow_counter) flow_counters;
+   /* Flow counters. */
struct {
uint32_t dev_gen; /* Generation number to flush local caches. */
rte_rwlock_t rwlock; /* MR Lock. */
diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index 4ad581ed9..9ce47 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -88,6 +88,7 @@ extern const struct eth_dev_ops mlx5_dev_ops_isolate;
 /* Modify a packet. */
 #define MLX5_FLOW_MOD_FLAG (1u << 0)
 #define MLX5_FLOW_MOD_MARK (1u << 1)
+#define MLX5_FLOW_MOD_COUNT (1u << 2)
 
 /* possible L3 layers protocols filtering. */
 #define MLX5_IP_PROTOCOL_TCP 6
@@ -249,6 +250,17 @@ struct mlx5_flow_verbs {
uint64_t hash_fields; /**< Verbs hash Rx queue hash fields. */
 };
 
+/* Counters information. */
+struct mlx5_flow_counter {
+   LIST_ENTRY(mlx5_flow_counter) next; /**< Pointer to the next counter. */
+   uint32_t shared:1; /**< Share counter ID with other flow rules. */
+   uint32_t ref_cnt:31; /**< Reference counter. */
+   uint32_t id; /**< Counter ID. */
+   struct ibv_counter_set *cs; /**< Holds the counters for the rule. */
+   uint64_t hits; /**< Number of packets matched by the rule. */
+   uint64_t bytes; /**< Number of bytes matched by the rule. */
+};
+
 /* Flow structure. */
 struct rte_flow {
TAILQ_ENTRY(rte_flow) next; /**< Pointer to the next flow structure. */
@@ -264,6 +276,7 @@ struct rte_flow {
LIST_HEAD(verbs, mlx5_flow_verbs) verbs; /**< Verbs flows list. */
struct mlx5_flow_verbs *cur_verbs;
/**< Current Verbs flow structure being filled. */
+   struct mlx5_flow_counter *counter; /**< Holds Verbs flow counter. */
struct rte_flow_action_rss rss;/**< RSS context. */
uint8_t key[MLX5_RSS_HASH_KEY_LEN]; /**< RSS hash key. */
uint16_t (*queue)[]; /**< Destination queues to redirect traffic to. */
@@ -275,6 +288,7 @@ static const struct rte_flow_ops mlx5_flow_ops = {
.destroy = mlx5_flow_destroy,
.flush = mlx5_flow_flush,
.isolate = mlx5_flow_isolate,
+   .query = mlx5_flow_query,
 };
 
 /* Convert FDIR request to Generic flow. */
@@ -454,6 +468,80 @@ mlx5_flow_adjust_priority(struct rte_eth_dev *dev, struct 
rte_flow *flow)
flow->cur_verbs->attr->priority = priority;
 }
 
+/**
+ * Get a flow counter.
+ *
+ * @param[in] dev
+ *   Pointer to Ethernet device.
+ * @param[in] shared
+ *   Indicate if this counter is shared with other flows.
+ * @param[in] id
+ *   Counter identifier.
+ *
+ * @return
+ *   A pointer to the counter, NULL otherwise and rte_errno is set.
+ */
+static struct mlx5_flow_counter *
+mlx5_flow_counter_new(struct rte_eth_dev *dev, uint32_t shared, uint32_t id)
+{
+   struct priv *priv = dev->data->dev_private;
+   struct mlx5_flow_counter *cnt;
+
+   LIST_FOREACH(cnt, &priv->flow_counters, next) {
+   if (cnt->shared != shared)
+   continue;
+   if (cnt->id != id)
+   continue;
+   cnt->ref_cnt++;
+   return cnt;
+   }
+#ifdef HAVE_IBV_DEVICE_COUNTERS_SET_SUPPORT
+
+   struct mlx5_flow_counter tmpl = {
+   .shared = shared,
+   .id = id,
+   .cs = mlx5_glue->create_counter_set
+   (priv->ctx,
+&(struct ibv_counter_set_init_attr){
+.counter_set_id = id,
+}),
+   .hits = 0,
+   .bytes = 0,
+   };
+
+   if (!tmpl.cs) {
+   rte_errno = errno;
+   return NULL;
+   }
+   cnt = rte_calloc(__func__, 1, sizeof(*cnt), 0);
+   if (!cnt) {
+   rte_errno = ENOMEM;
+   return NULL;
+   }
+   *cnt = tmpl;
+   LIST_INSERT_HEAD(&priv->flow_counters, cnt, next);
+   return cnt;
+#endif
+   rte_errno = ENOTSUP;
+   return NULL;
+}
+
+/**
+ * Release a flow counter.
+ *
+ * @param[in] counter
+ *   Pointer to the counter handler.
+ */
+static void
+mlx5_flow_counter_release(struct mlx5_flow_counter *counter)
+{
+   if (--counter->ref_cnt == 0) {
+   claim_zero(mlx5_glue->destroy_counter_set(counter->cs));
+   LIST_REMOVE(counter,

Re: [dpdk-dev] [PATCH] net/thunderx: Block sq door writes on zero pkts

2018-07-11 Thread Jerin Jacob
-Original Message-
> Date: Wed, 11 Jul 2018 12:52:03 +0530
> From: Kiran Kumar 
> To: dev@dpdk.org
> Cc: jerin.ja...@caviumnetworks.com, maciej.cze...@caviumnetworks.com, Kiran
>  Kumar 
> Subject: [dpdk-dev] [PATCH] net/thunderx: Block sq door writes on zero pkts
> X-Mailer: git-send-email 2.7.4

Use small letter to start the comment.(s/Block/block)

I think, we can change the subject to "avoid sq door bell writes on zero packet"

> 
> With current code, we are performing sq door writes even with 0 pkts.
> this will create pressure on register bus. This patch will block these
> writes.

Could rename to:

Avoid sq door bell write on zero packet case to reduce additional traffic
on register bus.

> 
> Fixes: 1c421f18e0 ("net/thunderx: add single and multi-segment Tx")

add
Cc: sta...@dpdk.org

> 
> Signed-off-by: Kiran Kumar 
> ---
>  drivers/net/thunderx/nicvf_rxtx.c | 12 +++-
>  1 file changed, 7 insertions(+), 5 deletions(-)
> 
> diff --git a/drivers/net/thunderx/nicvf_rxtx.c 
> b/drivers/net/thunderx/nicvf_rxtx.c
> index 72305d9..8075a8e 100644
> --- a/drivers/net/thunderx/nicvf_rxtx.c
> +++ b/drivers/net/thunderx/nicvf_rxtx.c
> @@ -162,12 +162,14 @@ nicvf_xmit_pkts(void *tx_queue, struct rte_mbuf 
> **tx_pkts, uint16_t nb_pkts)
>   free_desc -= TX_DESC_PER_PKT;
>   }
>  
> - sq->tail = tail;
> - sq->xmit_bufs += i;
> - rte_wmb();
> + if (likely(i)) {
> + sq->tail = tail;
> + sq->xmit_bufs += i;
> + rte_wmb();
>  
> - /* Inform HW to xmit the packets */
> - nicvf_addr_write(sq->sq_door, i * TX_DESC_PER_PKT);
> + /* Inform HW to xmit the packets */
> + nicvf_addr_write(sq->sq_door, i * TX_DESC_PER_PKT);
> + }


Please add same logic for nicvf_xmit_pkts_multiseg function.

With above changes:
Acked-by: Jerin Jacob 



>   return i;
>  }
>  
> -- 
> 2.7.4
> 


Re: [dpdk-dev] [PATCH v10 06/27] eal: introduce device class abstraction

2018-07-11 Thread Thomas Monjalon
05/07/2018 13:48, Gaetan Rivet:
> +/**
> + * @file
> + *
> + * DPDK device class interface.
> + *
> + * This file exposes API and interfaces of device classes.

What is the difference between "API" and "interfaces"?
I think you need to explain what is a device class.
You can mention examples ethdev and cryptodev.

> +#define RTE_REGISTER_CLASS(nm, cls) \
> +RTE_INIT_PRIO(classinitfn_ ##nm, CLASS); \

If you remove the ; you can avoid next line.
See this patch which removes such useless line:
https://patches.dpdk.org/patch/41228/

> +static void classinitfn_ ##nm(void) \
> +{\
> + (cls).name = RTE_STR(nm);\

Better to insert a space before \

> + rte_class_register(&cls); \
> +}





Re: [dpdk-dev] [PATCH v10 07/27] eal/class: register destructor

2018-07-11 Thread Thomas Monjalon
05/07/2018 13:48, Gaetan Rivet:
> Signed-off-by: Gaetan Rivet 

If there is no specific comment, this patch can be squashed.





Re: [dpdk-dev] [dpdk-users] Traffic doesn't forward on virtual devices

2018-07-11 Thread Loftus, Ciara
> > >
> > > Bala Sankaran  writes:
> > >
> > > > Perfect!
> > > >
> > > > Thanks for the help.
> > > >
> > > > - Original Message -
> > > >> From: "Keith Wiles" 
> > > >> To: "Bala Sankaran" 
> > > >> Cc: us...@dpdk.org, "Aaron Conole" 
> > > >> Sent: Thursday, July 5, 2018 11:41:46 AM
> > > >> Subject: Re: [dpdk-users] Traffic doesn't forward on virtual devices
> > > >>
> > > >>
> > > >>
> > > >> > On Jul 5, 2018, at 9:53 AM, Bala Sankaran 
> > > wrote:
> > > >> >
> > > >> > Greetings,
> > > >> >
> > > >> > I am currently using dpdk version 17.11.2. I see that there are a few
> > > other
> > > >> > revisions in 17.11.3, followed by the latest stable version of
> > > >> > 18.02.2.
> > > >> >
> > > >> > Based on the issues I have faced so far (see Original
> > > >> > Message below), would you suggest that  I go for
> > > >> > another version? If yes, which one? In essence, my question is,
> would
> > > >> > resorting to a different version of dpdk solve my current issue of
> > > >> > virtqueue id being invalid?
> > > >> >
> > > >> > Any help is much appreciated.
> > > >>
> > > >> From a support perspective using the latest version 18.05 or the long
> > > >> term
> > > >> supported version 17.11.3 is easier for most to help. I would pick the
> > > >> latest release 18.05 myself. As for fixing this problem I do not know.
> > > >> You
> > > >> can look into the MAINTAINERS file and find the maintainers of area(s)
> > > and
> > > >> include them in the CC line on your questions as sometimes they miss
> the
> > > >> emails as the volume can be high at times.
> > >
> > > Thanks Keith.
> > >
> > > I took a quick look and it seems like the queues are not setting up
> > > correctly between OvS and testpmd?  Probably there's a step missing
> > > somewhere, although nothing in either the netdev-dpdk.c from OvS nor
> the
> > > rte_ethdev was obvious to stand out to me.
> > >
> > > I've CC'd Maxime, Ian, and Ciara - maybe they have a better idea to try?
> >
> > Hi,
> >
> > I think the appropriate driver to use in this test on the test-pmd side 
> > might
> > be virtio-user.
> > Follow the same steps just change your vdev test-pmd arguments to:
> > --vdev='net_virtio_user0,path=/usr/local/var/run/openvswitch/vhu0'
> >
> > Thanks,
> > Ciara
> >
> 
> Thank you for your response.
> 
> I tried using virtio-user, but I face an error that says: Failed to prepare
> memory for vhost-user.
> The command I ran is as below:
> 
> [root@localhost openvswitch]# testpmd --socket-mem=1024 --
> vdev='net_virtio_user1,path=/usr/local/var/run/openvswitch/vhu1,server=
> 1'  --vdev='net_tap1,iface=tap1' --file-prefix page1 -- -i
> EAL: Detected 4 lcore(s)
> EAL: Detected 1 NUMA nodes
> EAL: Multi-process socket /var/run/dpdk/page1/mp_socket
> EAL: Probing VFIO support...
> EAL: WARNING: cpu flags constant_tsc=yes nonstop_tsc=no -> using
> unreliable clock cycles !
> rte_pmd_tap_probe(): Initializing pmd_tap for net_tap1 as tap1
> Interactive-mode selected
> Warning: NUMA should be configured manually by using --port-numa-config
> and --ring-numa-config parameters along with --numa.
> testpmd: create a new mbuf pool : n=171456,
> size=2176, socket=0
> testpmd: preferred mempool ops selected: ring_mp_mc
> Configuring Port 0 (socket 0)
> virtio_user_server_reconnect(): WARNING: Some features 0x1801 are not
> supported by vhost-user!
> get_hugepage_file_info(): Exceed maximum of 8
> prepare_vhost_memory_user(): Failed to prepare memory for vhost-user
> Port 0: DA:60:01:0C:4B:29
> Configuring Port 1 (socket 0)
> Port 1: D2:5A:94:68:AF:B3
> Checking link statuses...
> 
> Port 0: LSC event
> Done
> 
> I tried increasing the socket-memory, I checked /proc/meminfo and found
> there were over
> 1280 free hugepages.
> So my understanding is that this is not an issue where I don't have enough
> hugepages.
> 
> Can you provide leads on what's wrong here?

Hi,

The limitations section for Virtio User 
https://doc.dpdk.org/guides/howto/virtio_user_for_container_networking.html#limitations
 states that:

" Cannot work when there are more than VHOST_MEMORY_MAX_NREGIONS(8) hugepages. 
If you have more regions (especially when 2MB hugepages are used), the option, 
--single-file-segments, can help to reduce the number of shared files."

Suggest using the --single-file-segments option on the test-pmd command line or 
failing that increasing your hugepage size to 1G 
https://doc.dpdk.org/guides/linux_gsg/sys_reqs.html#use-of-hugepages-in-the-linux-environment

Thanks,
Ciara

> 
> > >
> > > >> >
> > > >> > Thanks
> > > >> >
> > > >> > - Original Message -
> > > >> >> From: "Bala Sankaran" 
> > > >> >> To: us...@dpdk.org
> > > >> >> Cc: "Aaron Conole" 
> > > >> >> Sent: Thursday, June 28, 2018 3:18:13 PM
> > > >> >> Subject: Traffic doesn't forward on virtual devices
> > > >> >>
> > > >> >>
> > > >> >> Hello team,
> > > >> >>
> > > >> >> I am working on a project to do PVP tests on dpdk. As a first step, 
> > > >> >> I
> > > >> >> would
> > > >> >> 

Re: [dpdk-dev] [PATCH v10 08/27] devargs: add function to parse device layers

2018-07-11 Thread Thomas Monjalon
05/07/2018 13:48, Gaetan Rivet:
> +/**
> + * @internal
> + * Parse a device string and store its information in an
> + * rte_devargs structure.

Please, explain what is a layer.

> + *
> + * Note: if the "data" field of da points to devstr,

Better to use "devargs" as variable name, instead of "da".

> + * then no dynamic allocation is performed and the rte_devargs
> + * can be safely discarded.
> + *
> + * Otherwise ``data`` will hold a workable copy of devstr, that will be
> + * used by layers descriptors within rte_devargs. In this case,
> + * any rte_devargs should be cleaned-up before being freed.
> + *
> + * @param da
> + *   rte_devargs structure to fill.
> + *
> + * @param devstr
> + *   Device string.
> + *
> + * @return
> + *   0 on success.
> + *   Negative errno values on error (rte_errno is set).
> + */
> +int
> +rte_devargs_layers_parse(struct rte_devargs *da,
> +  const char *devstr);
> +
>  #endif /* _EAL_PRIVATE_H_ */
> diff --git a/lib/librte_eal/common/include/rte_devargs.h 
> b/lib/librte_eal/common/include/rte_devargs.h
> index 6c3b6326b..148600258 100644
> --- a/lib/librte_eal/common/include/rte_devargs.h
> +++ b/lib/librte_eal/common/include/rte_devargs.h
> @@ -51,12 +51,19 @@ struct rte_devargs {
>   enum rte_devtype type;
>   /** Device policy. */
>   enum rte_dev_policy policy;
> - /** Bus handle for the device. */
> - struct rte_bus *bus;
>   /** Name of the device. */
>   char name[RTE_DEV_NAME_MAX_LEN];
> + RTE_STD_C11
> + union {
>   /** Arguments string as given by user or "" for no argument. */
> - char *args;
> + char *args;
> + const char *drvstr;
> + };
> + struct rte_bus *bus; /**< bus handle. */
> + struct rte_class *cls; /**< class handle. */

"class" is more readable than "cls"

> + const char *busstr; /**< bus-related part of device string. */

bus_str ?

> + const char *clsstr; /**< bus-related part of device string. */

class_str ?
+ there is a typo in the comment (copy/pasted "bus")

> + const char *data; /**< Device string storage. */
>  };





Re: [dpdk-dev] [PATCH v2 0/6] compress: add Octeontx ZIP compression PMD

2018-07-11 Thread De Lara Guarch, Pablo
Hi Shally,

> -Original Message-
> From: Verma, Shally [mailto:shally.ve...@cavium.com]
> Sent: Wednesday, July 11, 2018 6:53 AM
> To: Verma, Shally ; De Lara Guarch, Pablo
> 
> Cc: dev@dpdk.org; Athreya, Narayana Prasad
> ; Challa, Mahipal
> ; Jacob, Jerin
> ; Sahu, Sunila
> ; Gupta, Ashish 
> Subject: RE: [dpdk-dev] [PATCH v2 0/6] compress: add Octeontx ZIP
> compression PMD
> 
> Hi Pablo,
> 
> Can we close on this?  We're targeting it for 18.08.

Yes, this will go into RC2 to have more time for reviews.

> 
> Thanks
> Shally
> 
> 
> >-Original Message-
> >From: dev [mailto:dev-boun...@dpdk.org] On Behalf Of Shally Verma
> >Sent: 02 July 2018 22:25
> >To: pablo.de.lara.gua...@intel.com
> >Cc: dev@dpdk.org; Athreya, Narayana Prasad
> >; Challa, Mahipal
> >
> >Subject: [dpdk-dev] [PATCH v2 0/6] compress: add Octeontx ZIP
> >compression PMD
> >
> >External Email
> >
> >This patch series add compression PMD for cavium octeontx ZIP module in
> >DPDK compress drivers.
> >Currently PMD only tested for deflate, stateless compression and
> >decompression with direct memory buffers.
> >
> >Changes in v2:
> >- enable OCTEONTX_ZIPVF bydefault, and remove static debug flag
> >- fix meson build in compress/driver/octeontx
> >- fix 32-bit compiler error
> >- remove global declared data structure
> >- header files in alphabetical order
> >- update doc with correct feature support
> >
> >V1 support:
> >Octeontx ZIP PMD implementation
> >Confuguration and Makefile changes to build Octeontx ZIP PMD
> >Documentation ZIP PMD Build support in driver/compress meson for
> >Octeontx ZIP PMD
> >
> >TBDs:
> >Scatter Gather support,
> >Stateful compression/decompression,
> >test for performance
> >
> >This patchset is dependent upon compressdev API.
> >
> >Ashish Gupta (6):
> >  compress/octeontx: add octeontx zip PMD support
> >  compress/octeontx: add device setup PMD ops
> >  compress/octeontx: add xform and stream create support
> >  compress/octeontx: add ops enq deq apis
> >  doc: add Octeonx zip guide
> >  usertools: update devbind for octeontx zip device
> >
> > MAINTAINERS|   5 +
> > config/common_base |   5 +
> > doc/guides/compressdevs/features/octeontx.ini  |  22 +
> > doc/guides/compressdevs/index.rst  |   1 +
> > doc/guides/compressdevs/octeontx.rst   | 107 +++
> > drivers/compress/Makefile  |   1 +
> > drivers/compress/meson.build   |   2 +-
> > drivers/compress/octeontx/Makefile |  30 +
> > drivers/compress/octeontx/include/zip_regs.h   | 721
> +
> > drivers/compress/octeontx/meson.build  |  11 +
> > .../compress/octeontx/rte_pmd_octeontx_version.map |   3 +
> > drivers/compress/octeontx/zip_pmd.c| 625 ++
> > drivers/compress/octeontx/zipvf.c  | 185 ++
> > drivers/compress/octeontx/zipvf.h  | 354 ++
> > mk/rte.app.mk  |   1 +
> > usertools/dpdk-devbind.py  |   9 +
> > 16 files changed, 2081 insertions(+), 1 deletion(-) create mode 100644
> > doc/guides/compressdevs/features/octeontx.ini
> > create mode 100644 doc/guides/compressdevs/octeontx.rst
> > create mode 100644 drivers/compress/octeontx/Makefile
> > create mode 100644 drivers/compress/octeontx/include/zip_regs.h
> > create mode 100644 drivers/compress/octeontx/meson.build
> > create mode 100644
> > drivers/compress/octeontx/rte_pmd_octeontx_version.map
> > create mode 100644 drivers/compress/octeontx/zip_pmd.c
> > create mode 100644 drivers/compress/octeontx/zipvf.c create mode
> > 100644 drivers/compress/octeontx/zipvf.h
> >
> >--
> >2.14.3



[dpdk-dev] [PATCH v2] net/thunderx: avoid sq door bell writes on zero packets

2018-07-11 Thread Kiran Kumar
Avoid sq door bell write on zero packet case to reduce additional
traffic on register bus.

Fixes: 1c421f18e0 ("net/thunderx: add single and multi-segment Tx")
Cc: sta...@dpdk.org

Signed-off-by: Kiran Kumar 
---
 v2 Changes:
 - changed summery and description.
 - added same optimization to nicvf_xmit_pkts_multiseg

 drivers/net/thunderx/nicvf_rxtx.c | 24 ++--
 1 file changed, 14 insertions(+), 10 deletions(-)

diff --git a/drivers/net/thunderx/nicvf_rxtx.c 
b/drivers/net/thunderx/nicvf_rxtx.c
index 72305d9..6e075e2 100644
--- a/drivers/net/thunderx/nicvf_rxtx.c
+++ b/drivers/net/thunderx/nicvf_rxtx.c
@@ -162,12 +162,14 @@ nicvf_xmit_pkts(void *tx_queue, struct rte_mbuf 
**tx_pkts, uint16_t nb_pkts)
free_desc -= TX_DESC_PER_PKT;
}

-   sq->tail = tail;
-   sq->xmit_bufs += i;
-   rte_wmb();
+   if (likely(i)) {
+   sq->tail = tail;
+   sq->xmit_bufs += i;
+   rte_wmb();

-   /* Inform HW to xmit the packets */
-   nicvf_addr_write(sq->sq_door, i * TX_DESC_PER_PKT);
+   /* Inform HW to xmit the packets */
+   nicvf_addr_write(sq->sq_door, i * TX_DESC_PER_PKT);
+   }
return i;
 }

@@ -218,12 +220,14 @@ nicvf_xmit_pkts_multiseg(void *tx_queue, struct rte_mbuf 
**tx_pkts,
}
}

-   sq->tail = tail;
-   sq->xmit_bufs += used_bufs;
-   rte_wmb();
+   if (likely(used_desc)) {
+   sq->tail = tail;
+   sq->xmit_bufs += used_bufs;
+   rte_wmb();

-   /* Inform HW to xmit the packets */
-   nicvf_addr_write(sq->sq_door, used_desc);
+   /* Inform HW to xmit the packets */
+   nicvf_addr_write(sq->sq_door, used_desc);
+   }
return i;
 }

--
2.7.4



Re: [dpdk-dev] [PATCH v10 05/19] eal: enable hotplug on multi-process

2018-07-11 Thread Burakov, Anatoly

On 11-Jul-18 3:11 AM, Zhang, Qi Z wrote:

+/* Max length for a bus name */
+#define RTE_BUS_NAME_MAX_LEN 32


Is this enforced anywhere in the bus codebase? Can we guarantee that
bus name will never be bigger than this?


I think 32 should be enough for a bus name even in future.


Sorry, I missed your point, I think it is not enforced, we still can add a new 
bus exceed 32,
but for RTE_DEV_NAME_MAX_LEN which is used in rte_devargs to enforce all device 
name not exceed 64.
So, it's better to move RTE_BUS_NAME_MAX_LEN into hotplug_mp as internal , and 
this can be regarded as a limitation for hotplug so far, though it should be 
enough for all exist cases.
And same for RTE_DEV_ARGS_MAX_LEN.


Can we fix it in this patchset, or would that involve an ABI break of 
some sort?


--
Thanks,
Anatoly


Re: [dpdk-dev] [PATCH v10 08/27] devargs: add function to parse device layers

2018-07-11 Thread Gaëtan Rivet
Hi Thomas,

On Wed, Jul 11, 2018 at 10:19:07AM +0200, Thomas Monjalon wrote:
> 05/07/2018 13:48, Gaetan Rivet:
> > +/**
> > + * @internal
> > + * Parse a device string and store its information in an
> > + * rte_devargs structure.
> 
> Please, explain what is a layer.
> 
> > + *
> > + * Note: if the "data" field of da points to devstr,
> 
> Better to use "devargs" as variable name, instead of "da".
> 
> > + * then no dynamic allocation is performed and the rte_devargs
> > + * can be safely discarded.
> > + *
> > + * Otherwise ``data`` will hold a workable copy of devstr, that will be
> > + * used by layers descriptors within rte_devargs. In this case,
> > + * any rte_devargs should be cleaned-up before being freed.
> > + *
> > + * @param da
> > + *   rte_devargs structure to fill.
> > + *
> > + * @param devstr
> > + *   Device string.
> > + *
> > + * @return
> > + *   0 on success.
> > + *   Negative errno values on error (rte_errno is set).
> > + */
> > +int
> > +rte_devargs_layers_parse(struct rte_devargs *da,
> > +const char *devstr);
> > +
> >  #endif /* _EAL_PRIVATE_H_ */
> > diff --git a/lib/librte_eal/common/include/rte_devargs.h 
> > b/lib/librte_eal/common/include/rte_devargs.h
> > index 6c3b6326b..148600258 100644
> > --- a/lib/librte_eal/common/include/rte_devargs.h
> > +++ b/lib/librte_eal/common/include/rte_devargs.h
> > @@ -51,12 +51,19 @@ struct rte_devargs {
> > enum rte_devtype type;
> > /** Device policy. */
> > enum rte_dev_policy policy;
> > -   /** Bus handle for the device. */
> > -   struct rte_bus *bus;
> > /** Name of the device. */
> > char name[RTE_DEV_NAME_MAX_LEN];
> > +   RTE_STD_C11
> > +   union {
> > /** Arguments string as given by user or "" for no argument. */
> > -   char *args;
> > +   char *args;
> > +   const char *drvstr;
> > +   };
> > +   struct rte_bus *bus; /**< bus handle. */
> > +   struct rte_class *cls; /**< class handle. */
> 
> "class" is more readable than "cls"
> 

I was thinking that maybe this could be a problem when trying to build
with C++. The EAL headers in DPDK are meant to be included in C++ apps,
I think ``class`` would be an issue then.

If I'm mistaken, then sure, class is a better name.

> > +   const char *busstr; /**< bus-related part of device string. */
> 
> bus_str ?
> 
> > +   const char *clsstr; /**< bus-related part of device string. */
> 
> class_str ?
> + there is a typo in the comment (copy/pasted "bus")
> 
> > +   const char *data; /**< Device string storage. */
> >  };
> 
> 
> 

Otherwise, ok with the rest of the comments, will fix (as well as the
previous emails). Thanks for reading.

-- 
Gaëtan Rivet
6WIND


[dpdk-dev] [PATCH v2] ethdev: check queue stats mapping input arguments

2018-07-11 Thread Kiran Kumar
With current implementation, we are not checking for queue_id range
and stat_idx range in stats mapping function. This patch will add
check for queue_id and stat_idx range.

Fixes: 5de201df892 ("ethdev: add stats per queue")

Signed-off-by: Kiran Kumar 
Acked-by: Andrew Rybchenko 
---
 v2 changes:
 - changed summery

 lib/librte_ethdev/rte_ethdev.c | 10 ++
 1 file changed, 10 insertions(+)

diff --git a/lib/librte_ethdev/rte_ethdev.c b/lib/librte_ethdev/rte_ethdev.c
index a9977df..0849016 100644
--- a/lib/librte_ethdev/rte_ethdev.c
+++ b/lib/librte_ethdev/rte_ethdev.c
@@ -2457,6 +2457,16 @@ set_queue_stats_mapping(uint16_t port_id, uint16_t 
queue_id, uint8_t stat_idx,
dev = &rte_eth_devices[port_id];

RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->queue_stats_mapping_set, 
-ENOTSUP);
+
+   if (is_rx && (queue_id >= dev->data->nb_rx_queues))
+   return -EINVAL;
+
+   if (!is_rx && (queue_id >= dev->data->nb_tx_queues))
+   return -EINVAL;
+
+   if (stat_idx >= RTE_ETHDEV_QUEUE_STAT_CNTRS)
+   return -EINVAL;
+
return (*dev->dev_ops->queue_stats_mapping_set)
(dev, queue_id, stat_idx, is_rx);
 }
--
2.7.4



Re: [dpdk-dev] [PATCH v11 03/19] bus/pci: enable vfio unmap resource for secondary

2018-07-11 Thread Burakov, Anatoly

On 11-Jul-18 4:09 AM, Qi Zhang wrote:

Subroutine to unmap VFIO resource is shared by secondary and
primary, and it does not work on the secondary process. Since
for secondary process, it is not necessary to close interrupt
handler, set pci bus mastering and remove vfio_res from
vfio_res_list. So, the patch adds a dedicate function to handle
the situation when a device is unmapped on a secondary process.

Signed-off-by: Qi Zhang 
---

Reviewed-by: Anatoly Burakov 

--
Thanks,
Anatoly


[dpdk-dev] [PATCH] net/i40e: fix FDIR check programming status error

2018-07-11 Thread Wei Zhao
In i40e FDIR PMD code for checking programming status
function i40e_check_fdir_programming_status(), the initial value
of return value ret should be set to -1 not 0, because if DD bit of
I40E_RX_DESC_STATUS_DD is not write back, this function will return
0 to upper function, this give an error info to upper function, the
fact for this is it is time out for DD write back and it should return
-1.

Fixes: 05999aab4ca6 ("i40e: add or delete flow director")
Signed-off-by: Wei Zhao 
---
 drivers/net/i40e/i40e_fdir.c | 3 ++-
 1 file changed, 2 insertions(+), 1 deletion(-)

diff --git a/drivers/net/i40e/i40e_fdir.c b/drivers/net/i40e/i40e_fdir.c
index d41601a..b958bf6 100644
--- a/drivers/net/i40e/i40e_fdir.c
+++ b/drivers/net/i40e/i40e_fdir.c
@@ -1315,7 +1315,7 @@ i40e_check_fdir_programming_status(struct i40e_rx_queue 
*rxq)
uint32_t rx_status;
uint32_t len, id;
uint32_t error;
-   int ret = 0;
+   int ret = -1;
 
rxdp = &rxq->rx_ring[rxq->rx_tail];
qword1 = rte_le_to_cpu_64(rxdp->wb.qword1.status_error_len);
@@ -1360,6 +1360,7 @@ i40e_check_fdir_programming_status(struct i40e_rx_queue 
*rxq)
I40E_PCI_REG_WRITE(rxq->qrx_tail, rxq->nb_rx_desc - 1);
else
I40E_PCI_REG_WRITE(rxq->qrx_tail, rxq->rx_tail - 1);
+   ret = 0;
}
 
return ret;
-- 
2.7.5



Re: [dpdk-dev] [PATCH] compress/isal: fixes ISA-L PMD used with offsets in mbuf

2018-07-11 Thread De Lara Guarch, Pablo
Hi Lee,

> -Original Message-
> From: Daly, Lee
> Sent: Tuesday, July 10, 2018 1:44 PM
> To: De Lara Guarch, Pablo 
> Cc: dev@dpdk.org; Daly, Lee ; sta...@dpdk.org
> Subject: [PATCH] compress/isal: fixes ISA-L PMD used with offsets in mbuf
> 

No need to use ISA-L PMD in the title, as it is a duplicate.
Also, titles always start with verb in infinitive (fix).
So, maybe a suggestion could be: "compress/isal: fix offset usage"?


> This patch allows the ISA-L compression PMD, to be used with offsets in the
> mbuf.
> Offsets can now be used for source and destination buffers, during compression
> or decompression.
> 
> Fixes: dc49e6a "compress/isal: add ISA-L compression functionality"
> Fixes: 7bf4f06 "compress/isal: add ISA-L decomp functionality"
> Cc: sta...@dpdk.org
> 
> Signed-off-by: Lee Daly 

...

> + qp->stream->end_of_stream = 1; /* All input consumed in one go */
> + if ((op->src.length + op->src.offset) > op->m_src->data_len) {
> + ISAL_PMD_LOG(ERR, "Input mbuf not big enough for
> offset.\n");

I would rephrase this to "Input buffer not big enough for the length and offset 
provided".
Same for inflate.

> + op->status = RTE_COMP_OP_STATUS_INVALID_ARGS;
> + return -1;
> + }
> + /* Point compression stream to input buffer */
> + qp->stream->avail_in = op->src.length;
> + qp->stream->next_in = rte_pktmbuf_mtod_offset(op->m_src, uint8_t *,
> + op->src.offset);
> +
> + if (op->dst.offset > op->m_dst->data_len) {
> + ISAL_PMD_LOG(ERR, "Output mbuf not big enough for
> offset.\n");

I would rephrase this to "Output buffer not big enough for the offset provided".
Same for inflate.

> + op->status = RTE_COMP_OP_STATUS_INVALID_ARGS;
> + return -1;
> + }



Re: [dpdk-dev] [PATCH v4 1/4] ethdev: Add API to enable device event handler

2018-07-11 Thread Andrew Rybchenko

On 10.07.2018 15:51, Jeff Guo wrote:

Implement a couple of eal device event handler install/uninstall APIs
in ethdev, it could let PMDs have chance to manage the eal device event,
such as register device event callback, then monitor and process the
hotplug event.


I think it is moving in right direction, but my problem with the patch is
that I don't understand what prevents us to make it even more generic.
Right now it looks like that PCI driver flag RTE_PCI_DRV_INTR_RMV is
sufficient and everything else could be done on ethdev layer.
RTE_PCI_DRV_INTR_RMV  is mapped to RTE_ETH_DEV_INTR_RMV eth
device flag. The flag may be used as an indication in rte_eth_dev_create()
to register the callback.
May be I'm completely wrong above, but if so, I'd like to understand why
and prefer to see explanations in the patch description.


Signed-off-by: Jeff Guo 
---
v4->v3:
change to use eal device event handler install api
---
  doc/guides/rel_notes/release_18_08.rst | 12 +++
  lib/librte_ethdev/rte_ethdev.c | 59 ++
  lib/librte_ethdev/rte_ethdev_driver.h  | 32 ++
  3 files changed, 103 insertions(+)

diff --git a/doc/guides/rel_notes/release_18_08.rst 
b/doc/guides/rel_notes/release_18_08.rst
index bc01242..b6482ce 100644
--- a/doc/guides/rel_notes/release_18_08.rst
+++ b/doc/guides/rel_notes/release_18_08.rst
@@ -46,6 +46,18 @@ New Features
Flow API support has been added to CXGBE Poll Mode Driver to offload
flows to Chelsio T5/T6 NICs.
  
+* **Added eal device event process helper in ethdev.**

+
+  Implement a couple of eal device event handler install/uninstall APIs in
+  ethdev, these helper could let PMDs have chance to manage the eal device
+  event, such as register device event callback, then monitor and process
+  hotplug event.
+
+  * ``rte_eth_dev_event_handler_install`` for PMDs use to install the device
+event handler.
+  * ``rte_eth_dev_event_handler_uninstall`` for PMDs use to uninstall the 
device
+event handler.
+
  
  API Changes

  ---
diff --git a/lib/librte_ethdev/rte_ethdev.c b/lib/librte_ethdev/rte_ethdev.c
index a9977df..09ea02d 100644
--- a/lib/librte_ethdev/rte_ethdev.c
+++ b/lib/librte_ethdev/rte_ethdev.c
@@ -4518,6 +4518,65 @@ rte_eth_devargs_parse(const char *dargs, struct 
rte_eth_devargs *eth_da)
return result;
  }
  
+static void __rte_experimental

+eth_dev_event_callback(char *device_name, enum rte_dev_event_type type,
+void *arg)
+{
+   struct rte_eth_dev *eth_dev = (struct rte_eth_dev *)arg;
+
+   switch (type) {
+   case RTE_DEV_EVENT_REMOVE:
+   ethdev_log(INFO, "The device: %s has been removed!\n",


Colon after 'device' above looks strange for me. I'd suggest to remove it.
If so, similar below for ADD.


+   device_name);
+
+   if (!device_name || !eth_dev)
+   return;
+
+   if (!(eth_dev->data->dev_flags & RTE_ETH_EVENT_INTR_RMV))


It looks like a typo above. Shouldn't it be RTE_ETH_DEV_INTR_RMV?


+   return;
+
+   if (!strcmp(device_name, eth_dev->device->name))
+   _rte_eth_dev_callback_process(eth_dev,
+ RTE_ETH_EVENT_INTR_RMV,
+ NULL);
+   break;
+   case RTE_DEV_EVENT_ADD:
+   ethdev_log(INFO, "The device: %s has been added!\n",
+   device_name);
+   break;
+   default:
+   break;
+   }
+}
+
+int __rte_experimental
+rte_eth_dev_event_handler_install(struct rte_eth_dev *eth_dev)
+{
+   int result = 0;
+
+   result = rte_dev_event_callback_register(eth_dev->device->name,
+   eth_dev_event_callback, eth_dev);
+   if (result)
+   RTE_LOG(ERR, EAL, "device event callback register failed for "
+   "device:%s!\n", eth_dev->device->name);


Why ethdev_log() is used above, but here is RTE_LOG with EAL?
Similar question below.


+
+   return result;
+}
+
+int __rte_experimental
+rte_eth_dev_event_handler_uninstall(struct rte_eth_dev *eth_dev)
+{
+   int result = 0;
+
+   result = rte_dev_event_callback_unregister(eth_dev->device->name,
+   eth_dev_event_callback, eth_dev);
+   if (result)
+   RTE_LOG(ERR, EAL, "device event callback unregister failed for"
+   " device:%s!\n", eth_dev->device->name);
+
+   return result;
+}
+
  RTE_INIT(ethdev_init_log);
  static void
  ethdev_init_log(void)


<...>


Re: [dpdk-dev] [PATCH v2] net/thunderx: avoid sq door bell writes on zero packets

2018-07-11 Thread Jerin Jacob
-Original Message-
> Date: Wed, 11 Jul 2018 13:54:36 +0530
> From: Kiran Kumar 
> To: jerin.ja...@caviumnetworks.com, maciej.cze...@caviumnetworks.com,
>  ferruh.yi...@intel.com
> Cc: dev@dpdk.org, Kiran Kumar ,
>  sta...@dpdk.org
> Subject: [dpdk-dev] [PATCH v2] net/thunderx: avoid sq door bell writes on
>  zero packets
> X-Mailer: git-send-email 2.7.4
> 
> Avoid sq door bell write on zero packet case to reduce additional
> traffic on register bus.
> 
> Fixes: 1c421f18e0 ("net/thunderx: add single and multi-segment Tx")
> Cc: sta...@dpdk.org
> 
> Signed-off-by: Kiran Kumar 

Acked-by: Jerin Jacob 

> ---
>  v2 Changes:
>  - changed summery and description.
>  - added same optimization to nicvf_xmit_pkts_multiseg
> 
>  drivers/net/thunderx/nicvf_rxtx.c | 24 ++--
>  1 file changed, 14 insertions(+), 10 deletions(-)
> 
> diff --git a/drivers/net/thunderx/nicvf_rxtx.c 
> b/drivers/net/thunderx/nicvf_rxtx.c
> index 72305d9..6e075e2 100644
> --- a/drivers/net/thunderx/nicvf_rxtx.c
> +++ b/drivers/net/thunderx/nicvf_rxtx.c
> @@ -162,12 +162,14 @@ nicvf_xmit_pkts(void *tx_queue, struct rte_mbuf 
> **tx_pkts, uint16_t nb_pkts)
>   free_desc -= TX_DESC_PER_PKT;
>   }
> 
> - sq->tail = tail;
> - sq->xmit_bufs += i;
> - rte_wmb();
> + if (likely(i)) {
> + sq->tail = tail;
> + sq->xmit_bufs += i;
> + rte_wmb();
> 
> - /* Inform HW to xmit the packets */
> - nicvf_addr_write(sq->sq_door, i * TX_DESC_PER_PKT);
> + /* Inform HW to xmit the packets */
> + nicvf_addr_write(sq->sq_door, i * TX_DESC_PER_PKT);
> + }
>   return i;
>  }
> 
> @@ -218,12 +220,14 @@ nicvf_xmit_pkts_multiseg(void *tx_queue, struct 
> rte_mbuf **tx_pkts,
>   }
>   }
> 
> - sq->tail = tail;
> - sq->xmit_bufs += used_bufs;
> - rte_wmb();
> + if (likely(used_desc)) {
> + sq->tail = tail;
> + sq->xmit_bufs += used_bufs;
> + rte_wmb();
> 
> - /* Inform HW to xmit the packets */
> - nicvf_addr_write(sq->sq_door, used_desc);
> + /* Inform HW to xmit the packets */
> + nicvf_addr_write(sq->sq_door, used_desc);
> + }
>   return i;
>  }
> 
> --
> 2.7.4
> 


Re: [dpdk-dev] [PATCH v11 01/19] ethdev: add function to release port in local process

2018-07-11 Thread Andrew Rybchenko

On 11.07.2018 06:08, Qi Zhang wrote:

Add driver API rte_eth_release_port_private to support the
case when an ethdev need to be detached on a secondary process.
Local state is set to unused and shared data will not be reset
so the primary process can still use it.

Signed-off-by: Qi Zhang 
Reviewed-by: Andrew Rybchenko 
Acked-by: Remy Horton 
---
  lib/librte_ethdev/rte_ethdev.c| 12 
  lib/librte_ethdev/rte_ethdev_driver.h | 16 +++-
  lib/librte_ethdev/rte_ethdev_pci.h|  8 
  3 files changed, 35 insertions(+), 1 deletion(-)

diff --git a/lib/librte_ethdev/rte_ethdev.c b/lib/librte_ethdev/rte_ethdev.c
index a9977df97..52a97694c 100644
--- a/lib/librte_ethdev/rte_ethdev.c
+++ b/lib/librte_ethdev/rte_ethdev.c
@@ -359,6 +359,18 @@ rte_eth_dev_attach_secondary(const char *name)
  }
  
  int

+rte_eth_dev_release_port_private(struct rte_eth_dev *eth_dev)
+{
+   if (eth_dev == NULL)
+   return -EINVAL;
+
+   _rte_eth_dev_callback_process(eth_dev, RTE_ETH_EVENT_DESTROY, NULL);
+   eth_dev->state = RTE_ETH_DEV_UNUSED;
+
+   return 0;
+}
+
+int
  rte_eth_dev_release_port(struct rte_eth_dev *eth_dev)
  {
if (eth_dev == NULL)
diff --git a/lib/librte_ethdev/rte_ethdev_driver.h 
b/lib/librte_ethdev/rte_ethdev_driver.h
index c9c825e3f..269586d88 100644
--- a/lib/librte_ethdev/rte_ethdev_driver.h
+++ b/lib/librte_ethdev/rte_ethdev_driver.h
@@ -62,7 +62,7 @@ struct rte_eth_dev *rte_eth_dev_attach_secondary(const char 
*name);
   * Release the specified ethdev port.
   *
   * @param eth_dev
- * The *eth_dev* pointer is the address of the *rte_eth_dev* structure.
+ * Device to be detached.
   * @return
   *   - 0 on success, negative on error
   */
@@ -70,6 +70,20 @@ int rte_eth_dev_release_port(struct rte_eth_dev *eth_dev);
  
  /**

   * @internal
+ * Release the specified ethdev port in the local process.
+ * Only set ethdev state to unused, but not reset shared data since
+ * it assume other processes is still using it. typically it is
+ * called by a secondary process.
+ *
+ * @param eth_dev
+ * Device to be detached.
+ * @return
+ *   - 0 on success, negative on error
+ */
+int rte_eth_dev_release_port_private(struct rte_eth_dev *eth_dev);
+
+/**
+ * @internal
   * Release device queues and clear its configuration to force the user
   * application to reconfigure it. It is for internal use only.
   *
diff --git a/lib/librte_ethdev/rte_ethdev_pci.h 
b/lib/librte_ethdev/rte_ethdev_pci.h
index 2cfd37274..a46d9e182 100644
--- a/lib/librte_ethdev/rte_ethdev_pci.h
+++ b/lib/librte_ethdev/rte_ethdev_pci.h
@@ -197,6 +197,14 @@ rte_eth_dev_pci_generic_remove(struct rte_pci_device 
*pci_dev,
if (!eth_dev)
return -ENODEV;
  
+	/**

+* PCI device can only be globally detached directly by a
+* primary process. In secondary process, we only need to
+* release port.
+*/
+   if (rte_eal_process_type() != RTE_PROC_PRIMARY)
+   return rte_eth_dev_release_port_private(eth_dev);


I've realized that some uninit functions which will not be called anymore
in secondary processes have check for process type and handling of
secondary process case. It makes code inconsistent and should be fixed.


+
if (dev_uninit) {
ret = dev_uninit(eth_dev);
if (ret)




Re: [dpdk-dev] [PATCH v10 08/27] devargs: add function to parse device layers

2018-07-11 Thread Thomas Monjalon
11/07/2018 10:41, Gaëtan Rivet:
> > > + struct rte_bus *bus; /**< bus handle. */
> > > + struct rte_class *cls; /**< class handle. */
> > 
> > "class" is more readable than "cls"
> > 
> 
> I was thinking that maybe this could be a problem when trying to build
> with C++. The EAL headers in DPDK are meant to be included in C++ apps,
> I think ``class`` would be an issue then.

Yes, right.

So only options are "cls" or "devclass".
Up to you.

> If I'm mistaken, then sure, class is a better name.





[dpdk-dev] [PATCH] ethdev: fix port ID retrieval on vdev attach

2018-07-11 Thread Andrew Rybchenko
From: Ivan Malov 

Attaching a vdev port may result in multiple
ports actually added because a vdev port may
have slave devices to be attached implicitly.

Ethdev attach API has to fill in the port ID
to be read back by the user and what it does
is take the last assigned ID from the common
list after attach completion. Such an ID may
belong to a slave device and not to the vdev.

This mistake must be precluded by requesting
the port ID by name of device being attached.

Fixes: b0fb26685570 ("ethdev: convert to EAL hotplug")
Cc: David Marchand 
Cc: sta...@dpdk.org

Reported-by: Andrew Rybchenko 
Signed-off-by: Ivan Malov 
Signed-off-by: Andrew Rybchenko 
---
 lib/librte_ethdev/rte_ethdev.c | 26 --
 1 file changed, 4 insertions(+), 22 deletions(-)

diff --git a/lib/librte_ethdev/rte_ethdev.c b/lib/librte_ethdev/rte_ethdev.c
index 16b8258a7..dbb9244ec 100644
--- a/lib/librte_ethdev/rte_ethdev.c
+++ b/lib/librte_ethdev/rte_ethdev.c
@@ -46,7 +46,6 @@ int rte_eth_dev_logtype;
 
 static const char *MZ_RTE_ETH_DEV_DATA = "rte_eth_dev_data";
 struct rte_eth_dev rte_eth_devices[RTE_MAX_ETHPORTS];
-static uint16_t eth_dev_last_created_port;
 
 /* spinlock for eth device callbacks */
 static rte_spinlock_t rte_eth_dev_cb_lock = RTE_SPINLOCK_INITIALIZER;
@@ -284,8 +283,6 @@ eth_dev_get(uint16_t port_id)
 
eth_dev->data = &rte_eth_dev_shared_data->data[port_id];
 
-   eth_dev_last_created_port = port_id;
-
return eth_dev;
 }
 
@@ -646,7 +643,6 @@ eth_err(uint16_t port_id, int ret)
 int
 rte_eth_dev_attach(const char *devargs, uint16_t *port_id)
 {
-   int current = rte_eth_dev_count_total();
struct rte_devargs da;
int ret = -1;
 
@@ -665,24 +661,10 @@ rte_eth_dev_attach(const char *devargs, uint16_t *port_id)
if (ret < 0)
goto err;
 
-   /* no point looking at the port count if no port exists */
-   if (!rte_eth_dev_count_total()) {
-   RTE_ETHDEV_LOG(ERR, "No port found for device (%s)\n", da.name);
-   ret = -1;
-   goto err;
-   }
-
-   /* if nothing happened, there is a bug here, since some driver told us
-* it did attach a device, but did not create a port.
-* FIXME: race condition in case of plug-out of another device
-*/
-   if (current == rte_eth_dev_count_total()) {
-   ret = -1;
-   goto err;
-   }
-
-   *port_id = eth_dev_last_created_port;
-   ret = 0;
+   ret = rte_eth_dev_get_port_by_name(da.name, port_id);
+   if (ret != 0)
+   RTE_ETHDEV_LOG(ERR, "No port found for device (%s)\n",
+  da.name);
 
 err:
free(da.args);
-- 
2.17.1



Re: [dpdk-dev] [pull-request] next-crypto 18.08-rc1

2018-07-11 Thread Thomas Monjalon
10/07/2018 15:38, Pablo de Lara:
>   http://dpdk.org/git/next/dpdk-next-crypto 

Error when compiling x86_64-native-linuxapp-clang+shared+next+debug:
  LD l2fwd-crypto
undefined reference to `rte_cryptodev_scheduler_slaves_get'

seen on patch
examples/l2fwd-crypto: limit number of sessions





Re: [dpdk-dev] [PATCH v10 10/27] eal/class: add device iteration

2018-07-11 Thread Thomas Monjalon
05/07/2018 13:48, Gaetan Rivet:
> Signed-off-by: Gaetan Rivet 
> ---
>  lib/librte_eal/common/include/rte_class.h | 1 +
>  1 file changed, 1 insertion(+)
> 
> diff --git a/lib/librte_eal/common/include/rte_class.h 
> b/lib/librte_eal/common/include/rte_class.h
> index e8176f5e1..9d5b06807 100644
> --- a/lib/librte_eal/common/include/rte_class.h
> +++ b/lib/librte_eal/common/include/rte_class.h
> @@ -30,6 +30,7 @@ TAILQ_HEAD(rte_class_list, rte_class);
>  struct rte_class {
>   TAILQ_ENTRY(rte_class) next; /**< Next device class in linked list */
>   const char *name; /**< Name of the class */
> + rte_dev_iterate_t dev_iterate; /**< Device iterator. */
>  };

Patches 10 and 11 can be squashed in 9.




Re: [dpdk-dev] [PATCH v8 7/7] igb_uio: fix uio release issue for hotplug

2018-07-11 Thread Jeff Guo




On 7/11/2018 10:46 AM, Jeff Guo wrote:



On 7/11/2018 5:52 AM, Stephen Hemminger wrote:

On Tue, 10 Jul 2018 19:03:27 +0800
Jeff Guo  wrote:

When hotplug out device, the device resource will be released in 
kernel.
The fd sys file will disappear, and the irq will be released. At 
this time,
if igb uio driver still try to release this resource, it will cause 
kernel

crash. On the other hand, interrupt disabling do not automatically be
processed in kernel. If not handle it, this redundancy and dirty 
thing will

affect the interrupt resource be used by other device. So the igb_uio
driver have to check the hotplug status, and the corresponding process
should be taken in igb uio driver.

This patch propose to add enum rte_udev_state into struct 
rte_uio_pci_dev

of igb uio driver, which will record the state of uio device, such as
probed/opened/released/removed. When detect the unexpected removal 
which

cause of hotplug out behavior, it will corresponding disable interrupt
resource. For the part of releasement which kernel have already handle,
just skip it to avoid double free or null pointer crash issue.

Signed-off-by: Jeff Guo 

The PCI hotplug management is an important and potentially error prone
error of DPDK.

I realize that English is not your native language, but the commit 
messages
for this are hard to read. Perhaps you can get a volunteer or other 
person

in the community to reword them. The commit logs and comments contain
important information about the documentation of the code.


yes, i think that it might not be the whole thing to let you confused 
except something specific. But definitely it is my task to let 
reviewer most easily know what i want to propose before they will ack it,
especial for some complex case. let's try my best for check my word. I 
am also planning to go to more native English country and great 
conference study, anyway to improve my word : )



How does VFIO handle hotplug? We should direct all users to use VFIO
since it is supported and secure. Igb uio has always been a slightly
dangerous (as they say "running with scissors") way of accessing 
devices.


You exposure VFIO here to replace igo uio, maybe we should check if it 
is optional or not.
If we can fix all igb_uio issue it should be optional, if only vfio 
show stable we should go to vfio only.

What other else guy comment here?


Plus, the pci vfio hotplug enabling is still on the next plan, since the 
different framework between vifo and uio for uevent process.
Per vfio, when user space control  it will holding the resource so the 
uevent will be blocked to sent out from kernel.
Anyway that should be another story. We hope this patch set just fix 
hotplug failure issue for igb uio case.


Re: [dpdk-dev] [PATCH] ethdev: fix port ID retrieval on vdev attach

2018-07-11 Thread Thomas Monjalon
11/07/2018 11:49, Andrew Rybchenko:
> From: Ivan Malov 
> 
> Attaching a vdev port may result in multiple
> ports actually added because a vdev port may
> have slave devices to be attached implicitly.
> 
> Ethdev attach API has to fill in the port ID
> to be read back by the user and what it does
> is take the last assigned ID from the common
> list after attach completion. Such an ID may
> belong to a slave device and not to the vdev.
> 
> This mistake must be precluded by requesting
> the port ID by name of device being attached.

No, the real issue is using this function rte_eth_dev_attach.
It is broken since day 1.
Mixing EAL devargs and ethdev port cannot work by design.

If you want to attach a device, you must add it with
rte_eal_hotplug_add
and wait for the driver to probe the ports
(which can be ethdev or other classes).

We should take these actions in 18.08:
- deprecate rte_eth_dev_attach/rte_eth_dev_detach
- deprecate rte_eal_dev_attach/rte_eal_dev_detach
I did not take time to send the deprecation notices yet.
Feel free to deprecate these functions yourself.




Re: [dpdk-dev] [dpdk-stable] [PATCH v4 1/5] mem: add function for checking memsegs IOVAs addresses

2018-07-11 Thread Eelco Chaudron



On 10 Jul 2018, at 19:25, Alejandro Lucero wrote:

> A device can suffer addressing limitations. This functions checks
> memsegs have iovas within the supported range based on dma mask.
>
> PMD should use this during initialization if supported devices
> suffer addressing limitations, returning an error if this function
> returns memsegs out of range.
>
> Another potential usage is for emulated IOMMU hardware with addressing
> limitations.
>
> Applicable to v17.11.3 only.
>
> Signed-off-by: Alejandro Lucero 
> Acked-by: Anatoly Burakov 
> ---
Looks good to me.

Acked-by: Eelco Chaudron 



Re: [dpdk-dev] [PATCH] ethdev: fix port ID retrieval on vdev attach

2018-07-11 Thread Andrew Rybchenko

On 11.07.2018 13:02, Thomas Monjalon wrote:

11/07/2018 11:49, Andrew Rybchenko:

From: Ivan Malov 

Attaching a vdev port may result in multiple
ports actually added because a vdev port may
have slave devices to be attached implicitly.

Ethdev attach API has to fill in the port ID
to be read back by the user and what it does
is take the last assigned ID from the common
list after attach completion. Such an ID may
belong to a slave device and not to the vdev.

This mistake must be precluded by requesting
the port ID by name of device being attached.

No, the real issue is using this function rte_eth_dev_attach.
It is broken since day 1.
Mixing EAL devargs and ethdev port cannot work by design.

If you want to attach a device, you must add it with
rte_eal_hotplug_add
and wait for the driver to probe the ports
(which can be ethdev or other classes).


Yes, I totally agree. Doesn't it deserve to be fixed at least in stable?


We should take these actions in 18.08:
- deprecate rte_eth_dev_attach/rte_eth_dev_detach
- deprecate rte_eal_dev_attach/rte_eal_dev_detach
I did not take time to send the deprecation notices yet.
Feel free to deprecate these functions yourself.


Should we deprecate it right now in 18.08 and schedule removal to 18.11?
Or just add deprecation notice in 18.08, mark deprecated in 18.11 and
remove later?




Re: [dpdk-dev] [dpdk-stable] [PATCH v4 2/5] bus/pci: use IOVAs check when setting IOVA mode

2018-07-11 Thread Eelco Chaudron




On 10 Jul 2018, at 19:25, Alejandro Lucero wrote:


Although VT-d emulation currently only supports 39 bits, it could
be iovas being within that supported range. This patch allows
IOVA mode in such a case.

Indeed, memory initialization code can be modified for using lower
virtual addresses than those used by the kernel for 64 bits processes
by default, and therefore memsegs iovas can use 39 bits or less for
most system. And this is likely 100% true for VMs.

Applicable to v17.11.3 only.

Signed-off-by: Alejandro Lucero 
---
 drivers/bus/pci/linux/pci.c | 15 +++
 1 file changed, 11 insertions(+), 4 deletions(-)

diff --git a/drivers/bus/pci/linux/pci.c b/drivers/bus/pci/linux/pci.c
index 74deef3..792c819 100644
--- a/drivers/bus/pci/linux/pci.c
+++ b/drivers/bus/pci/linux/pci.c
@@ -43,6 +43,7 @@
 #include 
 #include 
 #include 
+#include 

 #include "eal_private.h"
 #include "eal_filesystem.h"
@@ -613,10 +614,12 @@
fclose(fp);

 	mgaw = ((vtd_cap_reg & VTD_CAP_MGAW_MASK) >> VTD_CAP_MGAW_SHIFT) + 
1;

-   if (mgaw < X86_VA_WIDTH)
+
+   if (!rte_eal_check_dma_mask(mgaw))
+   return true;
+   else
return false;

-   return true;
 }
 #elif defined(RTE_ARCH_PPC_64)
 static bool
@@ -640,13 +643,17 @@
 {
struct rte_pci_device *dev = NULL;
struct rte_pci_driver *drv = NULL;
+   int iommu_dma_mask_check_done = 0;

FOREACH_DRIVER_ON_PCIBUS(drv) {
FOREACH_DEVICE_ON_PCIBUS(dev) {
if (!rte_pci_match(drv, dev))
continue;
-   if (!pci_one_device_iommu_support_va(dev))
-   return false;
+   if (!iommu_dma_mask_check_done) {
+   if (pci_one_device_iommu_support_va(dev) < 0)
+   return false;
+   iommu_dma_mask_check_done  = 1;


As I do not know enough on what IOMMU hardware can coexist, I leave this 
patch for others to review.

Here is the previous question/answer:

Not sure why this change? Why do we only need to check one device on 
all the buses?


Because there is just one emulated IOMMU hardware. The limitation in 
this case is not in a specific PCI device. And I do not think it is 
possible to have two different (emulated or not) IOMMU hardware. Yes, 
you can have more than one controller but being same IOMMU type.


If the above is confirmed, you can consider this patch ack’ed.


+   }
}
}
return true;
--
1.9.1


Re: [dpdk-dev] [PATCH v2 1/5] compress/zlib: add ZLIB PMD support

2018-07-11 Thread De Lara Guarch, Pablo
Hi Shally/Ashish,

> -Original Message-
> From: Shally Verma [mailto:shally.ve...@caviumnetworks.com]
> Sent: Monday, July 2, 2018 5:57 PM
> To: De Lara Guarch, Pablo 
> Cc: dev@dpdk.org; pathr...@caviumnetworks.com;
> mcha...@caviumnetworks.com; Ashish Gupta
> ; Sunila Sahu
> 
> Subject: [PATCH v2 1/5] compress/zlib: add ZLIB PMD support
> 

Remove "support" from title.

Also, add the PMD in devtools/test-build.sh.
Since it depends on ZLIB, add a "sed" line below test "$DPDK_DEP_ZLIB" != y.

More comments inline below.

> From: Ashish Gupta 
> 
> Add sw zlib pmd support in compressdev driver.
> Add device probe and remove support.
> Add ZLIB build file support.
> 
> Signed-off-by: Sunila Sahu 
> Signed-off-by: Shally Verma 
> Signed-off-by: Ashish Gupta 

...

> +++ b/drivers/compress/zlib/meson.build
> @@ -0,0 +1,14 @@
> +# SPDX-License-Identifier: BSD-3-Clause # Copyright(c) 2018 Cavium
> +Networks
> +
> +dep = dependency('zlib', required: false) if not dep.found()
> + build = false
> +endif
> +
> +deps += 'bus_vdev'
> +sources = files('zlib_pmd.c', 'zlib_pmd_ops.c') ext_deps += dep

Zlib_pmd_ops.c is created in the next patch, so remove it from here.

> +pkgconfig_extra_libs += '-lz'
> +
> +allow_experimental_apis = true
> diff --git a/drivers/compress/zlib/rte_pmd_zlib_version.map
> b/drivers/compress/zlib/rte_pmd_zlib_version.map
> new file mode 100644
> index 000..1a99a33
> --- /dev/null
> +++ b/drivers/compress/zlib/rte_pmd_zlib_version.map
> @@ -0,0 +1,3 @@
> +18.08 {

DPDK_18.08.

> + local: *;
> +};
> diff --git a/drivers/compress/zlib/zlib_pmd.c 
> b/drivers/compress/zlib/zlib_pmd.c
> new file mode 100644
> index 000..f667ccc
> --- /dev/null
> +++ b/drivers/compress/zlib/zlib_pmd.c
> @@ -0,0 +1,81 @@
> +/* SPDX-License-Identifier: BSD-3-Clause
> + * Copyright(c) 2018 Cavium Networks
> + */
> +
> +#include 
> +#include 

Leave space between rte_common and zlib_pmd_private.h

> +#include "zlib_pmd_private.h"
> +
> +static int
> +zlib_create(const char *name,
> + struct rte_vdev_device *vdev,
> + struct rte_compressdev_pmd_init_params *init_params) {
> + struct rte_compressdev *dev;
> +
> + dev = rte_compressdev_pmd_create(name, &vdev->device,
> + sizeof(struct zlib_private), init_params);
> + if (dev == NULL) {
> + ZLIB_PMD_ERR("driver %s: create failed", init_params->name);
> + return -ENODEV;
> + }
> +
> + dev->feature_flags = RTE_COMP_FF_NONCOMPRESSED_BLOCKS;

This is an algorithm feature flag, so it should go in capabilities.

> +
> + return 0;
> +}
> +

...

> +RTE_PMD_REGISTER_VDEV(COMPRESSDEV_NAME_ZLIB_PMD, zlib_pmd_drv);
> +RTE_PMD_REGISTER_ALIAS(COMPRESSDEV_NAME_ZLIB_PMD,
> +compressdev_zlib_pmd);

No need to use an alias here. The convention now is driverType_driverName (e.g. 
compress_zlib).

> +
> +RTE_INIT(zlib_init_log);
> +
> +static void
> +zlib_init_log(void)
> +{
> + zlib_logtype_driver = rte_log_register("compress_zlib");
> + if (zlib_logtype_driver >= 0)
> + rte_log_set_level(zlib_logtype_driver, RTE_LOG_INFO); }



Re: [dpdk-dev] [dpdk-stable] [PATCH v4 2/5] bus/pci: use IOVAs check when setting IOVA mode

2018-07-11 Thread Burakov, Anatoly

On 11-Jul-18 11:18 AM, Eelco Chaudron wrote:



On 10 Jul 2018, at 19:25, Alejandro Lucero wrote:


Although VT-d emulation currently only supports 39 bits, it could
be iovas being within that supported range. This patch allows
IOVA mode in such a case.

Indeed, memory initialization code can be modified for using lower
virtual addresses than those used by the kernel for 64 bits processes
by default, and therefore memsegs iovas can use 39 bits or less for
most system. And this is likely 100% true for VMs.

Applicable to v17.11.3 only.

Signed-off-by: Alejandro Lucero 
---
 drivers/bus/pci/linux/pci.c | 15 +++
 1 file changed, 11 insertions(+), 4 deletions(-)

diff --git a/drivers/bus/pci/linux/pci.c b/drivers/bus/pci/linux/pci.c
index 74deef3..792c819 100644
--- a/drivers/bus/pci/linux/pci.c
+++ b/drivers/bus/pci/linux/pci.c
@@ -43,6 +43,7 @@
 #include 
 #include 
 #include 
+#include 

 #include "eal_private.h"
 #include "eal_filesystem.h"
@@ -613,10 +614,12 @@
 fclose(fp);

 mgaw = ((vtd_cap_reg & VTD_CAP_MGAW_MASK) >> VTD_CAP_MGAW_SHIFT) 
+ 1;

-    if (mgaw < X86_VA_WIDTH)
+
+    if (!rte_eal_check_dma_mask(mgaw))
+    return true;
+    else
 return false;

-    return true;
 }
 #elif defined(RTE_ARCH_PPC_64)
 static bool
@@ -640,13 +643,17 @@
 {
 struct rte_pci_device *dev = NULL;
 struct rte_pci_driver *drv = NULL;
+    int iommu_dma_mask_check_done = 0;

 FOREACH_DRIVER_ON_PCIBUS(drv) {
 FOREACH_DEVICE_ON_PCIBUS(dev) {
 if (!rte_pci_match(drv, dev))
 continue;
-    if (!pci_one_device_iommu_support_va(dev))
-    return false;
+    if (!iommu_dma_mask_check_done) {
+    if (pci_one_device_iommu_support_va(dev) < 0)
+    return false;
+    iommu_dma_mask_check_done  = 1;


As I do not know enough on what IOMMU hardware can coexist, I leave this 
patch for others to review.

Here is the previous question/answer:

Not sure why this change? Why do we only need to check one device on 
all the buses?


Because there is just one emulated IOMMU hardware. The limitation in 
this case is not in a specific PCI device. And I do not think it is 
possible to have two different (emulated or not) IOMMU hardware. Yes, 
you can have more than one controller but being same IOMMU type.


If the above is confirmed, you can consider this patch ack’ed.


As far as my knowledge goes, the above is correct. There can't be more 
than one IOMMU type per platform, and limitations specific for IOMMU 
controller apply to all devices, not just one.


I'm not sure if it's possible to have two IOMMU controllers of the same 
type with different address width, but i think even if it is possible, 
we can safely consider this to be unlikely :)





+    }
 }
 }
 return true;
--
1.9.1





--
Thanks,
Anatoly


[dpdk-dev] [PATCH v9 0/7] hotplug failure handle mechanism

2018-07-11 Thread Jeff Guo
As we know, hot plug is an importance feature, either use for the datacenter
device’s fail-safe, or use for SRIOV Live Migration in SDN/NFV. It could bring
the higher flexibility and continuality to the networking services in multiple
use cases in industry. So let we see, dpdk as an importance networking
framework, what can it help to implement hot plug solution for users.

We already have a general device event detect mechanism, failsafe driver,
bonding driver and hot plug/unplug api in framework, app could use these to
develop their hot plug solution.

let’s see the case of hot unplug, it can happen when a hardware device is
be removed physically, or when the software disables it.  App need to call
ether dev API to detach the device, to unplug the device at the bus level and
make access to the device invalid. But the problem is that, the removal of the
device from the software lists is not going to be instantaneous, at this time
if the data(fast) path still read/write the device, it will cause MMIO error
and result of the app crash out.

Seems that we have got fail-safe driver(or app) + RTE_ETH_EVENT_INTR_RMV +
kernel core driver solution to handle it, but still not have failsafe driver
(or app) + RTE_DEV_EVENT_REMOVE + PCIe pmd driver failure handle solution. So
there is an absence in dpdk hot plug solution right now.

Also, we know that kernel only guaranty hot plug on the kernel side, but not for
the user mode side. Firstly we can hardly have a gatekeeper for any MMIO for
multiple PMD driver. Secondly, no more specific 3rd tools such as udev/driverctl
have especially cover these hot plug failure processing. Third, the feasibility
of app’s implement for multiple user mode PMD driver is still a problem. Here,
a general hot plug failure handle mechanism in dpdk framework would be proposed,
it aim to guaranty that, when hot unplug occur, the system will not crash and
app will not be break out, and user space can normally stop and release any
relevant resources, then unplug of the device at the bus level cleanly.

The mechanism should be come across as bellow:

Firstly, app enabled the device event monitor and register the hot plug event’s
callback before running data path. Once the hot unplug behave occur, the
mechanism will detect the removal event and then accordingly do the failure
handle. In order to do that, below functional will be bring in.
 - Add a new bus ops “handle_hot_unplug” to handle bus read/write error, it is
   bus-specific and each kind of bus can implement its own logic.
 - Implement pci bus specific ops “pci_handle_hot_unplug”. It will base on the
   failure address to remap memory for the corresponding device that unplugged.

For the data path or other unexpected control from the control path when hot
unplug occur.
 - Implement a new sigbus handler, it is registered when start device even
   monitoring. The handler is per process. Base on the signal event principle,
   control path thread and data path thread will randomly receive the sigbus
   error, but will go to the common sigbus handler. Once the MMIO sigbus error
   exposure, it will trigger the above hot unplug operation. The sigbus will be
   check if it is cause of the hot unplug or not, if not will info exception as
   the original sigbus handler. If yes, will do memory remapping.

For the control path and the igb uio release:
 - When hot unplug device, the kernel will release the device resource in the
   kernel side, such as the fd sys file will disappear, and the irq will be
   released. At this time, if igb uio driver still try to release this resource,
   it will cause kernel crash.
   On the other hand, something like interrupt disable do not automatically
   process in kernel side. If not handler it, this redundancy and dirty thing
   will affect the interrupt resource be used by other device.
   So the igb_uio driver have to check the hot plug status and corresponding
   process should be taken in igb uio deriver.
   This patch propose to add structure of rte_udev_state into rte_uio_pci_dev
   of igb_uio kernel driver, which will record the state of uio device, such as
   probed/opened/released/removed/unplug. When detect the unexpected removal
   which cause of hot unplug behavior, it will corresponding disable interrupt
   resource, while for the part of releasement which kernel have already handle,
   just skip it to avoid double free or null pointer kernel crash issue.

The mechanism could be use for fail-safe driver and app which want to use hot
plug solution. let testpmd for example:
 - Enable device event monitor->device unplug->failure handle->stop forwarding->
   stop port->close port->detach port.

This process will not breaking the app/fail-safe running, and will not break
other irrelevance device. And app could plug in the device and restart the date
path again by below.
 - Device plug in->bind igb_uio driver ->attached device->start port->
   start forwarding.

patchset history:
v9->v8:
refine commit log

[dpdk-dev] [PATCH v9 1/7] bus: add hotplug failure handler

2018-07-11 Thread Jeff Guo
When device be hotplug out, if app still continue to access device by mmio,
it will cause of memory failure and result the system crash.

This patch introduces a bus ops to handle device hotplug failure, it is a
bus specific behavior, so each kind of bus can implement its own logic case
by case.

Signed-off-by: Jeff Guo 
Acked-by: Shaopeng He 
---
v9->v8:
no change
---
 lib/librte_eal/common/include/rte_bus.h | 16 
 1 file changed, 16 insertions(+)

diff --git a/lib/librte_eal/common/include/rte_bus.h 
b/lib/librte_eal/common/include/rte_bus.h
index eb9eded..e3a55a8 100644
--- a/lib/librte_eal/common/include/rte_bus.h
+++ b/lib/librte_eal/common/include/rte_bus.h
@@ -168,6 +168,20 @@ typedef int (*rte_bus_unplug_t)(struct rte_device *dev);
 typedef int (*rte_bus_parse_t)(const char *name, void *addr);
 
 /**
+ * Implementation a specific hotplug failure handler, which is responsible
+ * for handle the failure when the device be hotplug out from the bus. When
+ * hotplug removal event be detected, it could call this function to handle
+ * failure and guaranty the system would not crash in the case.
+ * @param dev
+ * Pointer of the device structure.
+ *
+ * @return
+ * 0 on success.
+ * !0 on error.
+ */
+typedef int (*rte_bus_hotplug_failure_handler_t)(struct rte_device *dev);
+
+/**
  * Bus scan policies
  */
 enum rte_bus_scan_mode {
@@ -211,6 +225,8 @@ struct rte_bus {
rte_bus_parse_t parse;   /**< Parse a device name */
struct rte_bus_conf conf;/**< Bus configuration */
rte_bus_get_iommu_class_t get_iommu_class; /**< Get iommu class */
+   rte_bus_hotplug_failure_handler_t hotplug_failure_handler;
+   /**< handle hotplug failure on bus */
 };
 
 /**
-- 
2.7.4



[dpdk-dev] [PATCH v9 4/7] bus/pci: implement sigbus handler operation

2018-07-11 Thread Jeff Guo
This patch implements the ops of sigbus handler for PCI bus, it is
functional to find the corresponding pci device which is been hotplug
out, and then call the bus ops of hotplug failure handler to handle
the failure for the device.

Signed-off-by: Jeff Guo 
Acked-by: Shaopeng He 
---
v9->v8:
no change
---
 drivers/bus/pci/pci_common.c | 49 
 1 file changed, 49 insertions(+)

diff --git a/drivers/bus/pci/pci_common.c b/drivers/bus/pci/pci_common.c
index d7abe6c..37ad266 100644
--- a/drivers/bus/pci/pci_common.c
+++ b/drivers/bus/pci/pci_common.c
@@ -407,6 +407,32 @@ pci_find_device(const struct rte_device *start, 
rte_dev_cmp_t cmp,
return NULL;
 }
 
+/* check the failure address belongs to which device. */
+static struct rte_pci_device *
+pci_find_device_by_addr(const void *failure_addr)
+{
+   struct rte_pci_device *pdev = NULL;
+   int i;
+
+   FOREACH_DEVICE_ON_PCIBUS(pdev) {
+   for (i = 0; i != RTE_DIM(pdev->mem_resource); i++) {
+   if ((uint64_t)(uintptr_t)failure_addr >=
+   (uint64_t)(uintptr_t)pdev->mem_resource[i].addr &&
+   (uint64_t)(uintptr_t)failure_addr <
+   (uint64_t)(uintptr_t)pdev->mem_resource[i].addr +
+   pdev->mem_resource[i].len) {
+   RTE_LOG(INFO, EAL, "Failure address "
+   "%16.16"PRIx64" belongs to "
+   "device %s!\n",
+   (uint64_t)(uintptr_t)failure_addr,
+   pdev->device.name);
+   return pdev;
+   }
+   }
+   }
+   return NULL;
+}
+
 static int
 pci_hotplug_failure_handler(struct rte_device *dev)
 {
@@ -435,6 +461,28 @@ pci_hotplug_failure_handler(struct rte_device *dev)
 }
 
 static int
+pci_sigbus_handler(const void *failure_addr)
+{
+   struct rte_pci_device *pdev = NULL;
+   int ret = 0;
+
+   pdev = pci_find_device_by_addr(failure_addr);
+   if (!pdev) {
+   /* It is a generic sigbus error, no bus would handle it. */
+   ret = 1;
+   } else {
+   /* The sigbus error is caused of hot removal. */
+   ret = pci_hotplug_failure_handler(&pdev->device);
+   if (ret) {
+   RTE_LOG(ERR, EAL, "Failed to handle hot plug for "
+   "device %s", pdev->name);
+   ret = -1;
+   }
+   }
+   return ret;
+}
+
+static int
 pci_plug(struct rte_device *dev)
 {
return pci_probe_all_drivers(RTE_DEV_TO_PCI(dev));
@@ -465,6 +513,7 @@ struct rte_pci_bus rte_pci_bus = {
.parse = pci_parse,
.get_iommu_class = rte_pci_get_iommu_class,
.hotplug_failure_handler = pci_hotplug_failure_handler,
+   .sigbus_handler = pci_sigbus_handler,
},
.device_list = TAILQ_HEAD_INITIALIZER(rte_pci_bus.device_list),
.driver_list = TAILQ_HEAD_INITIALIZER(rte_pci_bus.driver_list),
-- 
2.7.4



[dpdk-dev] [PATCH v9 2/7] bus/pci: implement hotplug failure handler ops

2018-07-11 Thread Jeff Guo
This patch implements the ops of hotplug failure handler for PCI bus,
it is functional to remap a new dummy memory which overlap to the
failure memory to avoid MMIO read/write error.

Signed-off-by: Jeff Guo 
Acked-by: Shaopeng He 
---
v9->v8:
no change
---
 drivers/bus/pci/pci_common.c | 28 
 drivers/bus/pci/pci_common_uio.c | 33 +
 drivers/bus/pci/private.h| 12 
 3 files changed, 73 insertions(+)

diff --git a/drivers/bus/pci/pci_common.c b/drivers/bus/pci/pci_common.c
index 94b0f41..d7abe6c 100644
--- a/drivers/bus/pci/pci_common.c
+++ b/drivers/bus/pci/pci_common.c
@@ -408,6 +408,33 @@ pci_find_device(const struct rte_device *start, 
rte_dev_cmp_t cmp,
 }
 
 static int
+pci_hotplug_failure_handler(struct rte_device *dev)
+{
+   struct rte_pci_device *pdev = NULL;
+   int ret = 0;
+
+   pdev = RTE_DEV_TO_PCI(dev);
+   if (!pdev)
+   return -1;
+
+   switch (pdev->kdrv) {
+   case RTE_KDRV_IGB_UIO:
+   case RTE_KDRV_UIO_GENERIC:
+   case RTE_KDRV_NIC_UIO:
+   /* mmio resource is invalid, remap it to be safe. */
+   ret = pci_uio_remap_resource(pdev);
+   break;
+   default:
+   RTE_LOG(DEBUG, EAL,
+   "Not managed by a supported kernel driver, skipped\n");
+   ret = -1;
+   break;
+   }
+
+   return ret;
+}
+
+static int
 pci_plug(struct rte_device *dev)
 {
return pci_probe_all_drivers(RTE_DEV_TO_PCI(dev));
@@ -437,6 +464,7 @@ struct rte_pci_bus rte_pci_bus = {
.unplug = pci_unplug,
.parse = pci_parse,
.get_iommu_class = rte_pci_get_iommu_class,
+   .hotplug_failure_handler = pci_hotplug_failure_handler,
},
.device_list = TAILQ_HEAD_INITIALIZER(rte_pci_bus.device_list),
.driver_list = TAILQ_HEAD_INITIALIZER(rte_pci_bus.driver_list),
diff --git a/drivers/bus/pci/pci_common_uio.c b/drivers/bus/pci/pci_common_uio.c
index 54bc20b..7ea73db 100644
--- a/drivers/bus/pci/pci_common_uio.c
+++ b/drivers/bus/pci/pci_common_uio.c
@@ -146,6 +146,39 @@ pci_uio_unmap(struct mapped_pci_resource *uio_res)
}
 }
 
+/* remap the PCI resource of a PCI device in anonymous virtual memory */
+int
+pci_uio_remap_resource(struct rte_pci_device *dev)
+{
+   int i;
+   void *map_address;
+
+   if (dev == NULL)
+   return -1;
+
+   /* Remap all BARs */
+   for (i = 0; i != PCI_MAX_RESOURCE; i++) {
+   /* skip empty BAR */
+   if (dev->mem_resource[i].phys_addr == 0)
+   continue;
+   map_address = mmap(dev->mem_resource[i].addr,
+   (size_t)dev->mem_resource[i].len,
+   PROT_READ | PROT_WRITE,
+   MAP_PRIVATE | MAP_ANONYMOUS | MAP_FIXED, -1, 0);
+   if (map_address == MAP_FAILED) {
+   RTE_LOG(ERR, EAL,
+   "Cannot remap resource for device %s\n",
+   dev->name);
+   return -1;
+   }
+   RTE_LOG(INFO, EAL,
+   "Successful remap resource for device %s\n",
+   dev->name);
+   }
+
+   return 0;
+}
+
 static struct mapped_pci_resource *
 pci_uio_find_resource(struct rte_pci_device *dev)
 {
diff --git a/drivers/bus/pci/private.h b/drivers/bus/pci/private.h
index 8ddd03e..6b312e5 100644
--- a/drivers/bus/pci/private.h
+++ b/drivers/bus/pci/private.h
@@ -123,6 +123,18 @@ void pci_uio_free_resource(struct rte_pci_device *dev,
struct mapped_pci_resource *uio_res);
 
 /**
+ * Remap the PCI resource of a PCI device in anonymous virtual memory.
+ *
+ * @param dev
+ *   Point to the struct rte pci device.
+ * @return
+ *   - On success, zero.
+ *   - On failure, a negative value.
+ */
+int
+pci_uio_remap_resource(struct rte_pci_device *dev);
+
+/**
  * Map device memory to uio resource
  *
  * This function is private to EAL.
-- 
2.7.4



[dpdk-dev] [PATCH v9 3/7] bus: add sigbus handler

2018-07-11 Thread Jeff Guo
When device be hotplug out, if data path still read/write device, the
sigbus error will occur, this error need to be handled. So a handler
need to be here to capture the signal and handle it correspondingly.

This patch introduces a bus ops to handle sigbus error, it is a bus
specific behavior, so that each kind of bus can implement its own logic
case by case.

Signed-off-by: Jeff Guo 
Acked-by: Shaopeng He 
---
v9->v8:
no change
---
 lib/librte_eal/common/include/rte_bus.h | 17 +
 1 file changed, 17 insertions(+)

diff --git a/lib/librte_eal/common/include/rte_bus.h 
b/lib/librte_eal/common/include/rte_bus.h
index e3a55a8..216ad1e 100644
--- a/lib/librte_eal/common/include/rte_bus.h
+++ b/lib/librte_eal/common/include/rte_bus.h
@@ -182,6 +182,21 @@ typedef int (*rte_bus_parse_t)(const char *name, void 
*addr);
 typedef int (*rte_bus_hotplug_failure_handler_t)(struct rte_device *dev);
 
 /**
+ * Implementation a specific sigbus handler, which is responsible for handle
+ * the sigbus error which is either original memory error, or specific memory
+ * error that caused of hot unplug. When sigbus error be captured, it could
+ * call this function to handle sigbus error.
+ * @param failure_addr
+ * Pointer of the fault address of the sigbus error.
+ *
+ * @return
+ * 0 for success handle the sigbus.
+ * 1 for no bus handle the sigbus.
+ * -1 for failed to handle the sigbus
+ */
+typedef int (*rte_bus_sigbus_handler_t)(const void *failure_addr);
+
+/**
  * Bus scan policies
  */
 enum rte_bus_scan_mode {
@@ -227,6 +242,8 @@ struct rte_bus {
rte_bus_get_iommu_class_t get_iommu_class; /**< Get iommu class */
rte_bus_hotplug_failure_handler_t hotplug_failure_handler;
/**< handle hotplug failure on bus */
+   rte_bus_sigbus_handler_t sigbus_handler; /**< handle sigbus error */
+
 };
 
 /**
-- 
2.7.4



[dpdk-dev] [PATCH v9 7/7] igb_uio: fix unexpected remove issue for hotplug

2018-07-11 Thread Jeff Guo
When device be hotplug out, the pci resource will be released in kernel,
the uio fd will disappear, and the irq will be released. At this time,
if igb uio driver still try to access or release these resource, it will
cause kernel crash.

On the other hand, uio_remove will be called unexpectedly before
uio_release. The uio_remove procedure will free resources which are needed
by uio_release. So there is no chance to disable interrupt which is defined
inside uio_release procedure. This will affect later usage of interrupt.

So the case of unexpectedly removal by hot unplug should be identify and
correspondingly processed.

This patch propose to add enum rte_udev_state in struct rte_uio_pci_dev,
that will keep the state of uio device as probed/opened/released/removed.

This patch also checks kobject’s remove_uevent_sent state to detect the
unexpectedly removal status which means hot unplug. Once hot unplug be
detected, it will call uio_release as soon as possible and set the uio
status to be “removed”. After that, uio will check this status in
uio_release function, if uio have already been removed, it will only free
the dirty uio resource.

Signed-off-by: Jeff Guo 
---
v9->v8:
refine commit log to be more readable.
---
 kernel/linux/igb_uio/igb_uio.c | 69 +-
 1 file changed, 55 insertions(+), 14 deletions(-)

diff --git a/kernel/linux/igb_uio/igb_uio.c b/kernel/linux/igb_uio/igb_uio.c
index 3398eac..d126371 100644
--- a/kernel/linux/igb_uio/igb_uio.c
+++ b/kernel/linux/igb_uio/igb_uio.c
@@ -19,6 +19,14 @@
 
 #include "compat.h"
 
+/* uio pci device state */
+enum rte_udev_state {
+   RTE_UDEV_PROBED,
+   RTE_UDEV_OPENNED,
+   RTE_UDEV_RELEASED,
+   RTE_UDEV_REMOVED,
+};
+
 /**
  * A structure describing the private information for a uio device.
  */
@@ -28,6 +36,7 @@ struct rte_uio_pci_dev {
enum rte_intr_mode mode;
struct mutex lock;
int refcnt;
+   enum rte_udev_state state;
 };
 
 static int wc_activate;
@@ -309,6 +318,17 @@ igbuio_pci_disable_interrupts(struct rte_uio_pci_dev *udev)
 #endif
 }
 
+/* Unmap previously ioremap'd resources */
+static void
+igbuio_pci_release_iomem(struct uio_info *info)
+{
+   int i;
+
+   for (i = 0; i < MAX_UIO_MAPS; i++) {
+   if (info->mem[i].internal_addr)
+   iounmap(info->mem[i].internal_addr);
+   }
+}
 
 /**
  * This gets called while opening uio device file.
@@ -331,20 +351,35 @@ igbuio_pci_open(struct uio_info *info, struct inode 
*inode)
 
/* enable interrupts */
err = igbuio_pci_enable_interrupts(udev);
-   mutex_unlock(&udev->lock);
if (err) {
dev_err(&dev->dev, "Enable interrupt fails\n");
+   pci_clear_master(dev);
+   mutex_unlock(&udev->lock);
return err;
}
+   udev->state = RTE_UDEV_OPENNED;
+   mutex_unlock(&udev->lock);
return 0;
 }
 
+/**
+ * This gets called while closing uio device file.
+ */
 static int
 igbuio_pci_release(struct uio_info *info, struct inode *inode)
 {
struct rte_uio_pci_dev *udev = info->priv;
struct pci_dev *dev = udev->pdev;
 
+   if (udev->state == RTE_UDEV_REMOVED) {
+   mutex_destroy(&udev->lock);
+   igbuio_pci_release_iomem(&udev->info);
+   pci_disable_device(dev);
+   pci_set_drvdata(dev, NULL);
+   kfree(udev);
+   return 0;
+   }
+
mutex_lock(&udev->lock);
if (--udev->refcnt > 0) {
mutex_unlock(&udev->lock);
@@ -356,7 +391,7 @@ igbuio_pci_release(struct uio_info *info, struct inode 
*inode)
 
/* stop the device from further DMA */
pci_clear_master(dev);
-
+   udev->state = RTE_UDEV_RELEASED;
mutex_unlock(&udev->lock);
return 0;
 }
@@ -414,18 +449,6 @@ igbuio_pci_setup_ioport(struct pci_dev *dev, struct 
uio_info *info,
return 0;
 }
 
-/* Unmap previously ioremap'd resources */
-static void
-igbuio_pci_release_iomem(struct uio_info *info)
-{
-   int i;
-
-   for (i = 0; i < MAX_UIO_MAPS; i++) {
-   if (info->mem[i].internal_addr)
-   iounmap(info->mem[i].internal_addr);
-   }
-}
-
 static int
 igbuio_setup_bars(struct pci_dev *dev, struct uio_info *info)
 {
@@ -562,6 +585,9 @@ igbuio_pci_probe(struct pci_dev *dev, const struct 
pci_device_id *id)
 (unsigned long long)map_dma_addr, map_addr);
}
 
+   mutex_lock(&udev->lock);
+   udev->state = RTE_UDEV_PROBED;
+   mutex_unlock(&udev->lock);
return 0;
 
 fail_remove_group:
@@ -579,6 +605,21 @@ static void
 igbuio_pci_remove(struct pci_dev *dev)
 {
struct rte_uio_pci_dev *udev = pci_get_drvdata(dev);
+   struct pci_dev *pdev = udev->pdev;
+   int ret;
+
+   /* handle unexpected removal */
+   if (udev->state == RTE_UDEV_OPENNED ||
+   (&pdev->dev.kobj)

[dpdk-dev] [PATCH v9 6/7] eal: add failure handle mechanism for hotplug

2018-07-11 Thread Jeff Guo
This patch introduces a failure handle mechanism to handle device
hotplug removal event.

First it can register sigbus handler when enable device event monitor. Once
sigbus error be captured, it will check the failure address and accordingly
remap the invalid memory for the corresponding device. Besed on this
mechanism, it could guaranty the application not crash when the device be
hotplug out.

Signed-off-by: Jeff Guo 
Acked-by: Shaopeng He 
---
v9->v8:
no change
---
 lib/librte_eal/linuxapp/eal/eal_dev.c | 113 +-
 1 file changed, 112 insertions(+), 1 deletion(-)

diff --git a/lib/librte_eal/linuxapp/eal/eal_dev.c 
b/lib/librte_eal/linuxapp/eal/eal_dev.c
index 1cf6aeb..1643b33 100644
--- a/lib/librte_eal/linuxapp/eal/eal_dev.c
+++ b/lib/librte_eal/linuxapp/eal/eal_dev.c
@@ -4,6 +4,8 @@
 
 #include 
 #include 
+#include 
+#include 
 #include 
 #include 
 
@@ -14,6 +16,10 @@
 #include 
 #include 
 #include 
+#include 
+#include 
+#include 
+#include 
 
 #include "eal_private.h"
 
@@ -23,6 +29,17 @@ static bool monitor_started;
 #define EAL_UEV_MSG_LEN 4096
 #define EAL_UEV_MSG_ELEM_LEN 128
 
+/*
+ * spinlock for device failure handle, if try to access bus or device,
+ * such as handle sigbus on bus or handle memory failure for device just use
+ * this lock. It could protect the bus and the device to avoid race condition.
+ */
+static rte_spinlock_t failure_handle_lock = RTE_SPINLOCK_INITIALIZER;
+
+static struct sigaction sigbus_action_old;
+
+static int sigbus_need_recover;
+
 static void dev_uev_handler(__rte_unused void *param);
 
 /* identify the system layer which reports this event. */
@@ -33,6 +50,49 @@ enum eal_dev_event_subsystem {
EAL_DEV_EVENT_SUBSYSTEM_MAX
 };
 
+static void
+sigbus_action_recover(void)
+{
+   if (sigbus_need_recover) {
+   sigaction(SIGBUS, &sigbus_action_old, NULL);
+   sigbus_need_recover = 0;
+   }
+}
+
+static void sigbus_handler(int signum, siginfo_t *info,
+   void *ctx __rte_unused)
+{
+   int ret;
+
+   RTE_LOG(INFO, EAL, "Thread[%d] catch SIGBUS, fault address:%p\n",
+   (int)pthread_self(), info->si_addr);
+
+   rte_spinlock_lock(&failure_handle_lock);
+   ret = rte_bus_sigbus_handler(info->si_addr);
+   rte_spinlock_unlock(&failure_handle_lock);
+   if (ret == -1) {
+   rte_exit(EXIT_FAILURE,
+"Failed to handle SIGBUS for hotplug, "
+"(rte_errno: %s)!", strerror(rte_errno));
+   } else if (ret == 1) {
+   if (sigbus_action_old.sa_handler)
+   (*(sigbus_action_old.sa_handler))(signum);
+   else
+   rte_exit(EXIT_FAILURE,
+"Failed to handle generic SIGBUS!");
+   }
+
+   RTE_LOG(INFO, EAL, "Success to handle SIGBUS for hotplug!\n");
+}
+
+static int cmp_dev_name(const struct rte_device *dev,
+   const void *_name)
+{
+   const char *name = _name;
+
+   return strcmp(dev->name, name);
+}
+
 static int
 dev_uev_socket_fd_create(void)
 {
@@ -147,6 +207,9 @@ dev_uev_handler(__rte_unused void *param)
struct rte_dev_event uevent;
int ret;
char buf[EAL_UEV_MSG_LEN];
+   struct rte_bus *bus;
+   struct rte_device *dev;
+   const char *busname = "";
 
memset(&uevent, 0, sizeof(struct rte_dev_event));
memset(buf, 0, EAL_UEV_MSG_LEN);
@@ -171,13 +234,50 @@ dev_uev_handler(__rte_unused void *param)
RTE_LOG(DEBUG, EAL, "receive uevent(name:%s, type:%d, subsystem:%d)\n",
uevent.devname, uevent.type, uevent.subsystem);
 
-   if (uevent.devname)
+   switch (uevent.subsystem) {
+   case EAL_DEV_EVENT_SUBSYSTEM_PCI:
+   case EAL_DEV_EVENT_SUBSYSTEM_UIO:
+   busname = "pci";
+   break;
+   default:
+   break;
+   }
+
+   if (uevent.devname) {
+   if (uevent.type == RTE_DEV_EVENT_REMOVE) {
+   rte_spinlock_lock(&failure_handle_lock);
+   bus = rte_bus_find_by_name(busname);
+   if (bus == NULL) {
+   RTE_LOG(ERR, EAL, "Cannot find bus (%s)\n",
+   busname);
+   return;
+   }
+
+   dev = bus->find_device(NULL, cmp_dev_name,
+  uevent.devname);
+   if (dev == NULL) {
+   RTE_LOG(ERR, EAL, "Cannot find device (%s) on "
+   "bus (%s)\n", uevent.devname, busname);
+   return;
+   }
+
+   ret = bus->hotplug_failure_handler(dev);
+   rte_spinlock_unlock(&failure_handle_lock);
+   if (ret) {
+ 

[dpdk-dev] [PATCH v9 5/7] bus: add helper to handle sigbus

2018-07-11 Thread Jeff Guo
This patch aim to add a helper to iterate all buses to find the
corresponding bus to handle the sigbus error.

Signed-off-by: Jeff Guo 
Acked-by: Shaopeng He 
---
v9->v8:
no change
---
 lib/librte_eal/common/eal_common_bus.c | 43 ++
 lib/librte_eal/common/eal_private.h| 12 ++
 2 files changed, 55 insertions(+)

diff --git a/lib/librte_eal/common/eal_common_bus.c 
b/lib/librte_eal/common/eal_common_bus.c
index 0943851..62b7318 100644
--- a/lib/librte_eal/common/eal_common_bus.c
+++ b/lib/librte_eal/common/eal_common_bus.c
@@ -37,6 +37,7 @@
 #include 
 #include 
 #include 
+#include 
 
 #include "eal_private.h"
 
@@ -242,3 +243,45 @@ rte_bus_get_iommu_class(void)
}
return mode;
 }
+
+static int
+bus_handle_sigbus(const struct rte_bus *bus,
+   const void *failure_addr)
+{
+   int ret;
+
+   if (!bus->sigbus_handler)
+   return -1;
+
+   ret = bus->sigbus_handler(failure_addr);
+
+   /* find bus but handle failed, keep the errno be set. */
+   if (ret < 0 && rte_errno == 0)
+   rte_errno = ENOTSUP;
+
+   return ret > 0;
+}
+
+int
+rte_bus_sigbus_handler(const void *failure_addr)
+{
+   struct rte_bus *bus;
+
+   int ret = 0;
+   int old_errno = rte_errno;
+
+   rte_errno = 0;
+
+   bus = rte_bus_find(NULL, bus_handle_sigbus, failure_addr);
+   /* can not find bus. */
+   if (!bus)
+   return 1;
+   /* find bus but handle failed, pass on the new errno. */
+   else if (rte_errno != 0)
+   return -1;
+
+   /* restore the old errno. */
+   rte_errno = old_errno;
+
+   return ret;
+}
diff --git a/lib/librte_eal/common/eal_private.h 
b/lib/librte_eal/common/eal_private.h
index bdadc4d..2337e71 100644
--- a/lib/librte_eal/common/eal_private.h
+++ b/lib/librte_eal/common/eal_private.h
@@ -258,4 +258,16 @@ int rte_mp_channel_init(void);
  */
 void dev_callback_process(char *device_name, enum rte_dev_event_type event);
 
+/**
+ * Iterate all buses to find the corresponding bus, to handle the sigbus error.
+ * @param failure_addr
+ * Pointer of the fault address of the sigbus error.
+ *
+ * @return
+ *  0 success to handle the sigbus.
+ * -1 failed to handle the sigbus
+ *  1 no bus can handler the sigbus
+ */
+int rte_bus_sigbus_handler(const void *failure_addr);
+
 #endif /* _EAL_PRIVATE_H_ */
-- 
2.7.4



Re: [dpdk-dev] [PATCH v5 01/23] net/softnic: restructuring

2018-07-11 Thread Thomas Monjalon
06/07/2018 19:20, Jasvinder Singh:
> +   dev = dev;

It is an error with clang:
drivers/net/softnic/rte_eth_softnic.c:206:6: fatal error: explicitly
  assigning value of variable of type 'struct rte_eth_dev *' to itself 
[-Wself-assign]
dev = dev;
~~~ ^ ~~~





Re: [dpdk-dev] [PATCH] ethdev: fix port ID retrieval on vdev attach

2018-07-11 Thread Thomas Monjalon
11/07/2018 12:15, Andrew Rybchenko:
> On 11.07.2018 13:02, Thomas Monjalon wrote:
> > 11/07/2018 11:49, Andrew Rybchenko:
> >> From: Ivan Malov 
> >>
> >> Attaching a vdev port may result in multiple
> >> ports actually added because a vdev port may
> >> have slave devices to be attached implicitly.
> >>
> >> Ethdev attach API has to fill in the port ID
> >> to be read back by the user and what it does
> >> is take the last assigned ID from the common
> >> list after attach completion. Such an ID may
> >> belong to a slave device and not to the vdev.
> >>
> >> This mistake must be precluded by requesting
> >> the port ID by name of device being attached.
> > No, the real issue is using this function rte_eth_dev_attach.
> > It is broken since day 1.
> > Mixing EAL devargs and ethdev port cannot work by design.
> >
> > If you want to attach a device, you must add it with
> > rte_eal_hotplug_add
> > and wait for the driver to probe the ports
> > (which can be ethdev or other classes).
> 
> Yes, I totally agree. Doesn't it deserve to be fixed at least in stable?

The concern with stable is to keep it stable :)
I'm afraid this change can introduce more regressions.

> > We should take these actions in 18.08:
> > - deprecate rte_eth_dev_attach/rte_eth_dev_detach
> > - deprecate rte_eal_dev_attach/rte_eal_dev_detach
> > I did not take time to send the deprecation notices yet.
> > Feel free to deprecate these functions yourself.
> 
> Should we deprecate it right now in 18.08 and schedule removal to 18.11?

Yes, we should mark it as deprecated in 18.08 and plan for removal in 18.11.
It is good to get it removed from the next LTS which is 18.11.

> Or just add deprecation notice in 18.08, mark deprecated in 18.11 and
> remove later?

Later is too late :)





[dpdk-dev] [PATCH 2/2] mempool: fold memory size calculation helper

2018-07-11 Thread Andrew Rybchenko
rte_mempool_calc_mem_size_helper() was introduced to avoid
code duplication and used in deprecated rte_mempool_mem_size() and
rte_mempool_op_calc_mem_size_default(). Now the first one is removed
and it is better to fold the helper into the second one to make it
more readable.

Signed-off-by: Andrew Rybchenko 
---
 lib/librte_mempool/rte_mempool.c | 25 
 lib/librte_mempool/rte_mempool.h | 22 -
 lib/librte_mempool/rte_mempool_ops_default.c | 25 +---
 3 files changed, 22 insertions(+), 50 deletions(-)

diff --git a/lib/librte_mempool/rte_mempool.c b/lib/librte_mempool/rte_mempool.c
index d48e53c7e..52bc97a75 100644
--- a/lib/librte_mempool/rte_mempool.c
+++ b/lib/librte_mempool/rte_mempool.c
@@ -225,31 +225,6 @@ rte_mempool_calc_obj_size(uint32_t elt_size, uint32_t 
flags,
return sz->total_size;
 }
 
-
-/*
- * Internal function to calculate required memory chunk size.
- */
-size_t
-rte_mempool_calc_mem_size_helper(uint32_t elt_num, size_t total_elt_sz,
-uint32_t pg_shift)
-{
-   size_t obj_per_page, pg_num, pg_sz;
-
-   if (total_elt_sz == 0)
-   return 0;
-
-   if (pg_shift == 0)
-   return total_elt_sz * elt_num;
-
-   pg_sz = (size_t)1 << pg_shift;
-   obj_per_page = pg_sz / total_elt_sz;
-   if (obj_per_page == 0)
-   return RTE_ALIGN_CEIL(total_elt_sz, pg_sz) * elt_num;
-
-   pg_num = (elt_num + obj_per_page - 1) / obj_per_page;
-   return pg_num << pg_shift;
-}
-
 /* free a memchunk allocated with rte_memzone_reserve() */
 static void
 rte_mempool_memchunk_mz_free(__rte_unused struct rte_mempool_memhdr *memhdr,
diff --git a/lib/librte_mempool/rte_mempool.h b/lib/librte_mempool/rte_mempool.h
index 5d1602555..7c9cd9a2f 100644
--- a/lib/librte_mempool/rte_mempool.h
+++ b/lib/librte_mempool/rte_mempool.h
@@ -487,28 +487,6 @@ ssize_t rte_mempool_op_calc_mem_size_default(const struct 
rte_mempool *mp,
uint32_t obj_num, uint32_t pg_shift,
size_t *min_chunk_size, size_t *align);
 
-/**
- * @internal Helper function to calculate memory size required to store
- * specified number of objects in assumption that the memory buffer will
- * be aligned at page boundary.
- *
- * Note that if object size is bigger than page size, then it assumes
- * that pages are grouped in subsets of physically continuous pages big
- * enough to store at least one object.
- *
- * @param elt_num
- *   Number of elements.
- * @param total_elt_sz
- *   The size of each element, including header and trailer, as returned
- *   by rte_mempool_calc_obj_size().
- * @param pg_shift
- *   LOG2 of the physical pages size. If set to 0, ignore page boundaries.
- * @return
- *   Required memory size aligned at page boundary.
- */
-size_t rte_mempool_calc_mem_size_helper(uint32_t elt_num, size_t total_elt_sz,
-   uint32_t pg_shift);
-
 /**
  * Function to be called for each populated object.
  *
diff --git a/lib/librte_mempool/rte_mempool_ops_default.c 
b/lib/librte_mempool/rte_mempool_ops_default.c
index fd63ca137..4e2bfc82d 100644
--- a/lib/librte_mempool/rte_mempool_ops_default.c
+++ b/lib/librte_mempool/rte_mempool_ops_default.c
@@ -12,12 +12,31 @@ rte_mempool_op_calc_mem_size_default(const struct 
rte_mempool *mp,
 size_t *min_chunk_size, size_t *align)
 {
size_t total_elt_sz;
+   size_t obj_per_page, pg_num, pg_sz;
size_t mem_size;
 
total_elt_sz = mp->header_size + mp->elt_size + mp->trailer_size;
-
-   mem_size = rte_mempool_calc_mem_size_helper(obj_num, total_elt_sz,
-   pg_shift);
+   if (total_elt_sz == 0) {
+   mem_size = 0;
+   } else if (pg_shift == 0) {
+   mem_size = total_elt_sz * obj_num;
+   } else {
+   pg_sz = (size_t)1 << pg_shift;
+   obj_per_page = pg_sz / total_elt_sz;
+   if (obj_per_page == 0) {
+   /*
+* Note that if object size is bigger than page size,
+* then it is assumed that pages are grouped in subsets
+* of physically continuous pages big enough to store
+* at least one object.
+*/
+   mem_size =
+   RTE_ALIGN_CEIL(total_elt_sz, pg_sz) * obj_num;
+   } else {
+   pg_num = (obj_num + obj_per_page - 1) / obj_per_page;
+   mem_size = pg_num << pg_shift;
+   }
+   }
 
*min_chunk_size = RTE_MAX((size_t)1 << pg_shift, total_elt_sz);
 
-- 
2.17.1



[dpdk-dev] [PATCH 1/2] mempool: remove deprecated functions

2018-07-11 Thread Andrew Rybchenko
Functions rte_mempool_populate_phys(), rte_mempool_virt2phy() and
rte_mempool_populate_phys_tab() are just wrappers for corresponding
IOVA functions and were deprecated in v17.11.

Functions rte_mempool_xmem_create(), rte_mempool_xmem_size(),
rte_mempool_xmem_usage() and rte_mempool_populate_iova_tab() were
deprecated in v18.05 and removal was announced earlier in v18.02.

Signed-off-by: Andrew Rybchenko 
---
 lib/librte_mempool/Makefile|   3 -
 lib/librte_mempool/meson.build |   4 -
 lib/librte_mempool/rte_mempool.c   | 181 +
 lib/librte_mempool/rte_mempool.h   | 179 
 lib/librte_mempool/rte_mempool_version.map |   6 -
 5 files changed, 1 insertion(+), 372 deletions(-)

diff --git a/lib/librte_mempool/Makefile b/lib/librte_mempool/Makefile
index e3c32b14f..1d4c21c0e 100644
--- a/lib/librte_mempool/Makefile
+++ b/lib/librte_mempool/Makefile
@@ -7,9 +7,6 @@ include $(RTE_SDK)/mk/rte.vars.mk
 LIB = librte_mempool.a
 
 CFLAGS += $(WERROR_FLAGS) -I$(SRCDIR) -O3
-# Allow deprecated symbol to use deprecated rte_mempool_populate_iova_tab()
-# from earlier deprecated rte_mempool_populate_phys_tab()
-CFLAGS += -Wno-deprecated-declarations
 CFLAGS += -DALLOW_EXPERIMENTAL_API
 LDLIBS += -lrte_eal -lrte_ring
 
diff --git a/lib/librte_mempool/meson.build b/lib/librte_mempool/meson.build
index d507e5511..5d4ab4169 100644
--- a/lib/librte_mempool/meson.build
+++ b/lib/librte_mempool/meson.build
@@ -5,10 +5,6 @@ allow_experimental_apis = true
 
 extra_flags = []
 
-# Allow deprecated symbol to use deprecated rte_mempool_populate_iova_tab()
-# from earlier deprecated rte_mempool_populate_phys_tab()
-extra_flags += '-Wno-deprecated-declarations'
-
 foreach flag: extra_flags
if cc.has_argument(flag)
cflags += flag
diff --git a/lib/librte_mempool/rte_mempool.c b/lib/librte_mempool/rte_mempool.c
index 8c8b9f809..d48e53c7e 100644
--- a/lib/librte_mempool/rte_mempool.c
+++ b/lib/librte_mempool/rte_mempool.c
@@ -227,9 +227,7 @@ rte_mempool_calc_obj_size(uint32_t elt_size, uint32_t flags,
 
 
 /*
- * Internal function to calculate required memory chunk size shared
- * by default implementation of the corresponding callback and
- * deprecated external function.
+ * Internal function to calculate required memory chunk size.
  */
 size_t
 rte_mempool_calc_mem_size_helper(uint32_t elt_num, size_t total_elt_sz,
@@ -252,66 +250,6 @@ rte_mempool_calc_mem_size_helper(uint32_t elt_num, size_t 
total_elt_sz,
return pg_num << pg_shift;
 }
 
-/*
- * Calculate maximum amount of memory required to store given number of 
objects.
- */
-size_t
-rte_mempool_xmem_size(uint32_t elt_num, size_t total_elt_sz, uint32_t pg_shift,
- __rte_unused unsigned int flags)
-{
-   return rte_mempool_calc_mem_size_helper(elt_num, total_elt_sz,
-   pg_shift);
-}
-
-/*
- * Calculate how much memory would be actually required with the
- * given memory footprint to store required number of elements.
- */
-ssize_t
-rte_mempool_xmem_usage(__rte_unused void *vaddr, uint32_t elt_num,
-   size_t total_elt_sz, const rte_iova_t iova[], uint32_t pg_num,
-   uint32_t pg_shift, __rte_unused unsigned int flags)
-{
-   uint32_t elt_cnt = 0;
-   rte_iova_t start, end;
-   uint32_t iova_idx;
-   size_t pg_sz = (size_t)1 << pg_shift;
-
-   /* if iova is NULL, assume contiguous memory */
-   if (iova == NULL) {
-   start = 0;
-   end = pg_sz * pg_num;
-   iova_idx = pg_num;
-   } else {
-   start = iova[0];
-   end = iova[0] + pg_sz;
-   iova_idx = 1;
-   }
-   while (elt_cnt < elt_num) {
-
-   if (end - start >= total_elt_sz) {
-   /* enough contiguous memory, add an object */
-   start += total_elt_sz;
-   elt_cnt++;
-   } else if (iova_idx < pg_num) {
-   /* no room to store one obj, add a page */
-   if (end == iova[iova_idx]) {
-   end += pg_sz;
-   } else {
-   start = iova[iova_idx];
-   end = iova[iova_idx] + pg_sz;
-   }
-   iova_idx++;
-
-   } else {
-   /* no more page, return how many elements fit */
-   return -(size_t)elt_cnt;
-   }
-   }
-
-   return (size_t)iova_idx << pg_shift;
-}
-
 /* free a memchunk allocated with rte_memzone_reserve() */
 static void
 rte_mempool_memchunk_mz_free(__rte_unused struct rte_mempool_memhdr *memhdr,
@@ -423,63 +361,6 @@ rte_mempool_populate_iova(struct rte_mempool *mp, char 
*vaddr,
return ret;
 }
 
-int
-rte_mempool_populate_phys(struct rte_mempool *mp, char *vaddr,
-   phys_addr_t paddr, size_t

Re: [dpdk-dev] [PATCH v2 2/5] compress/zlib: add device setup PMD ops

2018-07-11 Thread De Lara Guarch, Pablo



> -Original Message-
> From: Shally Verma [mailto:shally.ve...@caviumnetworks.com]
> Sent: Monday, July 2, 2018 5:57 PM
> To: De Lara Guarch, Pablo 
> Cc: dev@dpdk.org; pathr...@caviumnetworks.com;
> mcha...@caviumnetworks.com; Ashish Gupta
> ; Sunila Sahu
> 
> Subject: [PATCH v2 2/5] compress/zlib: add device setup PMD ops
> 
> From: Ashish Gupta 
> 
> Implement device configure and PMD ops
> 
> Signed-off-by: Sunila Sahu 
> Signed-off-by: Shally Verma 
> Signed-off-by: Ashish Gupta 

...

> b/drivers/compress/zlib/zlib_pmd_ops.c
> new file mode 100644
> index 000..03b6da5
> --- /dev/null
> +++ b/drivers/compress/zlib/zlib_pmd_ops.c
> @@ -0,0 +1,236 @@
> +/* SPDX-License-Identifier: BSD-3-Clause
> + * Copyright(c) 2018 Cavium Networks
> + */
> +
> +#include 
> +
> +#include 
> +#include 
> +
> +#include "zlib_pmd_private.h"
> +
> +static const struct rte_compressdev_capabilities zlib_pmd_capabilities[] = {
> + {   /* Deflate */
> + .algo = RTE_COMP_ALGO_DEFLATE,
> + .window_size = {
> + .min = 8,
> + .max = 15,
> + .increment = 1
> + },
> + },

Add comp_feature_flags here, including the new FIXED and DYNAMIC flags.

> +
> + RTE_COMP_END_OF_CAPABILITIES_LIST()
> +
> +};
> +

...

> +static int
> +zlib_pmd_qp_release(struct rte_compressdev *dev, uint16_t qp_id) {
> + struct zlib_qp *qp = dev->data->queue_pairs[qp_id];
> + struct rte_ring *r = NULL;
> +
> + if (qp != NULL) {
> + r = rte_ring_lookup(qp->name);
> + if (r)
> + rte_ring_free(r);

I think you can use rte_ring_free(qp->processed_pkts) here directly.

> + rte_free(qp);
> + dev->data->queue_pairs[qp_id] = NULL;
> + }
> + return 0;
> +}
> +
> +/** set a unique name for the queue pair based on it's name, dev_id and

Typo: "its"

...

> +
> +struct rte_compressdev_ops zlib_pmd_ops = {

...

> +
> + .private_xform_create   = NULL,
> + .private_xform_free = NULL,

Remove extra indentation here.

> +
> + .stream_create  = NULL,
> + .stream_free= NULL
> +};
> +
> +struct rte_compressdev_ops *rte_zlib_pmd_ops = &zlib_pmd_ops;



Re: [dpdk-dev] [PATCH] ethdev: fix port ID retrieval on vdev attach

2018-07-11 Thread Andrew Rybchenko

On 11.07.2018 13:52, Thomas Monjalon wrote:

11/07/2018 12:15, Andrew Rybchenko:

On 11.07.2018 13:02, Thomas Monjalon wrote:

11/07/2018 11:49, Andrew Rybchenko:

From: Ivan Malov 

Attaching a vdev port may result in multiple
ports actually added because a vdev port may
have slave devices to be attached implicitly.

Ethdev attach API has to fill in the port ID
to be read back by the user and what it does
is take the last assigned ID from the common
list after attach completion. Such an ID may
belong to a slave device and not to the vdev.

This mistake must be precluded by requesting
the port ID by name of device being attached.

No, the real issue is using this function rte_eth_dev_attach.
It is broken since day 1.
Mixing EAL devargs and ethdev port cannot work by design.

If you want to attach a device, you must add it with
rte_eal_hotplug_add
and wait for the driver to probe the ports
(which can be ethdev or other classes).

Yes, I totally agree. Doesn't it deserve to be fixed at least in stable?

The concern with stable is to keep it stable :)
I'm afraid this change can introduce more regressions.


Makes sense. I'll drop the patch.


We should take these actions in 18.08:
- deprecate rte_eth_dev_attach/rte_eth_dev_detach
- deprecate rte_eal_dev_attach/rte_eal_dev_detach
I did not take time to send the deprecation notices yet.
Feel free to deprecate these functions yourself.

Should we deprecate it right now in 18.08 and schedule removal to 18.11?

Yes, we should mark it as deprecated in 18.08 and plan for removal in 18.11.
It is good to get it removed from the next LTS which is 18.11.


Yes, I agree. I'll care about it.


Or just add deprecation notice in 18.08, mark deprecated in 18.11 and
remove later?

Later is too late :)




Re: [dpdk-dev] [PATCH v4 1/4] ethdev: Add API to enable device event handler

2018-07-11 Thread Jeff Guo




On 7/11/2018 4:49 PM, Andrew Rybchenko wrote:

On 10.07.2018 15:51, Jeff Guo wrote:

Implement a couple of eal device event handler install/uninstall APIs
in ethdev, it could let PMDs have chance to manage the eal device event,
such as register device event callback, then monitor and process the
hotplug event.


I think it is moving in right direction, but my problem with the patch is
that I don't understand what prevents us to make it even more generic.
Right now it looks like that PCI driver flag RTE_PCI_DRV_INTR_RMV is
sufficient and everything else could be done on ethdev layer.
RTE_PCI_DRV_INTR_RMV  is mapped to RTE_ETH_DEV_INTR_RMV eth
device flag. The flag may be used as an indication in 
rte_eth_dev_create()

to register the callback.
May be I'm completely wrong above, but if so, I'd like to understand why
and prefer to see explanations in the patch description.



Your acknowledgement is right, and it is make sense to check the reason 
why is the most generic but not other.
i think that let driver to decide if it support the RTE_PCI_DRV_INTR_RMV 
should be fine, that is first.
And second, the place of installer in driver is also fine, each ethdev 
driver install specific callback for each port, and could let
driver have change to manage the status of hotplug for further. So if 
there are no better place in ethdev here for more generic to install it,

that should be fine.


Signed-off-by: Jeff Guo 
---
v4->v3:
change to use eal device event handler install api
---
  doc/guides/rel_notes/release_18_08.rst | 12 +++
  lib/librte_ethdev/rte_ethdev.c | 59 
++

  lib/librte_ethdev/rte_ethdev_driver.h  | 32 ++
  3 files changed, 103 insertions(+)

diff --git a/doc/guides/rel_notes/release_18_08.rst 
b/doc/guides/rel_notes/release_18_08.rst

index bc01242..b6482ce 100644
--- a/doc/guides/rel_notes/release_18_08.rst
+++ b/doc/guides/rel_notes/release_18_08.rst
@@ -46,6 +46,18 @@ New Features
Flow API support has been added to CXGBE Poll Mode Driver to offload
flows to Chelsio T5/T6 NICs.
  +* **Added eal device event process helper in ethdev.**
+
+  Implement a couple of eal device event handler install/uninstall 
APIs in
+  ethdev, these helper could let PMDs have chance to manage the eal 
device
+  event, such as register device event callback, then monitor and 
process

+  hotplug event.
+
+  * ``rte_eth_dev_event_handler_install`` for PMDs use to install 
the device

+event handler.
+  * ``rte_eth_dev_event_handler_uninstall`` for PMDs use to 
uninstall the device

+event handler.
+
API Changes
  ---
diff --git a/lib/librte_ethdev/rte_ethdev.c 
b/lib/librte_ethdev/rte_ethdev.c

index a9977df..09ea02d 100644
--- a/lib/librte_ethdev/rte_ethdev.c
+++ b/lib/librte_ethdev/rte_ethdev.c
@@ -4518,6 +4518,65 @@ rte_eth_devargs_parse(const char *dargs, 
struct rte_eth_devargs *eth_da)

  return result;
  }
  +static void __rte_experimental
+eth_dev_event_callback(char *device_name, enum rte_dev_event_type type,
+ void *arg)
+{
+struct rte_eth_dev *eth_dev = (struct rte_eth_dev *)arg;
+
+switch (type) {
+case RTE_DEV_EVENT_REMOVE:
+ethdev_log(INFO, "The device: %s has been removed!\n",


Colon after 'device' above looks strange for me. I'd suggest to remove 
it.

If so, similar below for ADD.



ok, i am fine.


+device_name);
+
+if (!device_name || !eth_dev)
+return;
+
+if (!(eth_dev->data->dev_flags & RTE_ETH_EVENT_INTR_RMV))


It looks like a typo above. Shouldn't it be RTE_ETH_DEV_INTR_RMV?



you are right here, it is a typo.


+return;
+
+if (!strcmp(device_name, eth_dev->device->name))
+_rte_eth_dev_callback_process(eth_dev,
+  RTE_ETH_EVENT_INTR_RMV,
+  NULL);
+break;
+case RTE_DEV_EVENT_ADD:
+ethdev_log(INFO, "The device: %s has been added!\n",
+device_name);
+break;
+default:
+break;
+}
+}
+
+int __rte_experimental
+rte_eth_dev_event_handler_install(struct rte_eth_dev *eth_dev)
+{
+int result = 0;
+
+result = rte_dev_event_callback_register(eth_dev->device->name,
+eth_dev_event_callback, eth_dev);
+if (result)
+RTE_LOG(ERR, EAL, "device event callback register failed for "
+"device:%s!\n", eth_dev->device->name);


Why ethdev_log() is used above, but here is RTE_LOG with EAL?
Similar question below.



should be align to use ethdev_log.


+
+return result;
+}
+
+int __rte_experimental
+rte_eth_dev_event_handler_uninstall(struct rte_eth_dev *eth_dev)
+{
+int result = 0;
+
+result = rte_dev_event_callback_unregister(eth_dev->device->name,
+eth_dev_event_callback, eth_dev);
+if (result)
+RTE_LOG(ERR, EAL, "device event callback unregister failed for"
+" device:%s!\n", eth_dev->device->na

Re: [dpdk-dev] [PATCH v10 01/27] devargs: add non-variadic parsing function

2018-07-11 Thread Shreyansh Jain

On Thursday 05 July 2018 05:18 PM, Gaetan Rivet wrote:

rte_devargs_parse becomes non-variadic,
rte_devargs_parsef becomes the variadic version, to be used to compose
device strings.

Signed-off-by: Gaetan Rivet 
---
  drivers/net/failsafe/failsafe_args.c|  2 +-
  drivers/net/failsafe/failsafe_eal.c |  2 +-
  lib/librte_eal/common/eal_common_dev.c  |  4 +-
  lib/librte_eal/common/eal_common_devargs.c  | 42 -
  lib/librte_eal/common/include/rte_devargs.h | 40 +++-
  lib/librte_eal/rte_eal_version.map  |  1 +
  lib/librte_ethdev/rte_ethdev.c  |  2 +-
  7 files changed, 76 insertions(+), 17 deletions(-)



[...]


+__rte_experimental
+int
+rte_devargs_parsef(struct rte_devargs *da, const char *format, ...)
+{
+   va_list ap;
+   size_t len;
+   char *dev;
+
+   if (da == NULL)
+   return -EINVAL;
+
+   va_start(ap, format);
+   len = vsnprintf(NULL, 0, format, ap);
+   va_end(ap);
+
+   dev = calloc(1, len + 1);
+   if (dev == NULL) {
+   fprintf(stderr, "ERROR: not enough memory to parse device\n");


Should RTE_LOG be used here?


+   return -ENOMEM;
+   }
+
+   va_start(ap, format);
+   vsnprintf(dev, len, format, ap);
+   va_end(ap);
+
+   return rte_devargs_parse(da, dev);
+}
+
  int __rte_experimental
  rte_devargs_insert(struct rte_devargs *da)
  {


[...]

Except the comment above:

Acked-by: Shreyansh Jain 


[dpdk-dev] [Dpdk-pktgen]dpdk pktgen build failed on x86

2018-07-11 Thread khemendra kumar
Hi All,

Kindly help to check below compile error in DPDK Pkt-gen on x86.

I am following instructions from "
http://pktgen-dpdk.readthedocs.io/en/latest/getting_started.html";

*Below cmd I followed:*
sudo make config T=x86_64-native-linuxapp-gcc
sudo make

sudo make install
sudo make RTE_SDK=/home/root1/khem/dpdk

*Below is error log:*
  INSTALL-APP dpdk-test-eventdev
  INSTALL-MAP dpdk-test-eventdev.map
Build complete [x86_64-native-linuxapp-gcc]
root1@compute-201:~/khem/dpdk$ sudo make install
make[1]: Nothing to be done for 'pre_install'.
== Installing /usr/local/
Installation in /usr/local/ complete
root1@compute-201:~/khem/dpdk$ cd ../pktgen-dpdk/
root1@compute-201:~/khem/pktgen-dpdk$ sudo make
RTE_SDK=/home/root1/khem/dpdk
== lib
== common
== lua
== src
== cli
== app
  CC cli-functions.o
In file included from /home/root1/khem/pktgen-dpdk/app/pktgen.h:87:0,
 from /home/root1/khem/pktgen-dpdk/app/cli-functions.c:28:
/home/root1/khem/pktgen-dpdk/app/pktgen-port-cfg.h: In function
‘rte_eth_txconf_dump’:
/home/root1/khem/pktgen-dpdk/app/pktgen-port-cfg.h:289:5: error: ‘struct
rte_eth_txconf’ has no member named ‘txq_flags’
   tx->txq_flags);
 ^
/home/root1/khem/dpdk/mk/internal/rte.compile-pre.mk:114: recipe for target
'cli-functions.o' failed
make[2]: *** [cli-functions.o] Error 1
/home/root1/khem/dpdk/mk/rte.extapp.mk:19: recipe for target 'pktgen' failed
make[1]: *** [pktgen] Error 2
/home/root1/khem/dpdk/mk/rte.extsubdir.mk:21: recipe for target 'app' failed
make: *** [app] Error 2
root1@compute-201:~/khem/pktgen-dpdk$

Thanks & Regards
khem


[dpdk-dev] [PATCH v5 0/4] Install eal hotplug event handler in i40e/ixgbe

2018-07-11 Thread Jeff Guo
As we may know, we have eal event for rte device hotplug and ethdev event
for ethdev hotplug. Some ethdev need to use eal event to detect hotplug
behaviors, the privors way is register eal event callback in app, but seems
that it will have some race between these 2 event processes. In oder to fix
the it, it might be better to find a way to combind these 2 events detect.

This patch set introduce a way to combind these 2 event, by register the
ethdev eal event callback in ether dev and trigger the ethdev hotplug event
in the callback. That will let the ethdev device can easy process hotplug
by a common way.

Here let i40e/ixgbe pmd driver for example, other driver which support
hotplug feature could be use this way to detect and process hotplug.

patch history:
v5->v4:
refine some code style and typo

v4->v3:
change to use device event handler install api

v3->v2:
remove the callback from driver to ethdev for common.

v2->v1:
add ixgbe hotplug detect case.
refine some doc.

Jeff Guo (4):
  ethdev: Add eal device event callback
  net/ixgbe: install ethdev hotplug handler in ixgbe
  net/i40e: install hotplug handler in i40e
  testpmd: remove the dev event callback register

 app/test-pmd/testpmd.c   | 76 
 doc/guides/rel_notes/release_18_08.rst   | 12 +
 drivers/net/i40e/i40e_ethdev.c   |  8 +++-
 drivers/net/ixgbe/ixgbe_ethdev.c |  8 +++-
 lib/librte_ethdev/rte_ethdev.c   | 59 +
 lib/librte_ethdev/rte_ethdev_driver.h| 32 ++
 lib/librte_ethdev/rte_ethdev_version.map |  2 +
 7 files changed, 119 insertions(+), 78 deletions(-)

-- 
2.7.4



[dpdk-dev] [PATCH v5 1/4] ethdev: Add eal device event callback

2018-07-11 Thread Jeff Guo
Implement a couple of eal device event handler install/uninstall APIs
in ethdev. Each ethdev install the handler in PMDs, so each callback
corresponding with one port, and process the eal device event for specific
port. If PMDs install the handler when initial device, it could have
chance to manage the eal device event, such as register device event
callback, then monitor and process the hotplug event.

Signed-off-by: Jeff Guo 
Reviewed-by: Qi Zhang 
---
v5->v4:
refine some code style and typo
---
 doc/guides/rel_notes/release_18_08.rst   | 12 +++
 lib/librte_ethdev/rte_ethdev.c   | 59 
 lib/librte_ethdev/rte_ethdev_driver.h| 32 +
 lib/librte_ethdev/rte_ethdev_version.map |  2 ++
 4 files changed, 105 insertions(+)

diff --git a/doc/guides/rel_notes/release_18_08.rst 
b/doc/guides/rel_notes/release_18_08.rst
index bc01242..b6482ce 100644
--- a/doc/guides/rel_notes/release_18_08.rst
+++ b/doc/guides/rel_notes/release_18_08.rst
@@ -46,6 +46,18 @@ New Features
   Flow API support has been added to CXGBE Poll Mode Driver to offload
   flows to Chelsio T5/T6 NICs.
 
+* **Added eal device event process helper in ethdev.**
+
+  Implement a couple of eal device event handler install/uninstall APIs in
+  ethdev, these helper could let PMDs have chance to manage the eal device
+  event, such as register device event callback, then monitor and process
+  hotplug event.
+
+  * ``rte_eth_dev_event_handler_install`` for PMDs use to install the device
+event handler.
+  * ``rte_eth_dev_event_handler_uninstall`` for PMDs use to uninstall the 
device
+event handler.
+
 
 API Changes
 ---
diff --git a/lib/librte_ethdev/rte_ethdev.c b/lib/librte_ethdev/rte_ethdev.c
index a9977df..4d28db5 100644
--- a/lib/librte_ethdev/rte_ethdev.c
+++ b/lib/librte_ethdev/rte_ethdev.c
@@ -4518,6 +4518,65 @@ rte_eth_devargs_parse(const char *dargs, struct 
rte_eth_devargs *eth_da)
return result;
 }
 
+static void __rte_experimental
+eth_dev_event_callback(char *device_name, enum rte_dev_event_type type,
+void *arg)
+{
+   struct rte_eth_dev *eth_dev = (struct rte_eth_dev *)arg;
+
+   switch (type) {
+   case RTE_DEV_EVENT_REMOVE:
+   ethdev_log(INFO, "The device %s has been removed!\n",
+   device_name);
+
+   if (!device_name || !eth_dev)
+   return;
+
+   if (!(eth_dev->data->dev_flags & RTE_ETH_DEV_INTR_RMV))
+   return;
+
+   if (!strcmp(device_name, eth_dev->device->name))
+   _rte_eth_dev_callback_process(eth_dev,
+ RTE_ETH_EVENT_INTR_RMV,
+ NULL);
+   break;
+   case RTE_DEV_EVENT_ADD:
+   ethdev_log(INFO, "The device %s has been added!\n",
+   device_name);
+   break;
+   default:
+   break;
+   }
+}
+
+int __rte_experimental
+rte_eth_dev_event_handler_install(struct rte_eth_dev *eth_dev)
+{
+   int result = 0;
+
+   result = rte_dev_event_callback_register(eth_dev->device->name,
+   eth_dev_event_callback, eth_dev);
+   if (result)
+   ethdev_log(ERR, "device event callback register failed for "
+   "device %s!\n", eth_dev->device->name);
+
+   return result;
+}
+
+int __rte_experimental
+rte_eth_dev_event_handler_uninstall(struct rte_eth_dev *eth_dev)
+{
+   int result = 0;
+
+   result = rte_dev_event_callback_unregister(eth_dev->device->name,
+   eth_dev_event_callback, eth_dev);
+   if (result)
+   ethdev_log(ERR, "device event callback unregister failed for"
+   " device %s!\n", eth_dev->device->name);
+
+   return result;
+}
+
 RTE_INIT(ethdev_init_log);
 static void
 ethdev_init_log(void)
diff --git a/lib/librte_ethdev/rte_ethdev_driver.h 
b/lib/librte_ethdev/rte_ethdev_driver.h
index c9c825e..3de1cd4 100644
--- a/lib/librte_ethdev/rte_ethdev_driver.h
+++ b/lib/librte_ethdev/rte_ethdev_driver.h
@@ -82,6 +82,38 @@ int rte_eth_dev_release_port(struct rte_eth_dev *eth_dev);
 void _rte_eth_dev_reset(struct rte_eth_dev *dev);
 
 /**
+ * @warning
+ * @b EXPERIMENTAL: this API may change without prior notice.
+ *
+ * Implement a helper to install the device event handler for the specific
+ * ether device.
+ *
+ * @param dev
+ *  Pointer to struct rte_eth_dev.
+ *
+ * @return
+ *   - 0 on success, negative on error
+ */
+int __rte_experimental
+rte_eth_dev_event_handler_install(struct rte_eth_dev *dev);
+
+/**
+ * @warning
+ * @b EXPERIMENTAL: this API may change without prior notice.
+ *
+ * Implement a helper to uninstall the device event handler for the specific
+ * ether device.
+ *
+ * @param dev
+ *  Pointer to struct rte_eth_dev

[dpdk-dev] [PATCH v5 2/4] net/ixgbe: install ethdev hotplug handler in ixgbe

2018-07-11 Thread Jeff Guo
This patch aim to enable hotplug detect in ixgbe PMD. Firstly it
set the flags RTE_PCI_DRV_INTR_RMV in drv_flags to announce the hotplug
ability, and then use rte_eth_dev_event_handler_install to install
the hotplug event handler for ethdev. When eal detect the hotplug event,
it will call the ethdev callback to process it. If the event is hotplug
removal, it will trigger the RTE_ETH_EVENT_INTR_RMV event into ethdev
callback to let app process the hotplug for this ethdev.

This is an example for other driver, that if any driver support hotplug
feature could be use this way to install hotplug handler.

Signed-off-by: Jeff Guo 
Acked-by: Wenzhuo Lu 
---
v5->v4:
no change.
---
 drivers/net/ixgbe/ixgbe_ethdev.c | 8 +++-
 1 file changed, 7 insertions(+), 1 deletion(-)

diff --git a/drivers/net/ixgbe/ixgbe_ethdev.c b/drivers/net/ixgbe/ixgbe_ethdev.c
index 87d2ad0..e7ae9bf 100644
--- a/drivers/net/ixgbe/ixgbe_ethdev.c
+++ b/drivers/net/ixgbe/ixgbe_ethdev.c
@@ -1678,6 +1678,9 @@ eth_ixgbevf_dev_init(struct rte_eth_dev *eth_dev)
rte_intr_enable(intr_handle);
ixgbevf_intr_enable(eth_dev);
 
+   /* install the dev event handler for ethdev. */
+   rte_eth_dev_event_handler_install(eth_dev);
+
PMD_INIT_LOG(DEBUG, "port %d vendorID=0x%x deviceID=0x%x mac.type=%s",
 eth_dev->data->port_id, pci_dev->id.vendor_id,
 pci_dev->id.device_id, "ixgbe_mac_82599_vf");
@@ -1718,6 +1721,9 @@ eth_ixgbevf_dev_uninit(struct rte_eth_dev *eth_dev)
rte_intr_callback_unregister(intr_handle,
 ixgbevf_dev_interrupt_handler, eth_dev);
 
+   /* uninstall the dev event handler for ethdev. */
+   rte_eth_dev_event_handler_uninstall(eth_dev);
+
return 0;
 }
 
@@ -1801,7 +1807,7 @@ static int eth_ixgbe_pci_remove(struct rte_pci_device 
*pci_dev)
 static struct rte_pci_driver rte_ixgbe_pmd = {
.id_table = pci_id_ixgbe_map,
.drv_flags = RTE_PCI_DRV_NEED_MAPPING | RTE_PCI_DRV_INTR_LSC |
-RTE_PCI_DRV_IOVA_AS_VA,
+RTE_PCI_DRV_IOVA_AS_VA | RTE_PCI_DRV_INTR_RMV,
.probe = eth_ixgbe_pci_probe,
.remove = eth_ixgbe_pci_remove,
 };
-- 
2.7.4



[dpdk-dev] [PATCH v5 3/4] net/i40e: install hotplug handler in i40e

2018-07-11 Thread Jeff Guo
This patch aim to enable hotplug detect in i40e PMD. Firstly it
set the flags RTE_PCI_DRV_INTR_RMV in drv_flags to announce the hotplug
ability, and then use rte_eth_dev_event_handler_install to install
the hotplug event handler for ethdev. When eal detect the hotplug event,
it will call the ethdev callback to process it. If the event is hotplug
removal, it will trigger the RTE_ETH_EVENT_INTR_RMV event into ethdev
callback to let app process the hotplug for this ethdev.

This is an example for other driver, that if any driver support hotplug
feature could be use this way to install hotplug handler.

Signed-off-by: Jeff Guo 
Acked-by: Qi Zhang 
---
v5->v4:
no change.
---
 drivers/net/i40e/i40e_ethdev.c | 8 +++-
 1 file changed, 7 insertions(+), 1 deletion(-)

diff --git a/drivers/net/i40e/i40e_ethdev.c b/drivers/net/i40e/i40e_ethdev.c
index 13c5d32..8fccf04 100644
--- a/drivers/net/i40e/i40e_ethdev.c
+++ b/drivers/net/i40e/i40e_ethdev.c
@@ -688,7 +688,7 @@ static int eth_i40e_pci_remove(struct rte_pci_device 
*pci_dev)
 static struct rte_pci_driver rte_i40e_pmd = {
.id_table = pci_id_i40e_map,
.drv_flags = RTE_PCI_DRV_NEED_MAPPING | RTE_PCI_DRV_INTR_LSC |
-RTE_PCI_DRV_IOVA_AS_VA,
+RTE_PCI_DRV_IOVA_AS_VA | RTE_PCI_DRV_INTR_RMV,
.probe = eth_i40e_pci_probe,
.remove = eth_i40e_pci_remove,
 };
@@ -1442,6 +1442,9 @@ eth_i40e_dev_init(struct rte_eth_dev *dev, void 
*init_params __rte_unused)
rte_intr_callback_register(intr_handle,
   i40e_dev_interrupt_handler, dev);
 
+   /* install the dev event handler for ethdev. */
+   rte_eth_dev_event_handler_install(dev);
+
/* configure and enable device interrupt */
i40e_pf_config_irq0(hw, TRUE);
i40e_pf_enable_irq0(hw);
@@ -1674,6 +1677,9 @@ eth_i40e_dev_uninit(struct rte_eth_dev *dev)
/* Remove all Traffic Manager configuration */
i40e_tm_conf_uninit(dev);
 
+   /* uninstall the dev event handler for ethdev. */
+   rte_eth_dev_event_handler_uninstall(dev);
+
return 0;
 }
 
-- 
2.7.4



[dpdk-dev] [PATCH v5 4/4] testpmd: remove the dev event callback register

2018-07-11 Thread Jeff Guo
Since now we can use driver to management the eal event for hotplug,
so no need to register dev event callback in app anymore. This patch
remove the related code.

Signed-off-by: Jeff Guo 
Acked-by: Wenzhuo Lu 
---
v5->v4:
no change.
---
 app/test-pmd/testpmd.c | 76 --
 1 file changed, 76 deletions(-)

diff --git a/app/test-pmd/testpmd.c b/app/test-pmd/testpmd.c
index 24c1998..10ed660 100644
--- a/app/test-pmd/testpmd.c
+++ b/app/test-pmd/testpmd.c
@@ -400,12 +400,6 @@ static void check_all_ports_link_status(uint32_t 
port_mask);
 static int eth_event_callback(portid_t port_id,
  enum rte_eth_event_type type,
  void *param, void *ret_param);
-static void eth_dev_event_callback(char *device_name,
-   enum rte_dev_event_type type,
-   void *param);
-static int eth_dev_event_callback_register(void);
-static int eth_dev_event_callback_unregister(void);
-
 
 /*
  * Check if all the ports are started.
@@ -1915,39 +1909,6 @@ reset_port(portid_t pid)
printf("Done\n");
 }
 
-static int
-eth_dev_event_callback_register(void)
-{
-   int ret;
-
-   /* register the device event callback */
-   ret = rte_dev_event_callback_register(NULL,
-   eth_dev_event_callback, NULL);
-   if (ret) {
-   printf("Failed to register device event callback\n");
-   return -1;
-   }
-
-   return 0;
-}
-
-
-static int
-eth_dev_event_callback_unregister(void)
-{
-   int ret;
-
-   /* unregister the device event callback */
-   ret = rte_dev_event_callback_unregister(NULL,
-   eth_dev_event_callback, NULL);
-   if (ret < 0) {
-   printf("Failed to unregister device event callback\n");
-   return -1;
-   }
-
-   return 0;
-}
-
 void
 attach_port(char *identifier)
 {
@@ -2049,10 +2010,6 @@ pmd_test_exit(void)
RTE_LOG(ERR, EAL,
"fail to stop device event monitor.");
 
-   ret = eth_dev_event_callback_unregister();
-   if (ret)
-   RTE_LOG(ERR, EAL,
-   "fail to unregister all event callbacks.");
}
 
printf("\nBye...\n");
@@ -2191,37 +2148,6 @@ eth_event_callback(portid_t port_id, enum 
rte_eth_event_type type, void *param,
return 0;
 }
 
-/* This function is used by the interrupt thread */
-static void
-eth_dev_event_callback(char *device_name, enum rte_dev_event_type type,
-__rte_unused void *arg)
-{
-   if (type >= RTE_DEV_EVENT_MAX) {
-   fprintf(stderr, "%s called upon invalid event %d\n",
-   __func__, type);
-   fflush(stderr);
-   }
-
-   switch (type) {
-   case RTE_DEV_EVENT_REMOVE:
-   RTE_LOG(ERR, EAL, "The device: %s has been removed!\n",
-   device_name);
-   /* TODO: After finish failure handle, begin to stop
-* packet forward, stop port, close port, detach port.
-*/
-   break;
-   case RTE_DEV_EVENT_ADD:
-   RTE_LOG(ERR, EAL, "The device: %s has been added!\n",
-   device_name);
-   /* TODO: After finish kernel driver binding,
-* begin to attach port.
-*/
-   break;
-   default:
-   break;
-   }
-}
-
 static int
 set_tx_queue_stats_mapping_registers(portid_t port_id, struct rte_port *port)
 {
@@ -2735,8 +2661,6 @@ main(int argc, char** argv)
rte_errno = EINVAL;
return -1;
}
-   eth_dev_event_callback_register();
-
}
 
if (start_port(RTE_PORT_ALL) != 0)
-- 
2.7.4



Re: [dpdk-dev] [PATCH v10 04/27] kvargs: introduce a more flexible parsing function

2018-07-11 Thread Shreyansh Jain

On Friday 06 July 2018 03:30 AM, Thomas Monjalon wrote:

05/07/2018 13:48, Gaetan Rivet:

This function permits defining additional terminating characters,
ending the parsing to arbitrary delimiters.

[...]

+__rte_experimental
+struct rte_kvargs *rte_kvargs_parse2(const char *args,
+   const char *const valid_keys[],
+   const char *valid_ends);


The name is not ideal but I don't have a better idea.


Does 'rte_kvargs_parse_delim()' seem any better?
Or, a more obvious one 'rte_kvargs_parse_with_delim()'.

[...]

As I am already late in commenting on this, I am OK changing this later 
as well (using the 'experimental' blanket) when a better suggestion 
comes along. So, from me as well:


Acked-by: Shreyansh Jain 



[dpdk-dev] [PATCH v5 02/16] compress/qat: add makefiles for PMD

2018-07-11 Thread Fiona Trahe
Add Makefiles, directory and empty source files for compression PMD.
Handle cases for building either symmetric crypto PMD
or compression PMD or both and the common files both depend on.

Signed-off-by: Fiona Trahe 
Signed-off-by: Tomasz Jozwiak 
---
 MAINTAINERS |  4 +++
 config/common_base  |  3 +-
 drivers/common/qat/Makefile | 60 +++--
 drivers/compress/qat/qat_comp.c |  5 
 drivers/compress/qat/qat_comp.h | 14 +
 drivers/compress/qat/qat_comp_pmd.c |  5 
 drivers/compress/qat/qat_comp_pmd.h | 15 ++
 test/test/test_cryptodev.c  |  6 ++--
 8 files changed, 86 insertions(+), 26 deletions(-)
 create mode 100644 drivers/compress/qat/qat_comp.c
 create mode 100644 drivers/compress/qat/qat_comp.h
 create mode 100644 drivers/compress/qat/qat_comp_pmd.c
 create mode 100644 drivers/compress/qat/qat_comp_pmd.h

diff --git a/MAINTAINERS b/MAINTAINERS
index 8050b5d..50b2dff 100644
--- a/MAINTAINERS
+++ b/MAINTAINERS
@@ -852,6 +852,10 @@ F: drivers/compress/isal/
 F: doc/guides/compressdevs/isal.rst
 F: doc/guides/compressdevs/features/isal.ini
 
+Intel QuickAssist
+M: Fiona Trahe 
+F: drivers/compress/qat/
+F: drivers/common/qat/
 
 Eventdev Drivers
 
diff --git a/config/common_base b/config/common_base
index e4241db..1e340b4 100644
--- a/config/common_base
+++ b/config/common_base
@@ -480,7 +480,8 @@ CONFIG_RTE_LIBRTE_DPAA_MAX_CRYPTODEV=4
 #
 # Compile PMD for QuickAssist based devices
 #
-CONFIG_RTE_LIBRTE_PMD_QAT=n
+CONFIG_RTE_LIBRTE_PMD_QAT=y
+CONFIG_RTE_LIBRTE_PMD_QAT_SYM=n
 #
 # Max. number of QuickAssist devices, which can be detected and attached
 #
diff --git a/drivers/common/qat/Makefile b/drivers/common/qat/Makefile
index 02e83f9..2a4c99b 100644
--- a/drivers/common/qat/Makefile
+++ b/drivers/common/qat/Makefile
@@ -3,48 +3,64 @@
 
 include $(RTE_SDK)/mk/rte.vars.mk
 
-# library name
-LIB = librte_pmd_qat.a
-
-# library version
-LIBABIVER := 1
-
-# build flags
-CFLAGS += $(WERROR_FLAGS)
-CFLAGS += -O3
-
 # build directories
 QAT_CRYPTO_DIR := $(RTE_SDK)/drivers/crypto/qat
-VPATH=$(QAT_CRYPTO_DIR)
+QAT_COMPRESS_DIR := $(RTE_SDK)/drivers/compress/qat
+VPATH=$(QAT_CRYPTO_DIR):$(QAT_COMPRESS_DIR)
 
 # external library include paths
 CFLAGS += -I$(SRCDIR)/qat_adf
 CFLAGS += -I$(SRCDIR)
 CFLAGS += -I$(QAT_CRYPTO_DIR)
+CFLAGS += -I$(QAT_COMPRESS_DIR)
 
-# library common source files
-SRCS-y += qat_device.c
-SRCS-y += qat_common.c
-SRCS-y += qat_logs.c
-SRCS-y += qat_qp.c
+
+ifeq ($(CONFIG_RTE_LIBRTE_COMPRESSDEV),y)
+   CFLAGS += -DALLOW_EXPERIMENTAL_API
+   LDLIBS += -lrte_compressdev
+   SRCS-y += qat_comp.c
+   SRCS-y += qat_comp_pmd.c
+   build_qat = yes
+endif
 
 # library symmetric crypto source files
 ifeq ($(CONFIG_RTE_LIBRTE_CRYPTODEV),y)
+ifeq ($(CONFIG_RTE_LIBRTE_PMD_QAT_SYM),y)
LDLIBS += -lrte_cryptodev
LDLIBS += -lcrypto
CFLAGS += -DBUILD_QAT_SYM
SRCS-y += qat_sym.c
SRCS-y += qat_sym_session.c
SRCS-y += qat_sym_pmd.c
+   build_qat = yes
+endif
 endif
 
-LDLIBS += -lrte_eal -lrte_mbuf -lrte_mempool
-LDLIBS += -lrte_pci -lrte_bus_pci
+ifdef build_qat
 
-# export include files
-SYMLINK-y-include +=
+   # library name
+   LIB = librte_pmd_qat.a
 
-# versioning export map
-EXPORT_MAP := ../../crypto/qat/rte_pmd_qat_version.map
+   # library version
+   LIBABIVER := 1
+   # build flags
+   CFLAGS += $(WERROR_FLAGS)
+   CFLAGS += -O3
+
+   # library common source files
+   SRCS-y += qat_device.c
+   SRCS-y += qat_common.c
+   SRCS-y += qat_logs.c
+   SRCS-y += qat_qp.c
+
+   LDLIBS += -lrte_eal -lrte_mbuf -lrte_mempool
+   LDLIBS += -lrte_pci -lrte_bus_pci
+
+   # export include files
+   SYMLINK-y-include +=
+
+   # versioning export map
+   EXPORT_MAP := ../../crypto/qat/rte_pmd_qat_version.map
+endif
 
 include $(RTE_SDK)/mk/rte.lib.mk
diff --git a/drivers/compress/qat/qat_comp.c b/drivers/compress/qat/qat_comp.c
new file mode 100644
index 000..caa1158
--- /dev/null
+++ b/drivers/compress/qat/qat_comp.c
@@ -0,0 +1,5 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(c) 2018 Intel Corporation
+ */
+
+#include "qat_comp.h"
diff --git a/drivers/compress/qat/qat_comp.h b/drivers/compress/qat/qat_comp.h
new file mode 100644
index 000..89c475e
--- /dev/null
+++ b/drivers/compress/qat/qat_comp.h
@@ -0,0 +1,14 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(c) 2015-2018 Intel Corporation
+ */
+
+#ifndef _QAT_COMP_H_
+#define _QAT_COMP_H_
+
+#ifdef RTE_LIBRTE_COMPRESSDEV
+
+#include 
+#include 
+
+#endif
+#endif
diff --git a/drivers/compress/qat/qat_comp_pmd.c 
b/drivers/compress/qat/qat_comp_pmd.c
new file mode 100644
index 000..fb035d1
--- /dev/null
+++ b/drivers/compress/qat/qat_comp_pmd.c
@@ -0,0 +1,5 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(c) 2015-2018 Intel Corporation
+ */
+
+#inc

[dpdk-dev] [PATCH v5 01/16] common/qat: updated firmware headers

2018-07-11 Thread Fiona Trahe
Updated to latest firmware headers files for QuickAssist devices.
Includes updates for symmetric crypto, PKE and Compression services.

Signed-off-by: Fiona Trahe 
---
 drivers/common/qat/qat_adf/icp_qat_fw.h  |  69 +++-
 drivers/common/qat/qat_adf/icp_qat_fw_comp.h | 482 +++
 drivers/common/qat/qat_adf/icp_qat_hw.h  | 130 +++-
 3 files changed, 654 insertions(+), 27 deletions(-)
 create mode 100644 drivers/common/qat/qat_adf/icp_qat_fw_comp.h

diff --git a/drivers/common/qat/qat_adf/icp_qat_fw.h 
b/drivers/common/qat/qat_adf/icp_qat_fw.h
index ae39b7f..8f7cb37 100644
--- a/drivers/common/qat/qat_adf/icp_qat_fw.h
+++ b/drivers/common/qat/qat_adf/icp_qat_fw.h
@@ -117,6 +117,10 @@ struct icp_qat_fw_comn_resp {
 #define ICP_QAT_FW_COMN_VALID_FLAG_BITPOS 7
 #define ICP_QAT_FW_COMN_VALID_FLAG_MASK 0x1
 #define ICP_QAT_FW_COMN_HDR_RESRVD_FLD_MASK 0x7F
+#define ICP_QAT_FW_COMN_CNV_FLAG_BITPOS 6
+#define ICP_QAT_FW_COMN_CNV_FLAG_MASK 0x1
+#define ICP_QAT_FW_COMN_CNVNR_FLAG_BITPOS 5
+#define ICP_QAT_FW_COMN_CNVNR_FLAG_MASK 0x1
 
 #define ICP_QAT_FW_COMN_OV_SRV_TYPE_GET(icp_qat_fw_comn_req_hdr_t) \
icp_qat_fw_comn_req_hdr_t.service_type
@@ -133,6 +137,16 @@ struct icp_qat_fw_comn_resp {
 #define ICP_QAT_FW_COMN_HDR_VALID_FLAG_GET(hdr_t) \
ICP_QAT_FW_COMN_VALID_FLAG_GET(hdr_t.hdr_flags)
 
+#define ICP_QAT_FW_COMN_HDR_CNVNR_FLAG_GET(hdr_flags) \
+   QAT_FIELD_GET(hdr_flags, \
+   ICP_QAT_FW_COMN_CNVNR_FLAG_BITPOS, \
+   ICP_QAT_FW_COMN_CNVNR_FLAG_MASK)
+
+#define ICP_QAT_FW_COMN_HDR_CNV_FLAG_GET(hdr_flags) \
+   QAT_FIELD_GET(hdr_flags, \
+   ICP_QAT_FW_COMN_CNV_FLAG_BITPOS, \
+   ICP_QAT_FW_COMN_CNV_FLAG_MASK)
+
 #define ICP_QAT_FW_COMN_HDR_VALID_FLAG_SET(hdr_t, val) \
ICP_QAT_FW_COMN_VALID_FLAG_SET(hdr_t, val)
 
@@ -204,29 +218,44 @@ struct icp_qat_fw_comn_resp {
& ICP_QAT_FW_COMN_NEXT_ID_MASK) | \
((val) & ICP_QAT_FW_COMN_CURR_ID_MASK)); }
 
+#define ICP_QAT_FW_COMN_NEXT_ID_SET_2(next_curr_id, val)   
\
+   do {   \
+   (next_curr_id) =   \
+   (((next_curr_id) & ICP_QAT_FW_COMN_CURR_ID_MASK) | \
+(((val) << ICP_QAT_FW_COMN_NEXT_ID_BITPOS) &  \
+ ICP_QAT_FW_COMN_NEXT_ID_MASK))   \
+   } while (0)
+
+#define ICP_QAT_FW_COMN_CURR_ID_SET_2(next_curr_id, val)   
\
+   do {   \
+   (next_curr_id) =   \
+   (((next_curr_id) & ICP_QAT_FW_COMN_NEXT_ID_MASK) | \
+((val) & ICP_QAT_FW_COMN_CURR_ID_MASK))   \
+   } while (0)
+
 #define QAT_COMN_RESP_CRYPTO_STATUS_BITPOS 7
 #define QAT_COMN_RESP_CRYPTO_STATUS_MASK 0x1
+#define QAT_COMN_RESP_PKE_STATUS_BITPOS 6
+#define QAT_COMN_RESP_PKE_STATUS_MASK 0x1
 #define QAT_COMN_RESP_CMP_STATUS_BITPOS 5
 #define QAT_COMN_RESP_CMP_STATUS_MASK 0x1
 #define QAT_COMN_RESP_XLAT_STATUS_BITPOS 4
 #define QAT_COMN_RESP_XLAT_STATUS_MASK 0x1
 #define QAT_COMN_RESP_CMP_END_OF_LAST_BLK_BITPOS 3
 #define QAT_COMN_RESP_CMP_END_OF_LAST_BLK_MASK 0x1
-
-#define ICP_QAT_FW_COMN_RESP_STATUS_BUILD(crypto, comp, xlat, eolb) \
-   crypto) & QAT_COMN_RESP_CRYPTO_STATUS_MASK) << \
-   QAT_COMN_RESP_CRYPTO_STATUS_BITPOS) | \
-   (((comp) & QAT_COMN_RESP_CMP_STATUS_MASK) << \
-   QAT_COMN_RESP_CMP_STATUS_BITPOS) | \
-   (((xlat) & QAT_COMN_RESP_XLAT_STATUS_MASK) << \
-   QAT_COMN_RESP_XLAT_STATUS_BITPOS) | \
-   (((eolb) & QAT_COMN_RESP_CMP_END_OF_LAST_BLK_MASK) << \
-   QAT_COMN_RESP_CMP_END_OF_LAST_BLK_BITPOS))
+#define QAT_COMN_RESP_UNSUPPORTED_REQUEST_BITPOS 2
+#define QAT_COMN_RESP_UNSUPPORTED_REQUEST_MASK 0x1
+#define QAT_COMN_RESP_XLT_WA_APPLIED_BITPOS 0
+#define QAT_COMN_RESP_XLT_WA_APPLIED_MASK 0x1
 
 #define ICP_QAT_FW_COMN_RESP_CRYPTO_STAT_GET(status) \
QAT_FIELD_GET(status, QAT_COMN_RESP_CRYPTO_STATUS_BITPOS, \
QAT_COMN_RESP_CRYPTO_STATUS_MASK)
 
+#define ICP_QAT_FW_COMN_RESP_PKE_STAT_GET(status) \
+   QAT_FIELD_GET(status, QAT_COMN_RESP_PKE_STATUS_BITPOS, \
+   QAT_COMN_RESP_PKE_STATUS_MASK)
+
 #define ICP_QAT_FW_COMN_RESP_CMP_STAT_GET(status) \
QAT_FIELD_GET(status, QAT_COMN_RESP_CMP_STATUS_BITPOS, \
QAT_COMN_RESP_CMP_STATUS_MASK)
@@ -235,10 +264,18 @@ struct icp_qat_fw_comn_resp {
QAT_FIELD_GET(status, QAT_COMN_RESP_XLAT_STATUS_BITPOS, \
QAT_COMN_RESP_XLAT_STATUS_MASK)
 
+#define ICP_QAT_FW_COMN_RESP_XLT_WA_APPLIED_GET(status) \
+   QAT_FIELD_GET(status, QAT_COMN_RESP_XLT_WA_APPLIED_BITPOS, \
+   QAT_COMN_RESP_XLT_WA_APPLIED_MASK)
+
 #define ICP_QAT_FW_COMN_RESP_CMP_END_OF_LAST_BLK_FLAG_GET(status) \
QAT_FIELD_GET(

[dpdk-dev] [PATCH v5 04/16] compress/qat: add xform processing

2018-07-11 Thread Fiona Trahe
Add code to process compressdev rte_comp_xforms, creating
private qat_comp_xforms with prepared firmware message templates.

Signed-off-by: Fiona Trahe 
Signed-off-by: Tomasz Jozwiak 
---
 drivers/compress/qat/qat_comp.c | 239 
 drivers/compress/qat/qat_comp.h |  30 +
 drivers/compress/qat/qat_comp_pmd.h |  16 +++
 3 files changed, 285 insertions(+)

diff --git a/drivers/compress/qat/qat_comp.c b/drivers/compress/qat/qat_comp.c
index caa1158..cb2005a 100644
--- a/drivers/compress/qat/qat_comp.c
+++ b/drivers/compress/qat/qat_comp.c
@@ -2,4 +2,243 @@
  * Copyright(c) 2018 Intel Corporation
  */
 
+
+#include 
+#include 
+#include 
+#include 
+#include 
+#include 
+#include 
+#include 
+#include 
+#include 
+#include 
+
+#include "qat_logs.h"
 #include "qat_comp.h"
+#include "qat_comp_pmd.h"
+
+unsigned int
+qat_comp_xform_size(void)
+{
+   return RTE_ALIGN_CEIL(sizeof(struct qat_comp_xform), 8);
+}
+
+static void qat_comp_create_req_hdr(struct icp_qat_fw_comn_req_hdr *header,
+   enum qat_comp_request_type request)
+{
+   if (request == QAT_COMP_REQUEST_FIXED_COMP_STATELESS)
+   header->service_cmd_id = ICP_QAT_FW_COMP_CMD_STATIC;
+   else if (request == QAT_COMP_REQUEST_DYNAMIC_COMP_STATELESS)
+   header->service_cmd_id = ICP_QAT_FW_COMP_CMD_DYNAMIC;
+   else if (request == QAT_COMP_REQUEST_DECOMPRESS)
+   header->service_cmd_id = ICP_QAT_FW_COMP_CMD_DECOMPRESS;
+
+   header->service_type = ICP_QAT_FW_COMN_REQ_CPM_FW_COMP;
+   header->hdr_flags =
+   ICP_QAT_FW_COMN_HDR_FLAGS_BUILD(ICP_QAT_FW_COMN_REQ_FLAG_SET);
+
+   header->comn_req_flags = ICP_QAT_FW_COMN_FLAGS_BUILD(
+   QAT_COMN_CD_FLD_TYPE_16BYTE_DATA, QAT_COMN_PTR_TYPE_FLAT);
+}
+
+static int qat_comp_create_templates(struct qat_comp_xform *qat_xform,
+   const struct rte_memzone *interm_buff_mz __rte_unused,
+   const struct rte_comp_xform *xform)
+{
+   struct icp_qat_fw_comp_req *comp_req;
+   int comp_level, algo;
+   uint32_t req_par_flags;
+   int direction = ICP_QAT_HW_COMPRESSION_DIR_COMPRESS;
+
+   if (unlikely(qat_xform == NULL)) {
+   QAT_LOG(ERR, "Session was not created for this device");
+   return -EINVAL;
+   }
+
+   if (qat_xform->qat_comp_request_type == QAT_COMP_REQUEST_DECOMPRESS) {
+   direction = ICP_QAT_HW_COMPRESSION_DIR_DECOMPRESS;
+   comp_level = ICP_QAT_HW_COMPRESSION_DEPTH_1;
+   req_par_flags = ICP_QAT_FW_COMP_REQ_PARAM_FLAGS_BUILD(
+   ICP_QAT_FW_COMP_SOP, ICP_QAT_FW_COMP_EOP,
+   ICP_QAT_FW_COMP_BFINAL, ICP_QAT_FW_COMP_NO_CNV,
+   ICP_QAT_FW_COMP_NO_CNV_RECOVERY);
+
+   } else {
+   if (xform->compress.level == RTE_COMP_LEVEL_PMD_DEFAULT)
+   comp_level = ICP_QAT_HW_COMPRESSION_DEPTH_8;
+   else if (xform->compress.level == 1)
+   comp_level = ICP_QAT_HW_COMPRESSION_DEPTH_1;
+   else if (xform->compress.level == 2)
+   comp_level = ICP_QAT_HW_COMPRESSION_DEPTH_4;
+   else if (xform->compress.level == 3)
+   comp_level = ICP_QAT_HW_COMPRESSION_DEPTH_8;
+   else if (xform->compress.level >= 4 &&
+xform->compress.level <= 9)
+   comp_level = ICP_QAT_HW_COMPRESSION_DEPTH_16;
+   else {
+   QAT_LOG(ERR, "compression level not supported");
+   return -EINVAL;
+   }
+   req_par_flags = ICP_QAT_FW_COMP_REQ_PARAM_FLAGS_BUILD(
+   ICP_QAT_FW_COMP_SOP, ICP_QAT_FW_COMP_EOP,
+   ICP_QAT_FW_COMP_BFINAL, ICP_QAT_FW_COMP_CNV,
+   ICP_QAT_FW_COMP_CNV_RECOVERY);
+   }
+
+   switch (xform->compress.algo) {
+   case RTE_COMP_ALGO_DEFLATE:
+   algo = ICP_QAT_HW_COMPRESSION_ALGO_DEFLATE;
+   break;
+   case RTE_COMP_ALGO_LZS:
+   default:
+   /* RTE_COMP_NULL */
+   QAT_LOG(ERR, "compression algorithm not supported");
+   return -EINVAL;
+   }
+
+   comp_req = &qat_xform->qat_comp_req_tmpl;
+
+   /* Initialize header */
+   qat_comp_create_req_hdr(&comp_req->comn_hdr,
+   qat_xform->qat_comp_request_type);
+
+   comp_req->comn_hdr.serv_specif_flags = ICP_QAT_FW_COMP_FLAGS_BUILD(
+   ICP_QAT_FW_COMP_STATELESS_SESSION,
+   ICP_QAT_FW_COMP_NOT_AUTO_SELECT_BEST,
+   ICP_QAT_FW_COMP_NOT_ENH_AUTO_SELECT_BEST,
+   ICP_QAT_FW_COMP_NOT_DISABLE_TYPE0_ENH_AUTO_SELECT_BEST,
+   ICP_QAT_FW_COMP_DISABLE_SECURE_RAM_USED_AS_INTMD_BUF);
+
+   comp_req->cd_pars.sl.comp_slice

[dpdk-dev] [PATCH v5 03/16] compress/qat: add meson build

2018-07-11 Thread Fiona Trahe
Add meson build files.

Signed-off-by: Tomasz Jozwiak 
Signed-off-by: Fiona Trahe 
---
 drivers/common/qat/Makefile  |  2 +-
 drivers/compress/meson.build |  2 +-
 drivers/compress/qat/meson.build | 18 ++
 drivers/compress/qat/rte_pmd_qat_version.map |  3 +++
 drivers/crypto/qat/meson.build   | 10 ++
 drivers/crypto/qat/rte_pmd_qat_version.map   |  3 ---
 6 files changed, 25 insertions(+), 13 deletions(-)
 create mode 100644 drivers/compress/qat/meson.build
 create mode 100644 drivers/compress/qat/rte_pmd_qat_version.map
 delete mode 100644 drivers/crypto/qat/rte_pmd_qat_version.map

diff --git a/drivers/common/qat/Makefile b/drivers/common/qat/Makefile
index 2a4c99b..c68a032 100644
--- a/drivers/common/qat/Makefile
+++ b/drivers/common/qat/Makefile
@@ -60,7 +60,7 @@ ifdef build_qat
SYMLINK-y-include +=
 
# versioning export map
-   EXPORT_MAP := ../../crypto/qat/rte_pmd_qat_version.map
+   EXPORT_MAP := ../../compress/qat/rte_pmd_qat_version.map
 endif
 
 include $(RTE_SDK)/mk/rte.lib.mk
diff --git a/drivers/compress/meson.build b/drivers/compress/meson.build
index fb136e1..2352ad5 100644
--- a/drivers/compress/meson.build
+++ b/drivers/compress/meson.build
@@ -1,7 +1,7 @@
 # SPDX-License-Identifier: BSD-3-Clause
 # Copyright(c) 2018 Intel Corporation
 
-drivers = ['isal']
+drivers = ['isal', 'qat']
 
 std_deps = ['compressdev'] # compressdev pulls in all other needed deps
 config_flag_fmt = 'RTE_LIBRTE_@0@_PMD'
diff --git a/drivers/compress/qat/meson.build b/drivers/compress/qat/meson.build
new file mode 100644
index 000..9d15076
--- /dev/null
+++ b/drivers/compress/qat/meson.build
@@ -0,0 +1,18 @@
+# SPDX-License-Identifier: BSD-3-Clause
+# Copyright(c) 2017-2018 Intel Corporation
+
+
+# Add our sources files to the list
+allow_experimental_apis = true
+qat_sources += files('qat_comp_pmd.c',
+'qat_comp.c')
+qat_includes += include_directories('.')
+qat_deps += 'compressdev'
+qat_ext_deps += dep
+
+# build the whole driver
+sources += qat_sources
+cflags += qat_cflags
+deps += qat_deps
+ext_deps += qat_ext_deps
+includes += qat_includes
diff --git a/drivers/compress/qat/rte_pmd_qat_version.map 
b/drivers/compress/qat/rte_pmd_qat_version.map
new file mode 100644
index 000..ad6e191
--- /dev/null
+++ b/drivers/compress/qat/rte_pmd_qat_version.map
@@ -0,0 +1,3 @@
+DPDK_18.08 {
+   local: *;
+};
diff --git a/drivers/crypto/qat/meson.build b/drivers/crypto/qat/meson.build
index 2873637..d7cff68 100644
--- a/drivers/crypto/qat/meson.build
+++ b/drivers/crypto/qat/meson.build
@@ -1,6 +1,8 @@
 # SPDX-License-Identifier: BSD-3-Clause
 # Copyright(c) 2017-2018 Intel Corporation
 
+# this does not build the QAT driver, instead that is done in the compression
+# driver which comes later. Here we just add our sources files to the list
 build = false
 dep = dependency('libcrypto', required: false)
 if dep.found()
@@ -13,12 +15,4 @@ if dep.found()
qat_ext_deps += dep
pkgconfig_extra_libs += '-lcrypto'
qat_cflags += '-DBUILD_QAT_SYM'
-
-   # build the whole driver
-   sources += qat_sources
-   cflags += qat_cflags
-   deps += qat_deps
-   ext_deps += qat_ext_deps
-   includes += qat_includes
-   build = true
 endif
diff --git a/drivers/crypto/qat/rte_pmd_qat_version.map 
b/drivers/crypto/qat/rte_pmd_qat_version.map
deleted file mode 100644
index bbaf1c8..000
--- a/drivers/crypto/qat/rte_pmd_qat_version.map
+++ /dev/null
@@ -1,3 +0,0 @@
-DPDK_2.2 {
-   local: *;
-};
\ No newline at end of file
-- 
2.7.4



[dpdk-dev] [PATCH v5 00/16] compress/qat: add compression PMD

2018-07-11 Thread Fiona Trahe
Create compression PMD for Intel QuickAssist devices
Currently only the C62x and c3xxx devices are supported.

The qat comp PMD supports
 - stateless compression and
   decompression using the Deflate algorithm with Fixed Huffman
   encoding. Dynamic huffman encoding is not supported, it
   will be added in a later patch.
 - checksum generation: Adler32, CRC32 and combined.

The compression service is hosted on a QuickAssist VF PCI
device, which is managed by code in the
drivers/common/qat directory.

v5 changes:
 - rebased against latest r/n and features/default.ini 
 - fixed common/qat/Makefile so no build output files
   left hanging around in compress/qat src dir.

v4 changes:
 - corrected capabilities

v3 changes:
 - only commit message changes, i.e. removed ChangeId and fixed typos

v2 changes:
- Added check for correct firmware
- Split patchset
- Added documentation
- removed support for scatter-gather-lists and related config flag
- Removed support for Dynamic huffman encoding and related IM buffer config flag
- Removed support for DH895xcc device



Fiona Trahe (16):
  common/qat: updated firmware headers
  compress/qat: add makefiles for PMD
  compress/qat: add meson build
  compress/qat: add xform processing
  compress/qat: create fw request and process response
  compress/qat: check that correct firmware is in use
  compress/qat: add stats functions
  compress/qat: setup queue-pairs for compression service
  compress/qat: add fns to configure and clear device
  compress/qat: add fn to return device info
  compress/qat: add enqueue/dequeue functions
  compress/qat: add device start and stop fns
  compress/qat: create and populate the ops structure
  compress/qat: add fns to create and destroy the PMD
  compress/qat: prevent device usage if incorrect firmware
  docs/qat: refactor docs adding compression guide

 MAINTAINERS  |   4 +
 config/common_base   |   5 +-
 doc/guides/compressdevs/features/qat.ini |  24 ++
 doc/guides/compressdevs/index.rst|   1 +
 doc/guides/compressdevs/qat_comp.rst |  49 +++
 doc/guides/cryptodevs/qat.rst| 183 ++
 doc/guides/rel_notes/release_18_08.rst   |   5 +
 drivers/common/qat/Makefile  |  60 ++--
 drivers/common/qat/qat_adf/icp_qat_fw.h  |  69 +++-
 drivers/common/qat/qat_adf/icp_qat_fw_comp.h | 482 +++
 drivers/common/qat/qat_adf/icp_qat_hw.h  | 130 +++-
 drivers/common/qat/qat_device.h  |   4 +
 drivers/common/qat/qat_qp.c  |  11 +-
 drivers/common/qat/qat_qp.h  |   5 +
 drivers/compress/meson.build |   2 +-
 drivers/compress/qat/meson.build |  18 +
 drivers/compress/qat/qat_comp.c  | 359 
 drivers/compress/qat/qat_comp.h  |  56 
 drivers/compress/qat/qat_comp_pmd.c  | 407 ++
 drivers/compress/qat/qat_comp_pmd.h  |  39 +++
 drivers/compress/qat/rte_pmd_qat_version.map |   3 +
 drivers/crypto/qat/meson.build   |  10 +-
 drivers/crypto/qat/rte_pmd_qat_version.map   |   3 -
 test/test/test_cryptodev.c   |   6 +-
 24 files changed, 1809 insertions(+), 126 deletions(-)
 create mode 100644 doc/guides/compressdevs/features/qat.ini
 create mode 100644 doc/guides/compressdevs/qat_comp.rst
 create mode 100644 drivers/common/qat/qat_adf/icp_qat_fw_comp.h
 create mode 100644 drivers/compress/qat/meson.build
 create mode 100644 drivers/compress/qat/qat_comp.c
 create mode 100644 drivers/compress/qat/qat_comp.h
 create mode 100644 drivers/compress/qat/qat_comp_pmd.c
 create mode 100644 drivers/compress/qat/qat_comp_pmd.h
 create mode 100644 drivers/compress/qat/rte_pmd_qat_version.map
 delete mode 100644 drivers/crypto/qat/rte_pmd_qat_version.map

-- 
2.7.4



[dpdk-dev] [PATCH v5 10/16] compress/qat: add fn to return device info

2018-07-11 Thread Fiona Trahe
Add capabilities pointer to internal qat comp device
and function to return this and other info.

C
Signed-off-by: Fiona Trahe 
Signed-off-by: Tomasz Jozwiak 
---
 drivers/compress/qat/qat_comp_pmd.c | 19 +++
 drivers/compress/qat/qat_comp_pmd.h |  6 ++
 2 files changed, 25 insertions(+)

diff --git a/drivers/compress/qat/qat_comp_pmd.c 
b/drivers/compress/qat/qat_comp_pmd.c
index beab6e3..482ebd1 100644
--- a/drivers/compress/qat/qat_comp_pmd.c
+++ b/drivers/compress/qat/qat_comp_pmd.c
@@ -194,3 +194,22 @@ qat_comp_dev_close(struct rte_compressdev *dev)
 
return ret;
 }
+
+
+void
+qat_comp_dev_info_get(struct rte_compressdev *dev,
+   struct rte_compressdev_info *info)
+{
+   struct qat_comp_dev_private *comp_dev = dev->data->dev_private;
+   const struct qat_qp_hw_data *comp_hw_qps =
+   qat_gen_config[comp_dev->qat_dev->qat_dev_gen]
+ .qp_hw_data[QAT_SERVICE_COMPRESSION];
+
+   if (info != NULL) {
+   info->max_nb_queue_pairs =
+   qat_qps_per_service(comp_hw_qps,
+   QAT_SERVICE_COMPRESSION);
+   info->feature_flags = dev->feature_flags;
+   info->capabilities = comp_dev->qat_dev_capabilities;
+   }
+}
diff --git a/drivers/compress/qat/qat_comp_pmd.h 
b/drivers/compress/qat/qat_comp_pmd.h
index b10a66f..22576f4 100644
--- a/drivers/compress/qat/qat_comp_pmd.h
+++ b/drivers/compress/qat/qat_comp_pmd.h
@@ -21,6 +21,8 @@ struct qat_comp_dev_private {
/**< The qat pci device hosting the service */
struct rte_compressdev *compressdev;
/**< The pointer to this compression device structure */
+   const struct rte_compressdev_capabilities *qat_dev_capabilities;
+   /* QAT device compression capabilities */
const struct rte_memzone *interm_buff_mz;
/**< The device's memory for intermediate buffers */
struct rte_mempool *xformpool;
@@ -48,5 +50,9 @@ qat_comp_dev_config(struct rte_compressdev *dev,
 int
 qat_comp_dev_close(struct rte_compressdev *dev);
 
+void
+qat_comp_dev_info_get(struct rte_compressdev *dev,
+   struct rte_compressdev_info *info);
+
 #endif
 #endif /* _QAT_COMP_PMD_H_ */
-- 
2.7.4



[dpdk-dev] [PATCH v5 05/16] compress/qat: create fw request and process response

2018-07-11 Thread Fiona Trahe
Add functions to create the request message to send to
firmware and to process the firmware response.

Signed-off-by: Fiona Trahe 
Signed-off-by: Tomasz Jozwiak 
---
 drivers/compress/qat/qat_comp.c | 101 
 drivers/compress/qat/qat_comp.h |   8 +++
 drivers/compress/qat/qat_comp_pmd.h |   1 +
 3 files changed, 110 insertions(+)

diff --git a/drivers/compress/qat/qat_comp.c b/drivers/compress/qat/qat_comp.c
index cb2005a..a32d6ef 100644
--- a/drivers/compress/qat/qat_comp.c
+++ b/drivers/compress/qat/qat_comp.c
@@ -19,6 +19,107 @@
 #include "qat_comp.h"
 #include "qat_comp_pmd.h"
 
+
+int
+qat_comp_build_request(void *in_op, uint8_t *out_msg,
+  void *op_cookie __rte_unused,
+  enum qat_device_gen qat_dev_gen __rte_unused)
+{
+   struct rte_comp_op *op = in_op;
+   struct qat_comp_xform *qat_xform = op->private_xform;
+   const uint8_t *tmpl = (uint8_t *)&qat_xform->qat_comp_req_tmpl;
+   struct icp_qat_fw_comp_req *comp_req =
+   (struct icp_qat_fw_comp_req *)out_msg;
+
+   if (unlikely(op->op_type != RTE_COMP_OP_STATELESS)) {
+   QAT_DP_LOG(ERR, "QAT PMD only supports stateless compression "
+   "operation requests, op (%p) is not a "
+   "stateless operation.", op);
+   return -EINVAL;
+   }
+
+   rte_mov128(out_msg, tmpl);
+   comp_req->comn_mid.opaque_data = (uint64_t)(uintptr_t)op;
+
+   /* common for sgl and flat buffers */
+   comp_req->comp_pars.comp_len = op->src.length;
+   comp_req->comp_pars.out_buffer_sz = rte_pktmbuf_pkt_len(op->m_dst);
+
+   /* sgl */
+   if (op->m_src->next != NULL || op->m_dst->next != NULL) {
+   QAT_DP_LOG(ERR, "QAT PMD doesn't support scatter gather");
+   return -EINVAL;
+
+   } else {
+   ICP_QAT_FW_COMN_PTR_TYPE_SET(comp_req->comn_hdr.comn_req_flags,
+   QAT_COMN_PTR_TYPE_FLAT);
+   comp_req->comn_mid.src_length = rte_pktmbuf_data_len(op->m_src);
+   comp_req->comn_mid.dst_length = rte_pktmbuf_data_len(op->m_dst);
+
+   comp_req->comn_mid.src_data_addr =
+   rte_pktmbuf_mtophys_offset(op->m_src, op->src.offset);
+   comp_req->comn_mid.dest_data_addr =
+   rte_pktmbuf_mtophys_offset(op->m_dst, op->dst.offset);
+   }
+
+#if RTE_LOG_DP_LEVEL >= RTE_LOG_DEBUG
+   QAT_DP_LOG(DEBUG, "Direction: %s",
+   qat_xform->qat_comp_request_type == QAT_COMP_REQUEST_DECOMPRESS ?
+   "decompression" : "compression");
+   QAT_DP_HEXDUMP_LOG(DEBUG, "qat compression message:", comp_req,
+   sizeof(struct icp_qat_fw_comp_req));
+#endif
+   return 0;
+}
+
+int
+qat_comp_process_response(void **op, uint8_t *resp)
+{
+   struct icp_qat_fw_comp_resp *resp_msg =
+   (struct icp_qat_fw_comp_resp *)resp;
+   struct rte_comp_op *rx_op = (struct rte_comp_op *)(uintptr_t)
+   (resp_msg->opaque_data);
+
+#if RTE_LOG_DP_LEVEL >= RTE_LOG_DEBUG
+   QAT_DP_LOG(DEBUG, "Direction: %s",
+   qat_xform->qat_comp_request_type == QAT_COMP_REQUEST_DECOMPRESS ?
+   "decompression" : "compression");
+   QAT_DP_HEXDUMP_LOG(DEBUG,  "qat_response:", (uint8_t *)resp_msg,
+   sizeof(struct icp_qat_fw_comp_resp));
+#endif
+
+   if ((ICP_QAT_FW_COMN_RESP_CMP_STAT_GET(resp_msg->comn_resp.comn_status)
+   | ICP_QAT_FW_COMN_RESP_XLAT_STAT_GET(
+   resp_msg->comn_resp.comn_status)) !=
+   ICP_QAT_FW_COMN_STATUS_FLAG_OK) {
+
+   rx_op->status = RTE_COMP_OP_STATUS_ERROR;
+   rx_op->debug_status =
+   *((uint16_t *)(&resp_msg->comn_resp.comn_error));
+   } else {
+   struct qat_comp_xform *qat_xform = rx_op->private_xform;
+   struct icp_qat_fw_resp_comp_pars *comp_resp =
+ (struct icp_qat_fw_resp_comp_pars *)&resp_msg->comp_resp_pars;
+
+   rx_op->status = RTE_COMP_OP_STATUS_SUCCESS;
+   rx_op->consumed = comp_resp->input_byte_counter;
+   rx_op->produced = comp_resp->output_byte_counter;
+
+   if (qat_xform->checksum_type != RTE_COMP_CHECKSUM_NONE) {
+   if (qat_xform->checksum_type == RTE_COMP_CHECKSUM_CRC32)
+   rx_op->output_chksum = comp_resp->curr_crc32;
+   else if (qat_xform->checksum_type ==
+   RTE_COMP_CHECKSUM_ADLER32)
+   rx_op->output_chksum = comp_resp->curr_adler_32;
+   else
+   rx_op->output_chksum = comp_resp->curr_chksum;
+   }
+   }
+   *op = (void *)rx_op;
+
+   return 0;
+}
+
 unsi

[dpdk-dev] [PATCH v5 07/16] compress/qat: add stats functions

2018-07-11 Thread Fiona Trahe
Add functions to get and clear compression queue-pair statistics.


Signed-off-by: Fiona Trahe 
Signed-off-by: Tomasz Jozwiak 
---
 drivers/compress/qat/qat_comp_pmd.c | 35 +++
 drivers/compress/qat/qat_comp_pmd.h |  7 +++
 2 files changed, 42 insertions(+)

diff --git a/drivers/compress/qat/qat_comp_pmd.c 
b/drivers/compress/qat/qat_comp_pmd.c
index fb035d1..6feffb7 100644
--- a/drivers/compress/qat/qat_comp_pmd.c
+++ b/drivers/compress/qat/qat_comp_pmd.c
@@ -3,3 +3,38 @@
  */
 
 #include "qat_comp_pmd.h"
+
+void
+qat_comp_stats_get(struct rte_compressdev *dev,
+   struct rte_compressdev_stats *stats)
+{
+   struct qat_common_stats qat_stats = {0};
+   struct qat_comp_dev_private *qat_priv;
+
+   if (stats == NULL || dev == NULL) {
+   QAT_LOG(ERR, "invalid ptr: stats %p, dev %p", stats, dev);
+   return;
+   }
+   qat_priv = dev->data->dev_private;
+
+   qat_stats_get(qat_priv->qat_dev, &qat_stats, QAT_SERVICE_COMPRESSION);
+   stats->enqueued_count = qat_stats.enqueued_count;
+   stats->dequeued_count = qat_stats.dequeued_count;
+   stats->enqueue_err_count = qat_stats.enqueue_err_count;
+   stats->dequeue_err_count = qat_stats.dequeue_err_count;
+}
+
+void
+qat_comp_stats_reset(struct rte_compressdev *dev)
+{
+   struct qat_comp_dev_private *qat_priv;
+
+   if (dev == NULL) {
+   QAT_LOG(ERR, "invalid compressdev ptr %p", dev);
+   return;
+   }
+   qat_priv = dev->data->dev_private;
+
+   qat_stats_reset(qat_priv->qat_dev, QAT_SERVICE_COMPRESSION);
+
+}
diff --git a/drivers/compress/qat/qat_comp_pmd.h 
b/drivers/compress/qat/qat_comp_pmd.h
index cd04f11..27d84c8 100644
--- a/drivers/compress/qat/qat_comp_pmd.h
+++ b/drivers/compress/qat/qat_comp_pmd.h
@@ -28,5 +28,12 @@ struct qat_comp_dev_private {
 
 };
 
+void
+qat_comp_stats_reset(struct rte_compressdev *dev);
+
+void
+qat_comp_stats_get(struct rte_compressdev *dev,
+   struct rte_compressdev_stats *stats);
+
 #endif
 #endif /* _QAT_COMP_PMD_H_ */
-- 
2.7.4



[dpdk-dev] [PATCH v5 06/16] compress/qat: check that correct firmware is in use

2018-07-11 Thread Fiona Trahe
Check bit in response message to verify that correct firmware
is in use for compression. If not return an error.

Signed-off-by: Fiona Trahe 
Signed-off-by: Tomasz Jozwiak 
---
 drivers/compress/qat/qat_comp.c | 16 +++-
 drivers/compress/qat/qat_comp.h |  2 ++
 2 files changed, 17 insertions(+), 1 deletion(-)

diff --git a/drivers/compress/qat/qat_comp.c b/drivers/compress/qat/qat_comp.c
index a32d6ef..e8019eb 100644
--- a/drivers/compress/qat/qat_comp.c
+++ b/drivers/compress/qat/qat_comp.c
@@ -2,7 +2,6 @@
  * Copyright(c) 2018 Intel Corporation
  */
 
-
 #include 
 #include 
 #include 
@@ -79,6 +78,8 @@ qat_comp_process_response(void **op, uint8_t *resp)
(struct icp_qat_fw_comp_resp *)resp;
struct rte_comp_op *rx_op = (struct rte_comp_op *)(uintptr_t)
(resp_msg->opaque_data);
+   struct qat_comp_xform *qat_xform = (struct qat_comp_xform *)
+   (rx_op->private_xform);
 
 #if RTE_LOG_DP_LEVEL >= RTE_LOG_DEBUG
QAT_DP_LOG(DEBUG, "Direction: %s",
@@ -88,6 +89,19 @@ qat_comp_process_response(void **op, uint8_t *resp)
sizeof(struct icp_qat_fw_comp_resp));
 #endif
 
+   if (likely(qat_xform->qat_comp_request_type
+   != QAT_COMP_REQUEST_DECOMPRESS)) {
+   if (unlikely(ICP_QAT_FW_COMN_HDR_CNV_FLAG_GET(
+   resp_msg->comn_resp.hdr_flags)
+   == ICP_QAT_FW_COMP_NO_CNV)) {
+   rx_op->status = RTE_COMP_OP_STATUS_ERROR;
+   rx_op->debug_status = ERR_CODE_QAT_COMP_WRONG_FW;
+   *op = (void *)rx_op;
+   QAT_DP_LOG(ERR, "QAT has wrong firmware");
+   return 0;
+   }
+   }
+
if ((ICP_QAT_FW_COMN_RESP_CMP_STAT_GET(resp_msg->comn_resp.comn_status)
| ICP_QAT_FW_COMN_RESP_XLAT_STAT_GET(
resp_msg->comn_resp.comn_status)) !=
diff --git a/drivers/compress/qat/qat_comp.h b/drivers/compress/qat/qat_comp.h
index 46105b4..937f3c8 100644
--- a/drivers/compress/qat/qat_comp.h
+++ b/drivers/compress/qat/qat_comp.h
@@ -15,6 +15,8 @@
 #include "icp_qat_fw_comp.h"
 #include "icp_qat_fw_la.h"
 
+#define ERR_CODE_QAT_COMP_WRONG_FW -99
+
 enum qat_comp_request_type {
QAT_COMP_REQUEST_FIXED_COMP_STATELESS,
QAT_COMP_REQUEST_DYNAMIC_COMP_STATELESS,
-- 
2.7.4



[dpdk-dev] [PATCH v5 08/16] compress/qat: setup queue-pairs for compression service

2018-07-11 Thread Fiona Trahe
Setup and clear queue-pairs for handling compression
requests and responses.


Signed-off-by: Fiona Trahe 
Signed-off-by: Tomasz Jozwiak 
---
 drivers/compress/qat/qat_comp.h |  2 ++
 drivers/compress/qat/qat_comp_pmd.c | 61 +
 drivers/compress/qat/qat_comp_pmd.h |  6 
 3 files changed, 69 insertions(+)

diff --git a/drivers/compress/qat/qat_comp.h b/drivers/compress/qat/qat_comp.h
index 937f3c8..9e6861b 100644
--- a/drivers/compress/qat/qat_comp.h
+++ b/drivers/compress/qat/qat_comp.h
@@ -24,6 +24,8 @@ enum qat_comp_request_type {
REQ_COMP_END
 };
 
+struct qat_comp_op_cookie {
+};
 
 struct qat_comp_xform {
struct icp_qat_fw_comp_req qat_comp_req_tmpl;
diff --git a/drivers/compress/qat/qat_comp_pmd.c 
b/drivers/compress/qat/qat_comp_pmd.c
index 6feffb7..5ae6caf 100644
--- a/drivers/compress/qat/qat_comp_pmd.c
+++ b/drivers/compress/qat/qat_comp_pmd.c
@@ -2,6 +2,7 @@
  * Copyright(c) 2015-2018 Intel Corporation
  */
 
+#include "qat_comp.h"
 #include "qat_comp_pmd.h"
 
 void
@@ -38,3 +39,63 @@ qat_comp_stats_reset(struct rte_compressdev *dev)
qat_stats_reset(qat_priv->qat_dev, QAT_SERVICE_COMPRESSION);
 
 }
+
+int
+qat_comp_qp_release(struct rte_compressdev *dev, uint16_t queue_pair_id)
+{
+   struct qat_comp_dev_private *qat_private = dev->data->dev_private;
+
+   QAT_LOG(DEBUG, "Release comp qp %u on device %d",
+   queue_pair_id, dev->data->dev_id);
+
+   qat_private->qat_dev->qps_in_use[QAT_SERVICE_COMPRESSION][queue_pair_id]
+   = NULL;
+
+   return qat_qp_release((struct qat_qp **)
+   &(dev->data->queue_pairs[queue_pair_id]));
+}
+
+int
+qat_comp_qp_setup(struct rte_compressdev *dev, uint16_t qp_id,
+ uint32_t max_inflight_ops, int socket_id)
+{
+   int ret = 0;
+   struct qat_qp_config qat_qp_conf;
+
+   struct qat_qp **qp_addr =
+   (struct qat_qp **)&(dev->data->queue_pairs[qp_id]);
+   struct qat_comp_dev_private *qat_private = dev->data->dev_private;
+   const struct qat_qp_hw_data *comp_hw_qps =
+   qat_gen_config[qat_private->qat_dev->qat_dev_gen]
+ .qp_hw_data[QAT_SERVICE_COMPRESSION];
+   const struct qat_qp_hw_data *qp_hw_data = comp_hw_qps + qp_id;
+
+   /* If qp is already in use free ring memory and qp metadata. */
+   if (*qp_addr != NULL) {
+   ret = qat_comp_qp_release(dev, qp_id);
+   if (ret < 0)
+   return ret;
+   }
+   if (qp_id >= qat_qps_per_service(comp_hw_qps,
+QAT_SERVICE_COMPRESSION)) {
+   QAT_LOG(ERR, "qp_id %u invalid for this device", qp_id);
+   return -EINVAL;
+   }
+
+   qat_qp_conf.hw = qp_hw_data;
+   qat_qp_conf.build_request = qat_comp_build_request;
+   qat_qp_conf.cookie_size = sizeof(struct qat_comp_op_cookie);
+   qat_qp_conf.nb_descriptors = max_inflight_ops;
+   qat_qp_conf.socket_id = socket_id;
+   qat_qp_conf.service_str = "comp";
+
+   ret = qat_qp_setup(qat_private->qat_dev, qp_addr, qp_id, &qat_qp_conf);
+   if (ret != 0)
+   return ret;
+
+   /* store a link to the qp in the qat_pci_device */
+   qat_private->qat_dev->qps_in_use[QAT_SERVICE_COMPRESSION][qp_id]
+   = *qp_addr;
+
+   return ret;
+}
diff --git a/drivers/compress/qat/qat_comp_pmd.h 
b/drivers/compress/qat/qat_comp_pmd.h
index 27d84c8..5a4bc31 100644
--- a/drivers/compress/qat/qat_comp_pmd.h
+++ b/drivers/compress/qat/qat_comp_pmd.h
@@ -34,6 +34,12 @@ qat_comp_stats_reset(struct rte_compressdev *dev);
 void
 qat_comp_stats_get(struct rte_compressdev *dev,
struct rte_compressdev_stats *stats);
+int
+qat_comp_qp_release(struct rte_compressdev *dev, uint16_t queue_pair_id);
+
+int
+qat_comp_qp_setup(struct rte_compressdev *dev, uint16_t qp_id,
+ uint32_t max_inflight_ops, int socket_id);
 
 #endif
 #endif /* _QAT_COMP_PMD_H_ */
-- 
2.7.4



[dpdk-dev] [PATCH v5 09/16] compress/qat: add fns to configure and clear device

2018-07-11 Thread Fiona Trahe
Add functions to configure and clear the qat comp device,
including the creation and freeing of the xform pool
and the freeing of queue-pairs.

Signed-off-by: Fiona Trahe 
Signed-off-by: Tomasz Jozwiak 
---
 drivers/compress/qat/qat_comp_pmd.c | 95 +
 drivers/compress/qat/qat_comp_pmd.h |  7 +++
 2 files changed, 102 insertions(+)

diff --git a/drivers/compress/qat/qat_comp_pmd.c 
b/drivers/compress/qat/qat_comp_pmd.c
index 5ae6caf..beab6e3 100644
--- a/drivers/compress/qat/qat_comp_pmd.c
+++ b/drivers/compress/qat/qat_comp_pmd.c
@@ -99,3 +99,98 @@ qat_comp_qp_setup(struct rte_compressdev *dev, uint16_t 
qp_id,
 
return ret;
 }
+
+static struct rte_mempool *
+qat_comp_create_xform_pool(struct qat_comp_dev_private *comp_dev,
+ uint32_t num_elements)
+{
+   char xform_pool_name[RTE_MEMPOOL_NAMESIZE];
+   struct rte_mempool *mp;
+
+   snprintf(xform_pool_name, RTE_MEMPOOL_NAMESIZE,
+   "%s_xforms", comp_dev->qat_dev->name);
+
+   QAT_LOG(DEBUG, "xformpool: %s", xform_pool_name);
+   mp = rte_mempool_lookup(xform_pool_name);
+
+   if (mp != NULL) {
+   QAT_LOG(DEBUG, "xformpool already created");
+   if (mp->size != num_elements) {
+   QAT_LOG(DEBUG, "xformpool wrong size - delete it");
+   rte_mempool_free(mp);
+   mp = NULL;
+   comp_dev->xformpool = NULL;
+   }
+   }
+
+   if (mp == NULL)
+   mp = rte_mempool_create(xform_pool_name,
+   num_elements,
+   qat_comp_xform_size(), 0, 0,
+   NULL, NULL, NULL, NULL, rte_socket_id(),
+   0);
+   if (mp == NULL) {
+   QAT_LOG(ERR, "Err creating mempool %s w %d elements of size %d",
+   xform_pool_name, num_elements, qat_comp_xform_size());
+   return NULL;
+   }
+
+   return mp;
+}
+
+static void
+_qat_comp_dev_config_clear(struct qat_comp_dev_private *comp_dev)
+{
+   /* Free private_xform pool */
+   if (comp_dev->xformpool) {
+   /* Free internal mempool for private xforms */
+   rte_mempool_free(comp_dev->xformpool);
+   comp_dev->xformpool = NULL;
+   }
+}
+
+int
+qat_comp_dev_config(struct rte_compressdev *dev,
+   struct rte_compressdev_config *config)
+{
+   struct qat_comp_dev_private *comp_dev = dev->data->dev_private;
+   int ret = 0;
+
+   if (config->max_nb_streams != 0) {
+   QAT_LOG(ERR,
+   "QAT device does not support STATEFUL so max_nb_streams must be 0");
+   return -EINVAL;
+   }
+
+   comp_dev->xformpool = qat_comp_create_xform_pool(comp_dev,
+   config->max_nb_priv_xforms);
+   if (comp_dev->xformpool == NULL) {
+
+   ret = -ENOMEM;
+   goto error_out;
+   }
+   return 0;
+
+error_out:
+   _qat_comp_dev_config_clear(comp_dev);
+   return ret;
+}
+
+
+int
+qat_comp_dev_close(struct rte_compressdev *dev)
+{
+   int i;
+   int ret = 0;
+   struct qat_comp_dev_private *comp_dev = dev->data->dev_private;
+
+   for (i = 0; i < dev->data->nb_queue_pairs; i++) {
+   ret = qat_comp_qp_release(dev, i);
+   if (ret < 0)
+   return ret;
+   }
+
+   _qat_comp_dev_config_clear(comp_dev);
+
+   return ret;
+}
diff --git a/drivers/compress/qat/qat_comp_pmd.h 
b/drivers/compress/qat/qat_comp_pmd.h
index 5a4bc31..b10a66f 100644
--- a/drivers/compress/qat/qat_comp_pmd.h
+++ b/drivers/compress/qat/qat_comp_pmd.h
@@ -41,5 +41,12 @@ int
 qat_comp_qp_setup(struct rte_compressdev *dev, uint16_t qp_id,
  uint32_t max_inflight_ops, int socket_id);
 
+int
+qat_comp_dev_config(struct rte_compressdev *dev,
+   struct rte_compressdev_config *config);
+
+int
+qat_comp_dev_close(struct rte_compressdev *dev);
+
 #endif
 #endif /* _QAT_COMP_PMD_H_ */
-- 
2.7.4



[dpdk-dev] [PATCH v5 14/16] compress/qat: add fns to create and destroy the PMD

2018-07-11 Thread Fiona Trahe
Now that all the device operations are available,
add the functions to create and destroy the pmd.
Called on probe and remove of the qat pci device, these
register the device with the compressdev API
and plug in all the device functionality.


Signed-off-by: Fiona Trahe 
Signed-off-by: Tomasz Jozwiak 
---
 drivers/common/qat/qat_device.h |  4 ++
 drivers/common/qat/qat_qp.c | 11 -
 drivers/common/qat/qat_qp.h |  5 ++
 drivers/compress/qat/qat_comp_pmd.c | 98 +++--
 drivers/compress/qat/qat_comp_pmd.h | 11 ++---
 5 files changed, 117 insertions(+), 12 deletions(-)

diff --git a/drivers/common/qat/qat_device.h b/drivers/common/qat/qat_device.h
index 0cb370c..9599fc5 100644
--- a/drivers/common/qat/qat_device.h
+++ b/drivers/common/qat/qat_device.h
@@ -25,6 +25,8 @@
  *  - runtime data
  */
 struct qat_sym_dev_private;
+struct qat_comp_dev_private;
+
 struct qat_pci_device {
 
/* Data used by all services */
@@ -55,6 +57,8 @@ struct qat_pci_device {
 */
 
/* Data relating to compression service */
+   struct qat_comp_dev_private *comp_dev;
+   /**< link back to compressdev private data */
 
/* Data relating to asymmetric crypto service */
 
diff --git a/drivers/common/qat/qat_qp.c b/drivers/common/qat/qat_qp.c
index 32c1759..7ca7a45 100644
--- a/drivers/common/qat/qat_qp.c
+++ b/drivers/common/qat/qat_qp.c
@@ -15,6 +15,7 @@
 #include "qat_device.h"
 #include "qat_qp.h"
 #include "qat_sym.h"
+#include "qat_comp.h"
 #include "adf_transport_access_macros.h"
 
 
@@ -606,8 +607,8 @@ qat_dequeue_op_burst(void *qp, void **ops, uint16_t nb_ops)
 
if (tmp_qp->service_type == QAT_SERVICE_SYMMETRIC)
qat_sym_process_response(ops, resp_msg);
-   /* add qat_asym_process_response here */
-   /* add qat_comp_process_response here */
+   else if (tmp_qp->service_type == QAT_SERVICE_COMPRESSION)
+   qat_comp_process_response(ops, resp_msg);
 
head = adf_modulo(head + rx_queue->msg_size,
  rx_queue->modulo_mask);
@@ -633,3 +634,9 @@ qat_dequeue_op_burst(void *qp, void **ops, uint16_t nb_ops)
}
return resp_counter;
 }
+
+__attribute__((weak)) int
+qat_comp_process_response(void **op __rte_unused, uint8_t *resp __rte_unused)
+{
+   return  0;
+}
diff --git a/drivers/common/qat/qat_qp.h b/drivers/common/qat/qat_qp.h
index 59db945..69f8a61 100644
--- a/drivers/common/qat/qat_qp.h
+++ b/drivers/common/qat/qat_qp.h
@@ -103,4 +103,9 @@ qat_qp_setup(struct qat_pci_device *qat_dev,
 int
 qat_qps_per_service(const struct qat_qp_hw_data *qp_hw_data,
enum qat_service_type service);
+
+/* Needed for weak function*/
+int
+qat_comp_process_response(void **op __rte_unused, uint8_t *resp __rte_unused);
+
 #endif /* _QAT_QP_H_ */
diff --git a/drivers/compress/qat/qat_comp_pmd.c 
b/drivers/compress/qat/qat_comp_pmd.c
index 013ff6e..9bb9897 100644
--- a/drivers/compress/qat/qat_comp_pmd.c
+++ b/drivers/compress/qat/qat_comp_pmd.c
@@ -5,6 +5,18 @@
 #include "qat_comp.h"
 #include "qat_comp_pmd.h"
 
+static const struct rte_compressdev_capabilities qat_comp_gen_capabilities[] = 
{
+   {/* COMPRESSION - deflate */
+.algo = RTE_COMP_ALGO_DEFLATE,
+.comp_feature_flags = RTE_COMP_FF_MULTI_PKT_CHECKSUM |
+   RTE_COMP_FF_CRC32_CHECKSUM |
+   RTE_COMP_FF_ADLER32_CHECKSUM |
+   RTE_COMP_FF_CRC32_ADLER32_CHECKSUM |
+   RTE_COMP_FF_SHAREABLE_PRIV_XFORM |
+   RTE_COMP_FF_HUFFMAN_FIXED,
+.window_size = {.min = 15, .max = 15, .increment = 0} },
+   {RTE_COMP_ALGO_LIST_END, 0, {0, 0, 0} } };
+
 static void
 qat_comp_stats_get(struct rte_compressdev *dev,
struct rte_compressdev_stats *stats)
@@ -225,14 +237,14 @@ qat_comp_dev_info_get(struct rte_compressdev *dev,
}
 }
 
-uint16_t
+static uint16_t
 qat_comp_pmd_enqueue_op_burst(void *qp, struct rte_comp_op **ops,
uint16_t nb_ops)
 {
return qat_enqueue_op_burst(qp, (void **)ops, nb_ops);
 }
 
-uint16_t
+static uint16_t
 qat_comp_pmd_dequeue_op_burst(void *qp, struct rte_comp_op **ops,
  uint16_t nb_ops)
 {
@@ -240,7 +252,7 @@ qat_comp_pmd_dequeue_op_burst(void *qp, struct rte_comp_op 
**ops,
 }
 
 
-struct rte_compressdev_ops compress_qat_ops = {
+static struct rte_compressdev_ops compress_qat_ops = {
 
/* Device related operations */
.dev_configure  = qat_comp_dev_config,
@@ -258,3 +270,83 @@ struct rte_compressdev_ops compress_qat_ops = {
.private_xform_create   = qat_comp_private_xform_create,
.private_xform_free = qat_comp_private_xform_free
 };
+
+int
+qat_comp_dev_create(struct qat_pci_device *qat_pci_dev)
+{
+   if (qat_pci_dev->qat_dev_gen

[dpdk-dev] [PATCH v5 11/16] compress/qat: add enqueue/dequeue functions

2018-07-11 Thread Fiona Trahe
Wrap generic qat enqueue/dequeue functions with
compressdev enqueue and dequeue fns.


Signed-off-by: Fiona Trahe 
Signed-off-by: Tomasz Jozwiak 
---
 drivers/compress/qat/qat_comp_pmd.c | 14 ++
 drivers/compress/qat/qat_comp_pmd.h |  8 
 2 files changed, 22 insertions(+)

diff --git a/drivers/compress/qat/qat_comp_pmd.c 
b/drivers/compress/qat/qat_comp_pmd.c
index 482ebd1..086b6cf 100644
--- a/drivers/compress/qat/qat_comp_pmd.c
+++ b/drivers/compress/qat/qat_comp_pmd.c
@@ -213,3 +213,17 @@ qat_comp_dev_info_get(struct rte_compressdev *dev,
info->capabilities = comp_dev->qat_dev_capabilities;
}
 }
+
+uint16_t
+qat_comp_pmd_enqueue_op_burst(void *qp, struct rte_comp_op **ops,
+   uint16_t nb_ops)
+{
+   return qat_enqueue_op_burst(qp, (void **)ops, nb_ops);
+}
+
+uint16_t
+qat_comp_pmd_dequeue_op_burst(void *qp, struct rte_comp_op **ops,
+ uint16_t nb_ops)
+{
+   return qat_dequeue_op_burst(qp, (void **)ops, nb_ops);
+}
diff --git a/drivers/compress/qat/qat_comp_pmd.h 
b/drivers/compress/qat/qat_comp_pmd.h
index 22576f4..f360c29 100644
--- a/drivers/compress/qat/qat_comp_pmd.h
+++ b/drivers/compress/qat/qat_comp_pmd.h
@@ -54,5 +54,13 @@ void
 qat_comp_dev_info_get(struct rte_compressdev *dev,
struct rte_compressdev_info *info);
 
+uint16_t
+qat_comp_pmd_enqueue_op_burst(void *qp, struct rte_comp_op **ops,
+   uint16_t nb_ops);
+
+uint16_t
+qat_comp_pmd_dequeue_op_burst(void *qp, struct rte_comp_op **ops,
+   uint16_t nb_ops);
+
 #endif
 #endif /* _QAT_COMP_PMD_H_ */
-- 
2.7.4



[dpdk-dev] [PATCH v5 16/16] docs/qat: refactor docs adding compression guide

2018-07-11 Thread Fiona Trahe
Extend QAT guide to cover crypto and compression and common
information, particularly about kernel driver dependency.
Update release note.
Update compression feature list for qat.


Signed-off-by: Fiona Trahe 
---
 config/common_base   |   2 +-
 doc/guides/compressdevs/features/qat.ini |  24 
 doc/guides/compressdevs/index.rst|   1 +
 doc/guides/compressdevs/qat_comp.rst |  49 +
 doc/guides/cryptodevs/qat.rst| 183 +--
 doc/guides/rel_notes/release_18_08.rst   |   5 +
 6 files changed, 205 insertions(+), 59 deletions(-)
 create mode 100644 doc/guides/compressdevs/features/qat.ini
 create mode 100644 doc/guides/compressdevs/qat_comp.rst

diff --git a/config/common_base b/config/common_base
index 1e340b4..1380acf 100644
--- a/config/common_base
+++ b/config/common_base
@@ -478,7 +478,7 @@ CONFIG_RTE_LIBRTE_PMD_DPAA_SEC=n
 CONFIG_RTE_LIBRTE_DPAA_MAX_CRYPTODEV=4
 
 #
-# Compile PMD for QuickAssist based devices
+# Compile PMD for QuickAssist based devices - see docs for details
 #
 CONFIG_RTE_LIBRTE_PMD_QAT=y
 CONFIG_RTE_LIBRTE_PMD_QAT_SYM=n
diff --git a/doc/guides/compressdevs/features/qat.ini 
b/doc/guides/compressdevs/features/qat.ini
new file mode 100644
index 000..0d0e21d
--- /dev/null
+++ b/doc/guides/compressdevs/features/qat.ini
@@ -0,0 +1,24 @@
+;
+; Refer to default.ini for the full list of available PMD features.
+;
+; Supported features of 'QAT' compression driver.
+;
+[Features]
+HW Accelerated  = Y
+CPU SSE =
+CPU AVX =
+CPU AVX2=
+CPU AVX512  =
+CPU NEON=
+Stateful=
+Pass-through=
+OOP SGL In SGL Out  =
+OOP SGL In LB  Out  =
+OOP LB  In SGL Out  =
+Deflate = Y
+LZS =
+Adler32 = Y
+Crc32   = Y
+Adler32&Crc32   = Y
+Fixed   = Y
+Dynamic =
diff --git a/doc/guides/compressdevs/index.rst 
b/doc/guides/compressdevs/index.rst
index bc59ce8..4228768 100644
--- a/doc/guides/compressdevs/index.rst
+++ b/doc/guides/compressdevs/index.rst
@@ -11,3 +11,4 @@ Compression Device Drivers
 
 overview
 isal
+qat_comp
diff --git a/doc/guides/compressdevs/qat_comp.rst 
b/doc/guides/compressdevs/qat_comp.rst
new file mode 100644
index 000..167f816
--- /dev/null
+++ b/doc/guides/compressdevs/qat_comp.rst
@@ -0,0 +1,49 @@
+..  SPDX-License-Identifier: BSD-3-Clause
+Copyright(c) 2018 Intel Corporation.
+
+Intel(R) QuickAssist (QAT) Compression Poll Mode Driver
+===
+
+The QAT compression PMD provides poll mode compression & decompression driver
+support for the following hardware accelerator devices:
+
+* ``Intel QuickAssist Technology C62x``
+* ``Intel QuickAssist Technology C3xxx``
+
+
+Features
+
+
+QAT compression PMD has support for:
+
+Compression/Decompression algorithm:
+
+* DEFLATE
+
+Huffman code type:
+
+* FIXED
+
+Window size support:
+
+* 32K
+
+Checksum generation:
+
+* CRC32, Adler and combined checksum
+
+Limitations
+---
+
+* Chained mbufs are not yet supported, therefore max data size which can be 
passed to the PMD in a single mbuf is 64K - 1. If data is larger than this it 
will need to be split up and sent as multiple operations.
+
+* Compressdev level 0, no compression, is not supported.
+
+* Dynamic Huffman encoding is not yet supported.
+
+Installation
+
+
+The QAT compression PMD is built by default with a standard DPDK build.
+
+It depends on a QAT kernel driver, see :ref:`qat_kernel_installation`.
diff --git a/doc/guides/cryptodevs/qat.rst b/doc/guides/cryptodevs/qat.rst
index b899985..bdc58eb 100644
--- a/doc/guides/cryptodevs/qat.rst
+++ b/doc/guides/cryptodevs/qat.rst
@@ -68,12 +68,32 @@ Limitations
 * Queue pairs are not thread-safe (that is, within a single queue pair, RX and 
TX from different lcores is not supported).
 
 
-Installation
-
+Extra notes on KASUMI F9
+
+
+When using KASUMI F9 authentication algorithm, the input buffer must be
+constructed according to the 3GPP KASUMI specifications (section 4.4, page 13):
+``_.
+Input buffer has to have COUNT (4 bytes), FRESH (4 bytes), MESSAGE and 
DIRECTION (1 bit)
+concatenated. After the DIRECTION bit, a single '1' bit is appended, followed 
by
+between 0 and 7 '0' bits, so that the total length of the buffer is multiple 
of 8 bits.
+Note that the actual message can be any length, specified in bits.
 
-To enable QAT in DPDK, follow the instructions for modifying the compile-time
+Once this buffer is passed this way, when creating the crypto operation,
+length of data to authenticate (op.sym.auth.data.length) must be the length
+of all the items described above, including the padding at the end.
+Also, offset of data to authenticate (op.sym.auth.data.offset)
+must be such that points at the start of the CO

[dpdk-dev] [PATCH v5 13/16] compress/qat: create and populate the ops structure

2018-07-11 Thread Fiona Trahe
Create an ops structure and populate it with the
qat-specific functions.


Signed-off-by: Fiona Trahe 
Signed-off-by: Tomasz Jozwiak 
---
 drivers/compress/qat/qat_comp_pmd.c | 38 -
 drivers/compress/qat/qat_comp_pmd.h | 30 -
 2 files changed, 29 insertions(+), 39 deletions(-)

diff --git a/drivers/compress/qat/qat_comp_pmd.c 
b/drivers/compress/qat/qat_comp_pmd.c
index 1ab5cf7..013ff6e 100644
--- a/drivers/compress/qat/qat_comp_pmd.c
+++ b/drivers/compress/qat/qat_comp_pmd.c
@@ -5,7 +5,7 @@
 #include "qat_comp.h"
 #include "qat_comp_pmd.h"
 
-void
+static void
 qat_comp_stats_get(struct rte_compressdev *dev,
struct rte_compressdev_stats *stats)
 {
@@ -25,7 +25,7 @@ qat_comp_stats_get(struct rte_compressdev *dev,
stats->dequeue_err_count = qat_stats.dequeue_err_count;
 }
 
-void
+static void
 qat_comp_stats_reset(struct rte_compressdev *dev)
 {
struct qat_comp_dev_private *qat_priv;
@@ -40,7 +40,7 @@ qat_comp_stats_reset(struct rte_compressdev *dev)
 
 }
 
-int
+static int
 qat_comp_qp_release(struct rte_compressdev *dev, uint16_t queue_pair_id)
 {
struct qat_comp_dev_private *qat_private = dev->data->dev_private;
@@ -55,7 +55,7 @@ qat_comp_qp_release(struct rte_compressdev *dev, uint16_t 
queue_pair_id)
&(dev->data->queue_pairs[queue_pair_id]));
 }
 
-int
+static int
 qat_comp_qp_setup(struct rte_compressdev *dev, uint16_t qp_id,
  uint32_t max_inflight_ops, int socket_id)
 {
@@ -149,7 +149,7 @@ _qat_comp_dev_config_clear(struct qat_comp_dev_private 
*comp_dev)
}
 }
 
-int
+static int
 qat_comp_dev_config(struct rte_compressdev *dev,
struct rte_compressdev_config *config)
 {
@@ -176,19 +176,19 @@ qat_comp_dev_config(struct rte_compressdev *dev,
return ret;
 }
 
-int
+static int
 qat_comp_dev_start(struct rte_compressdev *dev __rte_unused)
 {
return 0;
 }
 
-void
+static void
 qat_comp_dev_stop(struct rte_compressdev *dev __rte_unused)
 {
 
 }
 
-int
+static int
 qat_comp_dev_close(struct rte_compressdev *dev)
 {
int i;
@@ -207,7 +207,7 @@ qat_comp_dev_close(struct rte_compressdev *dev)
 }
 
 
-void
+static void
 qat_comp_dev_info_get(struct rte_compressdev *dev,
struct rte_compressdev_info *info)
 {
@@ -238,3 +238,23 @@ qat_comp_pmd_dequeue_op_burst(void *qp, struct rte_comp_op 
**ops,
 {
return qat_dequeue_op_burst(qp, (void **)ops, nb_ops);
 }
+
+
+struct rte_compressdev_ops compress_qat_ops = {
+
+   /* Device related operations */
+   .dev_configure  = qat_comp_dev_config,
+   .dev_start  = qat_comp_dev_start,
+   .dev_stop   = qat_comp_dev_stop,
+   .dev_close  = qat_comp_dev_close,
+   .dev_infos_get  = qat_comp_dev_info_get,
+
+   .stats_get  = qat_comp_stats_get,
+   .stats_reset= qat_comp_stats_reset,
+   .queue_pair_setup   = qat_comp_qp_setup,
+   .queue_pair_release = qat_comp_qp_release,
+
+   /* Compression related operations */
+   .private_xform_create   = qat_comp_private_xform_create,
+   .private_xform_free = qat_comp_private_xform_free
+};
diff --git a/drivers/compress/qat/qat_comp_pmd.h 
b/drivers/compress/qat/qat_comp_pmd.h
index 22cbefb..7ba1b8d 100644
--- a/drivers/compress/qat/qat_comp_pmd.h
+++ b/drivers/compress/qat/qat_comp_pmd.h
@@ -30,30 +30,6 @@ struct qat_comp_dev_private {
 
 };
 
-void
-qat_comp_stats_reset(struct rte_compressdev *dev);
-
-void
-qat_comp_stats_get(struct rte_compressdev *dev,
-   struct rte_compressdev_stats *stats);
-int
-qat_comp_qp_release(struct rte_compressdev *dev, uint16_t queue_pair_id);
-
-int
-qat_comp_qp_setup(struct rte_compressdev *dev, uint16_t qp_id,
- uint32_t max_inflight_ops, int socket_id);
-
-int
-qat_comp_dev_config(struct rte_compressdev *dev,
-   struct rte_compressdev_config *config);
-
-int
-qat_comp_dev_close(struct rte_compressdev *dev);
-
-void
-qat_comp_dev_info_get(struct rte_compressdev *dev,
-   struct rte_compressdev_info *info);
-
 uint16_t
 qat_comp_pmd_enqueue_op_burst(void *qp, struct rte_comp_op **ops,
uint16_t nb_ops);
@@ -62,11 +38,5 @@ uint16_t
 qat_comp_pmd_dequeue_op_burst(void *qp, struct rte_comp_op **ops,
uint16_t nb_ops);
 
-int
-qat_comp_dev_start(struct rte_compressdev *dev __rte_unused);
-
-void
-qat_comp_dev_stop(struct rte_compressdev *dev __rte_unused);
-
 #endif
 #endif /* _QAT_COMP_PMD_H_ */
-- 
2.7.4



[dpdk-dev] [PATCH v5 12/16] compress/qat: add device start and stop fns

2018-07-11 Thread Fiona Trahe
There are no specific actions needed to start/stop a QAT comp device
so these are just trivial fns to satisfy the pmd API.


Signed-off-by: Fiona Trahe 
---
 drivers/compress/qat/qat_comp_pmd.c | 11 +++
 drivers/compress/qat/qat_comp_pmd.h |  6 ++
 2 files changed, 17 insertions(+)

diff --git a/drivers/compress/qat/qat_comp_pmd.c 
b/drivers/compress/qat/qat_comp_pmd.c
index 086b6cf..1ab5cf7 100644
--- a/drivers/compress/qat/qat_comp_pmd.c
+++ b/drivers/compress/qat/qat_comp_pmd.c
@@ -176,6 +176,17 @@ qat_comp_dev_config(struct rte_compressdev *dev,
return ret;
 }
 
+int
+qat_comp_dev_start(struct rte_compressdev *dev __rte_unused)
+{
+   return 0;
+}
+
+void
+qat_comp_dev_stop(struct rte_compressdev *dev __rte_unused)
+{
+
+}
 
 int
 qat_comp_dev_close(struct rte_compressdev *dev)
diff --git a/drivers/compress/qat/qat_comp_pmd.h 
b/drivers/compress/qat/qat_comp_pmd.h
index f360c29..22cbefb 100644
--- a/drivers/compress/qat/qat_comp_pmd.h
+++ b/drivers/compress/qat/qat_comp_pmd.h
@@ -62,5 +62,11 @@ uint16_t
 qat_comp_pmd_dequeue_op_burst(void *qp, struct rte_comp_op **ops,
uint16_t nb_ops);
 
+int
+qat_comp_dev_start(struct rte_compressdev *dev __rte_unused);
+
+void
+qat_comp_dev_stop(struct rte_compressdev *dev __rte_unused);
+
 #endif
 #endif /* _QAT_COMP_PMD_H_ */
-- 
2.7.4



[dpdk-dev] [PATCH v5 1/4] ethdev: Add eal device event handler APIs

2018-07-11 Thread Jeff Guo
Implement a couple of eal device event handler install/uninstall APIs
in ethdev. Each ethdev install the handler in PMDs, so each callback
corresponding with one port, and process the eal device event for specific
port. If PMDs install the handler when initial device, it could have
chance to manage the eal device event, such as register device event
callback, then monitor and process the hotplug event.

Signed-off-by: Jeff Guo 
Reviewed-by: Qi Zhang 
---
v5->v4:
refine some code style and typo
---
 doc/guides/rel_notes/release_18_08.rst   | 12 +++
 lib/librte_ethdev/rte_ethdev.c   | 59 
 lib/librte_ethdev/rte_ethdev_driver.h| 32 +
 lib/librte_ethdev/rte_ethdev_version.map |  2 ++
 4 files changed, 105 insertions(+)

diff --git a/doc/guides/rel_notes/release_18_08.rst 
b/doc/guides/rel_notes/release_18_08.rst
index bc01242..b6482ce 100644
--- a/doc/guides/rel_notes/release_18_08.rst
+++ b/doc/guides/rel_notes/release_18_08.rst
@@ -46,6 +46,18 @@ New Features
   Flow API support has been added to CXGBE Poll Mode Driver to offload
   flows to Chelsio T5/T6 NICs.
 
+* **Added eal device event process helper in ethdev.**
+
+  Implement a couple of eal device event handler install/uninstall APIs in
+  ethdev, these helper could let PMDs have chance to manage the eal device
+  event, such as register device event callback, then monitor and process
+  hotplug event.
+
+  * ``rte_eth_dev_event_handler_install`` for PMDs use to install the device
+event handler.
+  * ``rte_eth_dev_event_handler_uninstall`` for PMDs use to uninstall the 
device
+event handler.
+
 
 API Changes
 ---
diff --git a/lib/librte_ethdev/rte_ethdev.c b/lib/librte_ethdev/rte_ethdev.c
index a9977df..4d28db5 100644
--- a/lib/librte_ethdev/rte_ethdev.c
+++ b/lib/librte_ethdev/rte_ethdev.c
@@ -4518,6 +4518,65 @@ rte_eth_devargs_parse(const char *dargs, struct 
rte_eth_devargs *eth_da)
return result;
 }
 
+static void __rte_experimental
+eth_dev_event_callback(char *device_name, enum rte_dev_event_type type,
+void *arg)
+{
+   struct rte_eth_dev *eth_dev = (struct rte_eth_dev *)arg;
+
+   switch (type) {
+   case RTE_DEV_EVENT_REMOVE:
+   ethdev_log(INFO, "The device %s has been removed!\n",
+   device_name);
+
+   if (!device_name || !eth_dev)
+   return;
+
+   if (!(eth_dev->data->dev_flags & RTE_ETH_DEV_INTR_RMV))
+   return;
+
+   if (!strcmp(device_name, eth_dev->device->name))
+   _rte_eth_dev_callback_process(eth_dev,
+ RTE_ETH_EVENT_INTR_RMV,
+ NULL);
+   break;
+   case RTE_DEV_EVENT_ADD:
+   ethdev_log(INFO, "The device %s has been added!\n",
+   device_name);
+   break;
+   default:
+   break;
+   }
+}
+
+int __rte_experimental
+rte_eth_dev_event_handler_install(struct rte_eth_dev *eth_dev)
+{
+   int result = 0;
+
+   result = rte_dev_event_callback_register(eth_dev->device->name,
+   eth_dev_event_callback, eth_dev);
+   if (result)
+   ethdev_log(ERR, "device event callback register failed for "
+   "device %s!\n", eth_dev->device->name);
+
+   return result;
+}
+
+int __rte_experimental
+rte_eth_dev_event_handler_uninstall(struct rte_eth_dev *eth_dev)
+{
+   int result = 0;
+
+   result = rte_dev_event_callback_unregister(eth_dev->device->name,
+   eth_dev_event_callback, eth_dev);
+   if (result)
+   ethdev_log(ERR, "device event callback unregister failed for"
+   " device %s!\n", eth_dev->device->name);
+
+   return result;
+}
+
 RTE_INIT(ethdev_init_log);
 static void
 ethdev_init_log(void)
diff --git a/lib/librte_ethdev/rte_ethdev_driver.h 
b/lib/librte_ethdev/rte_ethdev_driver.h
index c9c825e..3de1cd4 100644
--- a/lib/librte_ethdev/rte_ethdev_driver.h
+++ b/lib/librte_ethdev/rte_ethdev_driver.h
@@ -82,6 +82,38 @@ int rte_eth_dev_release_port(struct rte_eth_dev *eth_dev);
 void _rte_eth_dev_reset(struct rte_eth_dev *dev);
 
 /**
+ * @warning
+ * @b EXPERIMENTAL: this API may change without prior notice.
+ *
+ * Implement a helper to install the device event handler for the specific
+ * ether device.
+ *
+ * @param dev
+ *  Pointer to struct rte_eth_dev.
+ *
+ * @return
+ *   - 0 on success, negative on error
+ */
+int __rte_experimental
+rte_eth_dev_event_handler_install(struct rte_eth_dev *dev);
+
+/**
+ * @warning
+ * @b EXPERIMENTAL: this API may change without prior notice.
+ *
+ * Implement a helper to uninstall the device event handler for the specific
+ * ether device.
+ *
+ * @param dev
+ *  Pointer to struct rte_eth_dev

[dpdk-dev] [PATCH v5 15/16] compress/qat: prevent device usage if incorrect firmware

2018-07-11 Thread Fiona Trahe
Previous check only causes op to fail on dequeue.
This extends so once first fail is detected, application can
no longer enqueue ops to the device and will also get an
appropriate error if trying to reconfigure or setup the device.


Signed-off-by: Tomasz Jozwiak 
Signed-off-by: Fiona Trahe 
---
 drivers/compress/qat/qat_comp_pmd.c | 57 -
 1 file changed, 56 insertions(+), 1 deletion(-)

diff --git a/drivers/compress/qat/qat_comp_pmd.c 
b/drivers/compress/qat/qat_comp_pmd.c
index 9bb9897..0a571b3 100644
--- a/drivers/compress/qat/qat_comp_pmd.c
+++ b/drivers/compress/qat/qat_comp_pmd.c
@@ -252,6 +252,61 @@ qat_comp_pmd_dequeue_op_burst(void *qp, struct rte_comp_op 
**ops,
 }
 
 
+static uint16_t
+qat_comp_pmd_enq_deq_dummy_op_burst(void *qp __rte_unused,
+   struct rte_comp_op **ops __rte_unused,
+   uint16_t nb_ops __rte_unused)
+{
+   QAT_DP_LOG(ERR, "QAT PMD detected wrong FW version !");
+   return 0;
+}
+
+static struct rte_compressdev_ops compress_qat_dummy_ops = {
+
+   /* Device related operations */
+   .dev_configure  = NULL,
+   .dev_start  = NULL,
+   .dev_stop   = qat_comp_dev_stop,
+   .dev_close  = qat_comp_dev_close,
+   .dev_infos_get  = NULL,
+
+   .stats_get  = NULL,
+   .stats_reset= qat_comp_stats_reset,
+   .queue_pair_setup   = NULL,
+   .queue_pair_release = qat_comp_qp_release,
+
+   /* Compression related operations */
+   .private_xform_create   = NULL,
+   .private_xform_free = qat_comp_private_xform_free
+};
+
+static uint16_t
+qat_comp_pmd_dequeue_frst_op_burst(void *qp, struct rte_comp_op **ops,
+  uint16_t nb_ops)
+{
+   uint16_t ret = qat_dequeue_op_burst(qp, (void **)ops, nb_ops);
+   struct qat_qp *tmp_qp = (struct qat_qp *)qp;
+
+   if (ret) {
+   if ((*ops)->debug_status ==
+   (uint64_t)ERR_CODE_QAT_COMP_WRONG_FW) {
+   tmp_qp->qat_dev->comp_dev->compressdev->enqueue_burst =
+   qat_comp_pmd_enq_deq_dummy_op_burst;
+   tmp_qp->qat_dev->comp_dev->compressdev->dequeue_burst =
+   qat_comp_pmd_enq_deq_dummy_op_burst;
+
+   tmp_qp->qat_dev->comp_dev->compressdev->dev_ops =
+   &compress_qat_dummy_ops;
+   QAT_LOG(ERR, "QAT PMD detected wrong FW version !");
+
+   } else {
+   tmp_qp->qat_dev->comp_dev->compressdev->dequeue_burst =
+   qat_comp_pmd_dequeue_op_burst;
+   }
+   }
+   return ret;
+}
+
 static struct rte_compressdev_ops compress_qat_ops = {
 
/* Device related operations */
@@ -302,7 +357,7 @@ qat_comp_dev_create(struct qat_pci_device *qat_pci_dev)
compressdev->dev_ops = &compress_qat_ops;
 
compressdev->enqueue_burst = qat_comp_pmd_enqueue_op_burst;
-   compressdev->dequeue_burst = qat_comp_pmd_dequeue_op_burst;
+   compressdev->dequeue_burst = qat_comp_pmd_dequeue_frst_op_burst;
 
compressdev->feature_flags = RTE_COMPDEV_FF_HW_ACCELERATED;
 
-- 
2.7.4



[dpdk-dev] [PATCH v5 0/4] Install eal hotplug event handler in i40e/ixgbe

2018-07-11 Thread Jeff Guo
As we may know, we have eal event for rte device hotplug and ethdev event
for ethdev hotplug. Some ethdev need to use eal event to detect hotplug
behaviors, the privors way is register eal event callback in app, but seems
that it will have some race between these 2 event processes. In oder to fix
the it, it might be better to find a way to combind these 2 events detect.

This patch set introduce a way to combind these 2 event, by register the
ethdev eal event callback in ether dev and trigger the ethdev hotplug event
in the callback. That will let the ethdev device can easy process hotplug
by a common way.

Here let i40e/ixgbe pmd driver for example, other driver which support
hotplug feature could be use this way to detect and process hotplug.

patch history:
v5->v4:
refine some code style and typo

v4->v3:
change to use device event handler install api

v3->v2:
remove the callback from driver to ethdev for common.

v2->v1:
add ixgbe hotplug detect case.
refine some doc.

Jeff Guo (4):
  ethdev: Add eal device event callback
  net/ixgbe: install ethdev hotplug handler in ixgbe
  net/i40e: install hotplug handler in i40e
  testpmd: remove the dev event callback register

 app/test-pmd/testpmd.c   | 76 
 doc/guides/rel_notes/release_18_08.rst   | 12 +
 drivers/net/i40e/i40e_ethdev.c   |  8 +++-
 drivers/net/ixgbe/ixgbe_ethdev.c |  8 +++-
 lib/librte_ethdev/rte_ethdev.c   | 59 +
 lib/librte_ethdev/rte_ethdev_driver.h| 32 ++
 lib/librte_ethdev/rte_ethdev_version.map |  2 +
 7 files changed, 119 insertions(+), 78 deletions(-)

-- 
2.7.4



[dpdk-dev] [PATCH v5 4/4] testpmd: remove the dev event callback register

2018-07-11 Thread Jeff Guo
Since now we can use driver to management the eal event for hotplug,
so no need to register dev event callback in app anymore. This patch
remove the related code.

Signed-off-by: Jeff Guo 
Acked-by: Wenzhuo Lu 
---
v5->v4:
no change.
---
 app/test-pmd/testpmd.c | 76 --
 1 file changed, 76 deletions(-)

diff --git a/app/test-pmd/testpmd.c b/app/test-pmd/testpmd.c
index 24c1998..10ed660 100644
--- a/app/test-pmd/testpmd.c
+++ b/app/test-pmd/testpmd.c
@@ -400,12 +400,6 @@ static void check_all_ports_link_status(uint32_t 
port_mask);
 static int eth_event_callback(portid_t port_id,
  enum rte_eth_event_type type,
  void *param, void *ret_param);
-static void eth_dev_event_callback(char *device_name,
-   enum rte_dev_event_type type,
-   void *param);
-static int eth_dev_event_callback_register(void);
-static int eth_dev_event_callback_unregister(void);
-
 
 /*
  * Check if all the ports are started.
@@ -1915,39 +1909,6 @@ reset_port(portid_t pid)
printf("Done\n");
 }
 
-static int
-eth_dev_event_callback_register(void)
-{
-   int ret;
-
-   /* register the device event callback */
-   ret = rte_dev_event_callback_register(NULL,
-   eth_dev_event_callback, NULL);
-   if (ret) {
-   printf("Failed to register device event callback\n");
-   return -1;
-   }
-
-   return 0;
-}
-
-
-static int
-eth_dev_event_callback_unregister(void)
-{
-   int ret;
-
-   /* unregister the device event callback */
-   ret = rte_dev_event_callback_unregister(NULL,
-   eth_dev_event_callback, NULL);
-   if (ret < 0) {
-   printf("Failed to unregister device event callback\n");
-   return -1;
-   }
-
-   return 0;
-}
-
 void
 attach_port(char *identifier)
 {
@@ -2049,10 +2010,6 @@ pmd_test_exit(void)
RTE_LOG(ERR, EAL,
"fail to stop device event monitor.");
 
-   ret = eth_dev_event_callback_unregister();
-   if (ret)
-   RTE_LOG(ERR, EAL,
-   "fail to unregister all event callbacks.");
}
 
printf("\nBye...\n");
@@ -2191,37 +2148,6 @@ eth_event_callback(portid_t port_id, enum 
rte_eth_event_type type, void *param,
return 0;
 }
 
-/* This function is used by the interrupt thread */
-static void
-eth_dev_event_callback(char *device_name, enum rte_dev_event_type type,
-__rte_unused void *arg)
-{
-   if (type >= RTE_DEV_EVENT_MAX) {
-   fprintf(stderr, "%s called upon invalid event %d\n",
-   __func__, type);
-   fflush(stderr);
-   }
-
-   switch (type) {
-   case RTE_DEV_EVENT_REMOVE:
-   RTE_LOG(ERR, EAL, "The device: %s has been removed!\n",
-   device_name);
-   /* TODO: After finish failure handle, begin to stop
-* packet forward, stop port, close port, detach port.
-*/
-   break;
-   case RTE_DEV_EVENT_ADD:
-   RTE_LOG(ERR, EAL, "The device: %s has been added!\n",
-   device_name);
-   /* TODO: After finish kernel driver binding,
-* begin to attach port.
-*/
-   break;
-   default:
-   break;
-   }
-}
-
 static int
 set_tx_queue_stats_mapping_registers(portid_t port_id, struct rte_port *port)
 {
@@ -2735,8 +2661,6 @@ main(int argc, char** argv)
rte_errno = EINVAL;
return -1;
}
-   eth_dev_event_callback_register();
-
}
 
if (start_port(RTE_PORT_ALL) != 0)
-- 
2.7.4



Re: [dpdk-dev] [PATCH v10 01/27] devargs: add non-variadic parsing function

2018-07-11 Thread Gaëtan Rivet
On Wed, Jul 11, 2018 at 05:16:15PM +0530, Shreyansh Jain wrote:
> On Thursday 05 July 2018 05:18 PM, Gaetan Rivet wrote:
> > rte_devargs_parse becomes non-variadic,
> > rte_devargs_parsef becomes the variadic version, to be used to compose
> > device strings.
> > 
> > Signed-off-by: Gaetan Rivet 
> > ---
> >   drivers/net/failsafe/failsafe_args.c|  2 +-
> >   drivers/net/failsafe/failsafe_eal.c |  2 +-
> >   lib/librte_eal/common/eal_common_dev.c  |  4 +-
> >   lib/librte_eal/common/eal_common_devargs.c  | 42 -
> >   lib/librte_eal/common/include/rte_devargs.h | 40 +++-
> >   lib/librte_eal/rte_eal_version.map  |  1 +
> >   lib/librte_ethdev/rte_ethdev.c  |  2 +-
> >   7 files changed, 76 insertions(+), 17 deletions(-)
> > 
> 
> [...]
> 
> > +__rte_experimental
> > +int
> > +rte_devargs_parsef(struct rte_devargs *da, const char *format, ...)
> > +{
> > +   va_list ap;
> > +   size_t len;
> > +   char *dev;
> > +
> > +   if (da == NULL)
> > +   return -EINVAL;
> > +
> > +   va_start(ap, format);
> > +   len = vsnprintf(NULL, 0, format, ap);
> > +   va_end(ap);
> > +
> > +   dev = calloc(1, len + 1);
> > +   if (dev == NULL) {
> > +   fprintf(stderr, "ERROR: not enough memory to parse device\n");
> 
> Should RTE_LOG be used here?
> 

Yes, actually, I think the whole rte_devargs should be changed to
RTE_LOG.

> > +   return -ENOMEM;
> > +   }
> > +
> > +   va_start(ap, format);
> > +   vsnprintf(dev, len, format, ap);
> > +   va_end(ap);
> > +
> > +   return rte_devargs_parse(da, dev);
> > +}
> > +
> >   int __rte_experimental
> >   rte_devargs_insert(struct rte_devargs *da)
> >   {
> 
> [...]
> 
> Except the comment above:
> 
> Acked-by: Shreyansh Jain 

-- 
Gaëtan Rivet
6WIND


[dpdk-dev] [PATCH v5 3/4] net/i40e: install hotplug handler in i40e

2018-07-11 Thread Jeff Guo
This patch aim to enable hotplug detect in i40e PMD. Firstly it
set the flags RTE_PCI_DRV_INTR_RMV in drv_flags to announce the hotplug
ability, and then use rte_eth_dev_event_handler_install to install
the hotplug event handler for ethdev. When eal detect the hotplug event,
it will call the ethdev callback to process it. If the event is hotplug
removal, it will trigger the RTE_ETH_EVENT_INTR_RMV event into ethdev
callback to let app process the hotplug for this ethdev.

This is an example for other driver, that if any driver support hotplug
feature could be use this way to install hotplug handler.

Signed-off-by: Jeff Guo 
Acked-by: Qi Zhang 
---
v5->v4:
no change.
---
 drivers/net/i40e/i40e_ethdev.c | 8 +++-
 1 file changed, 7 insertions(+), 1 deletion(-)

diff --git a/drivers/net/i40e/i40e_ethdev.c b/drivers/net/i40e/i40e_ethdev.c
index 13c5d32..8fccf04 100644
--- a/drivers/net/i40e/i40e_ethdev.c
+++ b/drivers/net/i40e/i40e_ethdev.c
@@ -688,7 +688,7 @@ static int eth_i40e_pci_remove(struct rte_pci_device 
*pci_dev)
 static struct rte_pci_driver rte_i40e_pmd = {
.id_table = pci_id_i40e_map,
.drv_flags = RTE_PCI_DRV_NEED_MAPPING | RTE_PCI_DRV_INTR_LSC |
-RTE_PCI_DRV_IOVA_AS_VA,
+RTE_PCI_DRV_IOVA_AS_VA | RTE_PCI_DRV_INTR_RMV,
.probe = eth_i40e_pci_probe,
.remove = eth_i40e_pci_remove,
 };
@@ -1442,6 +1442,9 @@ eth_i40e_dev_init(struct rte_eth_dev *dev, void 
*init_params __rte_unused)
rte_intr_callback_register(intr_handle,
   i40e_dev_interrupt_handler, dev);
 
+   /* install the dev event handler for ethdev. */
+   rte_eth_dev_event_handler_install(dev);
+
/* configure and enable device interrupt */
i40e_pf_config_irq0(hw, TRUE);
i40e_pf_enable_irq0(hw);
@@ -1674,6 +1677,9 @@ eth_i40e_dev_uninit(struct rte_eth_dev *dev)
/* Remove all Traffic Manager configuration */
i40e_tm_conf_uninit(dev);
 
+   /* uninstall the dev event handler for ethdev. */
+   rte_eth_dev_event_handler_uninstall(dev);
+
return 0;
 }
 
-- 
2.7.4



[dpdk-dev] [PATCH v5 2/4] net/ixgbe: install ethdev hotplug handler in ixgbe

2018-07-11 Thread Jeff Guo
This patch aim to enable hotplug detect in ixgbe PMD. Firstly it
set the flags RTE_PCI_DRV_INTR_RMV in drv_flags to announce the hotplug
ability, and then use rte_eth_dev_event_handler_install to install
the hotplug event handler for ethdev. When eal detect the hotplug event,
it will call the ethdev callback to process it. If the event is hotplug
removal, it will trigger the RTE_ETH_EVENT_INTR_RMV event into ethdev
callback to let app process the hotplug for this ethdev.

This is an example for other driver, that if any driver support hotplug
feature could be use this way to install hotplug handler.

Signed-off-by: Jeff Guo 
Acked-by: Wenzhuo Lu 
---
v5->v4:
no change.
---
 drivers/net/ixgbe/ixgbe_ethdev.c | 8 +++-
 1 file changed, 7 insertions(+), 1 deletion(-)

diff --git a/drivers/net/ixgbe/ixgbe_ethdev.c b/drivers/net/ixgbe/ixgbe_ethdev.c
index 87d2ad0..e7ae9bf 100644
--- a/drivers/net/ixgbe/ixgbe_ethdev.c
+++ b/drivers/net/ixgbe/ixgbe_ethdev.c
@@ -1678,6 +1678,9 @@ eth_ixgbevf_dev_init(struct rte_eth_dev *eth_dev)
rte_intr_enable(intr_handle);
ixgbevf_intr_enable(eth_dev);
 
+   /* install the dev event handler for ethdev. */
+   rte_eth_dev_event_handler_install(eth_dev);
+
PMD_INIT_LOG(DEBUG, "port %d vendorID=0x%x deviceID=0x%x mac.type=%s",
 eth_dev->data->port_id, pci_dev->id.vendor_id,
 pci_dev->id.device_id, "ixgbe_mac_82599_vf");
@@ -1718,6 +1721,9 @@ eth_ixgbevf_dev_uninit(struct rte_eth_dev *eth_dev)
rte_intr_callback_unregister(intr_handle,
 ixgbevf_dev_interrupt_handler, eth_dev);
 
+   /* uninstall the dev event handler for ethdev. */
+   rte_eth_dev_event_handler_uninstall(eth_dev);
+
return 0;
 }
 
@@ -1801,7 +1807,7 @@ static int eth_ixgbe_pci_remove(struct rte_pci_device 
*pci_dev)
 static struct rte_pci_driver rte_ixgbe_pmd = {
.id_table = pci_id_ixgbe_map,
.drv_flags = RTE_PCI_DRV_NEED_MAPPING | RTE_PCI_DRV_INTR_LSC |
-RTE_PCI_DRV_IOVA_AS_VA,
+RTE_PCI_DRV_IOVA_AS_VA | RTE_PCI_DRV_INTR_RMV,
.probe = eth_ixgbe_pci_probe,
.remove = eth_ixgbe_pci_remove,
 };
-- 
2.7.4



Re: [dpdk-dev] [PATCH v2 1/5] compress/zlib: add ZLIB PMD support

2018-07-11 Thread De Lara Guarch, Pablo
Hi,

> -Original Message-
> From: Shally Verma [mailto:shally.ve...@caviumnetworks.com]
> Sent: Monday, July 2, 2018 5:57 PM
> To: De Lara Guarch, Pablo 
> Cc: dev@dpdk.org; pathr...@caviumnetworks.com;
> mcha...@caviumnetworks.com; Ashish Gupta
> ; Sunila Sahu
> 
> Subject: [PATCH v2 1/5] compress/zlib: add ZLIB PMD support
> 
> From: Ashish Gupta 
> 
> Add sw zlib pmd support in compressdev driver.
> Add device probe and remove support.
> Add ZLIB build file support.
> 
> Signed-off-by: Sunila Sahu 
> Signed-off-by: Shally Verma 
> Signed-off-by: Ashish Gupta 

...

> +++ b/drivers/compress/zlib/zlib_pmd_private.h

...

> +
> +struct zlib_private {
> + char mp_name[RTE_MEMPOOL_NAMESIZE];

One extra comment. I think you can store the pointer to the mempool here,
instead of the name (it saves a lookup and space).

> +};



Re: [dpdk-dev] [PATCH v2 1/5] compress/zlib: add ZLIB PMD support

2018-07-11 Thread De Lara Guarch, Pablo
And the last comments, sorry for the multiple replies.

> -Original Message-
> From: Shally Verma [mailto:shally.ve...@caviumnetworks.com]
> Sent: Monday, July 2, 2018 5:57 PM
> To: De Lara Guarch, Pablo 
> Cc: dev@dpdk.org; pathr...@caviumnetworks.com;
> mcha...@caviumnetworks.com; Ashish Gupta
> ; Sunila Sahu
> 
> Subject: [PATCH v2 1/5] compress/zlib: add ZLIB PMD support
> 
> From: Ashish Gupta 
> 
> Add sw zlib pmd support in compressdev driver.
> Add device probe and remove support.
> Add ZLIB build file support.
> 
> Signed-off-by: Sunila Sahu 
> Signed-off-by: Shally Verma 
> Signed-off-by: Ashish Gupta 

...

> +++ b/drivers/compress/zlib/zlib_pmd.c

...

> +static void
> +zlib_init_log(void)
> +{
> + zlib_logtype_driver = rte_log_register("compress_zlib");

The standard for the name of the logtype for PMDs is 
"pmd.driverType.driverName",
so in this case it would be "pmd.compress.zlib".


> + if (zlib_logtype_driver >= 0)
> + rte_log_set_level(zlib_logtype_driver, RTE_LOG_INFO); }
> diff --git a/drivers/compress/zlib/zlib_pmd_private.h
> b/drivers/compress/zlib/zlib_pmd_private.h
> new file mode 100644
> index 000..d4c80b1
> --- /dev/null
> +++ b/drivers/compress/zlib/zlib_pmd_private.h

...

> +#define ZLIB_PMD_INFO(fmt, args...) \
> + ZLIB_PMD_LOG(INFO, fmt, ## args)
> +#define ZLIB_PMD_ERR(fmt, args...) \
> + ZLIB_PMD_LOG(ERR, fmt, ## args)
> +#define ZLIB_PMD_WARN(fmt, args...) \
> + ZLIB_PMD_LOG(WARNING, fmt, ## args)

What do you think of having a single macro ZLIB_LOG(level, fmt, args...)?




Re: [dpdk-dev] [PATCH v2] net/mlx5: add support for 32bit systems

2018-07-11 Thread Shahaf Shuler
Hi Ferruh,

Thursday, July 5, 2018 2:27 PM, Ferruh Yigit:
> Subject: Re: [dpdk-dev] [PATCH v2] net/mlx5: add support for 32bit systems
> 
> On 7/5/2018 11:09 AM, Mordechay Haimovsky wrote:
> > Hi,
> >  Didn’t see it in our setups (not an excuse),  Investigating 
> 
> Thanks. Perhaps it can be related to compiler version:
> gcc (GCC) 8.1.1 20180502 (Red Hat 8.1.1-1) (ICC 32bit also gave same build
> error.)
> 
> btw, to clarify rdma-core v19 build errors was not just for 32bit build, I 
> lost my
> log but I can reproduce if you require.

Thanks for reporting it we will fix.

Here is the plan to include the patchset upstream:
1. there is a fix on rmda-core v19+ for the 32b compilation issue seen
2. as you reported there is another compilation issue w/ v19 which needs to be 
solved (I will provide a fix patch).
3. #2 fix patch should be on top of series : 
https://patches.dpdk.org/project/dpdk/list/?series=512

So the plan is to integrate the series and after it the compilation fix patch 
and the 32b support one.
Do you agree? 

I hope the above series will be merge at the beginning of next week.


> 
> >
> > Moti
> >
> >> -Original Message-
> >> From: Ferruh Yigit [mailto:ferruh.yi...@intel.com]
> >> Sent: Wednesday, July 4, 2018 4:49 PM
> >> To: Mordechay Haimovsky ; Shahaf Shuler
> >> 
> >> Cc: Adrien Mazarguil ; dev@dpdk.org
> >> Subject: Re: [dpdk-dev] [PATCH v2] net/mlx5: add support for 32bit
> >> systems
> >>
> >> On 7/2/2018 12:11 PM, Moti Haimovsky wrote:
> >>> This patch adds support for building and running mlx5 PMD on 32bit
> >>> systems such as i686.
> >>>
> >>> The main issue to tackle was handling the 32bit access to the UAR as
> >>> quoted from the mlx5 PRM:
> >>> QP and CQ DoorBells require 64-bit writes. For best performance, it
> >>> is recommended to execute the QP/CQ DoorBell as a single 64-bit
> >>> write operation. For platforms that do not support 64 bit writes, it
> >>> is possible to issue the 64 bits DoorBells through two consecutive
> >>> writes, each write 32 bits, as described below:
> >>> * The order of writing each of the Dwords is from lower to upper
> >>>   addresses.
> >>> * No other DoorBell can be rung (or even start ringing) in the midst of
> >>>   an on-going write of a DoorBell over a given UAR page.
> >>> The last rule implies that in a multi-threaded environment, the
> >>> access to a UAR page (which can be accessible by all threads in the
> >>> process) must be synchronized (for example, using a semaphore)
> >>> unless an atomic write of 64 bits in a single bus operation is
> >>> guaranteed. Such a synchronization is not required for when ringing
> >>> DoorBells on different UAR pages.
> >>>
> >>> Signed-off-by: Moti Haimovsky 
> >>> ---
> >>> v2:
> >>> * Fixed coding style issues.
> >>> * Modified documentation according to review inputs.
> >>> * Fixed merge conflicts.
> >>> ---
> >>>  doc/guides/nics/features/mlx5.ini |  1 +
> >>>  doc/guides/nics/mlx5.rst  |  6 +++-
> >>>  drivers/net/mlx5/mlx5.c   |  8 -
> >>>  drivers/net/mlx5/mlx5.h   |  5 +++
> >>>  drivers/net/mlx5/mlx5_defs.h  | 18 --
> >>>  drivers/net/mlx5/mlx5_rxq.c   |  6 +++-
> >>>  drivers/net/mlx5/mlx5_rxtx.c  | 22 +++--
> >>>  drivers/net/mlx5/mlx5_rxtx.h  | 69
> >> ++-
> >>>  drivers/net/mlx5/mlx5_txq.c   | 13 +++-
> >>>  9 files changed, 131 insertions(+), 17 deletions(-)
> >>>
> >>> diff --git a/doc/guides/nics/features/mlx5.ini
> >>> b/doc/guides/nics/features/mlx5.ini
> >>> index e75b14b..b28b43e 100644
> >>> --- a/doc/guides/nics/features/mlx5.ini
> >>> +++ b/doc/guides/nics/features/mlx5.ini
> >>> @@ -43,5 +43,6 @@ Multiprocess aware   = Y
> >>>  Other kdrv   = Y
> >>>  ARMv8= Y
> >>>  Power8   = Y
> >>> +x86-32   = Y
> >>>  x86-64   = Y
> >>>  Usage doc= Y
> >>> diff --git a/doc/guides/nics/mlx5.rst b/doc/guides/nics/mlx5.rst
> >>> index
> >>> 7dd9c1c..5fbad60 100644
> >>> --- a/doc/guides/nics/mlx5.rst
> >>> +++ b/doc/guides/nics/mlx5.rst
> >>> @@ -49,7 +49,7 @@ libibverbs.
> >>>  Features
> >>>  
> >>>
> >>> -- Multi arch support: x86_64, POWER8, ARMv8.
> >>> +- Multi arch support: x86_64, POWER8, ARMv8, i686.
> >>>  - Multiple TX and RX queues.
> >>>  - Support for scattered TX and RX frames.
> >>>  - IPv4, IPv6, TCPv4, TCPv6, UDPv4 and UDPv6 RSS on any number of
> >> queues.
> >>> @@ -477,6 +477,10 @@ RMDA Core with Linux Kernel
> >>>  - Minimal kernel version : v4.14 or the most recent 4.14-rc (see
> >>> `Linux installation documentation`_)
> >>>  - Minimal rdma-core version: v15+ commit 0c5f5765213a ("Merge pull
> >> request #227 from yishaih/tm")
> >>>(see `RDMA Core installation documentation`_)
> >>> +- When building for i686 use:
> >>> +
> >>> +  - rdma-core version 18.0 or above built with 32bit support.
> >>
> >> related "or above" part, v19 giving build errors with mlx5, FYI.
> >>

  1   2   3   >