> -----Original Message----- > From: dev [mailto:dev-boun...@dpdk.org] On Behalf Of Xing, Beilei > Sent: Wednesday, March 29, 2017 10:46 AM > To: Iremonger, Bernard <bernard.iremon...@intel.com>; dev@dpdk.org; > Wu, Jingjing <jingjing...@intel.com> > Cc: Zhang, Helin <helin.zh...@intel.com>; Lu, Wenzhuo > <wenzhuo...@intel.com> > Subject: Re: [dpdk-dev] [PATCH v3 2/5] net/i40e: parse QinQ pattern > > Hi Bernard, > > > -----Original Message----- > > From: Iremonger, Bernard > > Sent: Wednesday, March 29, 2017 12:21 AM > > To: dev@dpdk.org; Xing, Beilei <beilei.x...@intel.com>; Wu, Jingjing > > <jingjing...@intel.com> > > Cc: Zhang, Helin <helin.zh...@intel.com>; Lu, Wenzhuo > > <wenzhuo...@intel.com>; Iremonger, Bernard > > <bernard.iremon...@intel.com> > > Subject: [PATCH v3 2/5] net/i40e: parse QinQ pattern > > > > add QinQ pattern. > > add i40e_flow_parse_qinq_pattern function. > > add i40e_flow_parse_qinq_filter function. > > > > Signed-off-by: Bernard Iremonger <bernard.iremon...@intel.com> > > --- > > drivers/net/i40e/i40e_flow.c | 187 > > ++++++++++++++++++++++++++++++++++++++++++- > > 1 file changed, 185 insertions(+), 2 deletions(-) > > > > diff --git a/drivers/net/i40e/i40e_flow.c > > b/drivers/net/i40e/i40e_flow.c index be243e172..39b09ead5 100644 > > --- a/drivers/net/i40e/i40e_flow.c > > +++ b/drivers/net/i40e/i40e_flow.c > > @@ -1,7 +1,7 @@ > > /*- > > * BSD LICENSE > > * > > - * Copyright (c) 2016 Intel Corporation. All rights reserved. > > + * Copyright (c) 2016-2017 Intel Corporation. All rights reserved. > > * > > * Redistribution and use in source and binary forms, with or without > > * modification, are permitted provided that the following conditions > > @@ -128,6 +128,18 @@ static int i40e_flow_destroy_tunnel_filter(struct > > i40e_pf *pf, static int i40e_flow_flush_fdir_filter(struct i40e_pf > > *pf); static int i40e_flow_flush_ethertype_filter(struct i40e_pf > > *pf); static int i40e_flow_flush_tunnel_filter(struct i40e_pf *pf); > > +static int > > +i40e_flow_parse_qinq_filter(struct rte_eth_dev *dev, > > + const struct rte_flow_attr *attr, > > + const struct rte_flow_item pattern[], > > + const struct rte_flow_action actions[], > > + struct rte_flow_error *error, > > + union i40e_filter_t *filter); static int > > +i40e_flow_parse_qinq_pattern(__rte_unused struct rte_eth_dev *dev, > > + const struct rte_flow_item *pattern, > > + struct rte_flow_error *error, > > + struct i40e_tunnel_filter_conf *filter); > > > > const struct rte_flow_ops i40e_flow_ops = { > > .validate = i40e_flow_validate, > > @@ -318,6 +330,14 @@ static enum rte_flow_item_type pattern_mpls_4[] > = > > { > > RTE_FLOW_ITEM_TYPE_END, > > }; > > > > +/* Pattern matched QINQ */ > > +static enum rte_flow_item_type pattern_qinq_1[] = { > > + RTE_FLOW_ITEM_TYPE_ETH, > > + RTE_FLOW_ITEM_TYPE_VLAN, > > + RTE_FLOW_ITEM_TYPE_VLAN, > > + RTE_FLOW_ITEM_TYPE_END, > > +}; > > + > > static struct i40e_valid_pattern i40e_supported_patterns[] = { > > /* Ethertype */ > > { pattern_ethertype, i40e_flow_parse_ethertype_filter }, @@ -348,6 > > +368,8 @@ static struct i40e_valid_pattern i40e_supported_patterns[] = > > +{ > > { pattern_mpls_2, i40e_flow_parse_mpls_filter }, > > { pattern_mpls_3, i40e_flow_parse_mpls_filter }, > > { pattern_mpls_4, i40e_flow_parse_mpls_filter }, > > + /* QINQ */ > > + { pattern_qinq_1, i40e_flow_parse_qinq_filter }, > > }; > > > > #define NEXT_ITEM_OF_ACTION(act, actions, index) \ > > @@ -1171,7 +1193,7 @@ i40e_flow_parse_fdir_filter(struct rte_eth_dev > > *dev, > > return 0; > > } > > > > -/* Parse to get the action info of a tunnle filter > > +/* Parse to get the action info of a tunnel filter > > * Tunnel action only supports PF, VF and QUEUE. > > */ > > static int > > @@ -1748,6 +1770,167 @@ i40e_flow_parse_mpls_filter(struct > rte_eth_dev > > *dev, } > > > > static int > > +i40e_flow_parse_qinq_pattern(__rte_unused struct rte_eth_dev *dev, > > + const struct rte_flow_item *pattern, > > + struct rte_flow_error *error, > > + struct i40e_tunnel_filter_conf *filter) { > > + const struct rte_flow_item *item = pattern; > > + const struct rte_flow_item_eth *eth_spec; > > + const struct rte_flow_item_eth *eth_mask; > > + const struct rte_flow_item_eth *i_eth_spec = NULL; > > + const struct rte_flow_item_eth *i_eth_mask = NULL; > > + const struct rte_flow_item_vlan *vlan_spec = NULL; > > + const struct rte_flow_item_vlan *vlan_mask = NULL; > > + const struct rte_flow_item_vlan *i_vlan_spec = NULL; > > + const struct rte_flow_item_vlan *o_vlan_spec = NULL; > > + > > + enum rte_flow_item_type item_type; > > + bool vlan_flag = 0; > > Seems vlan_flag is not used at last.
Ignore it, sorry for missing the info of distinguishing outer vlan and inner vlan. > > > + > > + for (; item->type != RTE_FLOW_ITEM_TYPE_END; item++) { > > + if (item->last) { > > + rte_flow_error_set(error, EINVAL, > > + RTE_FLOW_ERROR_TYPE_ITEM, > > + item, > > + "Not support range"); > > + return -rte_errno; > > + } > > + item_type = item->type; > > + switch (item_type) { > > + case RTE_FLOW_ITEM_TYPE_ETH: > > + eth_spec = (const struct rte_flow_item_eth *)item- > > >spec; > > + eth_mask = (const struct rte_flow_item_eth *)item- > > >mask; > > + if ((!eth_spec && eth_mask) || > > + (eth_spec && !eth_mask)) { > > + rte_flow_error_set(error, EINVAL, > > + > > RTE_FLOW_ERROR_TYPE_ITEM, > > + item, > > + "Invalid ether spec/mask"); > > + return -rte_errno; > > + } > > + > > + if (eth_spec && eth_mask) { > > + /* DST address of inner MAC shouldn't be > > masked. > > + * SRC address of Inner MAC should be > > masked. > > + */ > > + if (!is_broadcast_ether_addr(ð_mask- > > >dst) || > > + !is_zero_ether_addr(ð_mask->src) || > > + eth_mask->type) { > > + rte_flow_error_set(error, EINVAL, > > + > > RTE_FLOW_ERROR_TYPE_ITEM, > > + item, > > + "Invalid ether spec/mask"); > > + return -rte_errno; > > + } > > + > > + rte_memcpy(&filter->outer_mac, > > + ð_spec->dst, > > + ETHER_ADDR_LEN); > > + } > > + > > + i_eth_spec = eth_spec; > > + i_eth_mask = eth_mask; > > + break; > > + case RTE_FLOW_ITEM_TYPE_VLAN: > > + vlan_spec = > > + (const struct rte_flow_item_vlan *)item- > > >spec; > > + vlan_mask = > > + (const struct rte_flow_item_vlan *)item- > > >mask; > > + > > + if (!(vlan_spec && vlan_mask)) { > > + rte_flow_error_set(error, EINVAL, > > + RTE_FLOW_ERROR_TYPE_ITEM, > > + item, > > + "Invalid vlan item"); > > + return -rte_errno; > > + } > > + > > + if (!vlan_flag) { > > + o_vlan_spec = vlan_spec; > > + vlan_flag = 1; > > + } else { > > + i_vlan_spec = vlan_spec; > > + vlan_flag = 0; > > + } > > + break; > > + > > + default: > > + break; > > + } > > + } > > + > > + /* Check specification and mask to get the filter type */ > > + if (vlan_spec && vlan_mask && > > + (vlan_mask->tci == rte_cpu_to_be_16(I40E_TCI_MASK))) { > > + /* There is an inner and outer vlan */ > > + filter->outer_vlan = rte_be_to_cpu_16(o_vlan_spec->tci) > > + & I40E_TCI_MASK; > > + filter->inner_vlan = rte_be_to_cpu_16(i_vlan_spec->tci) > > + & I40E_TCI_MASK; > > + if (i_eth_spec && i_eth_mask) > > + filter->filter_type = > > + I40E_TUNNEL_FILTER_CUSTOM_QINQ; > > + else { > > + rte_flow_error_set(error, EINVAL, > > + RTE_FLOW_ERROR_TYPE_ITEM, > > + NULL, > > + "Invalid filter type"); > > + return -rte_errno; > > + } > > + } else if ((!vlan_spec && !vlan_mask) || > > + (vlan_spec && vlan_mask && vlan_mask->tci == 0x0)) { > > + if (i_eth_spec && i_eth_mask) { > > + filter->filter_type = > > I40E_TUNNEL_FILTER_CUSTOM_QINQ; > > + } else { > > + rte_flow_error_set(error, EINVAL, > > + RTE_FLOW_ERROR_TYPE_ITEM, NULL, > > + "Invalid filter type"); > > + return -rte_errno; > > + } > > + } else { > > + rte_flow_error_set(error, EINVAL, > > + RTE_FLOW_ERROR_TYPE_ITEM, NULL, > > + "Not supported by tunnel filter."); > > + return -rte_errno; > > + } > > + > > + filter->tunnel_type = I40E_TUNNEL_TYPE_QINQ; > > + > > + return 0; > > +} > > + > > +static int > > +i40e_flow_parse_qinq_filter(struct rte_eth_dev *dev, > > + const struct rte_flow_attr *attr, > > + const struct rte_flow_item pattern[], > > + const struct rte_flow_action actions[], > > + struct rte_flow_error *error, > > + union i40e_filter_t *filter) { > > + struct i40e_tunnel_filter_conf *tunnel_filter = > > + &filter->consistent_tunnel_filter; > > + int ret; > > + > > + ret = i40e_flow_parse_qinq_pattern(dev, pattern, > > + error, tunnel_filter); > > + if (ret) > > + return ret; > > + > > + ret = i40e_flow_parse_tunnel_action(dev, actions, error, > > tunnel_filter); > > + if (ret) > > + return ret; > > + > > + ret = i40e_flow_parse_attr(attr, error); > > + if (ret) > > + return ret; > > + > > + cons_filter_type = RTE_ETH_FILTER_TUNNEL; > > + > > + return ret; > > +} > > + > > +static int > > i40e_flow_validate(struct rte_eth_dev *dev, > > const struct rte_flow_attr *attr, > > const struct rte_flow_item pattern[], > > -- > > 2.11.0