[dpdk-dev] [PATCH v3 2/5] net/i40e: parse QinQ pattern
Xing, Beilei
beilei.xing at intel.com
Wed Mar 29 04:46:14 CEST 2017
Hi Bernard,
> -----Original Message-----
> From: Iremonger, Bernard
> Sent: Wednesday, March 29, 2017 12:21 AM
> To: dev at dpdk.org; Xing, Beilei <beilei.xing at intel.com>; Wu, Jingjing
> <jingjing.wu at intel.com>
> Cc: Zhang, Helin <helin.zhang at intel.com>; Lu, Wenzhuo
> <wenzhuo.lu at intel.com>; Iremonger, Bernard
> <bernard.iremonger at intel.com>
> Subject: [PATCH v3 2/5] net/i40e: parse QinQ pattern
>
> add QinQ pattern.
> add i40e_flow_parse_qinq_pattern function.
> add i40e_flow_parse_qinq_filter function.
>
> Signed-off-by: Bernard Iremonger <bernard.iremonger at intel.com>
> ---
> drivers/net/i40e/i40e_flow.c | 187
> ++++++++++++++++++++++++++++++++++++++++++-
> 1 file changed, 185 insertions(+), 2 deletions(-)
>
> diff --git a/drivers/net/i40e/i40e_flow.c b/drivers/net/i40e/i40e_flow.c
> index be243e172..39b09ead5 100644
> --- a/drivers/net/i40e/i40e_flow.c
> +++ b/drivers/net/i40e/i40e_flow.c
> @@ -1,7 +1,7 @@
> /*-
> * BSD LICENSE
> *
> - * Copyright (c) 2016 Intel Corporation. All rights reserved.
> + * Copyright (c) 2016-2017 Intel Corporation. All rights reserved.
> *
> * Redistribution and use in source and binary forms, with or without
> * modification, are permitted provided that the following conditions
> @@ -128,6 +128,18 @@ static int i40e_flow_destroy_tunnel_filter(struct
> i40e_pf *pf, static int i40e_flow_flush_fdir_filter(struct i40e_pf *pf); static
> int i40e_flow_flush_ethertype_filter(struct i40e_pf *pf); static int
> i40e_flow_flush_tunnel_filter(struct i40e_pf *pf);
> +static int
> +i40e_flow_parse_qinq_filter(struct rte_eth_dev *dev,
> + const struct rte_flow_attr *attr,
> + const struct rte_flow_item pattern[],
> + const struct rte_flow_action actions[],
> + struct rte_flow_error *error,
> + union i40e_filter_t *filter);
> +static int
> +i40e_flow_parse_qinq_pattern(__rte_unused struct rte_eth_dev *dev,
> + const struct rte_flow_item *pattern,
> + struct rte_flow_error *error,
> + struct i40e_tunnel_filter_conf *filter);
>
> const struct rte_flow_ops i40e_flow_ops = {
> .validate = i40e_flow_validate,
> @@ -318,6 +330,14 @@ static enum rte_flow_item_type pattern_mpls_4[] =
> {
> RTE_FLOW_ITEM_TYPE_END,
> };
>
> +/* Pattern matched QINQ */
> +static enum rte_flow_item_type pattern_qinq_1[] = {
> + RTE_FLOW_ITEM_TYPE_ETH,
> + RTE_FLOW_ITEM_TYPE_VLAN,
> + RTE_FLOW_ITEM_TYPE_VLAN,
> + RTE_FLOW_ITEM_TYPE_END,
> +};
> +
> static struct i40e_valid_pattern i40e_supported_patterns[] = {
> /* Ethertype */
> { pattern_ethertype, i40e_flow_parse_ethertype_filter }, @@ -348,6
> +368,8 @@ static struct i40e_valid_pattern i40e_supported_patterns[] = {
> { pattern_mpls_2, i40e_flow_parse_mpls_filter },
> { pattern_mpls_3, i40e_flow_parse_mpls_filter },
> { pattern_mpls_4, i40e_flow_parse_mpls_filter },
> + /* QINQ */
> + { pattern_qinq_1, i40e_flow_parse_qinq_filter },
> };
>
> #define NEXT_ITEM_OF_ACTION(act, actions, index) \
> @@ -1171,7 +1193,7 @@ i40e_flow_parse_fdir_filter(struct rte_eth_dev
> *dev,
> return 0;
> }
>
> -/* Parse to get the action info of a tunnle filter
> +/* Parse to get the action info of a tunnel filter
> * Tunnel action only supports PF, VF and QUEUE.
> */
> static int
> @@ -1748,6 +1770,167 @@ i40e_flow_parse_mpls_filter(struct rte_eth_dev
> *dev, }
>
> static int
> +i40e_flow_parse_qinq_pattern(__rte_unused struct rte_eth_dev *dev,
> + const struct rte_flow_item *pattern,
> + struct rte_flow_error *error,
> + struct i40e_tunnel_filter_conf *filter) {
> + const struct rte_flow_item *item = pattern;
> + const struct rte_flow_item_eth *eth_spec;
> + const struct rte_flow_item_eth *eth_mask;
> + const struct rte_flow_item_eth *i_eth_spec = NULL;
> + const struct rte_flow_item_eth *i_eth_mask = NULL;
> + const struct rte_flow_item_vlan *vlan_spec = NULL;
> + const struct rte_flow_item_vlan *vlan_mask = NULL;
> + const struct rte_flow_item_vlan *i_vlan_spec = NULL;
> + const struct rte_flow_item_vlan *o_vlan_spec = NULL;
> +
> + enum rte_flow_item_type item_type;
> + bool vlan_flag = 0;
Seems vlan_flag is not used at last.
> +
> + for (; item->type != RTE_FLOW_ITEM_TYPE_END; item++) {
> + if (item->last) {
> + rte_flow_error_set(error, EINVAL,
> + RTE_FLOW_ERROR_TYPE_ITEM,
> + item,
> + "Not support range");
> + return -rte_errno;
> + }
> + item_type = item->type;
> + switch (item_type) {
> + case RTE_FLOW_ITEM_TYPE_ETH:
> + eth_spec = (const struct rte_flow_item_eth *)item-
> >spec;
> + eth_mask = (const struct rte_flow_item_eth *)item-
> >mask;
> + if ((!eth_spec && eth_mask) ||
> + (eth_spec && !eth_mask)) {
> + rte_flow_error_set(error, EINVAL,
> +
> RTE_FLOW_ERROR_TYPE_ITEM,
> + item,
> + "Invalid ether spec/mask");
> + return -rte_errno;
> + }
> +
> + if (eth_spec && eth_mask) {
> + /* DST address of inner MAC shouldn't be
> masked.
> + * SRC address of Inner MAC should be
> masked.
> + */
> + if (!is_broadcast_ether_addr(ð_mask-
> >dst) ||
> + !is_zero_ether_addr(ð_mask->src) ||
> + eth_mask->type) {
> + rte_flow_error_set(error, EINVAL,
> +
> RTE_FLOW_ERROR_TYPE_ITEM,
> + item,
> + "Invalid ether spec/mask");
> + return -rte_errno;
> + }
> +
> + rte_memcpy(&filter->outer_mac,
> + ð_spec->dst,
> + ETHER_ADDR_LEN);
> + }
> +
> + i_eth_spec = eth_spec;
> + i_eth_mask = eth_mask;
> + break;
> + case RTE_FLOW_ITEM_TYPE_VLAN:
> + vlan_spec =
> + (const struct rte_flow_item_vlan *)item-
> >spec;
> + vlan_mask =
> + (const struct rte_flow_item_vlan *)item-
> >mask;
> +
> + if (!(vlan_spec && vlan_mask)) {
> + rte_flow_error_set(error, EINVAL,
> + RTE_FLOW_ERROR_TYPE_ITEM,
> + item,
> + "Invalid vlan item");
> + return -rte_errno;
> + }
> +
> + if (!vlan_flag) {
> + o_vlan_spec = vlan_spec;
> + vlan_flag = 1;
> + } else {
> + i_vlan_spec = vlan_spec;
> + vlan_flag = 0;
> + }
> + break;
> +
> + default:
> + break;
> + }
> + }
> +
> + /* Check specification and mask to get the filter type */
> + if (vlan_spec && vlan_mask &&
> + (vlan_mask->tci == rte_cpu_to_be_16(I40E_TCI_MASK))) {
> + /* There is an inner and outer vlan */
> + filter->outer_vlan = rte_be_to_cpu_16(o_vlan_spec->tci)
> + & I40E_TCI_MASK;
> + filter->inner_vlan = rte_be_to_cpu_16(i_vlan_spec->tci)
> + & I40E_TCI_MASK;
> + if (i_eth_spec && i_eth_mask)
> + filter->filter_type =
> + I40E_TUNNEL_FILTER_CUSTOM_QINQ;
> + else {
> + rte_flow_error_set(error, EINVAL,
> + RTE_FLOW_ERROR_TYPE_ITEM,
> + NULL,
> + "Invalid filter type");
> + return -rte_errno;
> + }
> + } else if ((!vlan_spec && !vlan_mask) ||
> + (vlan_spec && vlan_mask && vlan_mask->tci == 0x0)) {
> + if (i_eth_spec && i_eth_mask) {
> + filter->filter_type =
> I40E_TUNNEL_FILTER_CUSTOM_QINQ;
> + } else {
> + rte_flow_error_set(error, EINVAL,
> + RTE_FLOW_ERROR_TYPE_ITEM, NULL,
> + "Invalid filter type");
> + return -rte_errno;
> + }
> + } else {
> + rte_flow_error_set(error, EINVAL,
> + RTE_FLOW_ERROR_TYPE_ITEM, NULL,
> + "Not supported by tunnel filter.");
> + return -rte_errno;
> + }
> +
> + filter->tunnel_type = I40E_TUNNEL_TYPE_QINQ;
> +
> + return 0;
> +}
> +
> +static int
> +i40e_flow_parse_qinq_filter(struct rte_eth_dev *dev,
> + const struct rte_flow_attr *attr,
> + const struct rte_flow_item pattern[],
> + const struct rte_flow_action actions[],
> + struct rte_flow_error *error,
> + union i40e_filter_t *filter)
> +{
> + struct i40e_tunnel_filter_conf *tunnel_filter =
> + &filter->consistent_tunnel_filter;
> + int ret;
> +
> + ret = i40e_flow_parse_qinq_pattern(dev, pattern,
> + error, tunnel_filter);
> + if (ret)
> + return ret;
> +
> + ret = i40e_flow_parse_tunnel_action(dev, actions, error,
> tunnel_filter);
> + if (ret)
> + return ret;
> +
> + ret = i40e_flow_parse_attr(attr, error);
> + if (ret)
> + return ret;
> +
> + cons_filter_type = RTE_ETH_FILTER_TUNNEL;
> +
> + return ret;
> +}
> +
> +static int
> i40e_flow_validate(struct rte_eth_dev *dev,
> const struct rte_flow_attr *attr,
> const struct rte_flow_item pattern[],
> --
> 2.11.0
More information about the dev
mailing list