From: Mateusz Polchlopek <mateusz.polchlo...@intel.com>
Date: Tue, 30 Jul 2024 05:15:06 -0400

> From: Jacob Keller <jacob.e.kel...@intel.com>
> 
> Using VIRTCHNL_VF_OFFLOAD_FLEX_DESC, the iAVF driver is capable of
> negotiating to enable the advanced flexible descriptor layout. Add the
> flexible NIC layout (RXDID=2) as a member of the Rx descriptor union.

[...]

> -static void iavf_rx_checksum(struct iavf_vsi *vsi, struct sk_buff *skb,
> -                          struct iavf_rx_desc *rx_desc)
> +static void iavf_rx_csum(const struct iavf_vsi *vsi, struct sk_buff *skb,
> +                      struct libeth_rx_pt ptype,

You can save a couple lines of diff if you rename ptype -> decoded. And
it would also be more consistent to what you're defining below.

> +                      struct libeth_rx_csum_decoded csum_bits)
>  {
> -     struct libeth_rx_pt decoded;
> -     u32 rx_error, rx_status;
>       bool ipv4, ipv6;
> -     u64 qword;
> -     u8 ptype;
>  
>       skb->ip_summed = CHECKSUM_NONE;
>  
> -     qword = le64_to_cpu(rx_desc->qw1);
> -     ptype = FIELD_GET(IAVF_RXD_QW1_PTYPE_MASK, qword);
> -
> -     decoded = libie_rx_pt_parse(ptype);
> -     if (!libeth_rx_pt_has_checksum(vsi->netdev, decoded))
> -             return;
> -
> -     rx_error = FIELD_GET(IAVF_RXD_QW1_ERROR_MASK, qword);
> -     rx_status = FIELD_GET(IAVF_RXD_QW1_STATUS_MASK, qword);
> -
>       /* did the hardware decode the packet and checksum? */
> -     if (!(rx_status & BIT(IAVF_RX_DESC_STATUS_L3L4P_SHIFT)))
> +     if (unlikely(!csum_bits.l3l4p))
>               return;
>  
> -     ipv4 = libeth_rx_pt_get_ip_ver(decoded) == LIBETH_RX_PT_OUTER_IPV4;
> -     ipv6 = libeth_rx_pt_get_ip_ver(decoded) == LIBETH_RX_PT_OUTER_IPV6;
> +     ipv4 = libeth_rx_pt_get_ip_ver(ptype) == LIBETH_RX_PT_OUTER_IPV4;
> +     ipv6 = libeth_rx_pt_get_ip_ver(ptype) == LIBETH_RX_PT_OUTER_IPV6;
>  
> -     if (ipv4 &&
> -         (rx_error & (BIT(IAVF_RX_DESC_ERROR_IPE_SHIFT) |
> -                      BIT(IAVF_RX_DESC_ERROR_EIPE_SHIFT))))
> +     if (unlikely(ipv4 && (csum_bits.ipe || csum_bits.eipe)))
>               goto checksum_fail;
>  
>       /* likely incorrect csum if alternate IP extension headers found */
> -     if (ipv6 &&
> -         rx_status & BIT(IAVF_RX_DESC_STATUS_IPV6EXADD_SHIFT))
> -             /* don't increment checksum err here, non-fatal err */
> +     if (unlikely(ipv6 && csum_bits.ipv6exadd))
>               return;
>  
>       /* there was some L4 error, count error and punt packet to the stack */
> -     if (rx_error & BIT(IAVF_RX_DESC_ERROR_L4E_SHIFT))
> +     if (unlikely(csum_bits.l4e))
>               goto checksum_fail;
>  
>       /* handle packets that were not able to be checksummed due
>        * to arrival speed, in this case the stack can compute
>        * the csum.
>        */
> -     if (rx_error & BIT(IAVF_RX_DESC_ERROR_PPRS_SHIFT))
> +     if (unlikely(csum_bits.pprs))
>               return;
>  
>       skb->ip_summed = CHECKSUM_UNNECESSARY;
> @@ -959,30 +943,122 @@ static void iavf_rx_checksum(struct iavf_vsi *vsi, 
> struct sk_buff *skb,
>  }
>  
>  /**
> - * iavf_rx_hash - set the hash value in the skb
> + * iavf_legacy_rx_csum - Indicate in skb if hw indicated a good checksum
> + * @vsi: the VSI we care about
> + * @rx_desc: the receive descriptor
> + * @decoded: decoded packet type
> + *
> + * This function only operates on the VIRTCHNL_RXDID_1_32B_BASE legacy 32byte
> + * descriptor writeback format.
> + *
> + * Return: decoded checksum bits

(don't forget about periods at the end of sentences)

> + **/
> +static struct libeth_rx_csum_decoded
> +iavf_legacy_rx_csum(const struct iavf_vsi *vsi,
> +                 const struct iavf_rx_desc *rx_desc,
> +                 const struct libeth_rx_pt decoded)
> +{
> +     struct libeth_rx_csum_decoded csum_bits = {};
> +     u64 qw1 = le64_to_cpu(rx_desc->qw1);
> +
> +     if (!libeth_rx_pt_has_checksum(vsi->netdev, decoded))
> +             return csum_bits;
> +
> +     csum_bits.ipe = FIELD_GET(IAVF_RXD_LEGACY_QW1_IPE_M, qw1);
> +     csum_bits.eipe = FIELD_GET(IAVF_RXD_LEGACY_QW1_EIPE_M, qw1);
> +     csum_bits.l4e = FIELD_GET(IAVF_RXD_LEGACY_QW1_L4E_M, qw1);
> +     csum_bits.pprs = FIELD_GET(IAVF_RXD_LEGACY_QW1_PPRS_M, qw1);
> +     csum_bits.l3l4p = FIELD_GET(IAVF_RXD_LEGACY_QW1_L3L4P_M, qw1);
> +     csum_bits.ipv6exadd = FIELD_GET(IAVF_RXD_LEGACY_QW1_IPV6EXADD_M, qw1);
> +
> +     return csum_bits;
> +}
> +
> +/**
> + * iavf_flex_rx_csum - Indicate in skb if hw indicated a good checksum
> + * @vsi: the VSI we care about
> + * @rx_desc: the receive descriptor
> + * @decoded: decoded packet type
> + *
> + * This function only operates on the VIRTCHNL_RXDID_2_FLEX_SQ_NIC flexible
> + * descriptor writeback format.
> + *
> + * Return: decoded checksum bits
> + **/
> +static struct libeth_rx_csum_decoded
> +iavf_flex_rx_csum(const struct iavf_vsi *vsi,
> +               const struct iavf_rx_desc *rx_desc,
> +               const struct libeth_rx_pt decoded)
> +{
> +     struct libeth_rx_csum_decoded csum_bits = {};
> +     u64 qw1 = le64_to_cpu(rx_desc->qw1);
> +
> +     if (!libeth_rx_pt_has_checksum(vsi->netdev, decoded))
> +             return csum_bits;
> +
> +     csum_bits.ipe = FIELD_GET(IAVF_RXD_FLEX_QW1_XSUM_IPE_M, qw1);
> +     csum_bits.eipe = FIELD_GET(IAVF_RXD_FLEX_QW1_XSUM_EIPE_M, qw1);
> +     csum_bits.l4e = FIELD_GET(IAVF_RXD_FLEX_QW1_XSUM_L4E_M, qw1);
> +     csum_bits.eudpe = FIELD_GET(IAVF_RXD_FLEX_QW1_XSUM_EUDPE_M, qw1);
> +     csum_bits.l3l4p = FIELD_GET(IAVF_RXD_FLEX_QW1_L3L4P_M, qw1);
> +     csum_bits.ipv6exadd = FIELD_GET(IAVF_RXD_FLEX_QW1_IPV6EXADD_M, qw1);
> +     csum_bits.nat = FIELD_GET(IAVF_RXD_FLEX_QW2_NAT_M, qw1);
> +
> +     return csum_bits;
> +}
> +
> +/**
> + * iavf_legacy_rx_hash - set the hash value in the skb
>   * @ring: descriptor ring
>   * @rx_desc: specific descriptor
>   * @skb: skb currently being received and modified
> - * @rx_ptype: Rx packet type
> + * @decoded: decoded packet type
> + *
> + * This function only operates on the VIRTCHNL_RXDID_1_32B_BASE legacy 32byte
> + * descriptor writeback format.
>   **/
> -static void iavf_rx_hash(struct iavf_ring *ring,
> -                      struct iavf_rx_desc *rx_desc,
> -                      struct sk_buff *skb,
> -                      u8 rx_ptype)
> +static void iavf_legacy_rx_hash(const struct iavf_ring *ring,
> +                             const struct iavf_rx_desc *rx_desc,
> +                             struct sk_buff *skb,
> +                             const struct libeth_rx_pt decoded)
>  {
> -     struct libeth_rx_pt decoded;
> +     const __le64 rss_mask = cpu_to_le64(IAVF_RXD_LEGACY_QW1_FLTSTAT_M);
>       u32 hash;
> -     const __le64 rss_mask =
> -             cpu_to_le64((u64)IAVF_RX_DESC_FLTSTAT_RSS_HASH <<
> -                         IAVF_RX_DESC_STATUS_FLTSTAT_SHIFT);
>  
> -     decoded = libie_rx_pt_parse(rx_ptype);
>       if (!libeth_rx_pt_has_hash(ring->netdev, decoded))
>               return;
>  
>       if ((rx_desc->qw1 & rss_mask) == rss_mask) {
> -             hash = le64_get_bits(rx_desc->qw0,
> -                                  IAVF_RX_DESC_LEGACY_QW0_RSS_M);
> +             hash = le64_get_bits(rx_desc->qw0, IAVF_RXD_LEGACY_QW0_RSS_M);
> +             libeth_rx_pt_set_hash(skb, hash, decoded);
> +     }
> +}
> +
> +/**
> + * iavf_flex_rx_hash - set the hash value in the skb
> + * @ring: descriptor ring
> + * @rx_desc: specific descriptor
> + * @skb: skb currently being received and modified
> + * @decoded: decoded packet type
> + *
> + * This function only operates on the VIRTCHNL_RXDID_2_FLEX_SQ_NIC flexible
> + * descriptor writeback format.
> + **/
> +static void iavf_flex_rx_hash(const struct iavf_ring *ring,
> +                           const struct iavf_rx_desc *rx_desc,
> +                           struct sk_buff *skb,
> +                           const struct libeth_rx_pt decoded)
> +{
> +     __le64 qw1 = rx_desc->qw1;
> +     bool rss_valid;
> +     u32 hash;
> +
> +     if (!libeth_rx_pt_has_hash(ring->netdev, decoded))
> +             return;
> +
> +     rss_valid = le64_get_bits(qw1, IAVF_RXD_FLEX_QW1_RSS_VALID_M);
> +     if (rss_valid) {
> +             hash = le64_get_bits(qw1, IAVF_RXD_FLEX_QW1_RSS_HASH_M);
>               libeth_rx_pt_set_hash(skb, hash, decoded);
>       }
>  }
> @@ -998,14 +1074,23 @@ static void iavf_rx_hash(struct iavf_ring *ring,
>   * order to populate the hash, checksum, VLAN, protocol, and
>   * other fields within the skb.
>   **/
> -static void
> -iavf_process_skb_fields(struct iavf_ring *rx_ring,
> -                     struct iavf_rx_desc *rx_desc, struct sk_buff *skb,
> -                     u8 rx_ptype)
> +static void iavf_process_skb_fields(const struct iavf_ring *rx_ring,
> +                                 const struct iavf_rx_desc *rx_desc,
> +                                 struct sk_buff *skb, u32 rx_ptype)
>  {
> -     iavf_rx_hash(rx_ring, rx_desc, skb, rx_ptype);
> +     struct libeth_rx_csum_decoded csum_bits = {};

Since you assign @csum_bits unconditionally below, it's not needed to
initialize it here.

> +     struct libeth_rx_pt decoded;
>  
> -     iavf_rx_checksum(rx_ring->vsi, skb, rx_desc);
> +     decoded = libie_rx_pt_parse(rx_ptype);
> +
> +     if (rx_ring->rxdid == VIRTCHNL_RXDID_1_32B_BASE) {
> +             iavf_legacy_rx_hash(rx_ring, rx_desc, skb, decoded);
> +             csum_bits = iavf_legacy_rx_csum(rx_ring->vsi, rx_desc, decoded);
> +     } else {
> +             iavf_flex_rx_hash(rx_ring, rx_desc, skb, decoded);
> +             csum_bits = iavf_flex_rx_csum(rx_ring->vsi, rx_desc, decoded);
> +     }
> +     iavf_rx_csum(rx_ring->vsi, skb, decoded, csum_bits);
>  
>       skb_record_rx_queue(skb, rx_ring->queue_index);

[...]

> +static struct libeth_rx_extracted
> +iavf_extract_legacy_rx_fields(const struct iavf_ring *rx_ring,
> +                           const struct iavf_rx_desc *rx_desc)
> +{
> +     struct libeth_rx_extracted fields = {};
> +     __le64 qw0 = rx_desc->qw0;
> +     __le64 qw1 = rx_desc->qw1;
> +     __le64 qw2 = rx_desc->qw2;

Make them u64 right here with le64_to_cpu() and then just use
FIELD_GET()s instead of le64_get_bits() below. On BE systems, each
le64_get_bits() implies a byteswap.

> +     bool l2tag1p;
> +     bool l2tag2p;
> +
> +     fields.end_of_packet = le64_get_bits(qw1, IAVF_RXD_LEGACY_QW1_EOP_M);
> +     fields.size = le64_get_bits(qw1, IAVF_RXD_LEGACY_QW1_LENGTH_M);
> +     fields.rxe = le64_get_bits(qw1, IAVF_RXD_LEGACY_QW1_RXE_M);
> +     fields.rx_ptype = le64_get_bits(qw1, IAVF_RXD_LEGACY_QW1_PTYPE_M);
> +
> +     l2tag1p = le64_get_bits(qw1, IAVF_RXD_LEGACY_QW1_L2TAG1P_M);
> +     if (l2tag1p && (rx_ring->flags & IAVF_TXRX_FLAGS_VLAN_TAG_LOC_L2TAG1))
> +             fields.vlan_tag = le64_get_bits(qw0,
> +                                             IAVF_RXD_LEGACY_QW0_L2TAG1_M);
> +
> +     l2tag2p = le64_get_bits(qw2, IAVF_RXD_LEGACY_QW2_L2TAG2P_M);
> +     if (l2tag2p && (rx_ring->flags & IAVF_RXR_FLAGS_VLAN_TAG_LOC_L2TAG2_2))
> +             fields.vlan_tag = le64_get_bits(qw2,
> +                                             IAVF_RXD_LEGACY_QW2_L2TAG2_M);
> +
> +     return fields;

As I wrote in the previous reply, this needs to be split into several
functions as not all the fields are always needed.

> +}
> +
> +/**
> + * iavf_extract_flex_rx_fields - Extract fields from the Rx descriptor
> + * @rx_ring: rx descriptor ring
> + * @rx_desc: the descriptor to process
> + *
> + * Decode the Rx descriptor and extract relevant information including the
> + * size, VLAN tag, Rx packet type, end of packet field and RXE field value.
> + *
> + * This function only operates on the VIRTCHNL_RXDID_2_FLEX_SQ_NIC flexible
> + * descriptor writeback format.
> + *
> + * Return: fields extracted from the Rx descriptor
> + */
> +static struct libeth_rx_extracted
> +iavf_extract_flex_rx_fields(const struct iavf_ring *rx_ring,
> +                         const struct iavf_rx_desc *rx_desc)
> +{
> +     struct libeth_rx_extracted fields = {};
> +     __le64 qw0 = rx_desc->qw0;
> +     __le64 qw1 = rx_desc->qw1;
> +     __le64 qw2 = rx_desc->qw2;

Same here.

> +     bool l2tag1p, l2tag2p;
> +
> +     fields.size = le64_get_bits(qw0, IAVF_RXD_FLEX_QW0_PKT_LEN_M);
> +     fields.rx_ptype = le64_get_bits(qw0, IAVF_RXD_FLEX_QW0_PTYPE_M);
> +     fields.rxe = le64_get_bits(qw1, IAVF_RXD_FLEX_QW1_RXE_M);
> +     fields.end_of_packet = le64_get_bits(qw1, IAVF_RXD_FLEX_QW1_EOP_M);
> +
> +     l2tag1p = le64_get_bits(qw1, IAVF_RXD_FLEX_QW1_L2TAG1P_M);
> +     if (l2tag1p && (rx_ring->flags & IAVF_TXRX_FLAGS_VLAN_TAG_LOC_L2TAG1))
> +             fields.vlan_tag = le64_get_bits(qw1,
> +                                             IAVF_RXD_FLEX_QW1_L2TAG1_M);
> +
> +     l2tag2p = le64_get_bits(qw2, IAVF_RXD_FLEX_QW2_L2TAG2P_M);
> +     if (l2tag2p && (rx_ring->flags & IAVF_RXR_FLAGS_VLAN_TAG_LOC_L2TAG2_2))
> +             fields.vlan_tag = le64_get_bits(qw2,
> +                                             IAVF_RXD_FLEX_QW2_L2TAG2_2_M);
> +
> +     return fields;

Same here.

> +}
> +
> +static struct libeth_rx_extracted
> +iavf_extract_rx_fields(const struct iavf_ring *rx_ring,
> +                    const struct iavf_rx_desc *rx_desc)
> +{
> +     if (rx_ring->rxdid == VIRTCHNL_RXDID_1_32B_BASE)
> +             return iavf_extract_legacy_rx_fields(rx_ring, rx_desc);
> +     else
> +             return iavf_extract_flex_rx_fields(rx_ring, rx_desc);
> +}
> +
>  /**
>   * iavf_clean_rx_irq - Clean completed descriptors from Rx ring - bounce buf
>   * @rx_ring: rx descriptor ring to transact packets on
> @@ -1142,13 +1317,9 @@ static int iavf_clean_rx_irq(struct iavf_ring 
> *rx_ring, int budget)
>       bool failure = false;
>  
>       while (likely(total_rx_packets < (unsigned int)budget)) {
> +             struct libeth_rx_extracted fields = {};

Initialization is not needed since you always assign it.

>               struct libeth_fqe *rx_buffer;
>               struct iavf_rx_desc *rx_desc;
> -             u16 ext_status = 0;
> -             unsigned int size;
> -             u16 vlan_tag = 0;
> -             u8 rx_ptype;
> -             u64 qword;
>  
>               /* return some buffers to hardware, one at a time is too slow */
>               if (cleaned_count >= IAVF_RX_BUFFER_WRITE) {

[...]

> diff --git a/drivers/net/ethernet/intel/iavf/iavf_type.h 
> b/drivers/net/ethernet/intel/iavf/iavf_type.h
> index 07e54db0bd4d..498746a83d35 100644
> --- a/drivers/net/ethernet/intel/iavf/iavf_type.h
> +++ b/drivers/net/ethernet/intel/iavf/iavf_type.h
> @@ -179,39 +179,13 @@ struct iavf_hw {
>  };
>  
>  struct iavf_rx_desc {
> -     aligned_le64 qw0;
> -     aligned_le64 qw1;
> -     aligned_le64 qw2;
> -     aligned_le64 qw3;
> -} __aligned(4 * sizeof(__le64));;
> -
> -enum iavf_rx_desc_status_bits {
> -     /* Note: These are predefined bit offsets */
> -     IAVF_RX_DESC_STATUS_DD_SHIFT            = 0,
> -     IAVF_RX_DESC_STATUS_EOF_SHIFT           = 1,
> -     IAVF_RX_DESC_STATUS_L2TAG1P_SHIFT       = 2,
> -     IAVF_RX_DESC_STATUS_L3L4P_SHIFT         = 3,
> -     IAVF_RX_DESC_STATUS_CRCP_SHIFT          = 4,
> -     IAVF_RX_DESC_STATUS_TSYNINDX_SHIFT      = 5, /* 2 BITS */
> -     IAVF_RX_DESC_STATUS_TSYNVALID_SHIFT     = 7,
> -     /* Note: Bit 8 is reserved in X710 and XL710 */
> -     IAVF_RX_DESC_STATUS_EXT_UDP_0_SHIFT     = 8,
> -     IAVF_RX_DESC_STATUS_UMBCAST_SHIFT       = 9, /* 2 BITS */
> -     IAVF_RX_DESC_STATUS_FLM_SHIFT           = 11,
> -     IAVF_RX_DESC_STATUS_FLTSTAT_SHIFT       = 12, /* 2 BITS */
> -     IAVF_RX_DESC_STATUS_LPBK_SHIFT          = 14,
> -     IAVF_RX_DESC_STATUS_IPV6EXADD_SHIFT     = 15,
> -     IAVF_RX_DESC_STATUS_RESERVED_SHIFT      = 16, /* 2 BITS */
> -     /* Note: For non-tunnel packets INT_UDP_0 is the right status for
> -      * UDP header
> -      */
> -     IAVF_RX_DESC_STATUS_INT_UDP_0_SHIFT     = 18,
> -     IAVF_RX_DESC_STATUS_LAST /* this entry must be last!!! */
> +     __le64 qw0;
> +     __le64 qw1;
> +     __le64 qw2;
> +     __le64 qw3;
>  };

Some rebasing issues here. You redefine the struct you introduced in the
previous patch. I'd say the previous definition was more correct.

>  
> -#define IAVF_RXD_QW1_STATUS_SHIFT    0
> -#define IAVF_RXD_QW1_STATUS_MASK     ((BIT(IAVF_RX_DESC_STATUS_LAST) - 1) \
> -                                      << IAVF_RXD_QW1_STATUS_SHIFT)
> +#define IAVF_RXD_QW1_STATUS_MASK             GENMASK(18, 0)
>  
>  #define IAVF_RXD_QW1_STATUS_TSYNINDX_SHIFT IAVF_RX_DESC_STATUS_TSYNINDX_SHIFT
>  #define IAVF_RXD_QW1_STATUS_TSYNINDX_MASK  (0x3UL << \
> @@ -228,22 +202,6 @@ enum iavf_rx_desc_fltstat_values {
>       IAVF_RX_DESC_FLTSTAT_RSS_HASH   = 3,
>  };
>  
> -#define IAVF_RXD_QW1_ERROR_SHIFT     19
> -#define IAVF_RXD_QW1_ERROR_MASK              (0xFFUL << 
> IAVF_RXD_QW1_ERROR_SHIFT)
> -
> -enum iavf_rx_desc_error_bits {
> -     /* Note: These are predefined bit offsets */
> -     IAVF_RX_DESC_ERROR_RXE_SHIFT            = 0,
> -     IAVF_RX_DESC_ERROR_RECIPE_SHIFT         = 1,
> -     IAVF_RX_DESC_ERROR_HBO_SHIFT            = 2,
> -     IAVF_RX_DESC_ERROR_L3L4E_SHIFT          = 3, /* 3 BITS */
> -     IAVF_RX_DESC_ERROR_IPE_SHIFT            = 3,
> -     IAVF_RX_DESC_ERROR_L4E_SHIFT            = 4,
> -     IAVF_RX_DESC_ERROR_EIPE_SHIFT           = 5,
> -     IAVF_RX_DESC_ERROR_OVERSIZE_SHIFT       = 6,
> -     IAVF_RX_DESC_ERROR_PPRS_SHIFT           = 7
> -};
> -
>  enum iavf_rx_desc_error_l3l4e_fcoe_masks {
>       IAVF_RX_DESC_ERROR_L3L4E_NONE           = 0,
>       IAVF_RX_DESC_ERROR_L3L4E_PROT           = 1,
> @@ -252,13 +210,6 @@ enum iavf_rx_desc_error_l3l4e_fcoe_masks {
>       IAVF_RX_DESC_ERROR_L3L4E_DMAC_WARN      = 4
>  };
>  
> -#define IAVF_RXD_QW1_PTYPE_SHIFT     30
> -#define IAVF_RXD_QW1_PTYPE_MASK              (0xFFULL << 
> IAVF_RXD_QW1_PTYPE_SHIFT)
> -
> -#define IAVF_RXD_QW1_LENGTH_PBUF_SHIFT       38
> -#define IAVF_RXD_QW1_LENGTH_PBUF_MASK        (0x3FFFULL << \
> -                                      IAVF_RXD_QW1_LENGTH_PBUF_SHIFT)
> -
>  #define IAVF_RXD_QW1_LENGTH_HBUF_SHIFT       52
>  #define IAVF_RXD_QW1_LENGTH_HBUF_MASK        (0x7FFULL << \
>                                        IAVF_RXD_QW1_LENGTH_HBUF_SHIFT)
> @@ -505,9 +456,85 @@ struct iavf_eth_stats {
>       u64 tx_errors;                  /* tepc */
>  };
>  
> -#define IAVF_RX_DESC_LEGACY_QW0_RSS_M                GENMASK_ULL(63, 32)
> -#define IAVF_RX_DESC_LEGACY_QW0_L2TAG1_M     GENMASK_ULL(33, 16)
> -#define IAVF_RX_DESC_LEGACY_QW2_L2TAG2_2_M   GENMASK_ULL(63, 48)
> -#define IAVF_RX_DESC_LEGACY_QW2_EXT_STATUS_M GENMASK_ULL(11, 0)

Define these four correctly in the previous patch, so you wouldn't
redefine it here once again?

> +/* LEGACY DESCRIPTOR */
> +/* Quad Word 0 */
> +#define IAVF_RXD_LEGACY_QW0_RSS_M            GENMASK_ULL(63, 32)
> +#define IAVF_RXD_LEGACY_QW0_L2TAG1_M         GENMASK_ULL(31, 16)

[...]

> diff --git a/drivers/net/ethernet/intel/iavf/iavf_virtchnl.c 
> b/drivers/net/ethernet/intel/iavf/iavf_virtchnl.c
> index 4163dfe90b4a..d60fba84b109 100644
> --- a/drivers/net/ethernet/intel/iavf/iavf_virtchnl.c
> +++ b/drivers/net/ethernet/intel/iavf/iavf_virtchnl.c
> @@ -403,6 +403,7 @@ void iavf_configure_queues(struct iavf_adapter *adapter)
>       int pairs = adapter->num_active_queues;
>       struct virtchnl_queue_pair_info *vqpi;
>       u32 i, max_frame;
> +     u8 rx_flags = 0;
>       size_t len;
>  
>       max_frame = LIBIE_MAX_RX_FRM_LEN(adapter->rx_rings->pp->p.offset);
> @@ -420,6 +421,9 @@ void iavf_configure_queues(struct iavf_adapter *adapter)
>       if (!vqci)
>               return;
>  
> +     if (iavf_ptp_cap_supported(adapter, VIRTCHNL_1588_PTP_CAP_RX_TSTAMP))
> +             rx_flags |= VIRTCHNL_PTP_RX_TSTAMP;

This is not related to the Rx descriptor refactoring I'd say?

> +
>       vqci->vsi_id = adapter->vsi_res->vsi_id;
>       vqci->num_queue_pairs = pairs;
>       vqpi = vqci->qpair;
> @@ -442,6 +446,7 @@ void iavf_configure_queues(struct iavf_adapter *adapter)
>               if (CRC_OFFLOAD_ALLOWED(adapter))
>                       vqpi->rxq.crc_disable = !!(adapter->netdev->features &
>                                                  NETIF_F_RXFCS);
> +             vqpi->rxq.flags = rx_flags;
>               vqpi++;
>       }

Thanks,
Olek

Reply via email to