Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit 6b037cd4 authored by Alexander Duyck's avatar Alexander Duyck Committed by Jeff Kirsher
Browse files

i40e: Fix ATR in relation to tunnels



This patch contains a number of fixes to make certain that we are using
the correct protocols when parsing both the inner and outer headers of a
frame that is mixed between IPv4 and IPv6 for inner and outer.

Signed-off-by: default avatarAlexander Duyck <aduyck@mirantis.com>
Acked-by: default avatarKiran Patil <kiran.patil@intel.com>
Tested-by: default avatarAndrew Bowers <andrewx.bowers@intel.com>
Signed-off-by: default avatarJeff Kirsher <jeffrey.t.kirsher@intel.com>
parent 5453205c
Loading
Loading
Loading
Loading
+11 −17
Original line number Original line Diff line number Diff line
@@ -2030,10 +2030,9 @@ int i40e_napi_poll(struct napi_struct *napi, int budget)
 * @tx_ring:  ring to add programming descriptor to
 * @tx_ring:  ring to add programming descriptor to
 * @skb:      send buffer
 * @skb:      send buffer
 * @tx_flags: send tx flags
 * @tx_flags: send tx flags
 * @protocol: wire protocol
 **/
 **/
static void i40e_atr(struct i40e_ring *tx_ring, struct sk_buff *skb,
static void i40e_atr(struct i40e_ring *tx_ring, struct sk_buff *skb,
		     u32 tx_flags, __be16 protocol)
		     u32 tx_flags)
{
{
	struct i40e_filter_program_desc *fdir_desc;
	struct i40e_filter_program_desc *fdir_desc;
	struct i40e_pf *pf = tx_ring->vsi->back;
	struct i40e_pf *pf = tx_ring->vsi->back;
@@ -2045,6 +2044,7 @@ static void i40e_atr(struct i40e_ring *tx_ring, struct sk_buff *skb,
	struct tcphdr *th;
	struct tcphdr *th;
	unsigned int hlen;
	unsigned int hlen;
	u32 flex_ptype, dtype_cmd;
	u32 flex_ptype, dtype_cmd;
	u8 l4_proto;
	u16 i;
	u16 i;


	/* make sure ATR is enabled */
	/* make sure ATR is enabled */
@@ -2058,6 +2058,7 @@ static void i40e_atr(struct i40e_ring *tx_ring, struct sk_buff *skb,
	if (!tx_ring->atr_sample_rate)
	if (!tx_ring->atr_sample_rate)
		return;
		return;


	/* Currently only IPv4/IPv6 with TCP is supported */
	if (!(tx_flags & (I40E_TX_FLAGS_IPV4 | I40E_TX_FLAGS_IPV6)))
	if (!(tx_flags & (I40E_TX_FLAGS_IPV4 | I40E_TX_FLAGS_IPV6)))
		return;
		return;


@@ -2065,29 +2066,22 @@ static void i40e_atr(struct i40e_ring *tx_ring, struct sk_buff *skb,
		/* snag network header to get L4 type and address */
		/* snag network header to get L4 type and address */
		hdr.network = skb_network_header(skb);
		hdr.network = skb_network_header(skb);


		/* Currently only IPv4/IPv6 with TCP is supported
		/* access ihl as u8 to avoid unaligned access on ia64 */
		 * access ihl as u8 to avoid unaligned access on ia64
		 */
		if (tx_flags & I40E_TX_FLAGS_IPV4)
		if (tx_flags & I40E_TX_FLAGS_IPV4)
			hlen = (hdr.network[0] & 0x0F) << 2;
			hlen = (hdr.network[0] & 0x0F) << 2;
		else if (protocol == htons(ETH_P_IPV6))
			hlen = sizeof(struct ipv6hdr);
		else
		else
			return;
			hlen = sizeof(struct ipv6hdr);
	} else {
	} else {
		hdr.network = skb_inner_network_header(skb);
		hdr.network = skb_inner_network_header(skb);
		hlen = skb_inner_network_header_len(skb);
		hlen = skb_inner_network_header_len(skb);
	}
	}


	/* Currently only IPv4/IPv6 with TCP is supported
	/* Note: tx_flags gets modified to reflect inner protocols in
	 * Note: tx_flags gets modified to reflect inner protocols in
	 * tx_enable_csum function if encap is enabled.
	 * tx_enable_csum function if encap is enabled.
	 */
	 */
	if ((tx_flags & I40E_TX_FLAGS_IPV4) &&
	l4_proto = (tx_flags & I40E_TX_FLAGS_IPV4) ? hdr.ipv4->protocol :
	    (hdr.ipv4->protocol != IPPROTO_TCP))
						     hdr.ipv6->nexthdr;
		return;
	if (l4_proto != IPPROTO_TCP)
	else if ((tx_flags & I40E_TX_FLAGS_IPV6) &&
		 (hdr.ipv6->nexthdr != IPPROTO_TCP))
		return;
		return;


	th = (struct tcphdr *)(hdr.network + hlen);
	th = (struct tcphdr *)(hdr.network + hlen);
@@ -2124,7 +2118,7 @@ static void i40e_atr(struct i40e_ring *tx_ring, struct sk_buff *skb,


	flex_ptype = (tx_ring->queue_index << I40E_TXD_FLTR_QW0_QINDEX_SHIFT) &
	flex_ptype = (tx_ring->queue_index << I40E_TXD_FLTR_QW0_QINDEX_SHIFT) &
		      I40E_TXD_FLTR_QW0_QINDEX_MASK;
		      I40E_TXD_FLTR_QW0_QINDEX_MASK;
	flex_ptype |= (protocol == htons(ETH_P_IP)) ?
	flex_ptype |= (tx_flags & I40E_TX_FLAGS_IPV4) ?
		      (I40E_FILTER_PCTYPE_NONF_IPV4_TCP <<
		      (I40E_FILTER_PCTYPE_NONF_IPV4_TCP <<
		       I40E_TXD_FLTR_QW0_PCTYPE_SHIFT) :
		       I40E_TXD_FLTR_QW0_PCTYPE_SHIFT) :
		      (I40E_FILTER_PCTYPE_NONF_IPV6_TCP <<
		      (I40E_FILTER_PCTYPE_NONF_IPV6_TCP <<
@@ -2992,7 +2986,7 @@ static netdev_tx_t i40e_xmit_frame_ring(struct sk_buff *skb,
	 *
	 *
	 * NOTE: this must always be directly before the data descriptor.
	 * NOTE: this must always be directly before the data descriptor.
	 */
	 */
	i40e_atr(tx_ring, skb, tx_flags, protocol);
	i40e_atr(tx_ring, skb, tx_flags);


	i40e_tx_map(tx_ring, skb, first, tx_flags, hdr_len,
	i40e_tx_map(tx_ring, skb, first, tx_flags, hdr_len,
		    td_cmd, td_offset);
		    td_cmd, td_offset);