ASR_BASE

Change-Id: Icf3719cc0afe3eeb3edc7fa80a2eb5199ca9dda1
diff --git a/target/linux/layerscape/patches-5.4/701-net-0358-sdk_dpaa-ls1043a-errata-update-and-optimize-the-rest.patch b/target/linux/layerscape/patches-5.4/701-net-0358-sdk_dpaa-ls1043a-errata-update-and-optimize-the-rest.patch
new file mode 100644
index 0000000..684d04b
--- /dev/null
+++ b/target/linux/layerscape/patches-5.4/701-net-0358-sdk_dpaa-ls1043a-errata-update-and-optimize-the-rest.patch
@@ -0,0 +1,252 @@
+From e89db675171a7a12f19b6ec0089a9cc62807cdf1 Mon Sep 17 00:00:00 2001
+From: Camelia Groza <camelia.groza@nxp.com>
+Date: Tue, 29 Oct 2019 16:34:08 +0200
+Subject: [PATCH] sdk_dpaa: ls1043a errata: update and optimize the
+ restrictions
+
+An skb is in no danger of triggering the errata under the following
+conditions:
+- the paged data doesn't cross a 4K page boundary OR the linear data
+is aligned to 256 bytes when crossing a 4K page boundary
+- the linear and the paged data are 16 byte aligned
+- the paged data is a multiple of 16 bytes in size
+
+Optimize the detection for each skb that might trigger the errata. Parse
+the skb twice, at most, and realign it only once.
+
+Signed-off-by: Camelia Groza <camelia.groza@nxp.com>
+---
+ drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth.h |   2 +-
+ .../net/ethernet/freescale/sdk_dpaa/dpaa_eth_sg.c  | 147 +++++++++++++++------
+ 2 files changed, 111 insertions(+), 38 deletions(-)
+
+--- a/drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth.h
++++ b/drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth.h
+@@ -662,7 +662,7 @@ static inline void _dpa_bp_free_pf(void
+ #ifndef CONFIG_PPC
+ extern bool dpaa_errata_a010022; /* SoC affected by A010022 errata */
+ #define NONREC_MARK	0x01
+-#define HAS_DMA_ISSUE(start, size) \
++#define CROSS_4K(start, size) \
+ 	(((uintptr_t)(start) + (size)) > \
+ 	 (((uintptr_t)(start) + 0x1000) & ~0xFFF))
+ /* The headroom needs to accommodate our private data (64 bytes) but
+--- a/drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth_sg.c
++++ b/drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth_sg.c
+@@ -771,32 +771,73 @@ int __hot skb_to_contig_fd(struct dpa_pr
+ EXPORT_SYMBOL(skb_to_contig_fd);
+ 
+ #ifndef CONFIG_PPC
+-/* Verify the conditions that trigger the A010022 errata: data unaligned to
+- * 16 bytes, 4K memory address crossings and S/G fragments.
++/* Verify the conditions that trigger the A010022 errata:
++ * - 4K memory address boundary crossings when the data/SG fragments aren't
++ *   aligned to 256 bytes
++ * - data and SG fragments that aren't aligned to 16 bytes
++ * - SG fragments that aren't mod 16 bytes in size (except for the last
++ *   fragment)
+  */
+ static bool a010022_check_skb(struct sk_buff *skb, struct dpa_priv_s *priv)
+ {
+-	/* Check if the headroom is aligned */
+-	if (((uintptr_t)skb->data - priv->tx_headroom) %
+-	    priv->buf_layout[TX].data_align != 0)
+-		return true;
++	skb_frag_t *frag;
++	int i, nr_frags;
+ 
+-	/* Check for paged data in the skb. We do not support S/G fragments */
+-	if (skb_is_nonlinear(skb))
++	nr_frags = skb_shinfo(skb)->nr_frags;
++
++	/* Check if the linear data is 16 byte aligned */
++	if ((uintptr_t)skb->data % 16)
+ 		return true;
+ 
+-	/* Check if the headroom crosses a boundary */
+-	if (HAS_DMA_ISSUE(skb->head, skb_headroom(skb)))
++	/* Check if the needed headroom crosses a 4K address boundary without
++	 * being 256 byte aligned
++	 */
++	if (CROSS_4K(skb->data - priv->tx_headroom, priv->tx_headroom) &&
++	    (((uintptr_t)skb->data - priv->tx_headroom) % 256))
+ 		return true;
+ 
+-	/* Check if the non-paged data crosses a boundary */
+-	if (HAS_DMA_ISSUE(skb->data, skb_headlen(skb)))
++	/* Check if the linear data crosses a 4K address boundary without
++	 * being 256 byte aligned
++	 */
++	if (CROSS_4K(skb->data, skb_headlen(skb)) &&
++	    ((uintptr_t)skb->data % 256))
+ 		return true;
+ 
+-	/* Check if the entire linear skb crosses a boundary */
+-	if (HAS_DMA_ISSUE(skb->head, skb_end_offset(skb)))
++	/* When using Scatter/Gather, the linear data becomes the first
++	 * fragment in the list and must follow the same restrictions as the
++	 * other fragments.
++	 *
++	 * Check if the linear data is mod 16 bytes in size.
++	 */
++	if (nr_frags && (skb_headlen(skb) % 16))
+ 		return true;
+ 
++	/* Check the SG fragments. They must follow the same rules as the
++	 * linear data with and additional restriction: they must be multiple
++	 * of 16 bytes in size to account for the hardware carryover effect.
++	 */
++	for (i = 0; i < nr_frags; i++) {
++		frag = &skb_shinfo(skb)->frags[i];
++
++		/* Check if the fragment is a multiple of 16 bytes in size.
++		 * The last fragment is exempt from this restriction.
++		 */
++		if ((i != (nr_frags - 1)) && (skb_frag_size(frag) % 16))
++			return true;
++
++		/* Check if the fragment is 16 byte aligned */
++		if (skb_frag_off(frag) % 16)
++			return true;
++
++		/* Check if the fragment crosses a 4K address boundary. Since
++		 * the alignment of previous fragments can influence the
++		 * current fragment, checking for the 256 byte alignment
++		 * isn't relevant.
++		 */
++		if (CROSS_4K(skb_frag_off(frag), skb_frag_size(frag)))
++			return true;
++	}
++
+ 	return false;
+ }
+ 
+@@ -1062,10 +1103,24 @@ int __hot dpa_tx_extended(struct sk_buff
+ 	struct dpa_percpu_priv_s *percpu_priv;
+ 	struct rtnl_link_stats64 *percpu_stats;
+ 	int err = 0;
+-	bool nonlinear;
++	bool nonlinear, skb_changed, skb_need_wa;
+ 	int *countptr, offset = 0;
+ 	struct sk_buff *nskb;
+ 
++	/* Flags to help optimize the A010022 errata restriction checks.
++	 *
++	 * First flag marks if the skb changed between the first A010022 check
++	 * and the moment it's converted to an FD.
++	 *
++	 * The second flag marks if the skb needs to be realigned in order to
++	 * avoid the errata.
++	 *
++	 * The flags should have minimal impact on platforms not impacted by
++	 * the errata.
++	 */
++	skb_changed = false;
++	skb_need_wa = false;
++
+ 	priv = netdev_priv(net_dev);
+ 	/* Non-migratable context, safe to use raw_cpu_ptr */
+ 	percpu_priv = raw_cpu_ptr(priv->percpu_priv);
+@@ -1075,13 +1130,8 @@ int __hot dpa_tx_extended(struct sk_buff
+ 	clear_fd(&fd);
+ 
+ #ifndef CONFIG_PPC
+-	if (unlikely(dpaa_errata_a010022) && a010022_check_skb(skb, priv)) {
+-		nskb = a010022_realign_skb(skb, priv);
+-		if (!nskb)
+-			goto skb_to_fd_failed;
+-		dev_kfree_skb(skb);
+-		skb = nskb;
+-	}
++	if (unlikely(dpaa_errata_a010022) && a010022_check_skb(skb, priv))
++		skb_need_wa = true;
+ #endif
+ 
+ 	nonlinear = skb_is_nonlinear(skb);
+@@ -1102,8 +1152,8 @@ int __hot dpa_tx_extended(struct sk_buff
+ 	 * Btw, we're using the first sgt entry to store the linear part of
+ 	 * the skb, so we're one extra frag short.
+ 	 */
+-	if (nonlinear &&
+-		likely(skb_shinfo(skb)->nr_frags < DPA_SGT_MAX_ENTRIES)) {
++	if (nonlinear && !skb_need_wa &&
++	    likely(skb_shinfo(skb)->nr_frags < DPA_SGT_MAX_ENTRIES)) {
+ 		/* Just create a S/G fd based on the skb */
+ 		err = skb_to_sg_fd(priv, skb, &fd);
+ 		percpu_priv->tx_frag_skbuffs++;
+@@ -1128,39 +1178,62 @@ int __hot dpa_tx_extended(struct sk_buff
+ 
+ 			dev_kfree_skb(skb);
+ 			skb = skb_new;
++			skb_changed = true;
+ 		}
+ 
+ 		/* We're going to store the skb backpointer at the beginning
+ 		 * of the data buffer, so we need a privately owned skb
++		 *
++		 * Under the A010022 errata, we are going to have a privately
++		 * owned skb after realigning the current one, so no point in
++		 * copying it here in that case.
+ 		 */
+ 
+ 		/* Code borrowed from skb_unshare(). */
+-		if (skb_cloned(skb)) {
++		if (skb_cloned(skb) && !skb_need_wa) {
+ 			nskb = skb_copy(skb, GFP_ATOMIC);
+ 			kfree_skb(skb);
+ 			skb = nskb;
+-#ifndef CONFIG_PPC
+-			if (unlikely(dpaa_errata_a010022) &&
+-			    a010022_check_skb(skb, priv)) {
+-				nskb = a010022_realign_skb(skb, priv);
+-				if (!nskb)
+-					goto skb_to_fd_failed;
+-				dev_kfree_skb(skb);
+-				skb = nskb;
+-			}
+-#endif
++			skb_changed = true;
++
+ 			/* skb_copy() has now linearized the skbuff. */
+-		} else if (unlikely(nonlinear)) {
++		} else if (unlikely(nonlinear) && !skb_need_wa) {
+ 			/* We are here because the egress skb contains
+ 			 * more fragments than we support. In this case,
+ 			 * we have no choice but to linearize it ourselves.
+ 			 */
+-			err = __skb_linearize(skb);
++#ifndef CONFIG_PPC
++			/* No point in linearizing the skb now if we are going
++			 * to realign and linearize it again further down due
++			 * to the A010022 errata
++			 */
++			if (unlikely(dpaa_errata_a010022))
++				skb_need_wa = true;
++			else
++#endif
++				err = __skb_linearize(skb);
+ 		}
+ 		if (unlikely(!skb || err < 0))
+ 			/* Common out-of-memory error path */
+ 			goto enomem;
+ 
++#ifndef CONFIG_PPC
++		/* Verify the skb a second time if it has been updated since
++		 * the previous check
++		 */
++		if (unlikely(dpaa_errata_a010022) && skb_changed &&
++		    a010022_check_skb(skb, priv))
++			skb_need_wa = true;
++
++		if (unlikely(dpaa_errata_a010022) && skb_need_wa) {
++			nskb = a010022_realign_skb(skb, priv);
++			if (!nskb)
++				goto skb_to_fd_failed;
++			dev_kfree_skb(skb);
++			skb = nskb;
++		}
++#endif
++
+ 		err = skb_to_contig_fd(priv, skb, &fd, countptr, &offset);
+ 	}
+ 	if (unlikely(err < 0))