ASR_BASE

Change-Id: Icf3719cc0afe3eeb3edc7fa80a2eb5199ca9dda1
diff --git a/target/linux/layerscape/patches-5.4/701-net-0357-sdk_dpaa-ls1043a-errata-memory-related-fixes.patch b/target/linux/layerscape/patches-5.4/701-net-0357-sdk_dpaa-ls1043a-errata-memory-related-fixes.patch
new file mode 100644
index 0000000..6daee46
--- /dev/null
+++ b/target/linux/layerscape/patches-5.4/701-net-0357-sdk_dpaa-ls1043a-errata-memory-related-fixes.patch
@@ -0,0 +1,107 @@
+From 26d3cc476c26832e1e05db182ac27906f6c81f2d Mon Sep 17 00:00:00 2001
+From: Camelia Groza <camelia.groza@nxp.com>
+Date: Tue, 29 Oct 2019 16:12:18 +0200
+Subject: [PATCH] sdk_dpaa: ls1043a errata: memory related fixes
+
+Avoid a crash by verifying the allocation return status.
+
+Use the standard API for determining the page order needed for
+allocating Jumbo sized skbs.
+
+Explicitly remove the old skb outside the w/a, for both successful and
+unsuccessful realignments. Make sure the old skb's memory isn't leaked.
+
+Signed-off-by: Camelia Groza <camelia.groza@nxp.com>
+---
+ .../net/ethernet/freescale/sdk_dpaa/dpaa_eth_sg.c  | 30 ++++++++++++++--------
+ 1 file changed, 19 insertions(+), 11 deletions(-)
+
+--- a/drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth_sg.c
++++ b/drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth_sg.c
+@@ -809,8 +809,8 @@ static struct sk_buff *a010022_realign_s
+ {
+ 	int trans_offset = skb_transport_offset(skb);
+ 	int net_offset = skb_network_offset(skb);
+-	int nsize, headroom, npage_order;
+ 	struct sk_buff *nskb = NULL;
++	int nsize, headroom;
+ 	struct page *npage;
+ 	void *npage_addr;
+ 
+@@ -825,8 +825,7 @@ static struct sk_buff *a010022_realign_s
+ 		SKB_DATA_ALIGN(sizeof(struct skb_shared_info));
+ 
+ 	/* Reserve enough memory to accommodate Jumbo frames */
+-	npage_order = (nsize - 1) / PAGE_SIZE;
+-	npage = alloc_pages(GFP_ATOMIC | __GFP_COMP, npage_order);
++	npage = alloc_pages(GFP_ATOMIC | __GFP_COMP, get_order(nsize));
+ 	if (unlikely(!npage)) {
+ 		WARN_ONCE(1, "Memory allocation failure\n");
+ 		return NULL;
+@@ -869,7 +868,6 @@ static struct sk_buff *a010022_realign_s
+ 	/* We don't want the buffer to be recycled so we mark it accordingly */
+ 	nskb->mark = NONREC_MARK;
+ 
+-	dev_kfree_skb(skb);
+ 	return nskb;
+ 
+ err:
+@@ -911,8 +909,13 @@ int __hot skb_to_sg_fd(struct dpa_priv_s
+ 	 * is in place and we need to avoid crossing a 4k boundary.
+ 	 */
+ #ifndef CONFIG_PPC
+-	if (unlikely(dpaa_errata_a010022))
+-		sgt_buf = page_address(alloc_page(GFP_ATOMIC));
++	if (unlikely(dpaa_errata_a010022)) {
++		struct page *new_page = alloc_page(GFP_ATOMIC);
++
++		if (unlikely(!new_page))
++			return -ENOMEM;
++		sgt_buf = page_address(new_page);
++	}
+ 	else
+ #endif
+ 		sgt_buf = netdev_alloc_frag(priv->tx_headroom + sgt_size);
+@@ -1061,6 +1064,7 @@ int __hot dpa_tx_extended(struct sk_buff
+ 	int err = 0;
+ 	bool nonlinear;
+ 	int *countptr, offset = 0;
++	struct sk_buff *nskb;
+ 
+ 	priv = netdev_priv(net_dev);
+ 	/* Non-migratable context, safe to use raw_cpu_ptr */
+@@ -1072,9 +1076,11 @@ int __hot dpa_tx_extended(struct sk_buff
+ 
+ #ifndef CONFIG_PPC
+ 	if (unlikely(dpaa_errata_a010022) && a010022_check_skb(skb, priv)) {
+-		skb = a010022_realign_skb(skb, priv);
+-		if (!skb)
++		nskb = a010022_realign_skb(skb, priv);
++		if (!nskb)
+ 			goto skb_to_fd_failed;
++		dev_kfree_skb(skb);
++		skb = nskb;
+ 	}
+ #endif
+ 
+@@ -1130,15 +1136,17 @@ int __hot dpa_tx_extended(struct sk_buff
+ 
+ 		/* Code borrowed from skb_unshare(). */
+ 		if (skb_cloned(skb)) {
+-			struct sk_buff *nskb = skb_copy(skb, GFP_ATOMIC);
++			nskb = skb_copy(skb, GFP_ATOMIC);
+ 			kfree_skb(skb);
+ 			skb = nskb;
+ #ifndef CONFIG_PPC
+ 			if (unlikely(dpaa_errata_a010022) &&
+ 			    a010022_check_skb(skb, priv)) {
+-				skb = a010022_realign_skb(skb, priv);
+-				if (!skb)
++				nskb = a010022_realign_skb(skb, priv);
++				if (!nskb)
+ 					goto skb_to_fd_failed;
++				dev_kfree_skb(skb);
++				skb = nskb;
+ 			}
+ #endif
+ 			/* skb_copy() has now linearized the skbuff. */