[PATCH net v1] net: stmmac: TSO: Fix unbalanced DMA map/unmap for non-paged SKB data

Jon Hunter jonathanh at nvidia.com
Thu Dec 5 02:57:59 PST 2024


On 04/12/2024 18:18, Thierry Reding wrote:
> On Wed, Dec 04, 2024 at 05:45:43PM +0000, Russell King (Oracle) wrote:
>> On Wed, Dec 04, 2024 at 05:02:19PM +0000, Jon Hunter wrote:
>>> Hi Russell,
>>>
>>> On 04/12/2024 16:39, Russell King (Oracle) wrote:
>>>> On Wed, Dec 04, 2024 at 04:58:34PM +0100, Thierry Reding wrote:
>>>>> This doesn't match the location from earlier, but at least there's
>>>>> something afoot here that needs fixing. I suppose this could simply be
>>>>> hiding any subsequent errors, so once this is fixed we might see other
>>>>> similar issues.
>>>>
>>>> Well, having a quick look at this, the first thing which stands out is:
>>>>
>>>> In stmmac_tx_clean(), we have:
>>>>
>>>>                   if (likely(tx_q->tx_skbuff_dma[entry].buf &&
>>>>                              tx_q->tx_skbuff_dma[entry].buf_type != STMMAC_TXBUF_T
>>>> _XDP_TX)) {
>>>>                           if (tx_q->tx_skbuff_dma[entry].map_as_page)
>>>>                                   dma_unmap_page(priv->device,
>>>>                                                  tx_q->tx_skbuff_dma[entry].buf,
>>>>                                                  tx_q->tx_skbuff_dma[entry].len,
>>>>                                                  DMA_TO_DEVICE);
>>>>                           else
>>>>                                   dma_unmap_single(priv->device,
>>>>                                                    tx_q->tx_skbuff_dma[entry].buf,
>>>>                                                    tx_q->tx_skbuff_dma[entry].len,
>>>>                                                    DMA_TO_DEVICE);
>>>>                           tx_q->tx_skbuff_dma[entry].buf = 0;
>>>>                           tx_q->tx_skbuff_dma[entry].len = 0;
>>>>                           tx_q->tx_skbuff_dma[entry].map_as_page = false;
>>>>                   }
>>>>
>>>> So, tx_skbuff_dma[entry].buf is expected to point appropriately to the
>>>> DMA region.
>>>>
>>>> Now if we look at stmmac_tso_xmit():
>>>>
>>>>           des = dma_map_single(priv->device, skb->data, skb_headlen(skb),
>>>>                                DMA_TO_DEVICE);
>>>>           if (dma_mapping_error(priv->device, des))
>>>>                   goto dma_map_err;
>>>>
>>>>           if (priv->dma_cap.addr64 <= 32) {
>>>> ...
>>>>           } else {
>>>> ...
>>>>                   des += proto_hdr_len;
>>>> ...
>>>> 	}
>>>>
>>>>           tx_q->tx_skbuff_dma[tx_q->cur_tx].buf = des;
>>>>           tx_q->tx_skbuff_dma[tx_q->cur_tx].len = skb_headlen(skb);
>>>>           tx_q->tx_skbuff_dma[tx_q->cur_tx].map_as_page = false;
>>>>           tx_q->tx_skbuff_dma[tx_q->cur_tx].buf_type = STMMAC_TXBUF_T_SKB;
>>>>
>>>> This will result in stmmac_tx_clean() calling dma_unmap_single() using
>>>> "des" and "skb_headlen(skb)" as the buffer start and length.
>>>>
>>>> One of the requirements of the DMA mapping API is that the DMA handle
>>>> returned by the map operation will be passed into the unmap function.
>>>> Not something that was offset. The length will also be the same.
>>>>
>>>> We can clearly see above that there is a case where the DMA handle has
>>>> been offset by proto_hdr_len, and when this is so, the value that is
>>>> passed into the unmap operation no longer matches this requirement.
>>>>
>>>> So, a question to the reporter - what is the value of
>>>> priv->dma_cap.addr64 in your failing case? You should see the value
>>>> in the "Using %d/%d bits DMA host/device width" kernel message.
>>>
>>> It is ...
>>>
>>>   dwc-eth-dwmac 2490000.ethernet: Using 40/40 bits DMA host/device width
>>
>> So yes, "des" is being offset, which will upset the unmap operation.
>> Please try the following patch, thanks:
>>
>> diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
>> index 9b262cdad60b..c81ea8cdfe6e 100644
>> --- a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
>> +++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
>> @@ -4192,8 +4192,8 @@ static netdev_tx_t stmmac_tso_xmit(struct sk_buff *skb, struct net_device *dev)
>>   	struct stmmac_txq_stats *txq_stats;
>>   	struct stmmac_tx_queue *tx_q;
>>   	u32 pay_len, mss, queue;
>> +	dma_addr_t tso_des, des;
>>   	u8 proto_hdr_len, hdr;
>> -	dma_addr_t des;
>>   	bool set_ic;
>>   	int i;
>>   
>> @@ -4289,14 +4289,15 @@ static netdev_tx_t stmmac_tso_xmit(struct sk_buff *skb, struct net_device *dev)
>>   
>>   		/* If needed take extra descriptors to fill the remaining payload */
>>   		tmp_pay_len = pay_len - TSO_MAX_BUFF_SIZE;
>> +		tso_des = des;
>>   	} else {
>>   		stmmac_set_desc_addr(priv, first, des);
>>   		tmp_pay_len = pay_len;
>> -		des += proto_hdr_len;
>> +		tso_des = des + proto_hdr_len;
>>   		pay_len = 0;
>>   	}
>>   
>> -	stmmac_tso_allocator(priv, des, tmp_pay_len, (nfrags == 0), queue);
>> +	stmmac_tso_allocator(priv, tso_des, tmp_pay_len, (nfrags == 0), queue);
>>   
>>   	/* In case two or more DMA transmit descriptors are allocated for this
>>   	 * non-paged SKB data, the DMA buffer address should be saved to
> 
> I see, that makes sense. Looks like this has been broken for a few years
> (since commit 34c15202896d ("net: stmmac: Fix the problem of tso_xmit"))
> and Furong's patch ended up exposing it.
> 
> Anyway, this seems to fix it for me. I can usually trigger the issue
> within one or two iperf runs, with your patch I haven't seen it break
> after a dozen or so runs.
> 
> It may be good to have Jon's test results as well, but looks good so
> far.


I have been running tests on my side and so far so good too. I have not 
seen any more mapping failure cases.

Russell, if you are planning to send a fix for this, please add my ...

Tested-by: Jon Hunter <jonathanh at nvidia.com>

Thanks!
Jon

-- 
nvpublic




More information about the linux-arm-kernel mailing list