|
@@ -1449,7 +1449,7 @@ static void free_dma_rx_desc_resources(struct stmmac_priv *priv)
|
|
|
static void free_dma_tx_desc_resources(struct stmmac_priv *priv)
|
|
|
{
|
|
|
u32 tx_count = priv->plat->tx_queues_to_use;
|
|
|
- u32 queue = 0;
|
|
|
+ u32 queue;
|
|
|
|
|
|
/* Free TX queue resources */
|
|
|
for (queue = 0; queue < tx_count; queue++) {
|
|
@@ -1498,7 +1498,7 @@ static int alloc_dma_rx_desc_resources(struct stmmac_priv *priv)
|
|
|
sizeof(dma_addr_t),
|
|
|
GFP_KERNEL);
|
|
|
if (!rx_q->rx_skbuff_dma)
|
|
|
- return -ENOMEM;
|
|
|
+ goto err_dma;
|
|
|
|
|
|
rx_q->rx_skbuff = kmalloc_array(DMA_RX_SIZE,
|
|
|
sizeof(struct sk_buff *),
|
|
@@ -1561,13 +1561,13 @@ static int alloc_dma_tx_desc_resources(struct stmmac_priv *priv)
|
|
|
sizeof(*tx_q->tx_skbuff_dma),
|
|
|
GFP_KERNEL);
|
|
|
if (!tx_q->tx_skbuff_dma)
|
|
|
- return -ENOMEM;
|
|
|
+ goto err_dma;
|
|
|
|
|
|
tx_q->tx_skbuff = kmalloc_array(DMA_TX_SIZE,
|
|
|
sizeof(struct sk_buff *),
|
|
|
GFP_KERNEL);
|
|
|
if (!tx_q->tx_skbuff)
|
|
|
- goto err_dma_buffers;
|
|
|
+ goto err_dma;
|
|
|
|
|
|
if (priv->extend_desc) {
|
|
|
tx_q->dma_etx = dma_zalloc_coherent(priv->device,
|
|
@@ -1577,7 +1577,7 @@ static int alloc_dma_tx_desc_resources(struct stmmac_priv *priv)
|
|
|
&tx_q->dma_tx_phy,
|
|
|
GFP_KERNEL);
|
|
|
if (!tx_q->dma_etx)
|
|
|
- goto err_dma_buffers;
|
|
|
+ goto err_dma;
|
|
|
} else {
|
|
|
tx_q->dma_tx = dma_zalloc_coherent(priv->device,
|
|
|
DMA_TX_SIZE *
|
|
@@ -1586,13 +1586,13 @@ static int alloc_dma_tx_desc_resources(struct stmmac_priv *priv)
|
|
|
&tx_q->dma_tx_phy,
|
|
|
GFP_KERNEL);
|
|
|
if (!tx_q->dma_tx)
|
|
|
- goto err_dma_buffers;
|
|
|
+ goto err_dma;
|
|
|
}
|
|
|
}
|
|
|
|
|
|
return 0;
|
|
|
|
|
|
-err_dma_buffers:
|
|
|
+err_dma:
|
|
|
free_dma_tx_desc_resources(priv);
|
|
|
|
|
|
return ret;
|