|
@@ -438,14 +438,14 @@ static struct dma_async_tx_descriptor *edma_prep_slave_sg(
|
|
|
}
|
|
|
|
|
|
if (dev_width == DMA_SLAVE_BUSWIDTH_UNDEFINED) {
|
|
|
- dev_err(dev, "Undefined slave buswidth\n");
|
|
|
+ dev_err(dev, "%s: Undefined slave buswidth\n", __func__);
|
|
|
return NULL;
|
|
|
}
|
|
|
|
|
|
edesc = kzalloc(sizeof(*edesc) + sg_len *
|
|
|
sizeof(edesc->pset[0]), GFP_ATOMIC);
|
|
|
if (!edesc) {
|
|
|
- dev_dbg(dev, "Failed to allocate a descriptor\n");
|
|
|
+ dev_err(dev, "%s: Failed to allocate a descriptor\n", __func__);
|
|
|
return NULL;
|
|
|
}
|
|
|
|
|
@@ -461,7 +461,8 @@ static struct dma_async_tx_descriptor *edma_prep_slave_sg(
|
|
|
EDMA_SLOT_ANY);
|
|
|
if (echan->slot[i] < 0) {
|
|
|
kfree(edesc);
|
|
|
- dev_err(dev, "Failed to allocate slot\n");
|
|
|
+ dev_err(dev, "%s: Failed to allocate slot\n",
|
|
|
+ __func__);
|
|
|
return NULL;
|
|
|
}
|
|
|
}
|
|
@@ -530,7 +531,7 @@ static struct dma_async_tx_descriptor *edma_prep_dma_cyclic(
|
|
|
}
|
|
|
|
|
|
if (dev_width == DMA_SLAVE_BUSWIDTH_UNDEFINED) {
|
|
|
- dev_err(dev, "Undefined slave buswidth\n");
|
|
|
+ dev_err(dev, "%s: Undefined slave buswidth\n", __func__);
|
|
|
return NULL;
|
|
|
}
|
|
|
|
|
@@ -555,7 +556,7 @@ static struct dma_async_tx_descriptor *edma_prep_dma_cyclic(
|
|
|
edesc = kzalloc(sizeof(*edesc) + nslots *
|
|
|
sizeof(edesc->pset[0]), GFP_ATOMIC);
|
|
|
if (!edesc) {
|
|
|
- dev_dbg(dev, "Failed to allocate a descriptor\n");
|
|
|
+ dev_err(dev, "%s: Failed to allocate a descriptor\n", __func__);
|
|
|
return NULL;
|
|
|
}
|
|
|
|
|
@@ -573,7 +574,8 @@ static struct dma_async_tx_descriptor *edma_prep_dma_cyclic(
|
|
|
EDMA_SLOT_ANY);
|
|
|
if (echan->slot[i] < 0) {
|
|
|
kfree(edesc);
|
|
|
- dev_err(dev, "Failed to allocate slot\n");
|
|
|
+ dev_err(dev, "%s: Failed to allocate slot\n",
|
|
|
+ __func__);
|
|
|
return NULL;
|
|
|
}
|
|
|
}
|