|
@@ -382,27 +382,20 @@ static void spi_qup_write(struct spi_qup *controller)
|
|
|
} while (remainder);
|
|
|
}
|
|
|
|
|
|
-static int spi_qup_prep_sg(struct spi_master *master, struct spi_transfer *xfer,
|
|
|
- enum dma_transfer_direction dir,
|
|
|
+static int spi_qup_prep_sg(struct spi_master *master, struct scatterlist *sgl,
|
|
|
+ unsigned int nents, enum dma_transfer_direction dir,
|
|
|
dma_async_tx_callback callback)
|
|
|
{
|
|
|
struct spi_qup *qup = spi_master_get_devdata(master);
|
|
|
unsigned long flags = DMA_PREP_INTERRUPT | DMA_PREP_FENCE;
|
|
|
struct dma_async_tx_descriptor *desc;
|
|
|
- struct scatterlist *sgl;
|
|
|
struct dma_chan *chan;
|
|
|
dma_cookie_t cookie;
|
|
|
- unsigned int nents;
|
|
|
|
|
|
- if (dir == DMA_MEM_TO_DEV) {
|
|
|
+ if (dir == DMA_MEM_TO_DEV)
|
|
|
chan = master->dma_tx;
|
|
|
- nents = xfer->tx_sg.nents;
|
|
|
- sgl = xfer->tx_sg.sgl;
|
|
|
- } else {
|
|
|
+ else
|
|
|
chan = master->dma_rx;
|
|
|
- nents = xfer->rx_sg.nents;
|
|
|
- sgl = xfer->rx_sg.sgl;
|
|
|
- }
|
|
|
|
|
|
desc = dmaengine_prep_slave_sg(chan, sgl, nents, dir, flags);
|
|
|
if (IS_ERR_OR_NULL(desc))
|
|
@@ -451,7 +444,9 @@ static int spi_qup_do_dma(struct spi_device *spi, struct spi_transfer *xfer,
|
|
|
}
|
|
|
|
|
|
if (xfer->rx_buf) {
|
|
|
- ret = spi_qup_prep_sg(master, xfer, DMA_DEV_TO_MEM, rx_done);
|
|
|
+ ret = spi_qup_prep_sg(master, xfer->rx_sg.sgl,
|
|
|
+ xfer->rx_sg.nents, DMA_DEV_TO_MEM,
|
|
|
+ rx_done);
|
|
|
if (ret)
|
|
|
return ret;
|
|
|
|
|
@@ -459,7 +454,9 @@ static int spi_qup_do_dma(struct spi_device *spi, struct spi_transfer *xfer,
|
|
|
}
|
|
|
|
|
|
if (xfer->tx_buf) {
|
|
|
- ret = spi_qup_prep_sg(master, xfer, DMA_MEM_TO_DEV, tx_done);
|
|
|
+ ret = spi_qup_prep_sg(master, xfer->tx_sg.sgl,
|
|
|
+ xfer->tx_sg.nents, DMA_MEM_TO_DEV,
|
|
|
+ tx_done);
|
|
|
if (ret)
|
|
|
return ret;
|
|
|
|