|
@@ -313,8 +313,7 @@ static void txstate(struct musb *musb, struct musb_request *req)
|
|
|
|
|
|
/* MUSB_TXCSR_P_ISO is still set correctly */
|
|
|
|
|
|
-#if defined(CONFIG_USB_INVENTRA_DMA) || defined(CONFIG_USB_UX500_DMA)
|
|
|
- {
|
|
|
+ if (musb_dma_inventra(musb) || musb_dma_ux500(musb)) {
|
|
|
if (request_size < musb_ep->packet_sz)
|
|
|
musb_ep->dma->desired_mode = 0;
|
|
|
else
|
|
@@ -365,7 +364,6 @@ static void txstate(struct musb *musb, struct musb_request *req)
|
|
|
}
|
|
|
}
|
|
|
|
|
|
-#endif
|
|
|
if (is_cppi_enabled(musb)) {
|
|
|
/* program endpoint CSR first, then setup DMA */
|
|
|
csr &= ~(MUSB_TXCSR_P_UNDERRUN | MUSB_TXCSR_TXPKTRDY);
|
|
@@ -641,8 +639,10 @@ static void rxstate(struct musb *musb, struct musb_request *req)
|
|
|
use_mode_1 = 0;
|
|
|
|
|
|
if (request->actual < request->length) {
|
|
|
-#ifdef CONFIG_USB_INVENTRA_DMA
|
|
|
- if (is_buffer_mapped(req)) {
|
|
|
+ if (!is_buffer_mapped(req))
|
|
|
+ goto buffer_aint_mapped;
|
|
|
+
|
|
|
+ if (musb_dma_inventra(musb)) {
|
|
|
struct dma_controller *c;
|
|
|
struct dma_channel *channel;
|
|
|
int use_dma = 0;
|
|
@@ -716,8 +716,8 @@ static void rxstate(struct musb *musb, struct musb_request *req)
|
|
|
if (use_dma)
|
|
|
return;
|
|
|
}
|
|
|
-#elif defined(CONFIG_USB_UX500_DMA)
|
|
|
- if ((is_buffer_mapped(req)) &&
|
|
|
+
|
|
|
+ if ((musb_dma_ux500(musb)) &&
|
|
|
(request->actual < request->length)) {
|
|
|
|
|
|
struct dma_controller *c;
|
|
@@ -765,7 +765,6 @@ static void rxstate(struct musb *musb, struct musb_request *req)
|
|
|
|
|
|
return;
|
|
|
}
|
|
|
-#endif /* Mentor's DMA */
|
|
|
|
|
|
len = request->length - request->actual;
|
|
|
dev_dbg(musb->controller, "%s OUT/RX pio fifo %d/%d, maxpacket %d\n",
|
|
@@ -775,8 +774,7 @@ static void rxstate(struct musb *musb, struct musb_request *req)
|
|
|
|
|
|
fifo_count = min_t(unsigned, len, fifo_count);
|
|
|
|
|
|
-#ifdef CONFIG_USB_TUSB_OMAP_DMA
|
|
|
- if (tusb_dma_omap(musb) && is_buffer_mapped(req)) {
|
|
|
+ if (tusb_dma_omap(musb)) {
|
|
|
struct dma_controller *c = musb->dma_controller;
|
|
|
struct dma_channel *channel = musb_ep->dma;
|
|
|
u32 dma_addr = request->dma + request->actual;
|
|
@@ -790,23 +788,22 @@ static void rxstate(struct musb *musb, struct musb_request *req)
|
|
|
if (ret)
|
|
|
return;
|
|
|
}
|
|
|
-#endif
|
|
|
+
|
|
|
/*
|
|
|
* Unmap the dma buffer back to cpu if dma channel
|
|
|
* programming fails. This buffer is mapped if the
|
|
|
* channel allocation is successful
|
|
|
*/
|
|
|
- if (is_buffer_mapped(req)) {
|
|
|
- unmap_dma_buffer(req, musb);
|
|
|
-
|
|
|
- /*
|
|
|
- * Clear DMAENAB and AUTOCLEAR for the
|
|
|
- * PIO mode transfer
|
|
|
- */
|
|
|
- csr &= ~(MUSB_RXCSR_DMAENAB | MUSB_RXCSR_AUTOCLEAR);
|
|
|
- musb_writew(epio, MUSB_RXCSR, csr);
|
|
|
- }
|
|
|
+ unmap_dma_buffer(req, musb);
|
|
|
+
|
|
|
+ /*
|
|
|
+ * Clear DMAENAB and AUTOCLEAR for the
|
|
|
+ * PIO mode transfer
|
|
|
+ */
|
|
|
+ csr &= ~(MUSB_RXCSR_DMAENAB | MUSB_RXCSR_AUTOCLEAR);
|
|
|
+ musb_writew(epio, MUSB_RXCSR, csr);
|
|
|
|
|
|
+buffer_aint_mapped:
|
|
|
musb_read_fifo(musb_ep->hw_ep, fifo_count, (u8 *)
|
|
|
(request->buf + request->actual));
|
|
|
request->actual += fifo_count;
|