|
@@ -18,8 +18,8 @@
|
|
|
|
|
|
#if defined(CONFIG_MMU) && !defined(CONFIG_COLDFIRE)
|
|
#if defined(CONFIG_MMU) && !defined(CONFIG_COLDFIRE)
|
|
|
|
|
|
-void *dma_alloc_coherent(struct device *dev, size_t size,
|
|
|
|
- dma_addr_t *handle, gfp_t flag)
|
|
|
|
|
|
+static void *m68k_dma_alloc(struct device *dev, size_t size, dma_addr_t *handle,
|
|
|
|
+ gfp_t flag, struct dma_attrs *attrs)
|
|
{
|
|
{
|
|
struct page *page, **map;
|
|
struct page *page, **map;
|
|
pgprot_t pgprot;
|
|
pgprot_t pgprot;
|
|
@@ -61,8 +61,8 @@ void *dma_alloc_coherent(struct device *dev, size_t size,
|
|
return addr;
|
|
return addr;
|
|
}
|
|
}
|
|
|
|
|
|
-void dma_free_coherent(struct device *dev, size_t size,
|
|
|
|
- void *addr, dma_addr_t handle)
|
|
|
|
|
|
+static void m68k_dma_free(struct device *dev, size_t size, void *addr,
|
|
|
|
+ dma_addr_t handle, struct dma_attrs *attrs)
|
|
{
|
|
{
|
|
pr_debug("dma_free_coherent: %p, %x\n", addr, handle);
|
|
pr_debug("dma_free_coherent: %p, %x\n", addr, handle);
|
|
vfree(addr);
|
|
vfree(addr);
|
|
@@ -72,8 +72,8 @@ void dma_free_coherent(struct device *dev, size_t size,
|
|
|
|
|
|
#include <asm/cacheflush.h>
|
|
#include <asm/cacheflush.h>
|
|
|
|
|
|
-void *dma_alloc_coherent(struct device *dev, size_t size,
|
|
|
|
- dma_addr_t *dma_handle, gfp_t gfp)
|
|
|
|
|
|
+static void *m68k_dma_alloc(struct device *dev, size_t size,
|
|
|
|
+ dma_addr_t *dma_handle, gfp_t gfp, struct dma_attrs *attrs)
|
|
{
|
|
{
|
|
void *ret;
|
|
void *ret;
|
|
/* ignore region specifiers */
|
|
/* ignore region specifiers */
|
|
@@ -90,19 +90,16 @@ void *dma_alloc_coherent(struct device *dev, size_t size,
|
|
return ret;
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
|
|
-void dma_free_coherent(struct device *dev, size_t size,
|
|
|
|
- void *vaddr, dma_addr_t dma_handle)
|
|
|
|
|
|
+static void m68k_dma_free(struct device *dev, size_t size, void *vaddr,
|
|
|
|
+ dma_addr_t dma_handle, struct dma_attrs *attrs)
|
|
{
|
|
{
|
|
free_pages((unsigned long)vaddr, get_order(size));
|
|
free_pages((unsigned long)vaddr, get_order(size));
|
|
}
|
|
}
|
|
|
|
|
|
#endif /* CONFIG_MMU && !CONFIG_COLDFIRE */
|
|
#endif /* CONFIG_MMU && !CONFIG_COLDFIRE */
|
|
|
|
|
|
-EXPORT_SYMBOL(dma_alloc_coherent);
|
|
|
|
-EXPORT_SYMBOL(dma_free_coherent);
|
|
|
|
-
|
|
|
|
-void dma_sync_single_for_device(struct device *dev, dma_addr_t handle,
|
|
|
|
- size_t size, enum dma_data_direction dir)
|
|
|
|
|
|
+static void m68k_dma_sync_single_for_device(struct device *dev,
|
|
|
|
+ dma_addr_t handle, size_t size, enum dma_data_direction dir)
|
|
{
|
|
{
|
|
switch (dir) {
|
|
switch (dir) {
|
|
case DMA_BIDIRECTIONAL:
|
|
case DMA_BIDIRECTIONAL:
|
|
@@ -118,10 +115,9 @@ void dma_sync_single_for_device(struct device *dev, dma_addr_t handle,
|
|
break;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
-EXPORT_SYMBOL(dma_sync_single_for_device);
|
|
|
|
|
|
|
|
-void dma_sync_sg_for_device(struct device *dev, struct scatterlist *sglist,
|
|
|
|
- int nents, enum dma_data_direction dir)
|
|
|
|
|
|
+static void m68k_dma_sync_sg_for_device(struct device *dev,
|
|
|
|
+ struct scatterlist *sglist, int nents, enum dma_data_direction dir)
|
|
{
|
|
{
|
|
int i;
|
|
int i;
|
|
struct scatterlist *sg;
|
|
struct scatterlist *sg;
|
|
@@ -131,31 +127,19 @@ void dma_sync_sg_for_device(struct device *dev, struct scatterlist *sglist,
|
|
dir);
|
|
dir);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
-EXPORT_SYMBOL(dma_sync_sg_for_device);
|
|
|
|
-
|
|
|
|
-dma_addr_t dma_map_single(struct device *dev, void *addr, size_t size,
|
|
|
|
- enum dma_data_direction dir)
|
|
|
|
-{
|
|
|
|
- dma_addr_t handle = virt_to_bus(addr);
|
|
|
|
-
|
|
|
|
- dma_sync_single_for_device(dev, handle, size, dir);
|
|
|
|
- return handle;
|
|
|
|
-}
|
|
|
|
-EXPORT_SYMBOL(dma_map_single);
|
|
|
|
|
|
|
|
-dma_addr_t dma_map_page(struct device *dev, struct page *page,
|
|
|
|
- unsigned long offset, size_t size,
|
|
|
|
- enum dma_data_direction dir)
|
|
|
|
|
|
+static dma_addr_t m68k_dma_map_page(struct device *dev, struct page *page,
|
|
|
|
+ unsigned long offset, size_t size, enum dma_data_direction dir,
|
|
|
|
+ struct dma_attrs *attrs)
|
|
{
|
|
{
|
|
dma_addr_t handle = page_to_phys(page) + offset;
|
|
dma_addr_t handle = page_to_phys(page) + offset;
|
|
|
|
|
|
dma_sync_single_for_device(dev, handle, size, dir);
|
|
dma_sync_single_for_device(dev, handle, size, dir);
|
|
return handle;
|
|
return handle;
|
|
}
|
|
}
|
|
-EXPORT_SYMBOL(dma_map_page);
|
|
|
|
|
|
|
|
-int dma_map_sg(struct device *dev, struct scatterlist *sglist, int nents,
|
|
|
|
- enum dma_data_direction dir)
|
|
|
|
|
|
+static int m68k_dma_map_sg(struct device *dev, struct scatterlist *sglist,
|
|
|
|
+ int nents, enum dma_data_direction dir, struct dma_attrs *attrs)
|
|
{
|
|
{
|
|
int i;
|
|
int i;
|
|
struct scatterlist *sg;
|
|
struct scatterlist *sg;
|
|
@@ -167,4 +151,13 @@ int dma_map_sg(struct device *dev, struct scatterlist *sglist, int nents,
|
|
}
|
|
}
|
|
return nents;
|
|
return nents;
|
|
}
|
|
}
|
|
-EXPORT_SYMBOL(dma_map_sg);
|
|
|
|
|
|
+
|
|
|
|
+struct dma_map_ops m68k_dma_ops = {
|
|
|
|
+ .alloc = m68k_dma_alloc,
|
|
|
|
+ .free = m68k_dma_free,
|
|
|
|
+ .map_page = m68k_dma_map_page,
|
|
|
|
+ .map_sg = m68k_dma_map_sg,
|
|
|
|
+ .sync_single_for_device = m68k_dma_sync_single_for_device,
|
|
|
|
+ .sync_sg_for_device = m68k_dma_sync_sg_for_device,
|
|
|
|
+};
|
|
|
|
+EXPORT_SYMBOL(m68k_dma_ops);
|