|
@@ -159,33 +159,13 @@ static void octeon_dma_sync_sg_for_device(struct device *dev,
|
|
|
static void *octeon_dma_alloc_coherent(struct device *dev, size_t size,
|
|
|
dma_addr_t *dma_handle, gfp_t gfp, unsigned long attrs)
|
|
|
{
|
|
|
- void *ret;
|
|
|
-
|
|
|
- if (IS_ENABLED(CONFIG_ZONE_DMA) && dev == NULL)
|
|
|
- gfp |= __GFP_DMA;
|
|
|
- else if (IS_ENABLED(CONFIG_ZONE_DMA) &&
|
|
|
- dev->coherent_dma_mask <= DMA_BIT_MASK(24))
|
|
|
- gfp |= __GFP_DMA;
|
|
|
- else if (IS_ENABLED(CONFIG_ZONE_DMA32) &&
|
|
|
- dev->coherent_dma_mask <= DMA_BIT_MASK(32))
|
|
|
- gfp |= __GFP_DMA32;
|
|
|
-
|
|
|
- /* Don't invoke OOM killer */
|
|
|
- gfp |= __GFP_NORETRY;
|
|
|
-
|
|
|
- ret = swiotlb_alloc_coherent(dev, size, dma_handle, gfp);
|
|
|
+ void *ret = swiotlb_alloc(dev, size, dma_handle, gfp, attrs);
|
|
|
|
|
|
mb();
|
|
|
|
|
|
return ret;
|
|
|
}
|
|
|
|
|
|
-static void octeon_dma_free_coherent(struct device *dev, size_t size,
|
|
|
- void *vaddr, dma_addr_t dma_handle, unsigned long attrs)
|
|
|
-{
|
|
|
- swiotlb_free_coherent(dev, size, vaddr, dma_handle);
|
|
|
-}
|
|
|
-
|
|
|
static dma_addr_t octeon_unity_phys_to_dma(struct device *dev, phys_addr_t paddr)
|
|
|
{
|
|
|
return paddr;
|
|
@@ -225,7 +205,7 @@ EXPORT_SYMBOL(dma_to_phys);
|
|
|
static struct octeon_dma_map_ops octeon_linear_dma_map_ops = {
|
|
|
.dma_map_ops = {
|
|
|
.alloc = octeon_dma_alloc_coherent,
|
|
|
- .free = octeon_dma_free_coherent,
|
|
|
+ .free = swiotlb_free,
|
|
|
.map_page = octeon_dma_map_page,
|
|
|
.unmap_page = swiotlb_unmap_page,
|
|
|
.map_sg = octeon_dma_map_sg,
|
|
@@ -311,7 +291,7 @@ void __init plat_swiotlb_setup(void)
|
|
|
static struct octeon_dma_map_ops _octeon_pci_dma_map_ops = {
|
|
|
.dma_map_ops = {
|
|
|
.alloc = octeon_dma_alloc_coherent,
|
|
|
- .free = octeon_dma_free_coherent,
|
|
|
+ .free = swiotlb_free,
|
|
|
.map_page = octeon_dma_map_page,
|
|
|
.unmap_page = swiotlb_unmap_page,
|
|
|
.map_sg = octeon_dma_map_sg,
|