|
@@ -18,7 +18,7 @@
|
|
|
|
|
|
static int forbid_dac __read_mostly;
|
|
|
|
|
|
-const struct dma_map_ops *dma_ops = &nommu_dma_ops;
|
|
|
+const struct dma_map_ops *dma_ops = &dma_direct_ops;
|
|
|
EXPORT_SYMBOL(dma_ops);
|
|
|
|
|
|
static int iommu_sac_force __read_mostly;
|
|
@@ -76,60 +76,6 @@ void __init pci_iommu_alloc(void)
|
|
|
}
|
|
|
}
|
|
|
}
|
|
|
-void *dma_generic_alloc_coherent(struct device *dev, size_t size,
|
|
|
- dma_addr_t *dma_addr, gfp_t flag,
|
|
|
- unsigned long attrs)
|
|
|
-{
|
|
|
- struct page *page;
|
|
|
- unsigned int count = PAGE_ALIGN(size) >> PAGE_SHIFT;
|
|
|
- dma_addr_t addr;
|
|
|
-
|
|
|
-again:
|
|
|
- page = NULL;
|
|
|
- /* CMA can be used only in the context which permits sleeping */
|
|
|
- if (gfpflags_allow_blocking(flag)) {
|
|
|
- page = dma_alloc_from_contiguous(dev, count, get_order(size),
|
|
|
- flag);
|
|
|
- if (page) {
|
|
|
- addr = phys_to_dma(dev, page_to_phys(page));
|
|
|
- if (addr + size > dev->coherent_dma_mask) {
|
|
|
- dma_release_from_contiguous(dev, page, count);
|
|
|
- page = NULL;
|
|
|
- }
|
|
|
- }
|
|
|
- }
|
|
|
- /* fallback */
|
|
|
- if (!page)
|
|
|
- page = alloc_pages_node(dev_to_node(dev), flag, get_order(size));
|
|
|
- if (!page)
|
|
|
- return NULL;
|
|
|
-
|
|
|
- addr = phys_to_dma(dev, page_to_phys(page));
|
|
|
- if (addr + size > dev->coherent_dma_mask) {
|
|
|
- __free_pages(page, get_order(size));
|
|
|
-
|
|
|
- if (dev->coherent_dma_mask < DMA_BIT_MASK(32) &&
|
|
|
- !(flag & GFP_DMA)) {
|
|
|
- flag = (flag & ~GFP_DMA32) | GFP_DMA;
|
|
|
- goto again;
|
|
|
- }
|
|
|
-
|
|
|
- return NULL;
|
|
|
- }
|
|
|
- memset(page_address(page), 0, size);
|
|
|
- *dma_addr = addr;
|
|
|
- return page_address(page);
|
|
|
-}
|
|
|
-
|
|
|
-void dma_generic_free_coherent(struct device *dev, size_t size, void *vaddr,
|
|
|
- dma_addr_t dma_addr, unsigned long attrs)
|
|
|
-{
|
|
|
- unsigned int count = PAGE_ALIGN(size) >> PAGE_SHIFT;
|
|
|
- struct page *page = virt_to_page(vaddr);
|
|
|
-
|
|
|
- if (!dma_release_from_contiguous(dev, page, count))
|
|
|
- free_pages((unsigned long)vaddr, get_order(size));
|
|
|
-}
|
|
|
|
|
|
bool arch_dma_alloc_attrs(struct device **dev, gfp_t *gfp)
|
|
|
{
|
|
@@ -243,16 +189,6 @@ int arch_dma_supported(struct device *dev, u64 mask)
|
|
|
}
|
|
|
EXPORT_SYMBOL(arch_dma_supported);
|
|
|
|
|
|
-int x86_dma_supported(struct device *dev, u64 mask)
|
|
|
-{
|
|
|
- /* Copied from i386. Doesn't make much sense, because it will
|
|
|
- only work for pci_alloc_coherent.
|
|
|
- The caller just has to use GFP_DMA in this case. */
|
|
|
- if (mask < DMA_BIT_MASK(24))
|
|
|
- return 0;
|
|
|
- return 1;
|
|
|
-}
|
|
|
-
|
|
|
static int __init pci_iommu_init(void)
|
|
|
{
|
|
|
struct iommu_table_entry *p;
|