|
@@ -217,27 +217,29 @@ void dma_cleanup_tables(unsigned long *table)
|
|
dma_free_cpu_table(table);
|
|
dma_free_cpu_table(table);
|
|
}
|
|
}
|
|
|
|
|
|
-static unsigned long __dma_alloc_iommu(struct zpci_dev *zdev,
|
|
|
|
|
|
+static unsigned long __dma_alloc_iommu(struct device *dev,
|
|
unsigned long start, int size)
|
|
unsigned long start, int size)
|
|
{
|
|
{
|
|
|
|
+ struct zpci_dev *zdev = to_zpci(to_pci_dev(dev));
|
|
unsigned long boundary_size;
|
|
unsigned long boundary_size;
|
|
|
|
|
|
- boundary_size = ALIGN(dma_get_seg_boundary(&zdev->pdev->dev) + 1,
|
|
|
|
|
|
+ boundary_size = ALIGN(dma_get_seg_boundary(dev) + 1,
|
|
PAGE_SIZE) >> PAGE_SHIFT;
|
|
PAGE_SIZE) >> PAGE_SHIFT;
|
|
return iommu_area_alloc(zdev->iommu_bitmap, zdev->iommu_pages,
|
|
return iommu_area_alloc(zdev->iommu_bitmap, zdev->iommu_pages,
|
|
start, size, 0, boundary_size, 0);
|
|
start, size, 0, boundary_size, 0);
|
|
}
|
|
}
|
|
|
|
|
|
-static unsigned long dma_alloc_iommu(struct zpci_dev *zdev, int size)
|
|
|
|
|
|
+static unsigned long dma_alloc_iommu(struct device *dev, int size)
|
|
{
|
|
{
|
|
|
|
+ struct zpci_dev *zdev = to_zpci(to_pci_dev(dev));
|
|
unsigned long offset, flags;
|
|
unsigned long offset, flags;
|
|
int wrap = 0;
|
|
int wrap = 0;
|
|
|
|
|
|
spin_lock_irqsave(&zdev->iommu_bitmap_lock, flags);
|
|
spin_lock_irqsave(&zdev->iommu_bitmap_lock, flags);
|
|
- offset = __dma_alloc_iommu(zdev, zdev->next_bit, size);
|
|
|
|
|
|
+ offset = __dma_alloc_iommu(dev, zdev->next_bit, size);
|
|
if (offset == -1) {
|
|
if (offset == -1) {
|
|
/* wrap-around */
|
|
/* wrap-around */
|
|
- offset = __dma_alloc_iommu(zdev, 0, size);
|
|
|
|
|
|
+ offset = __dma_alloc_iommu(dev, 0, size);
|
|
wrap = 1;
|
|
wrap = 1;
|
|
}
|
|
}
|
|
|
|
|
|
@@ -251,8 +253,9 @@ static unsigned long dma_alloc_iommu(struct zpci_dev *zdev, int size)
|
|
return offset;
|
|
return offset;
|
|
}
|
|
}
|
|
|
|
|
|
-static void dma_free_iommu(struct zpci_dev *zdev, unsigned long offset, int size)
|
|
|
|
|
|
+static void dma_free_iommu(struct device *dev, unsigned long offset, int size)
|
|
{
|
|
{
|
|
|
|
+ struct zpci_dev *zdev = to_zpci(to_pci_dev(dev));
|
|
unsigned long flags;
|
|
unsigned long flags;
|
|
|
|
|
|
spin_lock_irqsave(&zdev->iommu_bitmap_lock, flags);
|
|
spin_lock_irqsave(&zdev->iommu_bitmap_lock, flags);
|
|
@@ -293,7 +296,7 @@ static dma_addr_t s390_dma_map_pages(struct device *dev, struct page *page,
|
|
|
|
|
|
/* This rounds up number of pages based on size and offset */
|
|
/* This rounds up number of pages based on size and offset */
|
|
nr_pages = iommu_num_pages(pa, size, PAGE_SIZE);
|
|
nr_pages = iommu_num_pages(pa, size, PAGE_SIZE);
|
|
- iommu_page_index = dma_alloc_iommu(zdev, nr_pages);
|
|
|
|
|
|
+ iommu_page_index = dma_alloc_iommu(dev, nr_pages);
|
|
if (iommu_page_index == -1) {
|
|
if (iommu_page_index == -1) {
|
|
ret = -ENOSPC;
|
|
ret = -ENOSPC;
|
|
goto out_err;
|
|
goto out_err;
|
|
@@ -319,7 +322,7 @@ static dma_addr_t s390_dma_map_pages(struct device *dev, struct page *page,
|
|
return dma_addr + (offset & ~PAGE_MASK);
|
|
return dma_addr + (offset & ~PAGE_MASK);
|
|
|
|
|
|
out_free:
|
|
out_free:
|
|
- dma_free_iommu(zdev, iommu_page_index, nr_pages);
|
|
|
|
|
|
+ dma_free_iommu(dev, iommu_page_index, nr_pages);
|
|
out_err:
|
|
out_err:
|
|
zpci_err("map error:\n");
|
|
zpci_err("map error:\n");
|
|
zpci_err_dma(ret, pa);
|
|
zpci_err_dma(ret, pa);
|
|
@@ -346,7 +349,7 @@ static void s390_dma_unmap_pages(struct device *dev, dma_addr_t dma_addr,
|
|
|
|
|
|
atomic64_add(npages, &zdev->unmapped_pages);
|
|
atomic64_add(npages, &zdev->unmapped_pages);
|
|
iommu_page_index = (dma_addr - zdev->start_dma) >> PAGE_SHIFT;
|
|
iommu_page_index = (dma_addr - zdev->start_dma) >> PAGE_SHIFT;
|
|
- dma_free_iommu(zdev, iommu_page_index, npages);
|
|
|
|
|
|
+ dma_free_iommu(dev, iommu_page_index, npages);
|
|
}
|
|
}
|
|
|
|
|
|
static void *s390_dma_alloc(struct device *dev, size_t size,
|
|
static void *s390_dma_alloc(struct device *dev, size_t size,
|