|
@@ -234,7 +234,7 @@ int amdgpu_gart_unbind(struct amdgpu_device *adev, uint64_t offset,
|
|
|
}
|
|
|
|
|
|
t = offset / AMDGPU_GPU_PAGE_SIZE;
|
|
|
- p = t / (PAGE_SIZE / AMDGPU_GPU_PAGE_SIZE);
|
|
|
+ p = t / AMDGPU_GPU_PAGES_IN_CPU_PAGE;
|
|
|
for (i = 0; i < pages; i++, p++) {
|
|
|
#ifdef CONFIG_DRM_AMDGPU_GART_DEBUGFS
|
|
|
adev->gart.pages[p] = NULL;
|
|
@@ -243,7 +243,7 @@ int amdgpu_gart_unbind(struct amdgpu_device *adev, uint64_t offset,
|
|
|
if (!adev->gart.ptr)
|
|
|
continue;
|
|
|
|
|
|
- for (j = 0; j < (PAGE_SIZE / AMDGPU_GPU_PAGE_SIZE); j++, t++) {
|
|
|
+ for (j = 0; j < AMDGPU_GPU_PAGES_IN_CPU_PAGE; j++, t++) {
|
|
|
amdgpu_gmc_set_pte_pde(adev, adev->gart.ptr,
|
|
|
t, page_base, flags);
|
|
|
page_base += AMDGPU_GPU_PAGE_SIZE;
|
|
@@ -282,7 +282,7 @@ int amdgpu_gart_map(struct amdgpu_device *adev, uint64_t offset,
|
|
|
|
|
|
for (i = 0; i < pages; i++) {
|
|
|
page_base = dma_addr[i];
|
|
|
- for (j = 0; j < (PAGE_SIZE / AMDGPU_GPU_PAGE_SIZE); j++, t++) {
|
|
|
+ for (j = 0; j < AMDGPU_GPU_PAGES_IN_CPU_PAGE; j++, t++) {
|
|
|
amdgpu_gmc_set_pte_pde(adev, dst, t, page_base, flags);
|
|
|
page_base += AMDGPU_GPU_PAGE_SIZE;
|
|
|
}
|
|
@@ -319,7 +319,7 @@ int amdgpu_gart_bind(struct amdgpu_device *adev, uint64_t offset,
|
|
|
|
|
|
#ifdef CONFIG_DRM_AMDGPU_GART_DEBUGFS
|
|
|
t = offset / AMDGPU_GPU_PAGE_SIZE;
|
|
|
- p = t / (PAGE_SIZE / AMDGPU_GPU_PAGE_SIZE);
|
|
|
+ p = t / AMDGPU_GPU_PAGES_IN_CPU_PAGE;
|
|
|
for (i = 0; i < pages; i++, p++)
|
|
|
adev->gart.pages[p] = pagelist ? pagelist[i] : NULL;
|
|
|
#endif
|