|
@@ -627,9 +627,14 @@ static int amdgpu_vm_update_ptes(struct amdgpu_device *adev,
|
|
|
{
|
|
|
uint64_t mask = AMDGPU_VM_PTE_COUNT - 1;
|
|
|
uint64_t last_pte = ~0, last_dst = ~0;
|
|
|
+ void *owner = AMDGPU_FENCE_OWNER_VM;
|
|
|
unsigned count = 0;
|
|
|
uint64_t addr;
|
|
|
|
|
|
+ /* sync to everything on unmapping */
|
|
|
+ if (!(flags & AMDGPU_PTE_VALID))
|
|
|
+ owner = AMDGPU_FENCE_OWNER_UNDEFINED;
|
|
|
+
|
|
|
/* walk over the address space and update the page tables */
|
|
|
for (addr = start; addr < end; ) {
|
|
|
uint64_t pt_idx = addr >> amdgpu_vm_block_size;
|
|
@@ -638,8 +643,7 @@ static int amdgpu_vm_update_ptes(struct amdgpu_device *adev,
|
|
|
uint64_t pte;
|
|
|
int r;
|
|
|
|
|
|
- amdgpu_sync_resv(adev, &ib->sync, pt->tbo.resv,
|
|
|
- AMDGPU_FENCE_OWNER_VM);
|
|
|
+ amdgpu_sync_resv(adev, &ib->sync, pt->tbo.resv, owner);
|
|
|
r = reservation_object_reserve_shared(pt->tbo.resv);
|
|
|
if (r)
|
|
|
return r;
|
|
@@ -790,17 +794,6 @@ static int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
|
|
|
|
|
|
ib->length_dw = 0;
|
|
|
|
|
|
- if (!(flags & AMDGPU_PTE_VALID)) {
|
|
|
- unsigned i;
|
|
|
-
|
|
|
- for (i = 0; i < AMDGPU_MAX_RINGS; ++i) {
|
|
|
- struct amdgpu_fence *f = vm->ids[i].last_id_use;
|
|
|
- r = amdgpu_sync_fence(adev, &ib->sync, &f->base);
|
|
|
- if (r)
|
|
|
- return r;
|
|
|
- }
|
|
|
- }
|
|
|
-
|
|
|
r = amdgpu_vm_update_ptes(adev, vm, ib, mapping->it.start,
|
|
|
mapping->it.last + 1, addr + mapping->offset,
|
|
|
flags, gtt_flags);
|