|
@@ -115,7 +115,7 @@ static int amdgpu_vmid_grab_reserved_locked(struct amdgpu_vm *vm,
|
|
|
|
|
|
flushed = id->flushed_updates;
|
|
|
if ((amdgpu_vmid_had_gpu_reset(adev, id)) ||
|
|
|
- (atomic64_read(&id->owner) != vm->client_id) ||
|
|
|
+ (atomic64_read(&id->owner) != vm->entity.fence_context) ||
|
|
|
(job->vm_pd_addr != id->pd_gpu_addr) ||
|
|
|
(updates && (!flushed || updates->context != flushed->context ||
|
|
|
dma_fence_is_later(updates, flushed))) ||
|
|
@@ -144,7 +144,7 @@ static int amdgpu_vmid_grab_reserved_locked(struct amdgpu_vm *vm,
|
|
|
id->flushed_updates = dma_fence_get(updates);
|
|
|
}
|
|
|
id->pd_gpu_addr = job->vm_pd_addr;
|
|
|
- atomic64_set(&id->owner, vm->client_id);
|
|
|
+ atomic64_set(&id->owner, vm->entity.fence_context);
|
|
|
job->vm_needs_flush = needs_flush;
|
|
|
if (needs_flush) {
|
|
|
dma_fence_put(id->last_flush);
|
|
@@ -242,7 +242,7 @@ int amdgpu_vmid_grab(struct amdgpu_vm *vm, struct amdgpu_ring *ring,
|
|
|
if (amdgpu_vmid_had_gpu_reset(adev, id))
|
|
|
continue;
|
|
|
|
|
|
- if (atomic64_read(&id->owner) != vm->client_id)
|
|
|
+ if (atomic64_read(&id->owner) != vm->entity.fence_context)
|
|
|
continue;
|
|
|
|
|
|
if (job->vm_pd_addr != id->pd_gpu_addr)
|
|
@@ -291,7 +291,7 @@ int amdgpu_vmid_grab(struct amdgpu_vm *vm, struct amdgpu_ring *ring,
|
|
|
id->pd_gpu_addr = job->vm_pd_addr;
|
|
|
dma_fence_put(id->flushed_updates);
|
|
|
id->flushed_updates = dma_fence_get(updates);
|
|
|
- atomic64_set(&id->owner, vm->client_id);
|
|
|
+ atomic64_set(&id->owner, vm->entity.fence_context);
|
|
|
|
|
|
needs_flush:
|
|
|
job->vm_needs_flush = true;
|