|
@@ -1138,6 +1138,9 @@ static void amdgpu_uvd_idle_work_handler(struct work_struct *work)
|
|
|
container_of(work, struct amdgpu_device, uvd.idle_work.work);
|
|
|
unsigned fences = amdgpu_fence_count_emitted(&adev->uvd.ring);
|
|
|
|
|
|
+ if (amdgpu_sriov_vf(adev))
|
|
|
+ return;
|
|
|
+
|
|
|
if (fences == 0) {
|
|
|
if (adev->pm.dpm_enabled) {
|
|
|
amdgpu_dpm_enable_uvd(adev, false);
|
|
@@ -1159,6 +1162,9 @@ void amdgpu_uvd_ring_begin_use(struct amdgpu_ring *ring)
|
|
|
struct amdgpu_device *adev = ring->adev;
|
|
|
bool set_clocks = !cancel_delayed_work_sync(&adev->uvd.idle_work);
|
|
|
|
|
|
+ if (amdgpu_sriov_vf(adev))
|
|
|
+ return;
|
|
|
+
|
|
|
if (set_clocks) {
|
|
|
if (adev->pm.dpm_enabled) {
|
|
|
amdgpu_dpm_enable_uvd(adev, true);
|