|
@@ -646,7 +646,7 @@ static void gfx_v9_0_kiq_free_ring(struct amdgpu_ring *ring,
|
|
}
|
|
}
|
|
|
|
|
|
/* create MQD for each compute queue */
|
|
/* create MQD for each compute queue */
|
|
-static int gfx_v9_0_compute_mqd_soft_init(struct amdgpu_device *adev)
|
|
|
|
|
|
+static int gfx_v9_0_compute_mqd_sw_init(struct amdgpu_device *adev)
|
|
{
|
|
{
|
|
struct amdgpu_ring *ring = NULL;
|
|
struct amdgpu_ring *ring = NULL;
|
|
int r, i;
|
|
int r, i;
|
|
@@ -684,7 +684,7 @@ static int gfx_v9_0_compute_mqd_soft_init(struct amdgpu_device *adev)
|
|
return 0;
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
|
|
-static void gfx_v9_0_compute_mqd_soft_fini(struct amdgpu_device *adev)
|
|
|
|
|
|
+static void gfx_v9_0_compute_mqd_sw_fini(struct amdgpu_device *adev)
|
|
{
|
|
{
|
|
struct amdgpu_ring *ring = NULL;
|
|
struct amdgpu_ring *ring = NULL;
|
|
int i;
|
|
int i;
|
|
@@ -1112,7 +1112,7 @@ static int gfx_v9_0_sw_init(void *handle)
|
|
return r;
|
|
return r;
|
|
|
|
|
|
/* create MQD for all compute queues as wel as KIQ for SRIOV case */
|
|
/* create MQD for all compute queues as wel as KIQ for SRIOV case */
|
|
- r = gfx_v9_0_compute_mqd_soft_init(adev);
|
|
|
|
|
|
+ r = gfx_v9_0_compute_mqd_sw_init(adev);
|
|
if (r)
|
|
if (r)
|
|
return r;
|
|
return r;
|
|
}
|
|
}
|
|
@@ -1163,7 +1163,7 @@ static int gfx_v9_0_sw_fini(void *handle)
|
|
amdgpu_ring_fini(&adev->gfx.compute_ring[i]);
|
|
amdgpu_ring_fini(&adev->gfx.compute_ring[i]);
|
|
|
|
|
|
if (amdgpu_sriov_vf(adev)) {
|
|
if (amdgpu_sriov_vf(adev)) {
|
|
- gfx_v9_0_compute_mqd_soft_fini(adev);
|
|
|
|
|
|
+ gfx_v9_0_compute_mqd_sw_fini(adev);
|
|
gfx_v9_0_kiq_free_ring(&adev->gfx.kiq.ring, &adev->gfx.kiq.irq);
|
|
gfx_v9_0_kiq_free_ring(&adev->gfx.kiq.ring, &adev->gfx.kiq.irq);
|
|
gfx_v9_0_kiq_fini(adev);
|
|
gfx_v9_0_kiq_fini(adev);
|
|
}
|
|
}
|