|
@@ -40,7 +40,6 @@
|
|
|
|
|
|
#include "bif/bif_5_0_d.h"
|
|
#include "bif/bif_5_0_d.h"
|
|
#include "bif/bif_5_0_sh_mask.h"
|
|
#include "bif/bif_5_0_sh_mask.h"
|
|
-
|
|
|
|
#include "gca/gfx_8_0_d.h"
|
|
#include "gca/gfx_8_0_d.h"
|
|
#include "gca/gfx_8_0_enum.h"
|
|
#include "gca/gfx_8_0_enum.h"
|
|
#include "gca/gfx_8_0_sh_mask.h"
|
|
#include "gca/gfx_8_0_sh_mask.h"
|
|
@@ -2100,7 +2099,7 @@ static int gfx_v8_0_sw_init(void *handle)
|
|
return r;
|
|
return r;
|
|
|
|
|
|
/* create MQD for all compute queues as well as KIQ for SRIOV case */
|
|
/* create MQD for all compute queues as well as KIQ for SRIOV case */
|
|
- r = amdgpu_gfx_compute_mqd_sw_init(adev, sizeof(struct vi_mqd));
|
|
|
|
|
|
+ r = amdgpu_gfx_compute_mqd_sw_init(adev, sizeof(struct vi_mqd_allocation));
|
|
if (r)
|
|
if (r)
|
|
return r;
|
|
return r;
|
|
|
|
|
|
@@ -4637,56 +4636,6 @@ static int gfx_v8_0_kiq_kcq_enable(struct amdgpu_device *adev)
|
|
return r;
|
|
return r;
|
|
}
|
|
}
|
|
|
|
|
|
-static int gfx_v8_0_kiq_kcq_disable(struct amdgpu_device *adev)
|
|
|
|
-{
|
|
|
|
- struct amdgpu_ring *kiq_ring = &adev->gfx.kiq.ring;
|
|
|
|
- uint32_t scratch, tmp = 0;
|
|
|
|
- int r, i;
|
|
|
|
-
|
|
|
|
- r = amdgpu_gfx_scratch_get(adev, &scratch);
|
|
|
|
- if (r) {
|
|
|
|
- DRM_ERROR("Failed to get scratch reg (%d).\n", r);
|
|
|
|
- return r;
|
|
|
|
- }
|
|
|
|
- WREG32(scratch, 0xCAFEDEAD);
|
|
|
|
-
|
|
|
|
- r = amdgpu_ring_alloc(kiq_ring, 6 + 3);
|
|
|
|
- if (r) {
|
|
|
|
- DRM_ERROR("Failed to lock KIQ (%d).\n", r);
|
|
|
|
- amdgpu_gfx_scratch_free(adev, scratch);
|
|
|
|
- return r;
|
|
|
|
- }
|
|
|
|
- /* unmap queues */
|
|
|
|
- amdgpu_ring_write(kiq_ring, PACKET3(PACKET3_UNMAP_QUEUES, 4));
|
|
|
|
- amdgpu_ring_write(kiq_ring,
|
|
|
|
- PACKET3_UNMAP_QUEUES_ACTION(1)| /* RESET_QUEUES */
|
|
|
|
- PACKET3_UNMAP_QUEUES_QUEUE_SEL(2)); /* select all queues */
|
|
|
|
- amdgpu_ring_write(kiq_ring, 0);
|
|
|
|
- amdgpu_ring_write(kiq_ring, 0);
|
|
|
|
- amdgpu_ring_write(kiq_ring, 0);
|
|
|
|
- amdgpu_ring_write(kiq_ring, 0);
|
|
|
|
- /* write to scratch for completion */
|
|
|
|
- amdgpu_ring_write(kiq_ring, PACKET3(PACKET3_SET_UCONFIG_REG, 1));
|
|
|
|
- amdgpu_ring_write(kiq_ring, (scratch - PACKET3_SET_UCONFIG_REG_START));
|
|
|
|
- amdgpu_ring_write(kiq_ring, 0xDEADBEEF);
|
|
|
|
- amdgpu_ring_commit(kiq_ring);
|
|
|
|
-
|
|
|
|
- for (i = 0; i < adev->usec_timeout; i++) {
|
|
|
|
- tmp = RREG32(scratch);
|
|
|
|
- if (tmp == 0xDEADBEEF)
|
|
|
|
- break;
|
|
|
|
- DRM_UDELAY(1);
|
|
|
|
- }
|
|
|
|
- if (i >= adev->usec_timeout) {
|
|
|
|
- DRM_ERROR("KCQ disabled failed (scratch(0x%04X)=0x%08X)\n",
|
|
|
|
- scratch, tmp);
|
|
|
|
- r = -EINVAL;
|
|
|
|
- }
|
|
|
|
- amdgpu_gfx_scratch_free(adev, scratch);
|
|
|
|
-
|
|
|
|
- return r;
|
|
|
|
-}
|
|
|
|
-
|
|
|
|
static int gfx_v8_0_deactivate_hqd(struct amdgpu_device *adev, u32 req)
|
|
static int gfx_v8_0_deactivate_hqd(struct amdgpu_device *adev, u32 req)
|
|
{
|
|
{
|
|
int i, r = 0;
|
|
int i, r = 0;
|
|
@@ -4715,9 +4664,6 @@ static int gfx_v8_0_mqd_init(struct amdgpu_ring *ring)
|
|
uint64_t hqd_gpu_addr, wb_gpu_addr, eop_base_addr;
|
|
uint64_t hqd_gpu_addr, wb_gpu_addr, eop_base_addr;
|
|
uint32_t tmp;
|
|
uint32_t tmp;
|
|
|
|
|
|
- /* init the mqd struct */
|
|
|
|
- memset(mqd, 0, sizeof(struct vi_mqd));
|
|
|
|
-
|
|
|
|
mqd->header = 0xC0310800;
|
|
mqd->header = 0xC0310800;
|
|
mqd->compute_pipelinestat_enable = 0x00000001;
|
|
mqd->compute_pipelinestat_enable = 0x00000001;
|
|
mqd->compute_static_thread_mgmt_se0 = 0xffffffff;
|
|
mqd->compute_static_thread_mgmt_se0 = 0xffffffff;
|
|
@@ -4725,7 +4671,12 @@ static int gfx_v8_0_mqd_init(struct amdgpu_ring *ring)
|
|
mqd->compute_static_thread_mgmt_se2 = 0xffffffff;
|
|
mqd->compute_static_thread_mgmt_se2 = 0xffffffff;
|
|
mqd->compute_static_thread_mgmt_se3 = 0xffffffff;
|
|
mqd->compute_static_thread_mgmt_se3 = 0xffffffff;
|
|
mqd->compute_misc_reserved = 0x00000003;
|
|
mqd->compute_misc_reserved = 0x00000003;
|
|
-
|
|
|
|
|
|
+ if (!(adev->flags & AMD_IS_APU)) {
|
|
|
|
+ mqd->dynamic_cu_mask_addr_lo = lower_32_bits(ring->mqd_gpu_addr
|
|
|
|
+ + offsetof(struct vi_mqd_allocation, dyamic_cu_mask));
|
|
|
|
+ mqd->dynamic_cu_mask_addr_hi = upper_32_bits(ring->mqd_gpu_addr
|
|
|
|
+ + offsetof(struct vi_mqd_allocation, dyamic_cu_mask));
|
|
|
|
+ }
|
|
eop_base_addr = ring->eop_gpu_addr >> 8;
|
|
eop_base_addr = ring->eop_gpu_addr >> 8;
|
|
mqd->cp_hqd_eop_base_addr_lo = eop_base_addr;
|
|
mqd->cp_hqd_eop_base_addr_lo = eop_base_addr;
|
|
mqd->cp_hqd_eop_base_addr_hi = upper_32_bits(eop_base_addr);
|
|
mqd->cp_hqd_eop_base_addr_hi = upper_32_bits(eop_base_addr);
|
|
@@ -4890,7 +4841,6 @@ int gfx_v8_0_mqd_commit(struct amdgpu_device *adev,
|
|
|
|
|
|
static int gfx_v8_0_kiq_init_queue(struct amdgpu_ring *ring)
|
|
static int gfx_v8_0_kiq_init_queue(struct amdgpu_ring *ring)
|
|
{
|
|
{
|
|
- int r = 0;
|
|
|
|
struct amdgpu_device *adev = ring->adev;
|
|
struct amdgpu_device *adev = ring->adev;
|
|
struct vi_mqd *mqd = ring->mqd_ptr;
|
|
struct vi_mqd *mqd = ring->mqd_ptr;
|
|
int mqd_idx = AMDGPU_MAX_COMPUTE_RINGS;
|
|
int mqd_idx = AMDGPU_MAX_COMPUTE_RINGS;
|
|
@@ -4900,44 +4850,32 @@ static int gfx_v8_0_kiq_init_queue(struct amdgpu_ring *ring)
|
|
if (adev->gfx.in_reset) { /* for GPU_RESET case */
|
|
if (adev->gfx.in_reset) { /* for GPU_RESET case */
|
|
/* reset MQD to a clean status */
|
|
/* reset MQD to a clean status */
|
|
if (adev->gfx.mec.mqd_backup[mqd_idx])
|
|
if (adev->gfx.mec.mqd_backup[mqd_idx])
|
|
- memcpy(mqd, adev->gfx.mec.mqd_backup[mqd_idx], sizeof(*mqd));
|
|
|
|
|
|
+ memcpy(mqd, adev->gfx.mec.mqd_backup[mqd_idx], sizeof(struct vi_mqd_allocation));
|
|
|
|
|
|
/* reset ring buffer */
|
|
/* reset ring buffer */
|
|
ring->wptr = 0;
|
|
ring->wptr = 0;
|
|
amdgpu_ring_clear_ring(ring);
|
|
amdgpu_ring_clear_ring(ring);
|
|
mutex_lock(&adev->srbm_mutex);
|
|
mutex_lock(&adev->srbm_mutex);
|
|
vi_srbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
|
|
vi_srbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
|
|
- r = gfx_v8_0_deactivate_hqd(adev, 1);
|
|
|
|
- if (r) {
|
|
|
|
- dev_err(adev->dev, "failed to deactivate ring %s\n", ring->name);
|
|
|
|
- goto out_unlock;
|
|
|
|
- }
|
|
|
|
gfx_v8_0_mqd_commit(adev, mqd);
|
|
gfx_v8_0_mqd_commit(adev, mqd);
|
|
vi_srbm_select(adev, 0, 0, 0, 0);
|
|
vi_srbm_select(adev, 0, 0, 0, 0);
|
|
mutex_unlock(&adev->srbm_mutex);
|
|
mutex_unlock(&adev->srbm_mutex);
|
|
} else {
|
|
} else {
|
|
|
|
+ memset((void *)mqd, 0, sizeof(struct vi_mqd_allocation));
|
|
|
|
+ ((struct vi_mqd_allocation *)mqd)->dyamic_cu_mask = 0xFFFFFFFF;
|
|
|
|
+ ((struct vi_mqd_allocation *)mqd)->dyamic_rb_mask = 0xFFFFFFFF;
|
|
mutex_lock(&adev->srbm_mutex);
|
|
mutex_lock(&adev->srbm_mutex);
|
|
vi_srbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
|
|
vi_srbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
|
|
gfx_v8_0_mqd_init(ring);
|
|
gfx_v8_0_mqd_init(ring);
|
|
- r = gfx_v8_0_deactivate_hqd(adev, 1);
|
|
|
|
- if (r) {
|
|
|
|
- dev_err(adev->dev, "failed to deactivate ring %s\n", ring->name);
|
|
|
|
- goto out_unlock;
|
|
|
|
- }
|
|
|
|
gfx_v8_0_mqd_commit(adev, mqd);
|
|
gfx_v8_0_mqd_commit(adev, mqd);
|
|
vi_srbm_select(adev, 0, 0, 0, 0);
|
|
vi_srbm_select(adev, 0, 0, 0, 0);
|
|
mutex_unlock(&adev->srbm_mutex);
|
|
mutex_unlock(&adev->srbm_mutex);
|
|
|
|
|
|
if (adev->gfx.mec.mqd_backup[mqd_idx])
|
|
if (adev->gfx.mec.mqd_backup[mqd_idx])
|
|
- memcpy(adev->gfx.mec.mqd_backup[mqd_idx], mqd, sizeof(*mqd));
|
|
|
|
|
|
+ memcpy(adev->gfx.mec.mqd_backup[mqd_idx], mqd, sizeof(struct vi_mqd_allocation));
|
|
}
|
|
}
|
|
|
|
|
|
- return r;
|
|
|
|
-
|
|
|
|
-out_unlock:
|
|
|
|
- vi_srbm_select(adev, 0, 0, 0, 0);
|
|
|
|
- mutex_unlock(&adev->srbm_mutex);
|
|
|
|
- return r;
|
|
|
|
|
|
+ return 0;
|
|
}
|
|
}
|
|
|
|
|
|
static int gfx_v8_0_kcq_init_queue(struct amdgpu_ring *ring)
|
|
static int gfx_v8_0_kcq_init_queue(struct amdgpu_ring *ring)
|
|
@@ -4947,6 +4885,9 @@ static int gfx_v8_0_kcq_init_queue(struct amdgpu_ring *ring)
|
|
int mqd_idx = ring - &adev->gfx.compute_ring[0];
|
|
int mqd_idx = ring - &adev->gfx.compute_ring[0];
|
|
|
|
|
|
if (!adev->gfx.in_reset && !adev->gfx.in_suspend) {
|
|
if (!adev->gfx.in_reset && !adev->gfx.in_suspend) {
|
|
|
|
+ memset((void *)mqd, 0, sizeof(struct vi_mqd_allocation));
|
|
|
|
+ ((struct vi_mqd_allocation *)mqd)->dyamic_cu_mask = 0xFFFFFFFF;
|
|
|
|
+ ((struct vi_mqd_allocation *)mqd)->dyamic_rb_mask = 0xFFFFFFFF;
|
|
mutex_lock(&adev->srbm_mutex);
|
|
mutex_lock(&adev->srbm_mutex);
|
|
vi_srbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
|
|
vi_srbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
|
|
gfx_v8_0_mqd_init(ring);
|
|
gfx_v8_0_mqd_init(ring);
|
|
@@ -4954,11 +4895,11 @@ static int gfx_v8_0_kcq_init_queue(struct amdgpu_ring *ring)
|
|
mutex_unlock(&adev->srbm_mutex);
|
|
mutex_unlock(&adev->srbm_mutex);
|
|
|
|
|
|
if (adev->gfx.mec.mqd_backup[mqd_idx])
|
|
if (adev->gfx.mec.mqd_backup[mqd_idx])
|
|
- memcpy(adev->gfx.mec.mqd_backup[mqd_idx], mqd, sizeof(*mqd));
|
|
|
|
|
|
+ memcpy(adev->gfx.mec.mqd_backup[mqd_idx], mqd, sizeof(struct vi_mqd_allocation));
|
|
} else if (adev->gfx.in_reset) { /* for GPU_RESET case */
|
|
} else if (adev->gfx.in_reset) { /* for GPU_RESET case */
|
|
/* reset MQD to a clean status */
|
|
/* reset MQD to a clean status */
|
|
if (adev->gfx.mec.mqd_backup[mqd_idx])
|
|
if (adev->gfx.mec.mqd_backup[mqd_idx])
|
|
- memcpy(mqd, adev->gfx.mec.mqd_backup[mqd_idx], sizeof(*mqd));
|
|
|
|
|
|
+ memcpy(mqd, adev->gfx.mec.mqd_backup[mqd_idx], sizeof(struct vi_mqd_allocation));
|
|
/* reset ring buffer */
|
|
/* reset ring buffer */
|
|
ring->wptr = 0;
|
|
ring->wptr = 0;
|
|
amdgpu_ring_clear_ring(ring);
|
|
amdgpu_ring_clear_ring(ring);
|
|
@@ -5138,7 +5079,6 @@ static int gfx_v8_0_hw_fini(void *handle)
|
|
pr_debug("For SRIOV client, shouldn't do anything.\n");
|
|
pr_debug("For SRIOV client, shouldn't do anything.\n");
|
|
return 0;
|
|
return 0;
|
|
}
|
|
}
|
|
- gfx_v8_0_kiq_kcq_disable(adev);
|
|
|
|
gfx_v8_0_cp_enable(adev, false);
|
|
gfx_v8_0_cp_enable(adev, false);
|
|
gfx_v8_0_rlc_stop(adev);
|
|
gfx_v8_0_rlc_stop(adev);
|
|
|
|
|
|
@@ -7080,7 +7020,9 @@ static void gfx_v8_0_get_cu_info(struct amdgpu_device *adev)
|
|
mask <<= 1;
|
|
mask <<= 1;
|
|
}
|
|
}
|
|
active_cu_number += counter;
|
|
active_cu_number += counter;
|
|
- ao_cu_mask |= (ao_bitmap << (i * 16 + j * 8));
|
|
|
|
|
|
+ if (i < 2 && j < 2)
|
|
|
|
+ ao_cu_mask |= (ao_bitmap << (i * 16 + j * 8));
|
|
|
|
+ cu_info->ao_cu_bitmap[i][j] = ao_bitmap;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
gfx_v8_0_select_se_sh(adev, 0xffffffff, 0xffffffff, 0xffffffff);
|
|
gfx_v8_0_select_se_sh(adev, 0xffffffff, 0xffffffff, 0xffffffff);
|