|
@@ -93,7 +93,10 @@ static int a3xx_hw_init(struct msm_gpu *gpu)
|
|
|
/* Set up AOOO: */
|
|
|
gpu_write(gpu, REG_A3XX_VBIF_OUT_AXI_AOOO_EN, 0x0000003c);
|
|
|
gpu_write(gpu, REG_A3XX_VBIF_OUT_AXI_AOOO, 0x003c003c);
|
|
|
-
|
|
|
+ } else if (adreno_is_a306(adreno_gpu)) {
|
|
|
+ gpu_write(gpu, REG_A3XX_VBIF_ROUND_ROBIN_QOS_ARB, 0x0003);
|
|
|
+ gpu_write(gpu, REG_A3XX_VBIF_OUT_RD_LIM_CONF0, 0x0000000a);
|
|
|
+ gpu_write(gpu, REG_A3XX_VBIF_OUT_WR_LIM_CONF0, 0x0000000a);
|
|
|
} else if (adreno_is_a320(adreno_gpu)) {
|
|
|
/* Set up 16 deep read/write request queues: */
|
|
|
gpu_write(gpu, REG_A3XX_VBIF_IN_RD_LIM_CONF0, 0x10101010);
|
|
@@ -186,7 +189,9 @@ static int a3xx_hw_init(struct msm_gpu *gpu)
|
|
|
gpu_write(gpu, REG_A3XX_UCHE_CACHE_MODE_CONTROL_REG, 0x00000001);
|
|
|
|
|
|
/* Enable Clock gating: */
|
|
|
- if (adreno_is_a320(adreno_gpu))
|
|
|
+ if (adreno_is_a306(adreno_gpu))
|
|
|
+ gpu_write(gpu, REG_A3XX_RBBM_CLOCK_CTL, 0xaaaaaaaa);
|
|
|
+ else if (adreno_is_a320(adreno_gpu))
|
|
|
gpu_write(gpu, REG_A3XX_RBBM_CLOCK_CTL, 0xbfffffff);
|
|
|
else if (adreno_is_a330v2(adreno_gpu))
|
|
|
gpu_write(gpu, REG_A3XX_RBBM_CLOCK_CTL, 0xaaaaaaaa);
|
|
@@ -271,7 +276,8 @@ static int a3xx_hw_init(struct msm_gpu *gpu)
|
|
|
gpu_write(gpu, REG_A3XX_CP_PFP_UCODE_DATA, ptr[i]);
|
|
|
|
|
|
/* CP ROQ queue sizes (bytes) - RB:16, ST:16, IB1:32, IB2:64 */
|
|
|
- if (adreno_is_a305(adreno_gpu) || adreno_is_a320(adreno_gpu)) {
|
|
|
+ if (adreno_is_a305(adreno_gpu) || adreno_is_a306(adreno_gpu) ||
|
|
|
+ adreno_is_a320(adreno_gpu)) {
|
|
|
gpu_write(gpu, REG_AXXX_CP_QUEUE_THRESHOLDS,
|
|
|
AXXX_CP_QUEUE_THRESHOLDS_CSQ_IB1_START(2) |
|
|
|
AXXX_CP_QUEUE_THRESHOLDS_CSQ_IB2_START(6) |
|