soc15.c 26 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938
  1. /*
  2. * Copyright 2016 Advanced Micro Devices, Inc.
  3. *
  4. * Permission is hereby granted, free of charge, to any person obtaining a
  5. * copy of this software and associated documentation files (the "Software"),
  6. * to deal in the Software without restriction, including without limitation
  7. * the rights to use, copy, modify, merge, publish, distribute, sublicense,
  8. * and/or sell copies of the Software, and to permit persons to whom the
  9. * Software is furnished to do so, subject to the following conditions:
  10. *
  11. * The above copyright notice and this permission notice shall be included in
  12. * all copies or substantial portions of the Software.
  13. *
  14. * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
  15. * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
  16. * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
  17. * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
  18. * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
  19. * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
  20. * OTHER DEALINGS IN THE SOFTWARE.
  21. *
  22. */
  23. #include <linux/firmware.h>
  24. #include <linux/slab.h>
  25. #include <linux/module.h>
  26. #include <drm/drmP.h>
  27. #include "amdgpu.h"
  28. #include "amdgpu_atomfirmware.h"
  29. #include "amdgpu_ih.h"
  30. #include "amdgpu_uvd.h"
  31. #include "amdgpu_vce.h"
  32. #include "amdgpu_ucode.h"
  33. #include "amdgpu_psp.h"
  34. #include "atom.h"
  35. #include "amd_pcie.h"
  36. #include "vega10/soc15ip.h"
  37. #include "vega10/UVD/uvd_7_0_offset.h"
  38. #include "vega10/GC/gc_9_0_offset.h"
  39. #include "vega10/GC/gc_9_0_sh_mask.h"
  40. #include "vega10/SDMA0/sdma0_4_0_offset.h"
  41. #include "vega10/SDMA1/sdma1_4_0_offset.h"
  42. #include "vega10/HDP/hdp_4_0_offset.h"
  43. #include "vega10/HDP/hdp_4_0_sh_mask.h"
  44. #include "vega10/MP/mp_9_0_offset.h"
  45. #include "vega10/MP/mp_9_0_sh_mask.h"
  46. #include "vega10/SMUIO/smuio_9_0_offset.h"
  47. #include "vega10/SMUIO/smuio_9_0_sh_mask.h"
  48. #include "soc15.h"
  49. #include "soc15_common.h"
  50. #include "gfx_v9_0.h"
  51. #include "gmc_v9_0.h"
  52. #include "gfxhub_v1_0.h"
  53. #include "mmhub_v1_0.h"
  54. #include "vega10_ih.h"
  55. #include "sdma_v4_0.h"
  56. #include "uvd_v7_0.h"
  57. #include "vce_v4_0.h"
  58. #include "vcn_v1_0.h"
  59. #include "amdgpu_powerplay.h"
  60. #include "dce_virtual.h"
  61. #include "mxgpu_ai.h"
  62. MODULE_FIRMWARE("amdgpu/vega10_smc.bin");
  63. #define mmFabricConfigAccessControl 0x0410
  64. #define mmFabricConfigAccessControl_BASE_IDX 0
  65. #define mmFabricConfigAccessControl_DEFAULT 0x00000000
  66. //FabricConfigAccessControl
  67. #define FabricConfigAccessControl__CfgRegInstAccEn__SHIFT 0x0
  68. #define FabricConfigAccessControl__CfgRegInstAccRegLock__SHIFT 0x1
  69. #define FabricConfigAccessControl__CfgRegInstID__SHIFT 0x10
  70. #define FabricConfigAccessControl__CfgRegInstAccEn_MASK 0x00000001L
  71. #define FabricConfigAccessControl__CfgRegInstAccRegLock_MASK 0x00000002L
  72. #define FabricConfigAccessControl__CfgRegInstID_MASK 0x00FF0000L
  73. #define mmDF_PIE_AON0_DfGlobalClkGater 0x00fc
  74. #define mmDF_PIE_AON0_DfGlobalClkGater_BASE_IDX 0
  75. //DF_PIE_AON0_DfGlobalClkGater
  76. #define DF_PIE_AON0_DfGlobalClkGater__MGCGMode__SHIFT 0x0
  77. #define DF_PIE_AON0_DfGlobalClkGater__MGCGMode_MASK 0x0000000FL
  78. enum {
  79. DF_MGCG_DISABLE = 0,
  80. DF_MGCG_ENABLE_00_CYCLE_DELAY =1,
  81. DF_MGCG_ENABLE_01_CYCLE_DELAY =2,
  82. DF_MGCG_ENABLE_15_CYCLE_DELAY =13,
  83. DF_MGCG_ENABLE_31_CYCLE_DELAY =14,
  84. DF_MGCG_ENABLE_63_CYCLE_DELAY =15
  85. };
  86. #define mmMP0_MISC_CGTT_CTRL0 0x01b9
  87. #define mmMP0_MISC_CGTT_CTRL0_BASE_IDX 0
  88. #define mmMP0_MISC_LIGHT_SLEEP_CTRL 0x01ba
  89. #define mmMP0_MISC_LIGHT_SLEEP_CTRL_BASE_IDX 0
  90. /*
  91. * Indirect registers accessor
  92. */
  93. static u32 soc15_pcie_rreg(struct amdgpu_device *adev, u32 reg)
  94. {
  95. unsigned long flags, address, data;
  96. u32 r;
  97. struct nbio_pcie_index_data *nbio_pcie_id;
  98. if (adev->flags & AMD_IS_APU)
  99. nbio_pcie_id = &nbio_v7_0_pcie_index_data;
  100. else
  101. nbio_pcie_id = &nbio_v6_1_pcie_index_data;
  102. address = nbio_pcie_id->index_offset;
  103. data = nbio_pcie_id->data_offset;
  104. spin_lock_irqsave(&adev->pcie_idx_lock, flags);
  105. WREG32(address, reg);
  106. (void)RREG32(address);
  107. r = RREG32(data);
  108. spin_unlock_irqrestore(&adev->pcie_idx_lock, flags);
  109. return r;
  110. }
  111. static void soc15_pcie_wreg(struct amdgpu_device *adev, u32 reg, u32 v)
  112. {
  113. unsigned long flags, address, data;
  114. struct nbio_pcie_index_data *nbio_pcie_id;
  115. if (adev->flags & AMD_IS_APU)
  116. nbio_pcie_id = &nbio_v7_0_pcie_index_data;
  117. else
  118. nbio_pcie_id = &nbio_v6_1_pcie_index_data;
  119. address = nbio_pcie_id->index_offset;
  120. data = nbio_pcie_id->data_offset;
  121. spin_lock_irqsave(&adev->pcie_idx_lock, flags);
  122. WREG32(address, reg);
  123. (void)RREG32(address);
  124. WREG32(data, v);
  125. (void)RREG32(data);
  126. spin_unlock_irqrestore(&adev->pcie_idx_lock, flags);
  127. }
  128. static u32 soc15_uvd_ctx_rreg(struct amdgpu_device *adev, u32 reg)
  129. {
  130. unsigned long flags, address, data;
  131. u32 r;
  132. address = SOC15_REG_OFFSET(UVD, 0, mmUVD_CTX_INDEX);
  133. data = SOC15_REG_OFFSET(UVD, 0, mmUVD_CTX_DATA);
  134. spin_lock_irqsave(&adev->uvd_ctx_idx_lock, flags);
  135. WREG32(address, ((reg) & 0x1ff));
  136. r = RREG32(data);
  137. spin_unlock_irqrestore(&adev->uvd_ctx_idx_lock, flags);
  138. return r;
  139. }
  140. static void soc15_uvd_ctx_wreg(struct amdgpu_device *adev, u32 reg, u32 v)
  141. {
  142. unsigned long flags, address, data;
  143. address = SOC15_REG_OFFSET(UVD, 0, mmUVD_CTX_INDEX);
  144. data = SOC15_REG_OFFSET(UVD, 0, mmUVD_CTX_DATA);
  145. spin_lock_irqsave(&adev->uvd_ctx_idx_lock, flags);
  146. WREG32(address, ((reg) & 0x1ff));
  147. WREG32(data, (v));
  148. spin_unlock_irqrestore(&adev->uvd_ctx_idx_lock, flags);
  149. }
  150. static u32 soc15_didt_rreg(struct amdgpu_device *adev, u32 reg)
  151. {
  152. unsigned long flags, address, data;
  153. u32 r;
  154. address = SOC15_REG_OFFSET(GC, 0, mmDIDT_IND_INDEX);
  155. data = SOC15_REG_OFFSET(GC, 0, mmDIDT_IND_DATA);
  156. spin_lock_irqsave(&adev->didt_idx_lock, flags);
  157. WREG32(address, (reg));
  158. r = RREG32(data);
  159. spin_unlock_irqrestore(&adev->didt_idx_lock, flags);
  160. return r;
  161. }
  162. static void soc15_didt_wreg(struct amdgpu_device *adev, u32 reg, u32 v)
  163. {
  164. unsigned long flags, address, data;
  165. address = SOC15_REG_OFFSET(GC, 0, mmDIDT_IND_INDEX);
  166. data = SOC15_REG_OFFSET(GC, 0, mmDIDT_IND_DATA);
  167. spin_lock_irqsave(&adev->didt_idx_lock, flags);
  168. WREG32(address, (reg));
  169. WREG32(data, (v));
  170. spin_unlock_irqrestore(&adev->didt_idx_lock, flags);
  171. }
  172. static u32 soc15_get_config_memsize(struct amdgpu_device *adev)
  173. {
  174. if (adev->flags & AMD_IS_APU)
  175. return nbio_v7_0_get_memsize(adev);
  176. else
  177. return nbio_v6_1_get_memsize(adev);
  178. }
  179. static const u32 vega10_golden_init[] =
  180. {
  181. };
  182. static const u32 raven_golden_init[] =
  183. {
  184. };
  185. static void soc15_init_golden_registers(struct amdgpu_device *adev)
  186. {
  187. /* Some of the registers might be dependent on GRBM_GFX_INDEX */
  188. mutex_lock(&adev->grbm_idx_mutex);
  189. switch (adev->asic_type) {
  190. case CHIP_VEGA10:
  191. amdgpu_program_register_sequence(adev,
  192. vega10_golden_init,
  193. (const u32)ARRAY_SIZE(vega10_golden_init));
  194. break;
  195. case CHIP_RAVEN:
  196. amdgpu_program_register_sequence(adev,
  197. raven_golden_init,
  198. (const u32)ARRAY_SIZE(raven_golden_init));
  199. break;
  200. default:
  201. break;
  202. }
  203. mutex_unlock(&adev->grbm_idx_mutex);
  204. }
  205. static u32 soc15_get_xclk(struct amdgpu_device *adev)
  206. {
  207. if (adev->asic_type == CHIP_VEGA10)
  208. return adev->clock.spll.reference_freq/4;
  209. else
  210. return adev->clock.spll.reference_freq;
  211. }
  212. void soc15_grbm_select(struct amdgpu_device *adev,
  213. u32 me, u32 pipe, u32 queue, u32 vmid)
  214. {
  215. u32 grbm_gfx_cntl = 0;
  216. grbm_gfx_cntl = REG_SET_FIELD(grbm_gfx_cntl, GRBM_GFX_CNTL, PIPEID, pipe);
  217. grbm_gfx_cntl = REG_SET_FIELD(grbm_gfx_cntl, GRBM_GFX_CNTL, MEID, me);
  218. grbm_gfx_cntl = REG_SET_FIELD(grbm_gfx_cntl, GRBM_GFX_CNTL, VMID, vmid);
  219. grbm_gfx_cntl = REG_SET_FIELD(grbm_gfx_cntl, GRBM_GFX_CNTL, QUEUEID, queue);
  220. WREG32(SOC15_REG_OFFSET(GC, 0, mmGRBM_GFX_CNTL), grbm_gfx_cntl);
  221. }
  222. static void soc15_vga_set_state(struct amdgpu_device *adev, bool state)
  223. {
  224. /* todo */
  225. }
  226. static bool soc15_read_disabled_bios(struct amdgpu_device *adev)
  227. {
  228. /* todo */
  229. return false;
  230. }
  231. static bool soc15_read_bios_from_rom(struct amdgpu_device *adev,
  232. u8 *bios, u32 length_bytes)
  233. {
  234. u32 *dw_ptr;
  235. u32 i, length_dw;
  236. if (bios == NULL)
  237. return false;
  238. if (length_bytes == 0)
  239. return false;
  240. /* APU vbios image is part of sbios image */
  241. if (adev->flags & AMD_IS_APU)
  242. return false;
  243. dw_ptr = (u32 *)bios;
  244. length_dw = ALIGN(length_bytes, 4) / 4;
  245. /* set rom index to 0 */
  246. WREG32(SOC15_REG_OFFSET(SMUIO, 0, mmROM_INDEX), 0);
  247. /* read out the rom data */
  248. for (i = 0; i < length_dw; i++)
  249. dw_ptr[i] = RREG32(SOC15_REG_OFFSET(SMUIO, 0, mmROM_DATA));
  250. return true;
  251. }
  252. static struct amdgpu_allowed_register_entry soc15_allowed_read_registers[] = {
  253. { SOC15_REG_OFFSET(GC, 0, mmGRBM_STATUS)},
  254. { SOC15_REG_OFFSET(GC, 0, mmGRBM_STATUS2)},
  255. { SOC15_REG_OFFSET(GC, 0, mmGRBM_STATUS_SE0)},
  256. { SOC15_REG_OFFSET(GC, 0, mmGRBM_STATUS_SE1)},
  257. { SOC15_REG_OFFSET(GC, 0, mmGRBM_STATUS_SE2)},
  258. { SOC15_REG_OFFSET(GC, 0, mmGRBM_STATUS_SE3)},
  259. { SOC15_REG_OFFSET(SDMA0, 0, mmSDMA0_STATUS_REG)},
  260. { SOC15_REG_OFFSET(SDMA1, 0, mmSDMA1_STATUS_REG)},
  261. { SOC15_REG_OFFSET(GC, 0, mmCP_STAT)},
  262. { SOC15_REG_OFFSET(GC, 0, mmCP_STALLED_STAT1)},
  263. { SOC15_REG_OFFSET(GC, 0, mmCP_STALLED_STAT2)},
  264. { SOC15_REG_OFFSET(GC, 0, mmCP_STALLED_STAT3)},
  265. { SOC15_REG_OFFSET(GC, 0, mmCP_CPF_BUSY_STAT)},
  266. { SOC15_REG_OFFSET(GC, 0, mmCP_CPF_STALLED_STAT1)},
  267. { SOC15_REG_OFFSET(GC, 0, mmCP_CPF_STATUS)},
  268. { SOC15_REG_OFFSET(GC, 0, mmCP_CPC_STALLED_STAT1)},
  269. { SOC15_REG_OFFSET(GC, 0, mmCP_CPC_STATUS)},
  270. { SOC15_REG_OFFSET(GC, 0, mmGB_ADDR_CONFIG)},
  271. };
  272. static uint32_t soc15_read_indexed_register(struct amdgpu_device *adev, u32 se_num,
  273. u32 sh_num, u32 reg_offset)
  274. {
  275. uint32_t val;
  276. mutex_lock(&adev->grbm_idx_mutex);
  277. if (se_num != 0xffffffff || sh_num != 0xffffffff)
  278. amdgpu_gfx_select_se_sh(adev, se_num, sh_num, 0xffffffff);
  279. val = RREG32(reg_offset);
  280. if (se_num != 0xffffffff || sh_num != 0xffffffff)
  281. amdgpu_gfx_select_se_sh(adev, 0xffffffff, 0xffffffff, 0xffffffff);
  282. mutex_unlock(&adev->grbm_idx_mutex);
  283. return val;
  284. }
  285. static uint32_t soc15_get_register_value(struct amdgpu_device *adev,
  286. bool indexed, u32 se_num,
  287. u32 sh_num, u32 reg_offset)
  288. {
  289. if (indexed) {
  290. return soc15_read_indexed_register(adev, se_num, sh_num, reg_offset);
  291. } else {
  292. switch (reg_offset) {
  293. case SOC15_REG_OFFSET(GC, 0, mmGB_ADDR_CONFIG):
  294. return adev->gfx.config.gb_addr_config;
  295. default:
  296. return RREG32(reg_offset);
  297. }
  298. }
  299. }
  300. static int soc15_read_register(struct amdgpu_device *adev, u32 se_num,
  301. u32 sh_num, u32 reg_offset, u32 *value)
  302. {
  303. uint32_t i;
  304. *value = 0;
  305. for (i = 0; i < ARRAY_SIZE(soc15_allowed_read_registers); i++) {
  306. if (reg_offset != soc15_allowed_read_registers[i].reg_offset)
  307. continue;
  308. *value = soc15_get_register_value(adev,
  309. soc15_allowed_read_registers[i].grbm_indexed,
  310. se_num, sh_num, reg_offset);
  311. return 0;
  312. }
  313. return -EINVAL;
  314. }
  315. static void soc15_gpu_pci_config_reset(struct amdgpu_device *adev)
  316. {
  317. u32 i;
  318. dev_info(adev->dev, "GPU pci config reset\n");
  319. /* disable BM */
  320. pci_clear_master(adev->pdev);
  321. /* reset */
  322. amdgpu_pci_config_reset(adev);
  323. udelay(100);
  324. /* wait for asic to come out of reset */
  325. for (i = 0; i < adev->usec_timeout; i++) {
  326. u32 memsize = (adev->flags & AMD_IS_APU) ?
  327. nbio_v7_0_get_memsize(adev) :
  328. nbio_v6_1_get_memsize(adev);
  329. if (memsize != 0xffffffff)
  330. break;
  331. udelay(1);
  332. }
  333. }
  334. static int soc15_asic_reset(struct amdgpu_device *adev)
  335. {
  336. amdgpu_atomfirmware_scratch_regs_engine_hung(adev, true);
  337. soc15_gpu_pci_config_reset(adev);
  338. amdgpu_atomfirmware_scratch_regs_engine_hung(adev, false);
  339. return 0;
  340. }
  341. /*static int soc15_set_uvd_clock(struct amdgpu_device *adev, u32 clock,
  342. u32 cntl_reg, u32 status_reg)
  343. {
  344. return 0;
  345. }*/
  346. static int soc15_set_uvd_clocks(struct amdgpu_device *adev, u32 vclk, u32 dclk)
  347. {
  348. /*int r;
  349. r = soc15_set_uvd_clock(adev, vclk, ixCG_VCLK_CNTL, ixCG_VCLK_STATUS);
  350. if (r)
  351. return r;
  352. r = soc15_set_uvd_clock(adev, dclk, ixCG_DCLK_CNTL, ixCG_DCLK_STATUS);
  353. */
  354. return 0;
  355. }
  356. static int soc15_set_vce_clocks(struct amdgpu_device *adev, u32 evclk, u32 ecclk)
  357. {
  358. /* todo */
  359. return 0;
  360. }
  361. static void soc15_pcie_gen3_enable(struct amdgpu_device *adev)
  362. {
  363. if (pci_is_root_bus(adev->pdev->bus))
  364. return;
  365. if (amdgpu_pcie_gen2 == 0)
  366. return;
  367. if (adev->flags & AMD_IS_APU)
  368. return;
  369. if (!(adev->pm.pcie_gen_mask & (CAIL_PCIE_LINK_SPEED_SUPPORT_GEN2 |
  370. CAIL_PCIE_LINK_SPEED_SUPPORT_GEN3)))
  371. return;
  372. /* todo */
  373. }
  374. static void soc15_program_aspm(struct amdgpu_device *adev)
  375. {
  376. if (amdgpu_aspm == 0)
  377. return;
  378. /* todo */
  379. }
  380. static void soc15_enable_doorbell_aperture(struct amdgpu_device *adev,
  381. bool enable)
  382. {
  383. if (adev->flags & AMD_IS_APU) {
  384. nbio_v7_0_enable_doorbell_aperture(adev, enable);
  385. } else {
  386. nbio_v6_1_enable_doorbell_aperture(adev, enable);
  387. nbio_v6_1_enable_doorbell_selfring_aperture(adev, enable);
  388. }
  389. }
  390. static const struct amdgpu_ip_block_version vega10_common_ip_block =
  391. {
  392. .type = AMD_IP_BLOCK_TYPE_COMMON,
  393. .major = 2,
  394. .minor = 0,
  395. .rev = 0,
  396. .funcs = &soc15_common_ip_funcs,
  397. };
  398. int soc15_set_ip_blocks(struct amdgpu_device *adev)
  399. {
  400. nbio_v6_1_detect_hw_virt(adev);
  401. if (amdgpu_sriov_vf(adev))
  402. adev->virt.ops = &xgpu_ai_virt_ops;
  403. switch (adev->asic_type) {
  404. case CHIP_VEGA10:
  405. amdgpu_ip_block_add(adev, &vega10_common_ip_block);
  406. amdgpu_ip_block_add(adev, &gmc_v9_0_ip_block);
  407. amdgpu_ip_block_add(adev, &vega10_ih_ip_block);
  408. if (amdgpu_fw_load_type == 2 || amdgpu_fw_load_type == -1)
  409. amdgpu_ip_block_add(adev, &psp_v3_1_ip_block);
  410. if (!amdgpu_sriov_vf(adev))
  411. amdgpu_ip_block_add(adev, &amdgpu_pp_ip_block);
  412. if (adev->enable_virtual_display || amdgpu_sriov_vf(adev))
  413. amdgpu_ip_block_add(adev, &dce_virtual_ip_block);
  414. amdgpu_ip_block_add(adev, &gfx_v9_0_ip_block);
  415. amdgpu_ip_block_add(adev, &sdma_v4_0_ip_block);
  416. amdgpu_ip_block_add(adev, &uvd_v7_0_ip_block);
  417. amdgpu_ip_block_add(adev, &vce_v4_0_ip_block);
  418. break;
  419. case CHIP_RAVEN:
  420. amdgpu_ip_block_add(adev, &vega10_common_ip_block);
  421. amdgpu_ip_block_add(adev, &gmc_v9_0_ip_block);
  422. amdgpu_ip_block_add(adev, &vega10_ih_ip_block);
  423. amdgpu_ip_block_add(adev, &psp_v10_0_ip_block);
  424. amdgpu_ip_block_add(adev, &amdgpu_pp_ip_block);
  425. if (adev->enable_virtual_display || amdgpu_sriov_vf(adev))
  426. amdgpu_ip_block_add(adev, &dce_virtual_ip_block);
  427. amdgpu_ip_block_add(adev, &gfx_v9_0_ip_block);
  428. amdgpu_ip_block_add(adev, &sdma_v4_0_ip_block);
  429. amdgpu_ip_block_add(adev, &vcn_v1_0_ip_block);
  430. break;
  431. default:
  432. return -EINVAL;
  433. }
  434. return 0;
  435. }
  436. static uint32_t soc15_get_rev_id(struct amdgpu_device *adev)
  437. {
  438. if (adev->flags & AMD_IS_APU)
  439. return nbio_v7_0_get_rev_id(adev);
  440. else
  441. return nbio_v6_1_get_rev_id(adev);
  442. }
  443. int gmc_v9_0_mc_wait_for_idle(struct amdgpu_device *adev)
  444. {
  445. /* to be implemented in MC IP*/
  446. return 0;
  447. }
  448. static const struct amdgpu_asic_funcs soc15_asic_funcs =
  449. {
  450. .read_disabled_bios = &soc15_read_disabled_bios,
  451. .read_bios_from_rom = &soc15_read_bios_from_rom,
  452. .read_register = &soc15_read_register,
  453. .reset = &soc15_asic_reset,
  454. .set_vga_state = &soc15_vga_set_state,
  455. .get_xclk = &soc15_get_xclk,
  456. .set_uvd_clocks = &soc15_set_uvd_clocks,
  457. .set_vce_clocks = &soc15_set_vce_clocks,
  458. .get_config_memsize = &soc15_get_config_memsize,
  459. };
  460. static int soc15_common_early_init(void *handle)
  461. {
  462. bool psp_enabled = false;
  463. struct amdgpu_device *adev = (struct amdgpu_device *)handle;
  464. adev->smc_rreg = NULL;
  465. adev->smc_wreg = NULL;
  466. adev->pcie_rreg = &soc15_pcie_rreg;
  467. adev->pcie_wreg = &soc15_pcie_wreg;
  468. adev->uvd_ctx_rreg = &soc15_uvd_ctx_rreg;
  469. adev->uvd_ctx_wreg = &soc15_uvd_ctx_wreg;
  470. adev->didt_rreg = &soc15_didt_rreg;
  471. adev->didt_wreg = &soc15_didt_wreg;
  472. adev->asic_funcs = &soc15_asic_funcs;
  473. if (amdgpu_get_ip_block(adev, AMD_IP_BLOCK_TYPE_PSP) &&
  474. (amdgpu_ip_block_mask & (1 << AMD_IP_BLOCK_TYPE_PSP)))
  475. psp_enabled = true;
  476. /*
  477. * nbio need be used for both sdma and gfx9, but only
  478. * initializes once
  479. */
  480. switch(adev->asic_type) {
  481. case CHIP_VEGA10:
  482. nbio_v6_1_init(adev);
  483. break;
  484. case CHIP_RAVEN:
  485. nbio_v7_0_init(adev);
  486. break;
  487. default:
  488. return -EINVAL;
  489. }
  490. adev->rev_id = soc15_get_rev_id(adev);
  491. adev->external_rev_id = 0xFF;
  492. switch (adev->asic_type) {
  493. case CHIP_VEGA10:
  494. adev->cg_flags = AMD_CG_SUPPORT_GFX_MGCG |
  495. AMD_CG_SUPPORT_GFX_MGLS |
  496. AMD_CG_SUPPORT_GFX_RLC_LS |
  497. AMD_CG_SUPPORT_GFX_CP_LS |
  498. AMD_CG_SUPPORT_GFX_3D_CGCG |
  499. AMD_CG_SUPPORT_GFX_3D_CGLS |
  500. AMD_CG_SUPPORT_GFX_CGCG |
  501. AMD_CG_SUPPORT_GFX_CGLS |
  502. AMD_CG_SUPPORT_BIF_MGCG |
  503. AMD_CG_SUPPORT_BIF_LS |
  504. AMD_CG_SUPPORT_HDP_LS |
  505. AMD_CG_SUPPORT_DRM_MGCG |
  506. AMD_CG_SUPPORT_DRM_LS |
  507. AMD_CG_SUPPORT_ROM_MGCG |
  508. AMD_CG_SUPPORT_DF_MGCG |
  509. AMD_CG_SUPPORT_SDMA_MGCG |
  510. AMD_CG_SUPPORT_SDMA_LS |
  511. AMD_CG_SUPPORT_MC_MGCG |
  512. AMD_CG_SUPPORT_MC_LS;
  513. adev->pg_flags = 0;
  514. adev->external_rev_id = 0x1;
  515. break;
  516. case CHIP_RAVEN:
  517. adev->cg_flags = AMD_CG_SUPPORT_GFX_MGCG |
  518. AMD_CG_SUPPORT_GFX_MGLS |
  519. AMD_CG_SUPPORT_GFX_RLC_LS |
  520. AMD_CG_SUPPORT_GFX_CP_LS |
  521. AMD_CG_SUPPORT_GFX_3D_CGCG |
  522. AMD_CG_SUPPORT_GFX_3D_CGLS |
  523. AMD_CG_SUPPORT_GFX_CGCG |
  524. AMD_CG_SUPPORT_GFX_CGLS |
  525. AMD_CG_SUPPORT_BIF_MGCG |
  526. AMD_CG_SUPPORT_BIF_LS |
  527. AMD_CG_SUPPORT_HDP_MGCG |
  528. AMD_CG_SUPPORT_HDP_LS |
  529. AMD_CG_SUPPORT_DRM_MGCG |
  530. AMD_CG_SUPPORT_DRM_LS |
  531. AMD_CG_SUPPORT_ROM_MGCG |
  532. AMD_CG_SUPPORT_MC_MGCG |
  533. AMD_CG_SUPPORT_MC_LS |
  534. AMD_CG_SUPPORT_SDMA_MGCG |
  535. AMD_CG_SUPPORT_SDMA_LS;
  536. adev->pg_flags = AMD_PG_SUPPORT_SDMA |
  537. AMD_PG_SUPPORT_MMHUB;
  538. adev->external_rev_id = 0x1;
  539. break;
  540. default:
  541. /* FIXME: not supported yet */
  542. return -EINVAL;
  543. }
  544. if (amdgpu_sriov_vf(adev)) {
  545. amdgpu_virt_init_setting(adev);
  546. xgpu_ai_mailbox_set_irq_funcs(adev);
  547. }
  548. adev->firmware.load_type = amdgpu_ucode_get_load_type(adev, amdgpu_fw_load_type);
  549. amdgpu_get_pcie_info(adev);
  550. return 0;
  551. }
  552. static int soc15_common_late_init(void *handle)
  553. {
  554. struct amdgpu_device *adev = (struct amdgpu_device *)handle;
  555. if (amdgpu_sriov_vf(adev))
  556. xgpu_ai_mailbox_get_irq(adev);
  557. return 0;
  558. }
  559. static int soc15_common_sw_init(void *handle)
  560. {
  561. struct amdgpu_device *adev = (struct amdgpu_device *)handle;
  562. if (amdgpu_sriov_vf(adev))
  563. xgpu_ai_mailbox_add_irq_id(adev);
  564. return 0;
  565. }
  566. static int soc15_common_sw_fini(void *handle)
  567. {
  568. return 0;
  569. }
  570. static int soc15_common_hw_init(void *handle)
  571. {
  572. struct amdgpu_device *adev = (struct amdgpu_device *)handle;
  573. /* move the golden regs per IP block */
  574. soc15_init_golden_registers(adev);
  575. /* enable pcie gen2/3 link */
  576. soc15_pcie_gen3_enable(adev);
  577. /* enable aspm */
  578. soc15_program_aspm(adev);
  579. /* enable the doorbell aperture */
  580. soc15_enable_doorbell_aperture(adev, true);
  581. return 0;
  582. }
  583. static int soc15_common_hw_fini(void *handle)
  584. {
  585. struct amdgpu_device *adev = (struct amdgpu_device *)handle;
  586. /* disable the doorbell aperture */
  587. soc15_enable_doorbell_aperture(adev, false);
  588. if (amdgpu_sriov_vf(adev))
  589. xgpu_ai_mailbox_put_irq(adev);
  590. return 0;
  591. }
  592. static int soc15_common_suspend(void *handle)
  593. {
  594. struct amdgpu_device *adev = (struct amdgpu_device *)handle;
  595. return soc15_common_hw_fini(adev);
  596. }
  597. static int soc15_common_resume(void *handle)
  598. {
  599. struct amdgpu_device *adev = (struct amdgpu_device *)handle;
  600. return soc15_common_hw_init(adev);
  601. }
  602. static bool soc15_common_is_idle(void *handle)
  603. {
  604. return true;
  605. }
  606. static int soc15_common_wait_for_idle(void *handle)
  607. {
  608. return 0;
  609. }
  610. static int soc15_common_soft_reset(void *handle)
  611. {
  612. return 0;
  613. }
  614. static void soc15_update_hdp_light_sleep(struct amdgpu_device *adev, bool enable)
  615. {
  616. uint32_t def, data;
  617. def = data = RREG32(SOC15_REG_OFFSET(HDP, 0, mmHDP_MEM_POWER_LS));
  618. if (enable && (adev->cg_flags & AMD_CG_SUPPORT_HDP_LS))
  619. data |= HDP_MEM_POWER_LS__LS_ENABLE_MASK;
  620. else
  621. data &= ~HDP_MEM_POWER_LS__LS_ENABLE_MASK;
  622. if (def != data)
  623. WREG32(SOC15_REG_OFFSET(HDP, 0, mmHDP_MEM_POWER_LS), data);
  624. }
  625. static void soc15_update_drm_clock_gating(struct amdgpu_device *adev, bool enable)
  626. {
  627. uint32_t def, data;
  628. def = data = RREG32(SOC15_REG_OFFSET(MP0, 0, mmMP0_MISC_CGTT_CTRL0));
  629. if (enable && (adev->cg_flags & AMD_CG_SUPPORT_DRM_MGCG))
  630. data &= ~(0x01000000 |
  631. 0x02000000 |
  632. 0x04000000 |
  633. 0x08000000 |
  634. 0x10000000 |
  635. 0x20000000 |
  636. 0x40000000 |
  637. 0x80000000);
  638. else
  639. data |= (0x01000000 |
  640. 0x02000000 |
  641. 0x04000000 |
  642. 0x08000000 |
  643. 0x10000000 |
  644. 0x20000000 |
  645. 0x40000000 |
  646. 0x80000000);
  647. if (def != data)
  648. WREG32(SOC15_REG_OFFSET(MP0, 0, mmMP0_MISC_CGTT_CTRL0), data);
  649. }
  650. static void soc15_update_drm_light_sleep(struct amdgpu_device *adev, bool enable)
  651. {
  652. uint32_t def, data;
  653. def = data = RREG32(SOC15_REG_OFFSET(MP0, 0, mmMP0_MISC_LIGHT_SLEEP_CTRL));
  654. if (enable && (adev->cg_flags & AMD_CG_SUPPORT_DRM_LS))
  655. data |= 1;
  656. else
  657. data &= ~1;
  658. if (def != data)
  659. WREG32(SOC15_REG_OFFSET(MP0, 0, mmMP0_MISC_LIGHT_SLEEP_CTRL), data);
  660. }
  661. static void soc15_update_rom_medium_grain_clock_gating(struct amdgpu_device *adev,
  662. bool enable)
  663. {
  664. uint32_t def, data;
  665. def = data = RREG32(SOC15_REG_OFFSET(SMUIO, 0, mmCGTT_ROM_CLK_CTRL0));
  666. if (enable && (adev->cg_flags & AMD_CG_SUPPORT_ROM_MGCG))
  667. data &= ~(CGTT_ROM_CLK_CTRL0__SOFT_OVERRIDE0_MASK |
  668. CGTT_ROM_CLK_CTRL0__SOFT_OVERRIDE1_MASK);
  669. else
  670. data |= CGTT_ROM_CLK_CTRL0__SOFT_OVERRIDE0_MASK |
  671. CGTT_ROM_CLK_CTRL0__SOFT_OVERRIDE1_MASK;
  672. if (def != data)
  673. WREG32(SOC15_REG_OFFSET(SMUIO, 0, mmCGTT_ROM_CLK_CTRL0), data);
  674. }
  675. static void soc15_update_df_medium_grain_clock_gating(struct amdgpu_device *adev,
  676. bool enable)
  677. {
  678. uint32_t data;
  679. /* Put DF on broadcast mode */
  680. data = RREG32(SOC15_REG_OFFSET(DF, 0, mmFabricConfigAccessControl));
  681. data &= ~FabricConfigAccessControl__CfgRegInstAccEn_MASK;
  682. WREG32(SOC15_REG_OFFSET(DF, 0, mmFabricConfigAccessControl), data);
  683. if (enable && (adev->cg_flags & AMD_CG_SUPPORT_DF_MGCG)) {
  684. data = RREG32(SOC15_REG_OFFSET(DF, 0, mmDF_PIE_AON0_DfGlobalClkGater));
  685. data &= ~DF_PIE_AON0_DfGlobalClkGater__MGCGMode_MASK;
  686. data |= DF_MGCG_ENABLE_15_CYCLE_DELAY;
  687. WREG32(SOC15_REG_OFFSET(DF, 0, mmDF_PIE_AON0_DfGlobalClkGater), data);
  688. } else {
  689. data = RREG32(SOC15_REG_OFFSET(DF, 0, mmDF_PIE_AON0_DfGlobalClkGater));
  690. data &= ~DF_PIE_AON0_DfGlobalClkGater__MGCGMode_MASK;
  691. data |= DF_MGCG_DISABLE;
  692. WREG32(SOC15_REG_OFFSET(DF, 0, mmDF_PIE_AON0_DfGlobalClkGater), data);
  693. }
  694. WREG32(SOC15_REG_OFFSET(DF, 0, mmFabricConfigAccessControl),
  695. mmFabricConfigAccessControl_DEFAULT);
  696. }
  697. static int soc15_common_set_clockgating_state(void *handle,
  698. enum amd_clockgating_state state)
  699. {
  700. struct amdgpu_device *adev = (struct amdgpu_device *)handle;
  701. if (amdgpu_sriov_vf(adev))
  702. return 0;
  703. switch (adev->asic_type) {
  704. case CHIP_VEGA10:
  705. nbio_v6_1_update_medium_grain_clock_gating(adev,
  706. state == AMD_CG_STATE_GATE ? true : false);
  707. nbio_v6_1_update_medium_grain_light_sleep(adev,
  708. state == AMD_CG_STATE_GATE ? true : false);
  709. soc15_update_hdp_light_sleep(adev,
  710. state == AMD_CG_STATE_GATE ? true : false);
  711. soc15_update_drm_clock_gating(adev,
  712. state == AMD_CG_STATE_GATE ? true : false);
  713. soc15_update_drm_light_sleep(adev,
  714. state == AMD_CG_STATE_GATE ? true : false);
  715. soc15_update_rom_medium_grain_clock_gating(adev,
  716. state == AMD_CG_STATE_GATE ? true : false);
  717. soc15_update_df_medium_grain_clock_gating(adev,
  718. state == AMD_CG_STATE_GATE ? true : false);
  719. break;
  720. case CHIP_RAVEN:
  721. nbio_v7_0_update_medium_grain_clock_gating(adev,
  722. state == AMD_CG_STATE_GATE ? true : false);
  723. nbio_v6_1_update_medium_grain_light_sleep(adev,
  724. state == AMD_CG_STATE_GATE ? true : false);
  725. soc15_update_hdp_light_sleep(adev,
  726. state == AMD_CG_STATE_GATE ? true : false);
  727. soc15_update_drm_clock_gating(adev,
  728. state == AMD_CG_STATE_GATE ? true : false);
  729. soc15_update_drm_light_sleep(adev,
  730. state == AMD_CG_STATE_GATE ? true : false);
  731. soc15_update_rom_medium_grain_clock_gating(adev,
  732. state == AMD_CG_STATE_GATE ? true : false);
  733. break;
  734. default:
  735. break;
  736. }
  737. return 0;
  738. }
  739. static void soc15_common_get_clockgating_state(void *handle, u32 *flags)
  740. {
  741. struct amdgpu_device *adev = (struct amdgpu_device *)handle;
  742. int data;
  743. if (amdgpu_sriov_vf(adev))
  744. *flags = 0;
  745. nbio_v6_1_get_clockgating_state(adev, flags);
  746. /* AMD_CG_SUPPORT_HDP_LS */
  747. data = RREG32(SOC15_REG_OFFSET(HDP, 0, mmHDP_MEM_POWER_LS));
  748. if (data & HDP_MEM_POWER_LS__LS_ENABLE_MASK)
  749. *flags |= AMD_CG_SUPPORT_HDP_LS;
  750. /* AMD_CG_SUPPORT_DRM_MGCG */
  751. data = RREG32(SOC15_REG_OFFSET(MP0, 0, mmMP0_MISC_CGTT_CTRL0));
  752. if (!(data & 0x01000000))
  753. *flags |= AMD_CG_SUPPORT_DRM_MGCG;
  754. /* AMD_CG_SUPPORT_DRM_LS */
  755. data = RREG32(SOC15_REG_OFFSET(MP0, 0, mmMP0_MISC_LIGHT_SLEEP_CTRL));
  756. if (data & 0x1)
  757. *flags |= AMD_CG_SUPPORT_DRM_LS;
  758. /* AMD_CG_SUPPORT_ROM_MGCG */
  759. data = RREG32(SOC15_REG_OFFSET(SMUIO, 0, mmCGTT_ROM_CLK_CTRL0));
  760. if (!(data & CGTT_ROM_CLK_CTRL0__SOFT_OVERRIDE0_MASK))
  761. *flags |= AMD_CG_SUPPORT_ROM_MGCG;
  762. /* AMD_CG_SUPPORT_DF_MGCG */
  763. data = RREG32(SOC15_REG_OFFSET(DF, 0, mmDF_PIE_AON0_DfGlobalClkGater));
  764. if (data & DF_MGCG_ENABLE_15_CYCLE_DELAY)
  765. *flags |= AMD_CG_SUPPORT_DF_MGCG;
  766. }
  767. static int soc15_common_set_powergating_state(void *handle,
  768. enum amd_powergating_state state)
  769. {
  770. /* todo */
  771. return 0;
  772. }
  773. const struct amd_ip_funcs soc15_common_ip_funcs = {
  774. .name = "soc15_common",
  775. .early_init = soc15_common_early_init,
  776. .late_init = soc15_common_late_init,
  777. .sw_init = soc15_common_sw_init,
  778. .sw_fini = soc15_common_sw_fini,
  779. .hw_init = soc15_common_hw_init,
  780. .hw_fini = soc15_common_hw_fini,
  781. .suspend = soc15_common_suspend,
  782. .resume = soc15_common_resume,
  783. .is_idle = soc15_common_is_idle,
  784. .wait_for_idle = soc15_common_wait_for_idle,
  785. .soft_reset = soc15_common_soft_reset,
  786. .set_clockgating_state = soc15_common_set_clockgating_state,
  787. .set_powergating_state = soc15_common_set_powergating_state,
  788. .get_clockgating_state= soc15_common_get_clockgating_state,
  789. };