amdgpu_vram_mgr.c 8.8 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349
  1. /*
  2. * Copyright 2016 Advanced Micro Devices, Inc.
  3. *
  4. * Permission is hereby granted, free of charge, to any person obtaining a
  5. * copy of this software and associated documentation files (the "Software"),
  6. * to deal in the Software without restriction, including without limitation
  7. * the rights to use, copy, modify, merge, publish, distribute, sublicense,
  8. * and/or sell copies of the Software, and to permit persons to whom the
  9. * Software is furnished to do so, subject to the following conditions:
  10. *
  11. * The above copyright notice and this permission notice shall be included in
  12. * all copies or substantial portions of the Software.
  13. *
  14. * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
  15. * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
  16. * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
  17. * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
  18. * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
  19. * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
  20. * OTHER DEALINGS IN THE SOFTWARE.
  21. *
  22. * Authors: Christian König
  23. */
  24. #include <drm/drmP.h>
  25. #include "amdgpu.h"
  26. struct amdgpu_vram_mgr {
  27. struct drm_mm mm;
  28. spinlock_t lock;
  29. atomic64_t usage;
  30. atomic64_t vis_usage;
  31. };
  32. /**
  33. * amdgpu_vram_mgr_init - init VRAM manager and DRM MM
  34. *
  35. * @man: TTM memory type manager
  36. * @p_size: maximum size of VRAM
  37. *
  38. * Allocate and initialize the VRAM manager.
  39. */
  40. static int amdgpu_vram_mgr_init(struct ttm_mem_type_manager *man,
  41. unsigned long p_size)
  42. {
  43. struct amdgpu_vram_mgr *mgr;
  44. mgr = kzalloc(sizeof(*mgr), GFP_KERNEL);
  45. if (!mgr)
  46. return -ENOMEM;
  47. drm_mm_init(&mgr->mm, 0, p_size);
  48. spin_lock_init(&mgr->lock);
  49. man->priv = mgr;
  50. return 0;
  51. }
  52. /**
  53. * amdgpu_vram_mgr_fini - free and destroy VRAM manager
  54. *
  55. * @man: TTM memory type manager
  56. *
  57. * Destroy and free the VRAM manager, returns -EBUSY if ranges are still
  58. * allocated inside it.
  59. */
  60. static int amdgpu_vram_mgr_fini(struct ttm_mem_type_manager *man)
  61. {
  62. struct amdgpu_vram_mgr *mgr = man->priv;
  63. spin_lock(&mgr->lock);
  64. drm_mm_takedown(&mgr->mm);
  65. spin_unlock(&mgr->lock);
  66. kfree(mgr);
  67. man->priv = NULL;
  68. return 0;
  69. }
  70. /**
  71. * amdgpu_vram_mgr_vis_size - Calculate visible node size
  72. *
  73. * @adev: amdgpu device structure
  74. * @node: MM node structure
  75. *
  76. * Calculate how many bytes of the MM node are inside visible VRAM
  77. */
  78. static u64 amdgpu_vram_mgr_vis_size(struct amdgpu_device *adev,
  79. struct drm_mm_node *node)
  80. {
  81. uint64_t start = node->start << PAGE_SHIFT;
  82. uint64_t end = (node->size + node->start) << PAGE_SHIFT;
  83. if (start >= adev->gmc.visible_vram_size)
  84. return 0;
  85. return (end > adev->gmc.visible_vram_size ?
  86. adev->gmc.visible_vram_size : end) - start;
  87. }
  88. /**
  89. * amdgpu_vram_mgr_bo_visible_size - CPU visible BO size
  90. *
  91. * @bo: &amdgpu_bo buffer object (must be in VRAM)
  92. *
  93. * Returns:
  94. * How much of the given &amdgpu_bo buffer object lies in CPU visible VRAM.
  95. */
  96. u64 amdgpu_vram_mgr_bo_visible_size(struct amdgpu_bo *bo)
  97. {
  98. struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
  99. struct ttm_mem_reg *mem = &bo->tbo.mem;
  100. struct drm_mm_node *nodes = mem->mm_node;
  101. unsigned pages = mem->num_pages;
  102. u64 usage;
  103. if (amdgpu_gmc_vram_full_visible(&adev->gmc))
  104. return amdgpu_bo_size(bo);
  105. if (mem->start >= adev->gmc.visible_vram_size >> PAGE_SHIFT)
  106. return 0;
  107. for (usage = 0; nodes && pages; pages -= nodes->size, nodes++)
  108. usage += amdgpu_vram_mgr_vis_size(adev, nodes);
  109. return usage;
  110. }
  111. /**
  112. * amdgpu_vram_mgr_virt_start - update virtual start address
  113. *
  114. * @mem: ttm_mem_reg to update
  115. * @node: just allocated node
  116. *
  117. * Calculate a virtual BO start address to easily check if everything is CPU
  118. * accessible.
  119. */
  120. static void amdgpu_vram_mgr_virt_start(struct ttm_mem_reg *mem,
  121. struct drm_mm_node *node)
  122. {
  123. unsigned long start;
  124. start = node->start + node->size;
  125. if (start > mem->num_pages)
  126. start -= mem->num_pages;
  127. else
  128. start = 0;
  129. mem->start = max(mem->start, start);
  130. }
  131. /**
  132. * amdgpu_vram_mgr_new - allocate new ranges
  133. *
  134. * @man: TTM memory type manager
  135. * @tbo: TTM BO we need this range for
  136. * @place: placement flags and restrictions
  137. * @mem: the resulting mem object
  138. *
  139. * Allocate VRAM for the given BO.
  140. */
  141. static int amdgpu_vram_mgr_new(struct ttm_mem_type_manager *man,
  142. struct ttm_buffer_object *tbo,
  143. const struct ttm_place *place,
  144. struct ttm_mem_reg *mem)
  145. {
  146. struct amdgpu_device *adev = amdgpu_ttm_adev(man->bdev);
  147. struct amdgpu_vram_mgr *mgr = man->priv;
  148. struct drm_mm *mm = &mgr->mm;
  149. struct drm_mm_node *nodes;
  150. enum drm_mm_insert_mode mode;
  151. unsigned long lpfn, num_nodes, pages_per_node, pages_left;
  152. uint64_t usage = 0, vis_usage = 0;
  153. unsigned i;
  154. int r;
  155. lpfn = place->lpfn;
  156. if (!lpfn)
  157. lpfn = man->size;
  158. if (place->flags & TTM_PL_FLAG_CONTIGUOUS ||
  159. amdgpu_vram_page_split == -1) {
  160. pages_per_node = ~0ul;
  161. num_nodes = 1;
  162. } else {
  163. pages_per_node = max((uint32_t)amdgpu_vram_page_split,
  164. mem->page_alignment);
  165. num_nodes = DIV_ROUND_UP(mem->num_pages, pages_per_node);
  166. }
  167. nodes = kvmalloc_array(num_nodes, sizeof(*nodes),
  168. GFP_KERNEL | __GFP_ZERO);
  169. if (!nodes)
  170. return -ENOMEM;
  171. mode = DRM_MM_INSERT_BEST;
  172. if (place->flags & TTM_PL_FLAG_TOPDOWN)
  173. mode = DRM_MM_INSERT_HIGH;
  174. mem->start = 0;
  175. pages_left = mem->num_pages;
  176. spin_lock(&mgr->lock);
  177. for (i = 0; pages_left >= pages_per_node; ++i) {
  178. unsigned long pages = rounddown_pow_of_two(pages_left);
  179. r = drm_mm_insert_node_in_range(mm, &nodes[i], pages,
  180. pages_per_node, 0,
  181. place->fpfn, lpfn,
  182. mode);
  183. if (unlikely(r))
  184. break;
  185. usage += nodes[i].size << PAGE_SHIFT;
  186. vis_usage += amdgpu_vram_mgr_vis_size(adev, &nodes[i]);
  187. amdgpu_vram_mgr_virt_start(mem, &nodes[i]);
  188. pages_left -= pages;
  189. }
  190. for (; pages_left; ++i) {
  191. unsigned long pages = min(pages_left, pages_per_node);
  192. uint32_t alignment = mem->page_alignment;
  193. if (pages == pages_per_node)
  194. alignment = pages_per_node;
  195. r = drm_mm_insert_node_in_range(mm, &nodes[i],
  196. pages, alignment, 0,
  197. place->fpfn, lpfn,
  198. mode);
  199. if (unlikely(r))
  200. goto error;
  201. usage += nodes[i].size << PAGE_SHIFT;
  202. vis_usage += amdgpu_vram_mgr_vis_size(adev, &nodes[i]);
  203. amdgpu_vram_mgr_virt_start(mem, &nodes[i]);
  204. pages_left -= pages;
  205. }
  206. spin_unlock(&mgr->lock);
  207. atomic64_add(usage, &mgr->usage);
  208. atomic64_add(vis_usage, &mgr->vis_usage);
  209. mem->mm_node = nodes;
  210. return 0;
  211. error:
  212. while (i--)
  213. drm_mm_remove_node(&nodes[i]);
  214. spin_unlock(&mgr->lock);
  215. kvfree(nodes);
  216. return r == -ENOSPC ? 0 : r;
  217. }
  218. /**
  219. * amdgpu_vram_mgr_del - free ranges
  220. *
  221. * @man: TTM memory type manager
  222. * @tbo: TTM BO we need this range for
  223. * @place: placement flags and restrictions
  224. * @mem: TTM memory object
  225. *
  226. * Free the allocated VRAM again.
  227. */
  228. static void amdgpu_vram_mgr_del(struct ttm_mem_type_manager *man,
  229. struct ttm_mem_reg *mem)
  230. {
  231. struct amdgpu_device *adev = amdgpu_ttm_adev(man->bdev);
  232. struct amdgpu_vram_mgr *mgr = man->priv;
  233. struct drm_mm_node *nodes = mem->mm_node;
  234. uint64_t usage = 0, vis_usage = 0;
  235. unsigned pages = mem->num_pages;
  236. if (!mem->mm_node)
  237. return;
  238. spin_lock(&mgr->lock);
  239. while (pages) {
  240. pages -= nodes->size;
  241. drm_mm_remove_node(nodes);
  242. usage += nodes->size << PAGE_SHIFT;
  243. vis_usage += amdgpu_vram_mgr_vis_size(adev, nodes);
  244. ++nodes;
  245. }
  246. spin_unlock(&mgr->lock);
  247. atomic64_sub(usage, &mgr->usage);
  248. atomic64_sub(vis_usage, &mgr->vis_usage);
  249. kvfree(mem->mm_node);
  250. mem->mm_node = NULL;
  251. }
  252. /**
  253. * amdgpu_vram_mgr_usage - how many bytes are used in this domain
  254. *
  255. * @man: TTM memory type manager
  256. *
  257. * Returns how many bytes are used in this domain.
  258. */
  259. uint64_t amdgpu_vram_mgr_usage(struct ttm_mem_type_manager *man)
  260. {
  261. struct amdgpu_vram_mgr *mgr = man->priv;
  262. return atomic64_read(&mgr->usage);
  263. }
  264. /**
  265. * amdgpu_vram_mgr_vis_usage - how many bytes are used in the visible part
  266. *
  267. * @man: TTM memory type manager
  268. *
  269. * Returns how many bytes are used in the visible part of VRAM
  270. */
  271. uint64_t amdgpu_vram_mgr_vis_usage(struct ttm_mem_type_manager *man)
  272. {
  273. struct amdgpu_vram_mgr *mgr = man->priv;
  274. return atomic64_read(&mgr->vis_usage);
  275. }
  276. /**
  277. * amdgpu_vram_mgr_debug - dump VRAM table
  278. *
  279. * @man: TTM memory type manager
  280. * @printer: DRM printer to use
  281. *
  282. * Dump the table content using printk.
  283. */
  284. static void amdgpu_vram_mgr_debug(struct ttm_mem_type_manager *man,
  285. struct drm_printer *printer)
  286. {
  287. struct amdgpu_vram_mgr *mgr = man->priv;
  288. spin_lock(&mgr->lock);
  289. drm_mm_print(&mgr->mm, printer);
  290. spin_unlock(&mgr->lock);
  291. drm_printf(printer, "man size:%llu pages, ram usage:%lluMB, vis usage:%lluMB\n",
  292. man->size, amdgpu_vram_mgr_usage(man) >> 20,
  293. amdgpu_vram_mgr_vis_usage(man) >> 20);
  294. }
  295. const struct ttm_mem_type_manager_func amdgpu_vram_mgr_func = {
  296. .init = amdgpu_vram_mgr_init,
  297. .takedown = amdgpu_vram_mgr_fini,
  298. .get_node = amdgpu_vram_mgr_new,
  299. .put_node = amdgpu_vram_mgr_del,
  300. .debug = amdgpu_vram_mgr_debug
  301. };