gtt.c 60 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372137313741375137613771378137913801381138213831384138513861387138813891390139113921393139413951396139713981399140014011402140314041405140614071408140914101411141214131414141514161417141814191420142114221423142414251426142714281429143014311432143314341435143614371438143914401441144214431444144514461447144814491450145114521453145414551456145714581459146014611462146314641465146614671468146914701471147214731474147514761477147814791480148114821483148414851486148714881489149014911492149314941495149614971498149915001501150215031504150515061507150815091510151115121513151415151516151715181519152015211522152315241525152615271528152915301531153215331534153515361537153815391540154115421543154415451546154715481549155015511552155315541555155615571558155915601561156215631564156515661567156815691570157115721573157415751576157715781579158015811582158315841585158615871588158915901591159215931594159515961597159815991600160116021603160416051606160716081609161016111612161316141615161616171618161916201621162216231624162516261627162816291630163116321633163416351636163716381639164016411642164316441645164616471648164916501651165216531654165516561657165816591660166116621663166416651666166716681669167016711672167316741675167616771678167916801681168216831684168516861687168816891690169116921693169416951696169716981699170017011702170317041705170617071708170917101711171217131714171517161717171817191720172117221723172417251726172717281729173017311732173317341735173617371738173917401741174217431744174517461747174817491750175117521753175417551756175717581759176017611762176317641765176617671768176917701771177217731774177517761777177817791780178117821783178417851786178717881789179017911792179317941795179617971798179918001801180218031804180518061807180818091810181118121813181418151816181718181819182018211822182318241825182618271828182918301831183218331834183518361837183818391840184118421843184418451846184718481849185018511852185318541855185618571858185918601861186218631864186518661867186818691870187118721873187418751876187718781879188018811882188318841885188618871888188918901891189218931894189518961897189818991900190119021903190419051906190719081909191019111912191319141915191619171918191919201921192219231924192519261927192819291930193119321933193419351936193719381939194019411942194319441945194619471948194919501951195219531954195519561957195819591960196119621963196419651966196719681969197019711972197319741975197619771978197919801981198219831984198519861987198819891990199119921993199419951996199719981999200020012002200320042005200620072008200920102011201220132014201520162017201820192020202120222023202420252026202720282029203020312032203320342035203620372038203920402041204220432044204520462047204820492050205120522053205420552056205720582059206020612062206320642065206620672068206920702071207220732074207520762077207820792080208120822083208420852086208720882089209020912092209320942095209620972098209921002101210221032104210521062107210821092110211121122113211421152116211721182119212021212122212321242125212621272128212921302131213221332134213521362137213821392140214121422143214421452146214721482149215021512152215321542155215621572158215921602161216221632164216521662167216821692170217121722173217421752176217721782179218021812182218321842185218621872188218921902191219221932194219521962197219821992200220122022203220422052206220722082209221022112212221322142215221622172218221922202221222222232224222522262227222822292230223122322233223422352236223722382239224022412242224322442245224622472248224922502251225222532254225522562257225822592260226122622263226422652266226722682269227022712272227322742275227622772278227922802281228222832284228522862287228822892290229122922293229422952296229722982299230023012302230323042305230623072308230923102311231223132314231523162317231823192320232123222323232423252326232723282329233023312332233323342335233623372338233923402341234223432344234523462347234823492350235123522353235423552356
  1. /*
  2. * GTT virtualization
  3. *
  4. * Copyright(c) 2011-2016 Intel Corporation. All rights reserved.
  5. *
  6. * Permission is hereby granted, free of charge, to any person obtaining a
  7. * copy of this software and associated documentation files (the "Software"),
  8. * to deal in the Software without restriction, including without limitation
  9. * the rights to use, copy, modify, merge, publish, distribute, sublicense,
  10. * and/or sell copies of the Software, and to permit persons to whom the
  11. * Software is furnished to do so, subject to the following conditions:
  12. *
  13. * The above copyright notice and this permission notice (including the next
  14. * paragraph) shall be included in all copies or substantial portions of the
  15. * Software.
  16. *
  17. * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
  18. * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
  19. * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
  20. * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
  21. * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
  22. * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
  23. * SOFTWARE.
  24. *
  25. * Authors:
  26. * Zhi Wang <zhi.a.wang@intel.com>
  27. * Zhenyu Wang <zhenyuw@linux.intel.com>
  28. * Xiao Zheng <xiao.zheng@intel.com>
  29. *
  30. * Contributors:
  31. * Min He <min.he@intel.com>
  32. * Bing Niu <bing.niu@intel.com>
  33. *
  34. */
  35. #include "i915_drv.h"
  36. #include "gvt.h"
  37. #include "i915_pvinfo.h"
  38. #include "trace.h"
  39. static bool enable_out_of_sync = false;
  40. static int preallocated_oos_pages = 8192;
  41. /*
  42. * validate a gm address and related range size,
  43. * translate it to host gm address
  44. */
  45. bool intel_gvt_ggtt_validate_range(struct intel_vgpu *vgpu, u64 addr, u32 size)
  46. {
  47. if ((!vgpu_gmadr_is_valid(vgpu, addr)) || (size
  48. && !vgpu_gmadr_is_valid(vgpu, addr + size - 1))) {
  49. gvt_vgpu_err("invalid range gmadr 0x%llx size 0x%x\n",
  50. addr, size);
  51. return false;
  52. }
  53. return true;
  54. }
  55. /* translate a guest gmadr to host gmadr */
  56. int intel_gvt_ggtt_gmadr_g2h(struct intel_vgpu *vgpu, u64 g_addr, u64 *h_addr)
  57. {
  58. if (WARN(!vgpu_gmadr_is_valid(vgpu, g_addr),
  59. "invalid guest gmadr %llx\n", g_addr))
  60. return -EACCES;
  61. if (vgpu_gmadr_is_aperture(vgpu, g_addr))
  62. *h_addr = vgpu_aperture_gmadr_base(vgpu)
  63. + (g_addr - vgpu_aperture_offset(vgpu));
  64. else
  65. *h_addr = vgpu_hidden_gmadr_base(vgpu)
  66. + (g_addr - vgpu_hidden_offset(vgpu));
  67. return 0;
  68. }
  69. /* translate a host gmadr to guest gmadr */
  70. int intel_gvt_ggtt_gmadr_h2g(struct intel_vgpu *vgpu, u64 h_addr, u64 *g_addr)
  71. {
  72. if (WARN(!gvt_gmadr_is_valid(vgpu->gvt, h_addr),
  73. "invalid host gmadr %llx\n", h_addr))
  74. return -EACCES;
  75. if (gvt_gmadr_is_aperture(vgpu->gvt, h_addr))
  76. *g_addr = vgpu_aperture_gmadr_base(vgpu)
  77. + (h_addr - gvt_aperture_gmadr_base(vgpu->gvt));
  78. else
  79. *g_addr = vgpu_hidden_gmadr_base(vgpu)
  80. + (h_addr - gvt_hidden_gmadr_base(vgpu->gvt));
  81. return 0;
  82. }
  83. int intel_gvt_ggtt_index_g2h(struct intel_vgpu *vgpu, unsigned long g_index,
  84. unsigned long *h_index)
  85. {
  86. u64 h_addr;
  87. int ret;
  88. ret = intel_gvt_ggtt_gmadr_g2h(vgpu, g_index << GTT_PAGE_SHIFT,
  89. &h_addr);
  90. if (ret)
  91. return ret;
  92. *h_index = h_addr >> GTT_PAGE_SHIFT;
  93. return 0;
  94. }
  95. int intel_gvt_ggtt_h2g_index(struct intel_vgpu *vgpu, unsigned long h_index,
  96. unsigned long *g_index)
  97. {
  98. u64 g_addr;
  99. int ret;
  100. ret = intel_gvt_ggtt_gmadr_h2g(vgpu, h_index << GTT_PAGE_SHIFT,
  101. &g_addr);
  102. if (ret)
  103. return ret;
  104. *g_index = g_addr >> GTT_PAGE_SHIFT;
  105. return 0;
  106. }
  107. #define gtt_type_is_entry(type) \
  108. (type > GTT_TYPE_INVALID && type < GTT_TYPE_PPGTT_ENTRY \
  109. && type != GTT_TYPE_PPGTT_PTE_ENTRY \
  110. && type != GTT_TYPE_PPGTT_ROOT_ENTRY)
  111. #define gtt_type_is_pt(type) \
  112. (type >= GTT_TYPE_PPGTT_PTE_PT && type < GTT_TYPE_MAX)
  113. #define gtt_type_is_pte_pt(type) \
  114. (type == GTT_TYPE_PPGTT_PTE_PT)
  115. #define gtt_type_is_root_pointer(type) \
  116. (gtt_type_is_entry(type) && type > GTT_TYPE_PPGTT_ROOT_ENTRY)
  117. #define gtt_init_entry(e, t, p, v) do { \
  118. (e)->type = t; \
  119. (e)->pdev = p; \
  120. memcpy(&(e)->val64, &v, sizeof(v)); \
  121. } while (0)
  122. /*
  123. * Mappings between GTT_TYPE* enumerations.
  124. * Following information can be found according to the given type:
  125. * - type of next level page table
  126. * - type of entry inside this level page table
  127. * - type of entry with PSE set
  128. *
  129. * If the given type doesn't have such a kind of information,
  130. * e.g. give a l4 root entry type, then request to get its PSE type,
  131. * give a PTE page table type, then request to get its next level page
  132. * table type, as we know l4 root entry doesn't have a PSE bit,
  133. * and a PTE page table doesn't have a next level page table type,
  134. * GTT_TYPE_INVALID will be returned. This is useful when traversing a
  135. * page table.
  136. */
  137. struct gtt_type_table_entry {
  138. int entry_type;
  139. int next_pt_type;
  140. int pse_entry_type;
  141. };
  142. #define GTT_TYPE_TABLE_ENTRY(type, e_type, npt_type, pse_type) \
  143. [type] = { \
  144. .entry_type = e_type, \
  145. .next_pt_type = npt_type, \
  146. .pse_entry_type = pse_type, \
  147. }
  148. static struct gtt_type_table_entry gtt_type_table[] = {
  149. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_PPGTT_ROOT_L4_ENTRY,
  150. GTT_TYPE_PPGTT_ROOT_L4_ENTRY,
  151. GTT_TYPE_PPGTT_PML4_PT,
  152. GTT_TYPE_INVALID),
  153. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_PPGTT_PML4_PT,
  154. GTT_TYPE_PPGTT_PML4_ENTRY,
  155. GTT_TYPE_PPGTT_PDP_PT,
  156. GTT_TYPE_INVALID),
  157. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_PPGTT_PML4_ENTRY,
  158. GTT_TYPE_PPGTT_PML4_ENTRY,
  159. GTT_TYPE_PPGTT_PDP_PT,
  160. GTT_TYPE_INVALID),
  161. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_PPGTT_PDP_PT,
  162. GTT_TYPE_PPGTT_PDP_ENTRY,
  163. GTT_TYPE_PPGTT_PDE_PT,
  164. GTT_TYPE_PPGTT_PTE_1G_ENTRY),
  165. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_PPGTT_ROOT_L3_ENTRY,
  166. GTT_TYPE_PPGTT_ROOT_L3_ENTRY,
  167. GTT_TYPE_PPGTT_PDE_PT,
  168. GTT_TYPE_PPGTT_PTE_1G_ENTRY),
  169. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_PPGTT_PDP_ENTRY,
  170. GTT_TYPE_PPGTT_PDP_ENTRY,
  171. GTT_TYPE_PPGTT_PDE_PT,
  172. GTT_TYPE_PPGTT_PTE_1G_ENTRY),
  173. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_PPGTT_PDE_PT,
  174. GTT_TYPE_PPGTT_PDE_ENTRY,
  175. GTT_TYPE_PPGTT_PTE_PT,
  176. GTT_TYPE_PPGTT_PTE_2M_ENTRY),
  177. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_PPGTT_PDE_ENTRY,
  178. GTT_TYPE_PPGTT_PDE_ENTRY,
  179. GTT_TYPE_PPGTT_PTE_PT,
  180. GTT_TYPE_PPGTT_PTE_2M_ENTRY),
  181. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_PPGTT_PTE_PT,
  182. GTT_TYPE_PPGTT_PTE_4K_ENTRY,
  183. GTT_TYPE_INVALID,
  184. GTT_TYPE_INVALID),
  185. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_PPGTT_PTE_4K_ENTRY,
  186. GTT_TYPE_PPGTT_PTE_4K_ENTRY,
  187. GTT_TYPE_INVALID,
  188. GTT_TYPE_INVALID),
  189. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_PPGTT_PTE_2M_ENTRY,
  190. GTT_TYPE_PPGTT_PDE_ENTRY,
  191. GTT_TYPE_INVALID,
  192. GTT_TYPE_PPGTT_PTE_2M_ENTRY),
  193. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_PPGTT_PTE_1G_ENTRY,
  194. GTT_TYPE_PPGTT_PDP_ENTRY,
  195. GTT_TYPE_INVALID,
  196. GTT_TYPE_PPGTT_PTE_1G_ENTRY),
  197. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_GGTT_PTE,
  198. GTT_TYPE_GGTT_PTE,
  199. GTT_TYPE_INVALID,
  200. GTT_TYPE_INVALID),
  201. };
  202. static inline int get_next_pt_type(int type)
  203. {
  204. return gtt_type_table[type].next_pt_type;
  205. }
  206. static inline int get_entry_type(int type)
  207. {
  208. return gtt_type_table[type].entry_type;
  209. }
  210. static inline int get_pse_type(int type)
  211. {
  212. return gtt_type_table[type].pse_entry_type;
  213. }
  214. static u64 read_pte64(struct drm_i915_private *dev_priv, unsigned long index)
  215. {
  216. void __iomem *addr = (gen8_pte_t __iomem *)dev_priv->ggtt.gsm + index;
  217. return readq(addr);
  218. }
  219. static void write_pte64(struct drm_i915_private *dev_priv,
  220. unsigned long index, u64 pte)
  221. {
  222. void __iomem *addr = (gen8_pte_t __iomem *)dev_priv->ggtt.gsm + index;
  223. writeq(pte, addr);
  224. I915_WRITE(GFX_FLSH_CNTL_GEN6, GFX_FLSH_CNTL_EN);
  225. POSTING_READ(GFX_FLSH_CNTL_GEN6);
  226. }
  227. static inline struct intel_gvt_gtt_entry *gtt_get_entry64(void *pt,
  228. struct intel_gvt_gtt_entry *e,
  229. unsigned long index, bool hypervisor_access, unsigned long gpa,
  230. struct intel_vgpu *vgpu)
  231. {
  232. const struct intel_gvt_device_info *info = &vgpu->gvt->device_info;
  233. int ret;
  234. if (WARN_ON(info->gtt_entry_size != 8))
  235. return e;
  236. if (hypervisor_access) {
  237. ret = intel_gvt_hypervisor_read_gpa(vgpu, gpa +
  238. (index << info->gtt_entry_size_shift),
  239. &e->val64, 8);
  240. WARN_ON(ret);
  241. } else if (!pt) {
  242. e->val64 = read_pte64(vgpu->gvt->dev_priv, index);
  243. } else {
  244. e->val64 = *((u64 *)pt + index);
  245. }
  246. return e;
  247. }
  248. static inline struct intel_gvt_gtt_entry *gtt_set_entry64(void *pt,
  249. struct intel_gvt_gtt_entry *e,
  250. unsigned long index, bool hypervisor_access, unsigned long gpa,
  251. struct intel_vgpu *vgpu)
  252. {
  253. const struct intel_gvt_device_info *info = &vgpu->gvt->device_info;
  254. int ret;
  255. if (WARN_ON(info->gtt_entry_size != 8))
  256. return e;
  257. if (hypervisor_access) {
  258. ret = intel_gvt_hypervisor_write_gpa(vgpu, gpa +
  259. (index << info->gtt_entry_size_shift),
  260. &e->val64, 8);
  261. WARN_ON(ret);
  262. } else if (!pt) {
  263. write_pte64(vgpu->gvt->dev_priv, index, e->val64);
  264. } else {
  265. *((u64 *)pt + index) = e->val64;
  266. }
  267. return e;
  268. }
  269. #define GTT_HAW 46
  270. #define ADDR_1G_MASK (((1UL << (GTT_HAW - 30 + 1)) - 1) << 30)
  271. #define ADDR_2M_MASK (((1UL << (GTT_HAW - 21 + 1)) - 1) << 21)
  272. #define ADDR_4K_MASK (((1UL << (GTT_HAW - 12 + 1)) - 1) << 12)
  273. static unsigned long gen8_gtt_get_pfn(struct intel_gvt_gtt_entry *e)
  274. {
  275. unsigned long pfn;
  276. if (e->type == GTT_TYPE_PPGTT_PTE_1G_ENTRY)
  277. pfn = (e->val64 & ADDR_1G_MASK) >> 12;
  278. else if (e->type == GTT_TYPE_PPGTT_PTE_2M_ENTRY)
  279. pfn = (e->val64 & ADDR_2M_MASK) >> 12;
  280. else
  281. pfn = (e->val64 & ADDR_4K_MASK) >> 12;
  282. return pfn;
  283. }
  284. static void gen8_gtt_set_pfn(struct intel_gvt_gtt_entry *e, unsigned long pfn)
  285. {
  286. if (e->type == GTT_TYPE_PPGTT_PTE_1G_ENTRY) {
  287. e->val64 &= ~ADDR_1G_MASK;
  288. pfn &= (ADDR_1G_MASK >> 12);
  289. } else if (e->type == GTT_TYPE_PPGTT_PTE_2M_ENTRY) {
  290. e->val64 &= ~ADDR_2M_MASK;
  291. pfn &= (ADDR_2M_MASK >> 12);
  292. } else {
  293. e->val64 &= ~ADDR_4K_MASK;
  294. pfn &= (ADDR_4K_MASK >> 12);
  295. }
  296. e->val64 |= (pfn << 12);
  297. }
  298. static bool gen8_gtt_test_pse(struct intel_gvt_gtt_entry *e)
  299. {
  300. /* Entry doesn't have PSE bit. */
  301. if (get_pse_type(e->type) == GTT_TYPE_INVALID)
  302. return false;
  303. e->type = get_entry_type(e->type);
  304. if (!(e->val64 & (1 << 7)))
  305. return false;
  306. e->type = get_pse_type(e->type);
  307. return true;
  308. }
  309. static bool gen8_gtt_test_present(struct intel_gvt_gtt_entry *e)
  310. {
  311. /*
  312. * i915 writes PDP root pointer registers without present bit,
  313. * it also works, so we need to treat root pointer entry
  314. * specifically.
  315. */
  316. if (e->type == GTT_TYPE_PPGTT_ROOT_L3_ENTRY
  317. || e->type == GTT_TYPE_PPGTT_ROOT_L4_ENTRY)
  318. return (e->val64 != 0);
  319. else
  320. return (e->val64 & (1 << 0));
  321. }
  322. static void gtt_entry_clear_present(struct intel_gvt_gtt_entry *e)
  323. {
  324. e->val64 &= ~(1 << 0);
  325. }
  326. /*
  327. * Per-platform GMA routines.
  328. */
  329. static unsigned long gma_to_ggtt_pte_index(unsigned long gma)
  330. {
  331. unsigned long x = (gma >> GTT_PAGE_SHIFT);
  332. trace_gma_index(__func__, gma, x);
  333. return x;
  334. }
  335. #define DEFINE_PPGTT_GMA_TO_INDEX(prefix, ename, exp) \
  336. static unsigned long prefix##_gma_to_##ename##_index(unsigned long gma) \
  337. { \
  338. unsigned long x = (exp); \
  339. trace_gma_index(__func__, gma, x); \
  340. return x; \
  341. }
  342. DEFINE_PPGTT_GMA_TO_INDEX(gen8, pte, (gma >> 12 & 0x1ff));
  343. DEFINE_PPGTT_GMA_TO_INDEX(gen8, pde, (gma >> 21 & 0x1ff));
  344. DEFINE_PPGTT_GMA_TO_INDEX(gen8, l3_pdp, (gma >> 30 & 0x3));
  345. DEFINE_PPGTT_GMA_TO_INDEX(gen8, l4_pdp, (gma >> 30 & 0x1ff));
  346. DEFINE_PPGTT_GMA_TO_INDEX(gen8, pml4, (gma >> 39 & 0x1ff));
  347. static struct intel_gvt_gtt_pte_ops gen8_gtt_pte_ops = {
  348. .get_entry = gtt_get_entry64,
  349. .set_entry = gtt_set_entry64,
  350. .clear_present = gtt_entry_clear_present,
  351. .test_present = gen8_gtt_test_present,
  352. .test_pse = gen8_gtt_test_pse,
  353. .get_pfn = gen8_gtt_get_pfn,
  354. .set_pfn = gen8_gtt_set_pfn,
  355. };
  356. static struct intel_gvt_gtt_gma_ops gen8_gtt_gma_ops = {
  357. .gma_to_ggtt_pte_index = gma_to_ggtt_pte_index,
  358. .gma_to_pte_index = gen8_gma_to_pte_index,
  359. .gma_to_pde_index = gen8_gma_to_pde_index,
  360. .gma_to_l3_pdp_index = gen8_gma_to_l3_pdp_index,
  361. .gma_to_l4_pdp_index = gen8_gma_to_l4_pdp_index,
  362. .gma_to_pml4_index = gen8_gma_to_pml4_index,
  363. };
  364. static int gtt_entry_p2m(struct intel_vgpu *vgpu, struct intel_gvt_gtt_entry *p,
  365. struct intel_gvt_gtt_entry *m)
  366. {
  367. struct intel_gvt_gtt_pte_ops *ops = vgpu->gvt->gtt.pte_ops;
  368. unsigned long gfn, mfn;
  369. *m = *p;
  370. if (!ops->test_present(p))
  371. return 0;
  372. gfn = ops->get_pfn(p);
  373. mfn = intel_gvt_hypervisor_gfn_to_mfn(vgpu, gfn);
  374. if (mfn == INTEL_GVT_INVALID_ADDR) {
  375. gvt_vgpu_err("fail to translate gfn: 0x%lx\n", gfn);
  376. return -ENXIO;
  377. }
  378. ops->set_pfn(m, mfn);
  379. return 0;
  380. }
  381. /*
  382. * MM helpers.
  383. */
  384. struct intel_gvt_gtt_entry *intel_vgpu_mm_get_entry(struct intel_vgpu_mm *mm,
  385. void *page_table, struct intel_gvt_gtt_entry *e,
  386. unsigned long index)
  387. {
  388. struct intel_gvt *gvt = mm->vgpu->gvt;
  389. struct intel_gvt_gtt_pte_ops *ops = gvt->gtt.pte_ops;
  390. e->type = mm->page_table_entry_type;
  391. ops->get_entry(page_table, e, index, false, 0, mm->vgpu);
  392. ops->test_pse(e);
  393. return e;
  394. }
  395. struct intel_gvt_gtt_entry *intel_vgpu_mm_set_entry(struct intel_vgpu_mm *mm,
  396. void *page_table, struct intel_gvt_gtt_entry *e,
  397. unsigned long index)
  398. {
  399. struct intel_gvt *gvt = mm->vgpu->gvt;
  400. struct intel_gvt_gtt_pte_ops *ops = gvt->gtt.pte_ops;
  401. return ops->set_entry(page_table, e, index, false, 0, mm->vgpu);
  402. }
  403. /*
  404. * PPGTT shadow page table helpers.
  405. */
  406. static inline struct intel_gvt_gtt_entry *ppgtt_spt_get_entry(
  407. struct intel_vgpu_ppgtt_spt *spt,
  408. void *page_table, int type,
  409. struct intel_gvt_gtt_entry *e, unsigned long index,
  410. bool guest)
  411. {
  412. struct intel_gvt *gvt = spt->vgpu->gvt;
  413. struct intel_gvt_gtt_pte_ops *ops = gvt->gtt.pte_ops;
  414. e->type = get_entry_type(type);
  415. if (WARN(!gtt_type_is_entry(e->type), "invalid entry type\n"))
  416. return e;
  417. ops->get_entry(page_table, e, index, guest,
  418. spt->guest_page.gfn << GTT_PAGE_SHIFT,
  419. spt->vgpu);
  420. ops->test_pse(e);
  421. return e;
  422. }
  423. static inline struct intel_gvt_gtt_entry *ppgtt_spt_set_entry(
  424. struct intel_vgpu_ppgtt_spt *spt,
  425. void *page_table, int type,
  426. struct intel_gvt_gtt_entry *e, unsigned long index,
  427. bool guest)
  428. {
  429. struct intel_gvt *gvt = spt->vgpu->gvt;
  430. struct intel_gvt_gtt_pte_ops *ops = gvt->gtt.pte_ops;
  431. if (WARN(!gtt_type_is_entry(e->type), "invalid entry type\n"))
  432. return e;
  433. return ops->set_entry(page_table, e, index, guest,
  434. spt->guest_page.gfn << GTT_PAGE_SHIFT,
  435. spt->vgpu);
  436. }
  437. #define ppgtt_get_guest_entry(spt, e, index) \
  438. ppgtt_spt_get_entry(spt, NULL, \
  439. spt->guest_page_type, e, index, true)
  440. #define ppgtt_set_guest_entry(spt, e, index) \
  441. ppgtt_spt_set_entry(spt, NULL, \
  442. spt->guest_page_type, e, index, true)
  443. #define ppgtt_get_shadow_entry(spt, e, index) \
  444. ppgtt_spt_get_entry(spt, spt->shadow_page.vaddr, \
  445. spt->shadow_page.type, e, index, false)
  446. #define ppgtt_set_shadow_entry(spt, e, index) \
  447. ppgtt_spt_set_entry(spt, spt->shadow_page.vaddr, \
  448. spt->shadow_page.type, e, index, false)
  449. /**
  450. * intel_vgpu_init_guest_page - init a guest page data structure
  451. * @vgpu: a vGPU
  452. * @p: a guest page data structure
  453. * @gfn: guest memory page frame number
  454. * @handler: function will be called when target guest memory page has
  455. * been modified.
  456. *
  457. * This function is called when user wants to track a guest memory page.
  458. *
  459. * Returns:
  460. * Zero on success, negative error code if failed.
  461. */
  462. int intel_vgpu_init_guest_page(struct intel_vgpu *vgpu,
  463. struct intel_vgpu_guest_page *p,
  464. unsigned long gfn,
  465. int (*handler)(void *, u64, void *, int),
  466. void *data)
  467. {
  468. INIT_HLIST_NODE(&p->node);
  469. p->writeprotection = false;
  470. p->gfn = gfn;
  471. p->handler = handler;
  472. p->data = data;
  473. p->oos_page = NULL;
  474. p->write_cnt = 0;
  475. hash_add(vgpu->gtt.guest_page_hash_table, &p->node, p->gfn);
  476. return 0;
  477. }
  478. static int detach_oos_page(struct intel_vgpu *vgpu,
  479. struct intel_vgpu_oos_page *oos_page);
  480. /**
  481. * intel_vgpu_clean_guest_page - release the resource owned by guest page data
  482. * structure
  483. * @vgpu: a vGPU
  484. * @p: a tracked guest page
  485. *
  486. * This function is called when user tries to stop tracking a guest memory
  487. * page.
  488. */
  489. void intel_vgpu_clean_guest_page(struct intel_vgpu *vgpu,
  490. struct intel_vgpu_guest_page *p)
  491. {
  492. if (!hlist_unhashed(&p->node))
  493. hash_del(&p->node);
  494. if (p->oos_page)
  495. detach_oos_page(vgpu, p->oos_page);
  496. if (p->writeprotection)
  497. intel_gvt_hypervisor_unset_wp_page(vgpu, p);
  498. }
  499. /**
  500. * intel_vgpu_find_guest_page - find a guest page data structure by GFN.
  501. * @vgpu: a vGPU
  502. * @gfn: guest memory page frame number
  503. *
  504. * This function is called when emulation logic wants to know if a trapped GFN
  505. * is a tracked guest page.
  506. *
  507. * Returns:
  508. * Pointer to guest page data structure, NULL if failed.
  509. */
  510. struct intel_vgpu_guest_page *intel_vgpu_find_guest_page(
  511. struct intel_vgpu *vgpu, unsigned long gfn)
  512. {
  513. struct intel_vgpu_guest_page *p;
  514. hash_for_each_possible(vgpu->gtt.guest_page_hash_table,
  515. p, node, gfn) {
  516. if (p->gfn == gfn)
  517. return p;
  518. }
  519. return NULL;
  520. }
  521. static inline int init_shadow_page(struct intel_vgpu *vgpu,
  522. struct intel_vgpu_shadow_page *p, int type)
  523. {
  524. struct device *kdev = &vgpu->gvt->dev_priv->drm.pdev->dev;
  525. dma_addr_t daddr;
  526. daddr = dma_map_page(kdev, p->page, 0, 4096, PCI_DMA_BIDIRECTIONAL);
  527. if (dma_mapping_error(kdev, daddr)) {
  528. gvt_vgpu_err("fail to map dma addr\n");
  529. return -EINVAL;
  530. }
  531. p->vaddr = page_address(p->page);
  532. p->type = type;
  533. INIT_HLIST_NODE(&p->node);
  534. p->mfn = daddr >> GTT_PAGE_SHIFT;
  535. hash_add(vgpu->gtt.shadow_page_hash_table, &p->node, p->mfn);
  536. return 0;
  537. }
  538. static inline void clean_shadow_page(struct intel_vgpu *vgpu,
  539. struct intel_vgpu_shadow_page *p)
  540. {
  541. struct device *kdev = &vgpu->gvt->dev_priv->drm.pdev->dev;
  542. dma_unmap_page(kdev, p->mfn << GTT_PAGE_SHIFT, 4096,
  543. PCI_DMA_BIDIRECTIONAL);
  544. if (!hlist_unhashed(&p->node))
  545. hash_del(&p->node);
  546. }
  547. static inline struct intel_vgpu_shadow_page *find_shadow_page(
  548. struct intel_vgpu *vgpu, unsigned long mfn)
  549. {
  550. struct intel_vgpu_shadow_page *p;
  551. hash_for_each_possible(vgpu->gtt.shadow_page_hash_table,
  552. p, node, mfn) {
  553. if (p->mfn == mfn)
  554. return p;
  555. }
  556. return NULL;
  557. }
  558. #define guest_page_to_ppgtt_spt(ptr) \
  559. container_of(ptr, struct intel_vgpu_ppgtt_spt, guest_page)
  560. #define shadow_page_to_ppgtt_spt(ptr) \
  561. container_of(ptr, struct intel_vgpu_ppgtt_spt, shadow_page)
  562. static void *alloc_spt(gfp_t gfp_mask)
  563. {
  564. struct intel_vgpu_ppgtt_spt *spt;
  565. spt = kzalloc(sizeof(*spt), gfp_mask);
  566. if (!spt)
  567. return NULL;
  568. spt->shadow_page.page = alloc_page(gfp_mask);
  569. if (!spt->shadow_page.page) {
  570. kfree(spt);
  571. return NULL;
  572. }
  573. return spt;
  574. }
  575. static void free_spt(struct intel_vgpu_ppgtt_spt *spt)
  576. {
  577. __free_page(spt->shadow_page.page);
  578. kfree(spt);
  579. }
  580. static void ppgtt_free_shadow_page(struct intel_vgpu_ppgtt_spt *spt)
  581. {
  582. trace_spt_free(spt->vgpu->id, spt, spt->shadow_page.type);
  583. clean_shadow_page(spt->vgpu, &spt->shadow_page);
  584. intel_vgpu_clean_guest_page(spt->vgpu, &spt->guest_page);
  585. list_del_init(&spt->post_shadow_list);
  586. free_spt(spt);
  587. }
  588. static void ppgtt_free_all_shadow_page(struct intel_vgpu *vgpu)
  589. {
  590. struct hlist_node *n;
  591. struct intel_vgpu_shadow_page *sp;
  592. int i;
  593. hash_for_each_safe(vgpu->gtt.shadow_page_hash_table, i, n, sp, node)
  594. ppgtt_free_shadow_page(shadow_page_to_ppgtt_spt(sp));
  595. }
  596. static int ppgtt_handle_guest_write_page_table_bytes(void *gp,
  597. u64 pa, void *p_data, int bytes);
  598. static int ppgtt_write_protection_handler(void *gp, u64 pa,
  599. void *p_data, int bytes)
  600. {
  601. struct intel_vgpu_guest_page *gpt = (struct intel_vgpu_guest_page *)gp;
  602. int ret;
  603. if (bytes != 4 && bytes != 8)
  604. return -EINVAL;
  605. if (!gpt->writeprotection)
  606. return -EINVAL;
  607. ret = ppgtt_handle_guest_write_page_table_bytes(gp,
  608. pa, p_data, bytes);
  609. if (ret)
  610. return ret;
  611. return ret;
  612. }
  613. static int reclaim_one_mm(struct intel_gvt *gvt);
  614. static struct intel_vgpu_ppgtt_spt *ppgtt_alloc_shadow_page(
  615. struct intel_vgpu *vgpu, int type, unsigned long gfn)
  616. {
  617. struct intel_vgpu_ppgtt_spt *spt = NULL;
  618. int ret;
  619. retry:
  620. spt = alloc_spt(GFP_KERNEL | __GFP_ZERO);
  621. if (!spt) {
  622. if (reclaim_one_mm(vgpu->gvt))
  623. goto retry;
  624. gvt_vgpu_err("fail to allocate ppgtt shadow page\n");
  625. return ERR_PTR(-ENOMEM);
  626. }
  627. spt->vgpu = vgpu;
  628. spt->guest_page_type = type;
  629. atomic_set(&spt->refcount, 1);
  630. INIT_LIST_HEAD(&spt->post_shadow_list);
  631. /*
  632. * TODO: guest page type may be different with shadow page type,
  633. * when we support PSE page in future.
  634. */
  635. ret = init_shadow_page(vgpu, &spt->shadow_page, type);
  636. if (ret) {
  637. gvt_vgpu_err("fail to initialize shadow page for spt\n");
  638. goto err;
  639. }
  640. ret = intel_vgpu_init_guest_page(vgpu, &spt->guest_page,
  641. gfn, ppgtt_write_protection_handler, NULL);
  642. if (ret) {
  643. gvt_vgpu_err("fail to initialize guest page for spt\n");
  644. goto err;
  645. }
  646. trace_spt_alloc(vgpu->id, spt, type, spt->shadow_page.mfn, gfn);
  647. return spt;
  648. err:
  649. ppgtt_free_shadow_page(spt);
  650. return ERR_PTR(ret);
  651. }
  652. static struct intel_vgpu_ppgtt_spt *ppgtt_find_shadow_page(
  653. struct intel_vgpu *vgpu, unsigned long mfn)
  654. {
  655. struct intel_vgpu_shadow_page *p = find_shadow_page(vgpu, mfn);
  656. if (p)
  657. return shadow_page_to_ppgtt_spt(p);
  658. gvt_vgpu_err("fail to find ppgtt shadow page: 0x%lx\n", mfn);
  659. return NULL;
  660. }
  661. #define pt_entry_size_shift(spt) \
  662. ((spt)->vgpu->gvt->device_info.gtt_entry_size_shift)
  663. #define pt_entries(spt) \
  664. (GTT_PAGE_SIZE >> pt_entry_size_shift(spt))
  665. #define for_each_present_guest_entry(spt, e, i) \
  666. for (i = 0; i < pt_entries(spt); i++) \
  667. if (spt->vgpu->gvt->gtt.pte_ops->test_present( \
  668. ppgtt_get_guest_entry(spt, e, i)))
  669. #define for_each_present_shadow_entry(spt, e, i) \
  670. for (i = 0; i < pt_entries(spt); i++) \
  671. if (spt->vgpu->gvt->gtt.pte_ops->test_present( \
  672. ppgtt_get_shadow_entry(spt, e, i)))
  673. static void ppgtt_get_shadow_page(struct intel_vgpu_ppgtt_spt *spt)
  674. {
  675. int v = atomic_read(&spt->refcount);
  676. trace_spt_refcount(spt->vgpu->id, "inc", spt, v, (v + 1));
  677. atomic_inc(&spt->refcount);
  678. }
  679. static int ppgtt_invalidate_shadow_page(struct intel_vgpu_ppgtt_spt *spt);
  680. static int ppgtt_invalidate_shadow_page_by_shadow_entry(struct intel_vgpu *vgpu,
  681. struct intel_gvt_gtt_entry *e)
  682. {
  683. struct intel_gvt_gtt_pte_ops *ops = vgpu->gvt->gtt.pte_ops;
  684. struct intel_vgpu_ppgtt_spt *s;
  685. intel_gvt_gtt_type_t cur_pt_type;
  686. if (WARN_ON(!gtt_type_is_pt(get_next_pt_type(e->type))))
  687. return -EINVAL;
  688. if (e->type != GTT_TYPE_PPGTT_ROOT_L3_ENTRY
  689. && e->type != GTT_TYPE_PPGTT_ROOT_L4_ENTRY) {
  690. cur_pt_type = get_next_pt_type(e->type) + 1;
  691. if (ops->get_pfn(e) ==
  692. vgpu->gtt.scratch_pt[cur_pt_type].page_mfn)
  693. return 0;
  694. }
  695. s = ppgtt_find_shadow_page(vgpu, ops->get_pfn(e));
  696. if (!s) {
  697. gvt_vgpu_err("fail to find shadow page: mfn: 0x%lx\n",
  698. ops->get_pfn(e));
  699. return -ENXIO;
  700. }
  701. return ppgtt_invalidate_shadow_page(s);
  702. }
  703. static int ppgtt_invalidate_shadow_page(struct intel_vgpu_ppgtt_spt *spt)
  704. {
  705. struct intel_vgpu *vgpu = spt->vgpu;
  706. struct intel_gvt_gtt_entry e;
  707. unsigned long index;
  708. int ret;
  709. int v = atomic_read(&spt->refcount);
  710. trace_spt_change(spt->vgpu->id, "die", spt,
  711. spt->guest_page.gfn, spt->shadow_page.type);
  712. trace_spt_refcount(spt->vgpu->id, "dec", spt, v, (v - 1));
  713. if (atomic_dec_return(&spt->refcount) > 0)
  714. return 0;
  715. if (gtt_type_is_pte_pt(spt->shadow_page.type))
  716. goto release;
  717. for_each_present_shadow_entry(spt, &e, index) {
  718. if (!gtt_type_is_pt(get_next_pt_type(e.type))) {
  719. gvt_vgpu_err("GVT doesn't support pse bit for now\n");
  720. return -EINVAL;
  721. }
  722. ret = ppgtt_invalidate_shadow_page_by_shadow_entry(
  723. spt->vgpu, &e);
  724. if (ret)
  725. goto fail;
  726. }
  727. release:
  728. trace_spt_change(spt->vgpu->id, "release", spt,
  729. spt->guest_page.gfn, spt->shadow_page.type);
  730. ppgtt_free_shadow_page(spt);
  731. return 0;
  732. fail:
  733. gvt_vgpu_err("fail: shadow page %p shadow entry 0x%llx type %d\n",
  734. spt, e.val64, e.type);
  735. return ret;
  736. }
  737. static int ppgtt_populate_shadow_page(struct intel_vgpu_ppgtt_spt *spt);
  738. static struct intel_vgpu_ppgtt_spt *ppgtt_populate_shadow_page_by_guest_entry(
  739. struct intel_vgpu *vgpu, struct intel_gvt_gtt_entry *we)
  740. {
  741. struct intel_gvt_gtt_pte_ops *ops = vgpu->gvt->gtt.pte_ops;
  742. struct intel_vgpu_ppgtt_spt *s = NULL;
  743. struct intel_vgpu_guest_page *g;
  744. int ret;
  745. if (WARN_ON(!gtt_type_is_pt(get_next_pt_type(we->type)))) {
  746. ret = -EINVAL;
  747. goto fail;
  748. }
  749. g = intel_vgpu_find_guest_page(vgpu, ops->get_pfn(we));
  750. if (g) {
  751. s = guest_page_to_ppgtt_spt(g);
  752. ppgtt_get_shadow_page(s);
  753. } else {
  754. int type = get_next_pt_type(we->type);
  755. s = ppgtt_alloc_shadow_page(vgpu, type, ops->get_pfn(we));
  756. if (IS_ERR(s)) {
  757. ret = PTR_ERR(s);
  758. goto fail;
  759. }
  760. ret = intel_gvt_hypervisor_set_wp_page(vgpu, &s->guest_page);
  761. if (ret)
  762. goto fail;
  763. ret = ppgtt_populate_shadow_page(s);
  764. if (ret)
  765. goto fail;
  766. trace_spt_change(vgpu->id, "new", s, s->guest_page.gfn,
  767. s->shadow_page.type);
  768. }
  769. return s;
  770. fail:
  771. gvt_vgpu_err("fail: shadow page %p guest entry 0x%llx type %d\n",
  772. s, we->val64, we->type);
  773. return ERR_PTR(ret);
  774. }
  775. static inline void ppgtt_generate_shadow_entry(struct intel_gvt_gtt_entry *se,
  776. struct intel_vgpu_ppgtt_spt *s, struct intel_gvt_gtt_entry *ge)
  777. {
  778. struct intel_gvt_gtt_pte_ops *ops = s->vgpu->gvt->gtt.pte_ops;
  779. se->type = ge->type;
  780. se->val64 = ge->val64;
  781. ops->set_pfn(se, s->shadow_page.mfn);
  782. }
  783. static int ppgtt_populate_shadow_page(struct intel_vgpu_ppgtt_spt *spt)
  784. {
  785. struct intel_vgpu *vgpu = spt->vgpu;
  786. struct intel_vgpu_ppgtt_spt *s;
  787. struct intel_gvt_gtt_entry se, ge;
  788. unsigned long i;
  789. int ret;
  790. trace_spt_change(spt->vgpu->id, "born", spt,
  791. spt->guest_page.gfn, spt->shadow_page.type);
  792. if (gtt_type_is_pte_pt(spt->shadow_page.type)) {
  793. for_each_present_guest_entry(spt, &ge, i) {
  794. ret = gtt_entry_p2m(vgpu, &ge, &se);
  795. if (ret)
  796. goto fail;
  797. ppgtt_set_shadow_entry(spt, &se, i);
  798. }
  799. return 0;
  800. }
  801. for_each_present_guest_entry(spt, &ge, i) {
  802. if (!gtt_type_is_pt(get_next_pt_type(ge.type))) {
  803. gvt_vgpu_err("GVT doesn't support pse bit now\n");
  804. ret = -EINVAL;
  805. goto fail;
  806. }
  807. s = ppgtt_populate_shadow_page_by_guest_entry(vgpu, &ge);
  808. if (IS_ERR(s)) {
  809. ret = PTR_ERR(s);
  810. goto fail;
  811. }
  812. ppgtt_get_shadow_entry(spt, &se, i);
  813. ppgtt_generate_shadow_entry(&se, s, &ge);
  814. ppgtt_set_shadow_entry(spt, &se, i);
  815. }
  816. return 0;
  817. fail:
  818. gvt_vgpu_err("fail: shadow page %p guest entry 0x%llx type %d\n",
  819. spt, ge.val64, ge.type);
  820. return ret;
  821. }
  822. static int ppgtt_handle_guest_entry_removal(struct intel_vgpu_guest_page *gpt,
  823. unsigned long index)
  824. {
  825. struct intel_vgpu_ppgtt_spt *spt = guest_page_to_ppgtt_spt(gpt);
  826. struct intel_vgpu_shadow_page *sp = &spt->shadow_page;
  827. struct intel_vgpu *vgpu = spt->vgpu;
  828. struct intel_gvt_gtt_pte_ops *ops = vgpu->gvt->gtt.pte_ops;
  829. struct intel_gvt_gtt_entry e;
  830. int ret;
  831. ppgtt_get_shadow_entry(spt, &e, index);
  832. trace_gpt_change(spt->vgpu->id, "remove", spt, sp->type, e.val64,
  833. index);
  834. if (!ops->test_present(&e))
  835. return 0;
  836. if (ops->get_pfn(&e) == vgpu->gtt.scratch_pt[sp->type].page_mfn)
  837. return 0;
  838. if (gtt_type_is_pt(get_next_pt_type(e.type))) {
  839. struct intel_vgpu_ppgtt_spt *s =
  840. ppgtt_find_shadow_page(vgpu, ops->get_pfn(&e));
  841. if (!s) {
  842. gvt_vgpu_err("fail to find guest page\n");
  843. ret = -ENXIO;
  844. goto fail;
  845. }
  846. ret = ppgtt_invalidate_shadow_page(s);
  847. if (ret)
  848. goto fail;
  849. }
  850. ops->set_pfn(&e, vgpu->gtt.scratch_pt[sp->type].page_mfn);
  851. ppgtt_set_shadow_entry(spt, &e, index);
  852. return 0;
  853. fail:
  854. gvt_vgpu_err("fail: shadow page %p guest entry 0x%llx type %d\n",
  855. spt, e.val64, e.type);
  856. return ret;
  857. }
  858. static int ppgtt_handle_guest_entry_add(struct intel_vgpu_guest_page *gpt,
  859. struct intel_gvt_gtt_entry *we, unsigned long index)
  860. {
  861. struct intel_vgpu_ppgtt_spt *spt = guest_page_to_ppgtt_spt(gpt);
  862. struct intel_vgpu_shadow_page *sp = &spt->shadow_page;
  863. struct intel_vgpu *vgpu = spt->vgpu;
  864. struct intel_gvt_gtt_entry m;
  865. struct intel_vgpu_ppgtt_spt *s;
  866. int ret;
  867. trace_gpt_change(spt->vgpu->id, "add", spt, sp->type,
  868. we->val64, index);
  869. if (gtt_type_is_pt(get_next_pt_type(we->type))) {
  870. s = ppgtt_populate_shadow_page_by_guest_entry(vgpu, we);
  871. if (IS_ERR(s)) {
  872. ret = PTR_ERR(s);
  873. goto fail;
  874. }
  875. ppgtt_get_shadow_entry(spt, &m, index);
  876. ppgtt_generate_shadow_entry(&m, s, we);
  877. ppgtt_set_shadow_entry(spt, &m, index);
  878. } else {
  879. ret = gtt_entry_p2m(vgpu, we, &m);
  880. if (ret)
  881. goto fail;
  882. ppgtt_set_shadow_entry(spt, &m, index);
  883. }
  884. return 0;
  885. fail:
  886. gvt_vgpu_err("fail: spt %p guest entry 0x%llx type %d\n",
  887. spt, we->val64, we->type);
  888. return ret;
  889. }
  890. static int sync_oos_page(struct intel_vgpu *vgpu,
  891. struct intel_vgpu_oos_page *oos_page)
  892. {
  893. const struct intel_gvt_device_info *info = &vgpu->gvt->device_info;
  894. struct intel_gvt *gvt = vgpu->gvt;
  895. struct intel_gvt_gtt_pte_ops *ops = gvt->gtt.pte_ops;
  896. struct intel_vgpu_ppgtt_spt *spt =
  897. guest_page_to_ppgtt_spt(oos_page->guest_page);
  898. struct intel_gvt_gtt_entry old, new, m;
  899. int index;
  900. int ret;
  901. trace_oos_change(vgpu->id, "sync", oos_page->id,
  902. oos_page->guest_page, spt->guest_page_type);
  903. old.type = new.type = get_entry_type(spt->guest_page_type);
  904. old.val64 = new.val64 = 0;
  905. for (index = 0; index < (GTT_PAGE_SIZE >> info->gtt_entry_size_shift);
  906. index++) {
  907. ops->get_entry(oos_page->mem, &old, index, false, 0, vgpu);
  908. ops->get_entry(NULL, &new, index, true,
  909. oos_page->guest_page->gfn << PAGE_SHIFT, vgpu);
  910. if (old.val64 == new.val64
  911. && !test_and_clear_bit(index, spt->post_shadow_bitmap))
  912. continue;
  913. trace_oos_sync(vgpu->id, oos_page->id,
  914. oos_page->guest_page, spt->guest_page_type,
  915. new.val64, index);
  916. ret = gtt_entry_p2m(vgpu, &new, &m);
  917. if (ret)
  918. return ret;
  919. ops->set_entry(oos_page->mem, &new, index, false, 0, vgpu);
  920. ppgtt_set_shadow_entry(spt, &m, index);
  921. }
  922. oos_page->guest_page->write_cnt = 0;
  923. list_del_init(&spt->post_shadow_list);
  924. return 0;
  925. }
  926. static int detach_oos_page(struct intel_vgpu *vgpu,
  927. struct intel_vgpu_oos_page *oos_page)
  928. {
  929. struct intel_gvt *gvt = vgpu->gvt;
  930. struct intel_vgpu_ppgtt_spt *spt =
  931. guest_page_to_ppgtt_spt(oos_page->guest_page);
  932. trace_oos_change(vgpu->id, "detach", oos_page->id,
  933. oos_page->guest_page, spt->guest_page_type);
  934. oos_page->guest_page->write_cnt = 0;
  935. oos_page->guest_page->oos_page = NULL;
  936. oos_page->guest_page = NULL;
  937. list_del_init(&oos_page->vm_list);
  938. list_move_tail(&oos_page->list, &gvt->gtt.oos_page_free_list_head);
  939. return 0;
  940. }
  941. static int attach_oos_page(struct intel_vgpu *vgpu,
  942. struct intel_vgpu_oos_page *oos_page,
  943. struct intel_vgpu_guest_page *gpt)
  944. {
  945. struct intel_gvt *gvt = vgpu->gvt;
  946. int ret;
  947. ret = intel_gvt_hypervisor_read_gpa(vgpu, gpt->gfn << GTT_PAGE_SHIFT,
  948. oos_page->mem, GTT_PAGE_SIZE);
  949. if (ret)
  950. return ret;
  951. oos_page->guest_page = gpt;
  952. gpt->oos_page = oos_page;
  953. list_move_tail(&oos_page->list, &gvt->gtt.oos_page_use_list_head);
  954. trace_oos_change(vgpu->id, "attach", gpt->oos_page->id,
  955. gpt, guest_page_to_ppgtt_spt(gpt)->guest_page_type);
  956. return 0;
  957. }
  958. static int ppgtt_set_guest_page_sync(struct intel_vgpu *vgpu,
  959. struct intel_vgpu_guest_page *gpt)
  960. {
  961. int ret;
  962. ret = intel_gvt_hypervisor_set_wp_page(vgpu, gpt);
  963. if (ret)
  964. return ret;
  965. trace_oos_change(vgpu->id, "set page sync", gpt->oos_page->id,
  966. gpt, guest_page_to_ppgtt_spt(gpt)->guest_page_type);
  967. list_del_init(&gpt->oos_page->vm_list);
  968. return sync_oos_page(vgpu, gpt->oos_page);
  969. }
  970. static int ppgtt_allocate_oos_page(struct intel_vgpu *vgpu,
  971. struct intel_vgpu_guest_page *gpt)
  972. {
  973. struct intel_gvt *gvt = vgpu->gvt;
  974. struct intel_gvt_gtt *gtt = &gvt->gtt;
  975. struct intel_vgpu_oos_page *oos_page = gpt->oos_page;
  976. int ret;
  977. WARN(oos_page, "shadow PPGTT page has already has a oos page\n");
  978. if (list_empty(&gtt->oos_page_free_list_head)) {
  979. oos_page = container_of(gtt->oos_page_use_list_head.next,
  980. struct intel_vgpu_oos_page, list);
  981. ret = ppgtt_set_guest_page_sync(vgpu, oos_page->guest_page);
  982. if (ret)
  983. return ret;
  984. ret = detach_oos_page(vgpu, oos_page);
  985. if (ret)
  986. return ret;
  987. } else
  988. oos_page = container_of(gtt->oos_page_free_list_head.next,
  989. struct intel_vgpu_oos_page, list);
  990. return attach_oos_page(vgpu, oos_page, gpt);
  991. }
  992. static int ppgtt_set_guest_page_oos(struct intel_vgpu *vgpu,
  993. struct intel_vgpu_guest_page *gpt)
  994. {
  995. struct intel_vgpu_oos_page *oos_page = gpt->oos_page;
  996. if (WARN(!oos_page, "shadow PPGTT page should have a oos page\n"))
  997. return -EINVAL;
  998. trace_oos_change(vgpu->id, "set page out of sync", gpt->oos_page->id,
  999. gpt, guest_page_to_ppgtt_spt(gpt)->guest_page_type);
  1000. list_add_tail(&oos_page->vm_list, &vgpu->gtt.oos_page_list_head);
  1001. return intel_gvt_hypervisor_unset_wp_page(vgpu, gpt);
  1002. }
  1003. /**
  1004. * intel_vgpu_sync_oos_pages - sync all the out-of-synced shadow for vGPU
  1005. * @vgpu: a vGPU
  1006. *
  1007. * This function is called before submitting a guest workload to host,
  1008. * to sync all the out-of-synced shadow for vGPU
  1009. *
  1010. * Returns:
  1011. * Zero on success, negative error code if failed.
  1012. */
  1013. int intel_vgpu_sync_oos_pages(struct intel_vgpu *vgpu)
  1014. {
  1015. struct list_head *pos, *n;
  1016. struct intel_vgpu_oos_page *oos_page;
  1017. int ret;
  1018. if (!enable_out_of_sync)
  1019. return 0;
  1020. list_for_each_safe(pos, n, &vgpu->gtt.oos_page_list_head) {
  1021. oos_page = container_of(pos,
  1022. struct intel_vgpu_oos_page, vm_list);
  1023. ret = ppgtt_set_guest_page_sync(vgpu, oos_page->guest_page);
  1024. if (ret)
  1025. return ret;
  1026. }
  1027. return 0;
  1028. }
  1029. /*
  1030. * The heart of PPGTT shadow page table.
  1031. */
  1032. static int ppgtt_handle_guest_write_page_table(
  1033. struct intel_vgpu_guest_page *gpt,
  1034. struct intel_gvt_gtt_entry *we, unsigned long index)
  1035. {
  1036. struct intel_vgpu_ppgtt_spt *spt = guest_page_to_ppgtt_spt(gpt);
  1037. struct intel_vgpu *vgpu = spt->vgpu;
  1038. struct intel_gvt_gtt_pte_ops *ops = vgpu->gvt->gtt.pte_ops;
  1039. int ret;
  1040. int new_present;
  1041. new_present = ops->test_present(we);
  1042. ret = ppgtt_handle_guest_entry_removal(gpt, index);
  1043. if (ret)
  1044. goto fail;
  1045. if (new_present) {
  1046. ret = ppgtt_handle_guest_entry_add(gpt, we, index);
  1047. if (ret)
  1048. goto fail;
  1049. }
  1050. return 0;
  1051. fail:
  1052. gvt_vgpu_err("fail: shadow page %p guest entry 0x%llx type %d.\n",
  1053. spt, we->val64, we->type);
  1054. return ret;
  1055. }
  1056. static inline bool can_do_out_of_sync(struct intel_vgpu_guest_page *gpt)
  1057. {
  1058. return enable_out_of_sync
  1059. && gtt_type_is_pte_pt(
  1060. guest_page_to_ppgtt_spt(gpt)->guest_page_type)
  1061. && gpt->write_cnt >= 2;
  1062. }
  1063. static void ppgtt_set_post_shadow(struct intel_vgpu_ppgtt_spt *spt,
  1064. unsigned long index)
  1065. {
  1066. set_bit(index, spt->post_shadow_bitmap);
  1067. if (!list_empty(&spt->post_shadow_list))
  1068. return;
  1069. list_add_tail(&spt->post_shadow_list,
  1070. &spt->vgpu->gtt.post_shadow_list_head);
  1071. }
  1072. /**
  1073. * intel_vgpu_flush_post_shadow - flush the post shadow transactions
  1074. * @vgpu: a vGPU
  1075. *
  1076. * This function is called before submitting a guest workload to host,
  1077. * to flush all the post shadows for a vGPU.
  1078. *
  1079. * Returns:
  1080. * Zero on success, negative error code if failed.
  1081. */
  1082. int intel_vgpu_flush_post_shadow(struct intel_vgpu *vgpu)
  1083. {
  1084. struct list_head *pos, *n;
  1085. struct intel_vgpu_ppgtt_spt *spt;
  1086. struct intel_gvt_gtt_entry ge;
  1087. unsigned long index;
  1088. int ret;
  1089. list_for_each_safe(pos, n, &vgpu->gtt.post_shadow_list_head) {
  1090. spt = container_of(pos, struct intel_vgpu_ppgtt_spt,
  1091. post_shadow_list);
  1092. for_each_set_bit(index, spt->post_shadow_bitmap,
  1093. GTT_ENTRY_NUM_IN_ONE_PAGE) {
  1094. ppgtt_get_guest_entry(spt, &ge, index);
  1095. ret = ppgtt_handle_guest_write_page_table(
  1096. &spt->guest_page, &ge, index);
  1097. if (ret)
  1098. return ret;
  1099. clear_bit(index, spt->post_shadow_bitmap);
  1100. }
  1101. list_del_init(&spt->post_shadow_list);
  1102. }
  1103. return 0;
  1104. }
  1105. static int ppgtt_handle_guest_write_page_table_bytes(void *gp,
  1106. u64 pa, void *p_data, int bytes)
  1107. {
  1108. struct intel_vgpu_guest_page *gpt = (struct intel_vgpu_guest_page *)gp;
  1109. struct intel_vgpu_ppgtt_spt *spt = guest_page_to_ppgtt_spt(gpt);
  1110. struct intel_vgpu *vgpu = spt->vgpu;
  1111. struct intel_gvt_gtt_pte_ops *ops = vgpu->gvt->gtt.pte_ops;
  1112. const struct intel_gvt_device_info *info = &vgpu->gvt->device_info;
  1113. struct intel_gvt_gtt_entry we;
  1114. unsigned long index;
  1115. int ret;
  1116. index = (pa & (PAGE_SIZE - 1)) >> info->gtt_entry_size_shift;
  1117. ppgtt_get_guest_entry(spt, &we, index);
  1118. ops->test_pse(&we);
  1119. if (bytes == info->gtt_entry_size) {
  1120. ret = ppgtt_handle_guest_write_page_table(gpt, &we, index);
  1121. if (ret)
  1122. return ret;
  1123. } else {
  1124. if (!test_bit(index, spt->post_shadow_bitmap)) {
  1125. ret = ppgtt_handle_guest_entry_removal(gpt, index);
  1126. if (ret)
  1127. return ret;
  1128. }
  1129. ppgtt_set_post_shadow(spt, index);
  1130. }
  1131. if (!enable_out_of_sync)
  1132. return 0;
  1133. gpt->write_cnt++;
  1134. if (gpt->oos_page)
  1135. ops->set_entry(gpt->oos_page->mem, &we, index,
  1136. false, 0, vgpu);
  1137. if (can_do_out_of_sync(gpt)) {
  1138. if (!gpt->oos_page)
  1139. ppgtt_allocate_oos_page(vgpu, gpt);
  1140. ret = ppgtt_set_guest_page_oos(vgpu, gpt);
  1141. if (ret < 0)
  1142. return ret;
  1143. }
  1144. return 0;
  1145. }
  1146. /*
  1147. * mm page table allocation policy for bdw+
  1148. * - for ggtt, only virtual page table will be allocated.
  1149. * - for ppgtt, dedicated virtual/shadow page table will be allocated.
  1150. */
  1151. static int gen8_mm_alloc_page_table(struct intel_vgpu_mm *mm)
  1152. {
  1153. struct intel_vgpu *vgpu = mm->vgpu;
  1154. struct intel_gvt *gvt = vgpu->gvt;
  1155. const struct intel_gvt_device_info *info = &gvt->device_info;
  1156. void *mem;
  1157. if (mm->type == INTEL_GVT_MM_PPGTT) {
  1158. mm->page_table_entry_cnt = 4;
  1159. mm->page_table_entry_size = mm->page_table_entry_cnt *
  1160. info->gtt_entry_size;
  1161. mem = kzalloc(mm->has_shadow_page_table ?
  1162. mm->page_table_entry_size * 2
  1163. : mm->page_table_entry_size, GFP_KERNEL);
  1164. if (!mem)
  1165. return -ENOMEM;
  1166. mm->virtual_page_table = mem;
  1167. if (!mm->has_shadow_page_table)
  1168. return 0;
  1169. mm->shadow_page_table = mem + mm->page_table_entry_size;
  1170. } else if (mm->type == INTEL_GVT_MM_GGTT) {
  1171. mm->page_table_entry_cnt =
  1172. (gvt_ggtt_gm_sz(gvt) >> GTT_PAGE_SHIFT);
  1173. mm->page_table_entry_size = mm->page_table_entry_cnt *
  1174. info->gtt_entry_size;
  1175. mem = vzalloc(mm->page_table_entry_size);
  1176. if (!mem)
  1177. return -ENOMEM;
  1178. mm->virtual_page_table = mem;
  1179. }
  1180. return 0;
  1181. }
  1182. static void gen8_mm_free_page_table(struct intel_vgpu_mm *mm)
  1183. {
  1184. if (mm->type == INTEL_GVT_MM_PPGTT) {
  1185. kfree(mm->virtual_page_table);
  1186. } else if (mm->type == INTEL_GVT_MM_GGTT) {
  1187. if (mm->virtual_page_table)
  1188. vfree(mm->virtual_page_table);
  1189. }
  1190. mm->virtual_page_table = mm->shadow_page_table = NULL;
  1191. }
  1192. static void invalidate_mm(struct intel_vgpu_mm *mm)
  1193. {
  1194. struct intel_vgpu *vgpu = mm->vgpu;
  1195. struct intel_gvt *gvt = vgpu->gvt;
  1196. struct intel_gvt_gtt *gtt = &gvt->gtt;
  1197. struct intel_gvt_gtt_pte_ops *ops = gtt->pte_ops;
  1198. struct intel_gvt_gtt_entry se;
  1199. int i;
  1200. if (WARN_ON(!mm->has_shadow_page_table || !mm->shadowed))
  1201. return;
  1202. for (i = 0; i < mm->page_table_entry_cnt; i++) {
  1203. ppgtt_get_shadow_root_entry(mm, &se, i);
  1204. if (!ops->test_present(&se))
  1205. continue;
  1206. ppgtt_invalidate_shadow_page_by_shadow_entry(
  1207. vgpu, &se);
  1208. se.val64 = 0;
  1209. ppgtt_set_shadow_root_entry(mm, &se, i);
  1210. trace_gpt_change(vgpu->id, "destroy root pointer",
  1211. NULL, se.type, se.val64, i);
  1212. }
  1213. mm->shadowed = false;
  1214. }
  1215. /**
  1216. * intel_vgpu_destroy_mm - destroy a mm object
  1217. * @mm: a kref object
  1218. *
  1219. * This function is used to destroy a mm object for vGPU
  1220. *
  1221. */
  1222. void intel_vgpu_destroy_mm(struct kref *mm_ref)
  1223. {
  1224. struct intel_vgpu_mm *mm = container_of(mm_ref, typeof(*mm), ref);
  1225. struct intel_vgpu *vgpu = mm->vgpu;
  1226. struct intel_gvt *gvt = vgpu->gvt;
  1227. struct intel_gvt_gtt *gtt = &gvt->gtt;
  1228. if (!mm->initialized)
  1229. goto out;
  1230. list_del(&mm->list);
  1231. list_del(&mm->lru_list);
  1232. if (mm->has_shadow_page_table)
  1233. invalidate_mm(mm);
  1234. gtt->mm_free_page_table(mm);
  1235. out:
  1236. kfree(mm);
  1237. }
  1238. static int shadow_mm(struct intel_vgpu_mm *mm)
  1239. {
  1240. struct intel_vgpu *vgpu = mm->vgpu;
  1241. struct intel_gvt *gvt = vgpu->gvt;
  1242. struct intel_gvt_gtt *gtt = &gvt->gtt;
  1243. struct intel_gvt_gtt_pte_ops *ops = gtt->pte_ops;
  1244. struct intel_vgpu_ppgtt_spt *spt;
  1245. struct intel_gvt_gtt_entry ge, se;
  1246. int i;
  1247. int ret;
  1248. if (WARN_ON(!mm->has_shadow_page_table || mm->shadowed))
  1249. return 0;
  1250. mm->shadowed = true;
  1251. for (i = 0; i < mm->page_table_entry_cnt; i++) {
  1252. ppgtt_get_guest_root_entry(mm, &ge, i);
  1253. if (!ops->test_present(&ge))
  1254. continue;
  1255. trace_gpt_change(vgpu->id, __func__, NULL,
  1256. ge.type, ge.val64, i);
  1257. spt = ppgtt_populate_shadow_page_by_guest_entry(vgpu, &ge);
  1258. if (IS_ERR(spt)) {
  1259. gvt_vgpu_err("fail to populate guest root pointer\n");
  1260. ret = PTR_ERR(spt);
  1261. goto fail;
  1262. }
  1263. ppgtt_generate_shadow_entry(&se, spt, &ge);
  1264. ppgtt_set_shadow_root_entry(mm, &se, i);
  1265. trace_gpt_change(vgpu->id, "populate root pointer",
  1266. NULL, se.type, se.val64, i);
  1267. }
  1268. return 0;
  1269. fail:
  1270. invalidate_mm(mm);
  1271. return ret;
  1272. }
  1273. /**
  1274. * intel_vgpu_create_mm - create a mm object for a vGPU
  1275. * @vgpu: a vGPU
  1276. * @mm_type: mm object type, should be PPGTT or GGTT
  1277. * @virtual_page_table: page table root pointers. Could be NULL if user wants
  1278. * to populate shadow later.
  1279. * @page_table_level: describe the page table level of the mm object
  1280. * @pde_base_index: pde root pointer base in GGTT MMIO.
  1281. *
  1282. * This function is used to create a mm object for a vGPU.
  1283. *
  1284. * Returns:
  1285. * Zero on success, negative error code in pointer if failed.
  1286. */
  1287. struct intel_vgpu_mm *intel_vgpu_create_mm(struct intel_vgpu *vgpu,
  1288. int mm_type, void *virtual_page_table, int page_table_level,
  1289. u32 pde_base_index)
  1290. {
  1291. struct intel_gvt *gvt = vgpu->gvt;
  1292. struct intel_gvt_gtt *gtt = &gvt->gtt;
  1293. struct intel_vgpu_mm *mm;
  1294. int ret;
  1295. mm = kzalloc(sizeof(*mm), GFP_KERNEL);
  1296. if (!mm) {
  1297. ret = -ENOMEM;
  1298. goto fail;
  1299. }
  1300. mm->type = mm_type;
  1301. if (page_table_level == 1)
  1302. mm->page_table_entry_type = GTT_TYPE_GGTT_PTE;
  1303. else if (page_table_level == 3)
  1304. mm->page_table_entry_type = GTT_TYPE_PPGTT_ROOT_L3_ENTRY;
  1305. else if (page_table_level == 4)
  1306. mm->page_table_entry_type = GTT_TYPE_PPGTT_ROOT_L4_ENTRY;
  1307. else {
  1308. WARN_ON(1);
  1309. ret = -EINVAL;
  1310. goto fail;
  1311. }
  1312. mm->page_table_level = page_table_level;
  1313. mm->pde_base_index = pde_base_index;
  1314. mm->vgpu = vgpu;
  1315. mm->has_shadow_page_table = !!(mm_type == INTEL_GVT_MM_PPGTT);
  1316. kref_init(&mm->ref);
  1317. atomic_set(&mm->pincount, 0);
  1318. INIT_LIST_HEAD(&mm->list);
  1319. INIT_LIST_HEAD(&mm->lru_list);
  1320. list_add_tail(&mm->list, &vgpu->gtt.mm_list_head);
  1321. ret = gtt->mm_alloc_page_table(mm);
  1322. if (ret) {
  1323. gvt_vgpu_err("fail to allocate page table for mm\n");
  1324. goto fail;
  1325. }
  1326. mm->initialized = true;
  1327. if (virtual_page_table)
  1328. memcpy(mm->virtual_page_table, virtual_page_table,
  1329. mm->page_table_entry_size);
  1330. if (mm->has_shadow_page_table) {
  1331. ret = shadow_mm(mm);
  1332. if (ret)
  1333. goto fail;
  1334. list_add_tail(&mm->lru_list, &gvt->gtt.mm_lru_list_head);
  1335. }
  1336. return mm;
  1337. fail:
  1338. gvt_vgpu_err("fail to create mm\n");
  1339. if (mm)
  1340. intel_gvt_mm_unreference(mm);
  1341. return ERR_PTR(ret);
  1342. }
  1343. /**
  1344. * intel_vgpu_unpin_mm - decrease the pin count of a vGPU mm object
  1345. * @mm: a vGPU mm object
  1346. *
  1347. * This function is called when user doesn't want to use a vGPU mm object
  1348. */
  1349. void intel_vgpu_unpin_mm(struct intel_vgpu_mm *mm)
  1350. {
  1351. if (WARN_ON(mm->type != INTEL_GVT_MM_PPGTT))
  1352. return;
  1353. atomic_dec(&mm->pincount);
  1354. }
  1355. /**
  1356. * intel_vgpu_pin_mm - increase the pin count of a vGPU mm object
  1357. * @vgpu: a vGPU
  1358. *
  1359. * This function is called when user wants to use a vGPU mm object. If this
  1360. * mm object hasn't been shadowed yet, the shadow will be populated at this
  1361. * time.
  1362. *
  1363. * Returns:
  1364. * Zero on success, negative error code if failed.
  1365. */
  1366. int intel_vgpu_pin_mm(struct intel_vgpu_mm *mm)
  1367. {
  1368. int ret;
  1369. if (WARN_ON(mm->type != INTEL_GVT_MM_PPGTT))
  1370. return 0;
  1371. atomic_inc(&mm->pincount);
  1372. if (!mm->shadowed) {
  1373. ret = shadow_mm(mm);
  1374. if (ret)
  1375. return ret;
  1376. }
  1377. list_del_init(&mm->lru_list);
  1378. list_add_tail(&mm->lru_list, &mm->vgpu->gvt->gtt.mm_lru_list_head);
  1379. return 0;
  1380. }
  1381. static int reclaim_one_mm(struct intel_gvt *gvt)
  1382. {
  1383. struct intel_vgpu_mm *mm;
  1384. struct list_head *pos, *n;
  1385. list_for_each_safe(pos, n, &gvt->gtt.mm_lru_list_head) {
  1386. mm = container_of(pos, struct intel_vgpu_mm, lru_list);
  1387. if (mm->type != INTEL_GVT_MM_PPGTT)
  1388. continue;
  1389. if (atomic_read(&mm->pincount))
  1390. continue;
  1391. list_del_init(&mm->lru_list);
  1392. invalidate_mm(mm);
  1393. return 1;
  1394. }
  1395. return 0;
  1396. }
  1397. /*
  1398. * GMA translation APIs.
  1399. */
  1400. static inline int ppgtt_get_next_level_entry(struct intel_vgpu_mm *mm,
  1401. struct intel_gvt_gtt_entry *e, unsigned long index, bool guest)
  1402. {
  1403. struct intel_vgpu *vgpu = mm->vgpu;
  1404. struct intel_gvt_gtt_pte_ops *ops = vgpu->gvt->gtt.pte_ops;
  1405. struct intel_vgpu_ppgtt_spt *s;
  1406. if (WARN_ON(!mm->has_shadow_page_table))
  1407. return -EINVAL;
  1408. s = ppgtt_find_shadow_page(vgpu, ops->get_pfn(e));
  1409. if (!s)
  1410. return -ENXIO;
  1411. if (!guest)
  1412. ppgtt_get_shadow_entry(s, e, index);
  1413. else
  1414. ppgtt_get_guest_entry(s, e, index);
  1415. return 0;
  1416. }
  1417. /**
  1418. * intel_vgpu_gma_to_gpa - translate a gma to GPA
  1419. * @mm: mm object. could be a PPGTT or GGTT mm object
  1420. * @gma: graphics memory address in this mm object
  1421. *
  1422. * This function is used to translate a graphics memory address in specific
  1423. * graphics memory space to guest physical address.
  1424. *
  1425. * Returns:
  1426. * Guest physical address on success, INTEL_GVT_INVALID_ADDR if failed.
  1427. */
  1428. unsigned long intel_vgpu_gma_to_gpa(struct intel_vgpu_mm *mm, unsigned long gma)
  1429. {
  1430. struct intel_vgpu *vgpu = mm->vgpu;
  1431. struct intel_gvt *gvt = vgpu->gvt;
  1432. struct intel_gvt_gtt_pte_ops *pte_ops = gvt->gtt.pte_ops;
  1433. struct intel_gvt_gtt_gma_ops *gma_ops = gvt->gtt.gma_ops;
  1434. unsigned long gpa = INTEL_GVT_INVALID_ADDR;
  1435. unsigned long gma_index[4];
  1436. struct intel_gvt_gtt_entry e;
  1437. int i, index;
  1438. int ret;
  1439. if (mm->type != INTEL_GVT_MM_GGTT && mm->type != INTEL_GVT_MM_PPGTT)
  1440. return INTEL_GVT_INVALID_ADDR;
  1441. if (mm->type == INTEL_GVT_MM_GGTT) {
  1442. if (!vgpu_gmadr_is_valid(vgpu, gma))
  1443. goto err;
  1444. ggtt_get_guest_entry(mm, &e,
  1445. gma_ops->gma_to_ggtt_pte_index(gma));
  1446. gpa = (pte_ops->get_pfn(&e) << GTT_PAGE_SHIFT)
  1447. + (gma & ~GTT_PAGE_MASK);
  1448. trace_gma_translate(vgpu->id, "ggtt", 0, 0, gma, gpa);
  1449. return gpa;
  1450. }
  1451. switch (mm->page_table_level) {
  1452. case 4:
  1453. ppgtt_get_shadow_root_entry(mm, &e, 0);
  1454. gma_index[0] = gma_ops->gma_to_pml4_index(gma);
  1455. gma_index[1] = gma_ops->gma_to_l4_pdp_index(gma);
  1456. gma_index[2] = gma_ops->gma_to_pde_index(gma);
  1457. gma_index[3] = gma_ops->gma_to_pte_index(gma);
  1458. index = 4;
  1459. break;
  1460. case 3:
  1461. ppgtt_get_shadow_root_entry(mm, &e,
  1462. gma_ops->gma_to_l3_pdp_index(gma));
  1463. gma_index[0] = gma_ops->gma_to_pde_index(gma);
  1464. gma_index[1] = gma_ops->gma_to_pte_index(gma);
  1465. index = 2;
  1466. break;
  1467. case 2:
  1468. ppgtt_get_shadow_root_entry(mm, &e,
  1469. gma_ops->gma_to_pde_index(gma));
  1470. gma_index[0] = gma_ops->gma_to_pte_index(gma);
  1471. index = 1;
  1472. break;
  1473. default:
  1474. WARN_ON(1);
  1475. goto err;
  1476. }
  1477. /* walk into the shadow page table and get gpa from guest entry */
  1478. for (i = 0; i < index; i++) {
  1479. ret = ppgtt_get_next_level_entry(mm, &e, gma_index[i],
  1480. (i == index - 1));
  1481. if (ret)
  1482. goto err;
  1483. }
  1484. gpa = (pte_ops->get_pfn(&e) << GTT_PAGE_SHIFT)
  1485. + (gma & ~GTT_PAGE_MASK);
  1486. trace_gma_translate(vgpu->id, "ppgtt", 0,
  1487. mm->page_table_level, gma, gpa);
  1488. return gpa;
  1489. err:
  1490. gvt_vgpu_err("invalid mm type: %d gma %lx\n", mm->type, gma);
  1491. return INTEL_GVT_INVALID_ADDR;
  1492. }
  1493. static int emulate_gtt_mmio_read(struct intel_vgpu *vgpu,
  1494. unsigned int off, void *p_data, unsigned int bytes)
  1495. {
  1496. struct intel_vgpu_mm *ggtt_mm = vgpu->gtt.ggtt_mm;
  1497. const struct intel_gvt_device_info *info = &vgpu->gvt->device_info;
  1498. unsigned long index = off >> info->gtt_entry_size_shift;
  1499. struct intel_gvt_gtt_entry e;
  1500. if (bytes != 4 && bytes != 8)
  1501. return -EINVAL;
  1502. ggtt_get_guest_entry(ggtt_mm, &e, index);
  1503. memcpy(p_data, (void *)&e.val64 + (off & (info->gtt_entry_size - 1)),
  1504. bytes);
  1505. return 0;
  1506. }
  1507. /**
  1508. * intel_vgpu_emulate_gtt_mmio_read - emulate GTT MMIO register read
  1509. * @vgpu: a vGPU
  1510. * @off: register offset
  1511. * @p_data: data will be returned to guest
  1512. * @bytes: data length
  1513. *
  1514. * This function is used to emulate the GTT MMIO register read
  1515. *
  1516. * Returns:
  1517. * Zero on success, error code if failed.
  1518. */
  1519. int intel_vgpu_emulate_gtt_mmio_read(struct intel_vgpu *vgpu, unsigned int off,
  1520. void *p_data, unsigned int bytes)
  1521. {
  1522. const struct intel_gvt_device_info *info = &vgpu->gvt->device_info;
  1523. int ret;
  1524. if (bytes != 4 && bytes != 8)
  1525. return -EINVAL;
  1526. off -= info->gtt_start_offset;
  1527. ret = emulate_gtt_mmio_read(vgpu, off, p_data, bytes);
  1528. return ret;
  1529. }
  1530. static int emulate_gtt_mmio_write(struct intel_vgpu *vgpu, unsigned int off,
  1531. void *p_data, unsigned int bytes)
  1532. {
  1533. struct intel_gvt *gvt = vgpu->gvt;
  1534. const struct intel_gvt_device_info *info = &gvt->device_info;
  1535. struct intel_vgpu_mm *ggtt_mm = vgpu->gtt.ggtt_mm;
  1536. struct intel_gvt_gtt_pte_ops *ops = gvt->gtt.pte_ops;
  1537. unsigned long g_gtt_index = off >> info->gtt_entry_size_shift;
  1538. unsigned long gma;
  1539. struct intel_gvt_gtt_entry e, m;
  1540. int ret;
  1541. if (bytes != 4 && bytes != 8)
  1542. return -EINVAL;
  1543. gma = g_gtt_index << GTT_PAGE_SHIFT;
  1544. /* the VM may configure the whole GM space when ballooning is used */
  1545. if (!vgpu_gmadr_is_valid(vgpu, gma))
  1546. return 0;
  1547. ggtt_get_guest_entry(ggtt_mm, &e, g_gtt_index);
  1548. memcpy((void *)&e.val64 + (off & (info->gtt_entry_size - 1)), p_data,
  1549. bytes);
  1550. if (ops->test_present(&e)) {
  1551. ret = gtt_entry_p2m(vgpu, &e, &m);
  1552. if (ret) {
  1553. gvt_vgpu_err("fail to translate guest gtt entry\n");
  1554. /* guest driver may read/write the entry when partial
  1555. * update the entry in this situation p2m will fail
  1556. * settting the shadow entry to point to a scratch page
  1557. */
  1558. ops->set_pfn(&m, gvt->gtt.scratch_ggtt_mfn);
  1559. }
  1560. } else {
  1561. m = e;
  1562. ops->set_pfn(&m, gvt->gtt.scratch_ggtt_mfn);
  1563. }
  1564. ggtt_set_shadow_entry(ggtt_mm, &m, g_gtt_index);
  1565. ggtt_set_guest_entry(ggtt_mm, &e, g_gtt_index);
  1566. return 0;
  1567. }
  1568. /*
  1569. * intel_vgpu_emulate_gtt_mmio_write - emulate GTT MMIO register write
  1570. * @vgpu: a vGPU
  1571. * @off: register offset
  1572. * @p_data: data from guest write
  1573. * @bytes: data length
  1574. *
  1575. * This function is used to emulate the GTT MMIO register write
  1576. *
  1577. * Returns:
  1578. * Zero on success, error code if failed.
  1579. */
  1580. int intel_vgpu_emulate_gtt_mmio_write(struct intel_vgpu *vgpu, unsigned int off,
  1581. void *p_data, unsigned int bytes)
  1582. {
  1583. const struct intel_gvt_device_info *info = &vgpu->gvt->device_info;
  1584. int ret;
  1585. if (bytes != 4 && bytes != 8)
  1586. return -EINVAL;
  1587. off -= info->gtt_start_offset;
  1588. ret = emulate_gtt_mmio_write(vgpu, off, p_data, bytes);
  1589. return ret;
  1590. }
  1591. static int alloc_scratch_pages(struct intel_vgpu *vgpu,
  1592. intel_gvt_gtt_type_t type)
  1593. {
  1594. struct intel_vgpu_gtt *gtt = &vgpu->gtt;
  1595. struct intel_gvt_gtt_pte_ops *ops = vgpu->gvt->gtt.pte_ops;
  1596. int page_entry_num = GTT_PAGE_SIZE >>
  1597. vgpu->gvt->device_info.gtt_entry_size_shift;
  1598. void *scratch_pt;
  1599. int i;
  1600. struct device *dev = &vgpu->gvt->dev_priv->drm.pdev->dev;
  1601. dma_addr_t daddr;
  1602. if (WARN_ON(type < GTT_TYPE_PPGTT_PTE_PT || type >= GTT_TYPE_MAX))
  1603. return -EINVAL;
  1604. scratch_pt = (void *)get_zeroed_page(GFP_KERNEL);
  1605. if (!scratch_pt) {
  1606. gvt_vgpu_err("fail to allocate scratch page\n");
  1607. return -ENOMEM;
  1608. }
  1609. daddr = dma_map_page(dev, virt_to_page(scratch_pt), 0,
  1610. 4096, PCI_DMA_BIDIRECTIONAL);
  1611. if (dma_mapping_error(dev, daddr)) {
  1612. gvt_vgpu_err("fail to dmamap scratch_pt\n");
  1613. __free_page(virt_to_page(scratch_pt));
  1614. return -ENOMEM;
  1615. }
  1616. gtt->scratch_pt[type].page_mfn =
  1617. (unsigned long)(daddr >> GTT_PAGE_SHIFT);
  1618. gtt->scratch_pt[type].page = virt_to_page(scratch_pt);
  1619. gvt_dbg_mm("vgpu%d create scratch_pt: type %d mfn=0x%lx\n",
  1620. vgpu->id, type, gtt->scratch_pt[type].page_mfn);
  1621. /* Build the tree by full filled the scratch pt with the entries which
  1622. * point to the next level scratch pt or scratch page. The
  1623. * scratch_pt[type] indicate the scratch pt/scratch page used by the
  1624. * 'type' pt.
  1625. * e.g. scratch_pt[GTT_TYPE_PPGTT_PDE_PT] is used by
  1626. * GTT_TYPE_PPGTT_PDE_PT level pt, that means this scratch_pt it self
  1627. * is GTT_TYPE_PPGTT_PTE_PT, and full filled by scratch page mfn.
  1628. */
  1629. if (type > GTT_TYPE_PPGTT_PTE_PT && type < GTT_TYPE_MAX) {
  1630. struct intel_gvt_gtt_entry se;
  1631. memset(&se, 0, sizeof(struct intel_gvt_gtt_entry));
  1632. se.type = get_entry_type(type - 1);
  1633. ops->set_pfn(&se, gtt->scratch_pt[type - 1].page_mfn);
  1634. /* The entry parameters like present/writeable/cache type
  1635. * set to the same as i915's scratch page tree.
  1636. */
  1637. se.val64 |= _PAGE_PRESENT | _PAGE_RW;
  1638. if (type == GTT_TYPE_PPGTT_PDE_PT)
  1639. se.val64 |= PPAT_CACHED_INDEX;
  1640. for (i = 0; i < page_entry_num; i++)
  1641. ops->set_entry(scratch_pt, &se, i, false, 0, vgpu);
  1642. }
  1643. return 0;
  1644. }
  1645. static int release_scratch_page_tree(struct intel_vgpu *vgpu)
  1646. {
  1647. int i;
  1648. struct device *dev = &vgpu->gvt->dev_priv->drm.pdev->dev;
  1649. dma_addr_t daddr;
  1650. for (i = GTT_TYPE_PPGTT_PTE_PT; i < GTT_TYPE_MAX; i++) {
  1651. if (vgpu->gtt.scratch_pt[i].page != NULL) {
  1652. daddr = (dma_addr_t)(vgpu->gtt.scratch_pt[i].page_mfn <<
  1653. GTT_PAGE_SHIFT);
  1654. dma_unmap_page(dev, daddr, 4096, PCI_DMA_BIDIRECTIONAL);
  1655. __free_page(vgpu->gtt.scratch_pt[i].page);
  1656. vgpu->gtt.scratch_pt[i].page = NULL;
  1657. vgpu->gtt.scratch_pt[i].page_mfn = 0;
  1658. }
  1659. }
  1660. return 0;
  1661. }
  1662. static int create_scratch_page_tree(struct intel_vgpu *vgpu)
  1663. {
  1664. int i, ret;
  1665. for (i = GTT_TYPE_PPGTT_PTE_PT; i < GTT_TYPE_MAX; i++) {
  1666. ret = alloc_scratch_pages(vgpu, i);
  1667. if (ret)
  1668. goto err;
  1669. }
  1670. return 0;
  1671. err:
  1672. release_scratch_page_tree(vgpu);
  1673. return ret;
  1674. }
  1675. /**
  1676. * intel_vgpu_init_gtt - initialize per-vGPU graphics memory virulization
  1677. * @vgpu: a vGPU
  1678. *
  1679. * This function is used to initialize per-vGPU graphics memory virtualization
  1680. * components.
  1681. *
  1682. * Returns:
  1683. * Zero on success, error code if failed.
  1684. */
  1685. int intel_vgpu_init_gtt(struct intel_vgpu *vgpu)
  1686. {
  1687. struct intel_vgpu_gtt *gtt = &vgpu->gtt;
  1688. struct intel_vgpu_mm *ggtt_mm;
  1689. hash_init(gtt->guest_page_hash_table);
  1690. hash_init(gtt->shadow_page_hash_table);
  1691. INIT_LIST_HEAD(&gtt->mm_list_head);
  1692. INIT_LIST_HEAD(&gtt->oos_page_list_head);
  1693. INIT_LIST_HEAD(&gtt->post_shadow_list_head);
  1694. intel_vgpu_reset_ggtt(vgpu);
  1695. ggtt_mm = intel_vgpu_create_mm(vgpu, INTEL_GVT_MM_GGTT,
  1696. NULL, 1, 0);
  1697. if (IS_ERR(ggtt_mm)) {
  1698. gvt_vgpu_err("fail to create mm for ggtt.\n");
  1699. return PTR_ERR(ggtt_mm);
  1700. }
  1701. gtt->ggtt_mm = ggtt_mm;
  1702. return create_scratch_page_tree(vgpu);
  1703. }
  1704. static void intel_vgpu_free_mm(struct intel_vgpu *vgpu, int type)
  1705. {
  1706. struct list_head *pos, *n;
  1707. struct intel_vgpu_mm *mm;
  1708. list_for_each_safe(pos, n, &vgpu->gtt.mm_list_head) {
  1709. mm = container_of(pos, struct intel_vgpu_mm, list);
  1710. if (mm->type == type) {
  1711. vgpu->gvt->gtt.mm_free_page_table(mm);
  1712. list_del(&mm->list);
  1713. list_del(&mm->lru_list);
  1714. kfree(mm);
  1715. }
  1716. }
  1717. }
  1718. /**
  1719. * intel_vgpu_clean_gtt - clean up per-vGPU graphics memory virulization
  1720. * @vgpu: a vGPU
  1721. *
  1722. * This function is used to clean up per-vGPU graphics memory virtualization
  1723. * components.
  1724. *
  1725. * Returns:
  1726. * Zero on success, error code if failed.
  1727. */
  1728. void intel_vgpu_clean_gtt(struct intel_vgpu *vgpu)
  1729. {
  1730. ppgtt_free_all_shadow_page(vgpu);
  1731. release_scratch_page_tree(vgpu);
  1732. intel_vgpu_free_mm(vgpu, INTEL_GVT_MM_PPGTT);
  1733. intel_vgpu_free_mm(vgpu, INTEL_GVT_MM_GGTT);
  1734. }
  1735. static void clean_spt_oos(struct intel_gvt *gvt)
  1736. {
  1737. struct intel_gvt_gtt *gtt = &gvt->gtt;
  1738. struct list_head *pos, *n;
  1739. struct intel_vgpu_oos_page *oos_page;
  1740. WARN(!list_empty(&gtt->oos_page_use_list_head),
  1741. "someone is still using oos page\n");
  1742. list_for_each_safe(pos, n, &gtt->oos_page_free_list_head) {
  1743. oos_page = container_of(pos, struct intel_vgpu_oos_page, list);
  1744. list_del(&oos_page->list);
  1745. kfree(oos_page);
  1746. }
  1747. }
  1748. static int setup_spt_oos(struct intel_gvt *gvt)
  1749. {
  1750. struct intel_gvt_gtt *gtt = &gvt->gtt;
  1751. struct intel_vgpu_oos_page *oos_page;
  1752. int i;
  1753. int ret;
  1754. INIT_LIST_HEAD(&gtt->oos_page_free_list_head);
  1755. INIT_LIST_HEAD(&gtt->oos_page_use_list_head);
  1756. for (i = 0; i < preallocated_oos_pages; i++) {
  1757. oos_page = kzalloc(sizeof(*oos_page), GFP_KERNEL);
  1758. if (!oos_page) {
  1759. ret = -ENOMEM;
  1760. goto fail;
  1761. }
  1762. INIT_LIST_HEAD(&oos_page->list);
  1763. INIT_LIST_HEAD(&oos_page->vm_list);
  1764. oos_page->id = i;
  1765. list_add_tail(&oos_page->list, &gtt->oos_page_free_list_head);
  1766. }
  1767. gvt_dbg_mm("%d oos pages preallocated\n", i);
  1768. return 0;
  1769. fail:
  1770. clean_spt_oos(gvt);
  1771. return ret;
  1772. }
  1773. /**
  1774. * intel_vgpu_find_ppgtt_mm - find a PPGTT mm object
  1775. * @vgpu: a vGPU
  1776. * @page_table_level: PPGTT page table level
  1777. * @root_entry: PPGTT page table root pointers
  1778. *
  1779. * This function is used to find a PPGTT mm object from mm object pool
  1780. *
  1781. * Returns:
  1782. * pointer to mm object on success, NULL if failed.
  1783. */
  1784. struct intel_vgpu_mm *intel_vgpu_find_ppgtt_mm(struct intel_vgpu *vgpu,
  1785. int page_table_level, void *root_entry)
  1786. {
  1787. struct list_head *pos;
  1788. struct intel_vgpu_mm *mm;
  1789. u64 *src, *dst;
  1790. list_for_each(pos, &vgpu->gtt.mm_list_head) {
  1791. mm = container_of(pos, struct intel_vgpu_mm, list);
  1792. if (mm->type != INTEL_GVT_MM_PPGTT)
  1793. continue;
  1794. if (mm->page_table_level != page_table_level)
  1795. continue;
  1796. src = root_entry;
  1797. dst = mm->virtual_page_table;
  1798. if (page_table_level == 3) {
  1799. if (src[0] == dst[0]
  1800. && src[1] == dst[1]
  1801. && src[2] == dst[2]
  1802. && src[3] == dst[3])
  1803. return mm;
  1804. } else {
  1805. if (src[0] == dst[0])
  1806. return mm;
  1807. }
  1808. }
  1809. return NULL;
  1810. }
  1811. /**
  1812. * intel_vgpu_g2v_create_ppgtt_mm - create a PPGTT mm object from
  1813. * g2v notification
  1814. * @vgpu: a vGPU
  1815. * @page_table_level: PPGTT page table level
  1816. *
  1817. * This function is used to create a PPGTT mm object from a guest to GVT-g
  1818. * notification.
  1819. *
  1820. * Returns:
  1821. * Zero on success, negative error code if failed.
  1822. */
  1823. int intel_vgpu_g2v_create_ppgtt_mm(struct intel_vgpu *vgpu,
  1824. int page_table_level)
  1825. {
  1826. u64 *pdp = (u64 *)&vgpu_vreg64(vgpu, vgtif_reg(pdp[0]));
  1827. struct intel_vgpu_mm *mm;
  1828. if (WARN_ON((page_table_level != 4) && (page_table_level != 3)))
  1829. return -EINVAL;
  1830. mm = intel_vgpu_find_ppgtt_mm(vgpu, page_table_level, pdp);
  1831. if (mm) {
  1832. intel_gvt_mm_reference(mm);
  1833. } else {
  1834. mm = intel_vgpu_create_mm(vgpu, INTEL_GVT_MM_PPGTT,
  1835. pdp, page_table_level, 0);
  1836. if (IS_ERR(mm)) {
  1837. gvt_vgpu_err("fail to create mm\n");
  1838. return PTR_ERR(mm);
  1839. }
  1840. }
  1841. return 0;
  1842. }
  1843. /**
  1844. * intel_vgpu_g2v_destroy_ppgtt_mm - destroy a PPGTT mm object from
  1845. * g2v notification
  1846. * @vgpu: a vGPU
  1847. * @page_table_level: PPGTT page table level
  1848. *
  1849. * This function is used to create a PPGTT mm object from a guest to GVT-g
  1850. * notification.
  1851. *
  1852. * Returns:
  1853. * Zero on success, negative error code if failed.
  1854. */
  1855. int intel_vgpu_g2v_destroy_ppgtt_mm(struct intel_vgpu *vgpu,
  1856. int page_table_level)
  1857. {
  1858. u64 *pdp = (u64 *)&vgpu_vreg64(vgpu, vgtif_reg(pdp[0]));
  1859. struct intel_vgpu_mm *mm;
  1860. if (WARN_ON((page_table_level != 4) && (page_table_level != 3)))
  1861. return -EINVAL;
  1862. mm = intel_vgpu_find_ppgtt_mm(vgpu, page_table_level, pdp);
  1863. if (!mm) {
  1864. gvt_vgpu_err("fail to find ppgtt instance.\n");
  1865. return -EINVAL;
  1866. }
  1867. intel_gvt_mm_unreference(mm);
  1868. return 0;
  1869. }
  1870. /**
  1871. * intel_gvt_init_gtt - initialize mm components of a GVT device
  1872. * @gvt: GVT device
  1873. *
  1874. * This function is called at the initialization stage, to initialize
  1875. * the mm components of a GVT device.
  1876. *
  1877. * Returns:
  1878. * zero on success, negative error code if failed.
  1879. */
  1880. int intel_gvt_init_gtt(struct intel_gvt *gvt)
  1881. {
  1882. int ret;
  1883. void *page;
  1884. struct device *dev = &gvt->dev_priv->drm.pdev->dev;
  1885. dma_addr_t daddr;
  1886. gvt_dbg_core("init gtt\n");
  1887. if (IS_BROADWELL(gvt->dev_priv) || IS_SKYLAKE(gvt->dev_priv)
  1888. || IS_KABYLAKE(gvt->dev_priv)) {
  1889. gvt->gtt.pte_ops = &gen8_gtt_pte_ops;
  1890. gvt->gtt.gma_ops = &gen8_gtt_gma_ops;
  1891. gvt->gtt.mm_alloc_page_table = gen8_mm_alloc_page_table;
  1892. gvt->gtt.mm_free_page_table = gen8_mm_free_page_table;
  1893. } else {
  1894. return -ENODEV;
  1895. }
  1896. page = (void *)get_zeroed_page(GFP_KERNEL);
  1897. if (!page) {
  1898. gvt_err("fail to allocate scratch ggtt page\n");
  1899. return -ENOMEM;
  1900. }
  1901. daddr = dma_map_page(dev, virt_to_page(page), 0,
  1902. 4096, PCI_DMA_BIDIRECTIONAL);
  1903. if (dma_mapping_error(dev, daddr)) {
  1904. gvt_err("fail to dmamap scratch ggtt page\n");
  1905. __free_page(virt_to_page(page));
  1906. return -ENOMEM;
  1907. }
  1908. gvt->gtt.scratch_ggtt_page = virt_to_page(page);
  1909. gvt->gtt.scratch_ggtt_mfn = (unsigned long)(daddr >> GTT_PAGE_SHIFT);
  1910. if (enable_out_of_sync) {
  1911. ret = setup_spt_oos(gvt);
  1912. if (ret) {
  1913. gvt_err("fail to initialize SPT oos\n");
  1914. return ret;
  1915. }
  1916. }
  1917. INIT_LIST_HEAD(&gvt->gtt.mm_lru_list_head);
  1918. return 0;
  1919. }
  1920. /**
  1921. * intel_gvt_clean_gtt - clean up mm components of a GVT device
  1922. * @gvt: GVT device
  1923. *
  1924. * This function is called at the driver unloading stage, to clean up the
  1925. * the mm components of a GVT device.
  1926. *
  1927. */
  1928. void intel_gvt_clean_gtt(struct intel_gvt *gvt)
  1929. {
  1930. struct device *dev = &gvt->dev_priv->drm.pdev->dev;
  1931. dma_addr_t daddr = (dma_addr_t)(gvt->gtt.scratch_ggtt_mfn <<
  1932. GTT_PAGE_SHIFT);
  1933. dma_unmap_page(dev, daddr, 4096, PCI_DMA_BIDIRECTIONAL);
  1934. __free_page(gvt->gtt.scratch_ggtt_page);
  1935. if (enable_out_of_sync)
  1936. clean_spt_oos(gvt);
  1937. }
  1938. /**
  1939. * intel_vgpu_reset_ggtt - reset the GGTT entry
  1940. * @vgpu: a vGPU
  1941. *
  1942. * This function is called at the vGPU create stage
  1943. * to reset all the GGTT entries.
  1944. *
  1945. */
  1946. void intel_vgpu_reset_ggtt(struct intel_vgpu *vgpu)
  1947. {
  1948. struct intel_gvt *gvt = vgpu->gvt;
  1949. struct drm_i915_private *dev_priv = gvt->dev_priv;
  1950. struct intel_gvt_gtt_pte_ops *ops = vgpu->gvt->gtt.pte_ops;
  1951. u32 index;
  1952. u32 offset;
  1953. u32 num_entries;
  1954. struct intel_gvt_gtt_entry e;
  1955. intel_runtime_pm_get(dev_priv);
  1956. memset(&e, 0, sizeof(struct intel_gvt_gtt_entry));
  1957. e.type = GTT_TYPE_GGTT_PTE;
  1958. ops->set_pfn(&e, gvt->gtt.scratch_ggtt_mfn);
  1959. e.val64 |= _PAGE_PRESENT;
  1960. index = vgpu_aperture_gmadr_base(vgpu) >> PAGE_SHIFT;
  1961. num_entries = vgpu_aperture_sz(vgpu) >> PAGE_SHIFT;
  1962. for (offset = 0; offset < num_entries; offset++)
  1963. ops->set_entry(NULL, &e, index + offset, false, 0, vgpu);
  1964. index = vgpu_hidden_gmadr_base(vgpu) >> PAGE_SHIFT;
  1965. num_entries = vgpu_hidden_sz(vgpu) >> PAGE_SHIFT;
  1966. for (offset = 0; offset < num_entries; offset++)
  1967. ops->set_entry(NULL, &e, index + offset, false, 0, vgpu);
  1968. intel_runtime_pm_put(dev_priv);
  1969. }
  1970. /**
  1971. * intel_vgpu_reset_gtt - reset the all GTT related status
  1972. * @vgpu: a vGPU
  1973. * @dmlr: true for vGPU Device Model Level Reset, false for GT Reset
  1974. *
  1975. * This function is called from vfio core to reset reset all
  1976. * GTT related status, including GGTT, PPGTT, scratch page.
  1977. *
  1978. */
  1979. void intel_vgpu_reset_gtt(struct intel_vgpu *vgpu, bool dmlr)
  1980. {
  1981. int i;
  1982. ppgtt_free_all_shadow_page(vgpu);
  1983. /* Shadow pages are only created when there is no page
  1984. * table tracking data, so remove page tracking data after
  1985. * removing the shadow pages.
  1986. */
  1987. intel_vgpu_free_mm(vgpu, INTEL_GVT_MM_PPGTT);
  1988. if (!dmlr)
  1989. return;
  1990. intel_vgpu_reset_ggtt(vgpu);
  1991. /* clear scratch page for security */
  1992. for (i = GTT_TYPE_PPGTT_PTE_PT; i < GTT_TYPE_MAX; i++) {
  1993. if (vgpu->gtt.scratch_pt[i].page != NULL)
  1994. memset(page_address(vgpu->gtt.scratch_pt[i].page),
  1995. 0, PAGE_SIZE);
  1996. }
  1997. }