gtt.c 59 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310131113121313131413151316131713181319132013211322132313241325132613271328132913301331133213331334133513361337133813391340134113421343134413451346134713481349135013511352135313541355135613571358135913601361136213631364136513661367136813691370137113721373137413751376137713781379138013811382138313841385138613871388138913901391139213931394139513961397139813991400140114021403140414051406140714081409141014111412141314141415141614171418141914201421142214231424142514261427142814291430143114321433143414351436143714381439144014411442144314441445144614471448144914501451145214531454145514561457145814591460146114621463146414651466146714681469147014711472147314741475147614771478147914801481148214831484148514861487148814891490149114921493149414951496149714981499150015011502150315041505150615071508150915101511151215131514151515161517151815191520152115221523152415251526152715281529153015311532153315341535153615371538153915401541154215431544154515461547154815491550155115521553155415551556155715581559156015611562156315641565156615671568156915701571157215731574157515761577157815791580158115821583158415851586158715881589159015911592159315941595159615971598159916001601160216031604160516061607160816091610161116121613161416151616161716181619162016211622162316241625162616271628162916301631163216331634163516361637163816391640164116421643164416451646164716481649165016511652165316541655165616571658165916601661166216631664166516661667166816691670167116721673167416751676167716781679168016811682168316841685168616871688168916901691169216931694169516961697169816991700170117021703170417051706170717081709171017111712171317141715171617171718171917201721172217231724172517261727172817291730173117321733173417351736173717381739174017411742174317441745174617471748174917501751175217531754175517561757175817591760176117621763176417651766176717681769177017711772177317741775177617771778177917801781178217831784178517861787178817891790179117921793179417951796179717981799180018011802180318041805180618071808180918101811181218131814181518161817181818191820182118221823182418251826182718281829183018311832183318341835183618371838183918401841184218431844184518461847184818491850185118521853185418551856185718581859186018611862186318641865186618671868186918701871187218731874187518761877187818791880188118821883188418851886188718881889189018911892189318941895189618971898189919001901190219031904190519061907190819091910191119121913191419151916191719181919192019211922192319241925192619271928192919301931193219331934193519361937193819391940194119421943194419451946194719481949195019511952195319541955195619571958195919601961196219631964196519661967196819691970197119721973197419751976197719781979198019811982198319841985198619871988198919901991199219931994199519961997199819992000200120022003200420052006200720082009201020112012201320142015201620172018201920202021202220232024202520262027202820292030203120322033203420352036203720382039204020412042204320442045204620472048204920502051205220532054205520562057205820592060206120622063206420652066206720682069207020712072207320742075207620772078207920802081208220832084208520862087208820892090209120922093209420952096209720982099210021012102210321042105210621072108210921102111211221132114211521162117211821192120212121222123212421252126212721282129213021312132213321342135213621372138213921402141214221432144214521462147214821492150215121522153215421552156215721582159216021612162216321642165216621672168216921702171217221732174217521762177217821792180218121822183218421852186218721882189219021912192219321942195219621972198219922002201220222032204220522062207220822092210221122122213221422152216221722182219222022212222222322242225222622272228222922302231223222332234223522362237223822392240224122422243224422452246224722482249225022512252225322542255225622572258225922602261226222632264226522662267226822692270227122722273227422752276227722782279228022812282228322842285228622872288228922902291229222932294229522962297229822992300230123022303230423052306
  1. /*
  2. * GTT virtualization
  3. *
  4. * Copyright(c) 2011-2016 Intel Corporation. All rights reserved.
  5. *
  6. * Permission is hereby granted, free of charge, to any person obtaining a
  7. * copy of this software and associated documentation files (the "Software"),
  8. * to deal in the Software without restriction, including without limitation
  9. * the rights to use, copy, modify, merge, publish, distribute, sublicense,
  10. * and/or sell copies of the Software, and to permit persons to whom the
  11. * Software is furnished to do so, subject to the following conditions:
  12. *
  13. * The above copyright notice and this permission notice (including the next
  14. * paragraph) shall be included in all copies or substantial portions of the
  15. * Software.
  16. *
  17. * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
  18. * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
  19. * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
  20. * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
  21. * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
  22. * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
  23. * SOFTWARE.
  24. *
  25. * Authors:
  26. * Zhi Wang <zhi.a.wang@intel.com>
  27. * Zhenyu Wang <zhenyuw@linux.intel.com>
  28. * Xiao Zheng <xiao.zheng@intel.com>
  29. *
  30. * Contributors:
  31. * Min He <min.he@intel.com>
  32. * Bing Niu <bing.niu@intel.com>
  33. *
  34. */
  35. #include "i915_drv.h"
  36. #include "gvt.h"
  37. #include "i915_pvinfo.h"
  38. #include "trace.h"
  39. static bool enable_out_of_sync = false;
  40. static int preallocated_oos_pages = 8192;
  41. /*
  42. * validate a gm address and related range size,
  43. * translate it to host gm address
  44. */
  45. bool intel_gvt_ggtt_validate_range(struct intel_vgpu *vgpu, u64 addr, u32 size)
  46. {
  47. if ((!vgpu_gmadr_is_valid(vgpu, addr)) || (size
  48. && !vgpu_gmadr_is_valid(vgpu, addr + size - 1))) {
  49. gvt_err("vgpu%d: invalid range gmadr 0x%llx size 0x%x\n",
  50. vgpu->id, addr, size);
  51. return false;
  52. }
  53. return true;
  54. }
  55. /* translate a guest gmadr to host gmadr */
  56. int intel_gvt_ggtt_gmadr_g2h(struct intel_vgpu *vgpu, u64 g_addr, u64 *h_addr)
  57. {
  58. if (WARN(!vgpu_gmadr_is_valid(vgpu, g_addr),
  59. "invalid guest gmadr %llx\n", g_addr))
  60. return -EACCES;
  61. if (vgpu_gmadr_is_aperture(vgpu, g_addr))
  62. *h_addr = vgpu_aperture_gmadr_base(vgpu)
  63. + (g_addr - vgpu_aperture_offset(vgpu));
  64. else
  65. *h_addr = vgpu_hidden_gmadr_base(vgpu)
  66. + (g_addr - vgpu_hidden_offset(vgpu));
  67. return 0;
  68. }
  69. /* translate a host gmadr to guest gmadr */
  70. int intel_gvt_ggtt_gmadr_h2g(struct intel_vgpu *vgpu, u64 h_addr, u64 *g_addr)
  71. {
  72. if (WARN(!gvt_gmadr_is_valid(vgpu->gvt, h_addr),
  73. "invalid host gmadr %llx\n", h_addr))
  74. return -EACCES;
  75. if (gvt_gmadr_is_aperture(vgpu->gvt, h_addr))
  76. *g_addr = vgpu_aperture_gmadr_base(vgpu)
  77. + (h_addr - gvt_aperture_gmadr_base(vgpu->gvt));
  78. else
  79. *g_addr = vgpu_hidden_gmadr_base(vgpu)
  80. + (h_addr - gvt_hidden_gmadr_base(vgpu->gvt));
  81. return 0;
  82. }
  83. int intel_gvt_ggtt_index_g2h(struct intel_vgpu *vgpu, unsigned long g_index,
  84. unsigned long *h_index)
  85. {
  86. u64 h_addr;
  87. int ret;
  88. ret = intel_gvt_ggtt_gmadr_g2h(vgpu, g_index << GTT_PAGE_SHIFT,
  89. &h_addr);
  90. if (ret)
  91. return ret;
  92. *h_index = h_addr >> GTT_PAGE_SHIFT;
  93. return 0;
  94. }
  95. int intel_gvt_ggtt_h2g_index(struct intel_vgpu *vgpu, unsigned long h_index,
  96. unsigned long *g_index)
  97. {
  98. u64 g_addr;
  99. int ret;
  100. ret = intel_gvt_ggtt_gmadr_h2g(vgpu, h_index << GTT_PAGE_SHIFT,
  101. &g_addr);
  102. if (ret)
  103. return ret;
  104. *g_index = g_addr >> GTT_PAGE_SHIFT;
  105. return 0;
  106. }
  107. #define gtt_type_is_entry(type) \
  108. (type > GTT_TYPE_INVALID && type < GTT_TYPE_PPGTT_ENTRY \
  109. && type != GTT_TYPE_PPGTT_PTE_ENTRY \
  110. && type != GTT_TYPE_PPGTT_ROOT_ENTRY)
  111. #define gtt_type_is_pt(type) \
  112. (type >= GTT_TYPE_PPGTT_PTE_PT && type < GTT_TYPE_MAX)
  113. #define gtt_type_is_pte_pt(type) \
  114. (type == GTT_TYPE_PPGTT_PTE_PT)
  115. #define gtt_type_is_root_pointer(type) \
  116. (gtt_type_is_entry(type) && type > GTT_TYPE_PPGTT_ROOT_ENTRY)
  117. #define gtt_init_entry(e, t, p, v) do { \
  118. (e)->type = t; \
  119. (e)->pdev = p; \
  120. memcpy(&(e)->val64, &v, sizeof(v)); \
  121. } while (0)
  122. /*
  123. * Mappings between GTT_TYPE* enumerations.
  124. * Following information can be found according to the given type:
  125. * - type of next level page table
  126. * - type of entry inside this level page table
  127. * - type of entry with PSE set
  128. *
  129. * If the given type doesn't have such a kind of information,
  130. * e.g. give a l4 root entry type, then request to get its PSE type,
  131. * give a PTE page table type, then request to get its next level page
  132. * table type, as we know l4 root entry doesn't have a PSE bit,
  133. * and a PTE page table doesn't have a next level page table type,
  134. * GTT_TYPE_INVALID will be returned. This is useful when traversing a
  135. * page table.
  136. */
  137. struct gtt_type_table_entry {
  138. int entry_type;
  139. int next_pt_type;
  140. int pse_entry_type;
  141. };
  142. #define GTT_TYPE_TABLE_ENTRY(type, e_type, npt_type, pse_type) \
  143. [type] = { \
  144. .entry_type = e_type, \
  145. .next_pt_type = npt_type, \
  146. .pse_entry_type = pse_type, \
  147. }
  148. static struct gtt_type_table_entry gtt_type_table[] = {
  149. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_PPGTT_ROOT_L4_ENTRY,
  150. GTT_TYPE_PPGTT_ROOT_L4_ENTRY,
  151. GTT_TYPE_PPGTT_PML4_PT,
  152. GTT_TYPE_INVALID),
  153. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_PPGTT_PML4_PT,
  154. GTT_TYPE_PPGTT_PML4_ENTRY,
  155. GTT_TYPE_PPGTT_PDP_PT,
  156. GTT_TYPE_INVALID),
  157. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_PPGTT_PML4_ENTRY,
  158. GTT_TYPE_PPGTT_PML4_ENTRY,
  159. GTT_TYPE_PPGTT_PDP_PT,
  160. GTT_TYPE_INVALID),
  161. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_PPGTT_PDP_PT,
  162. GTT_TYPE_PPGTT_PDP_ENTRY,
  163. GTT_TYPE_PPGTT_PDE_PT,
  164. GTT_TYPE_PPGTT_PTE_1G_ENTRY),
  165. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_PPGTT_ROOT_L3_ENTRY,
  166. GTT_TYPE_PPGTT_ROOT_L3_ENTRY,
  167. GTT_TYPE_PPGTT_PDE_PT,
  168. GTT_TYPE_PPGTT_PTE_1G_ENTRY),
  169. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_PPGTT_PDP_ENTRY,
  170. GTT_TYPE_PPGTT_PDP_ENTRY,
  171. GTT_TYPE_PPGTT_PDE_PT,
  172. GTT_TYPE_PPGTT_PTE_1G_ENTRY),
  173. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_PPGTT_PDE_PT,
  174. GTT_TYPE_PPGTT_PDE_ENTRY,
  175. GTT_TYPE_PPGTT_PTE_PT,
  176. GTT_TYPE_PPGTT_PTE_2M_ENTRY),
  177. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_PPGTT_PDE_ENTRY,
  178. GTT_TYPE_PPGTT_PDE_ENTRY,
  179. GTT_TYPE_PPGTT_PTE_PT,
  180. GTT_TYPE_PPGTT_PTE_2M_ENTRY),
  181. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_PPGTT_PTE_PT,
  182. GTT_TYPE_PPGTT_PTE_4K_ENTRY,
  183. GTT_TYPE_INVALID,
  184. GTT_TYPE_INVALID),
  185. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_PPGTT_PTE_4K_ENTRY,
  186. GTT_TYPE_PPGTT_PTE_4K_ENTRY,
  187. GTT_TYPE_INVALID,
  188. GTT_TYPE_INVALID),
  189. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_PPGTT_PTE_2M_ENTRY,
  190. GTT_TYPE_PPGTT_PDE_ENTRY,
  191. GTT_TYPE_INVALID,
  192. GTT_TYPE_PPGTT_PTE_2M_ENTRY),
  193. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_PPGTT_PTE_1G_ENTRY,
  194. GTT_TYPE_PPGTT_PDP_ENTRY,
  195. GTT_TYPE_INVALID,
  196. GTT_TYPE_PPGTT_PTE_1G_ENTRY),
  197. GTT_TYPE_TABLE_ENTRY(GTT_TYPE_GGTT_PTE,
  198. GTT_TYPE_GGTT_PTE,
  199. GTT_TYPE_INVALID,
  200. GTT_TYPE_INVALID),
  201. };
  202. static inline int get_next_pt_type(int type)
  203. {
  204. return gtt_type_table[type].next_pt_type;
  205. }
  206. static inline int get_entry_type(int type)
  207. {
  208. return gtt_type_table[type].entry_type;
  209. }
  210. static inline int get_pse_type(int type)
  211. {
  212. return gtt_type_table[type].pse_entry_type;
  213. }
  214. static u64 read_pte64(struct drm_i915_private *dev_priv, unsigned long index)
  215. {
  216. void __iomem *addr = (gen8_pte_t __iomem *)dev_priv->ggtt.gsm + index;
  217. return readq(addr);
  218. }
  219. static void write_pte64(struct drm_i915_private *dev_priv,
  220. unsigned long index, u64 pte)
  221. {
  222. void __iomem *addr = (gen8_pte_t __iomem *)dev_priv->ggtt.gsm + index;
  223. writeq(pte, addr);
  224. I915_WRITE(GFX_FLSH_CNTL_GEN6, GFX_FLSH_CNTL_EN);
  225. POSTING_READ(GFX_FLSH_CNTL_GEN6);
  226. }
  227. static inline struct intel_gvt_gtt_entry *gtt_get_entry64(void *pt,
  228. struct intel_gvt_gtt_entry *e,
  229. unsigned long index, bool hypervisor_access, unsigned long gpa,
  230. struct intel_vgpu *vgpu)
  231. {
  232. const struct intel_gvt_device_info *info = &vgpu->gvt->device_info;
  233. int ret;
  234. if (WARN_ON(info->gtt_entry_size != 8))
  235. return e;
  236. if (hypervisor_access) {
  237. ret = intel_gvt_hypervisor_read_gpa(vgpu, gpa +
  238. (index << info->gtt_entry_size_shift),
  239. &e->val64, 8);
  240. WARN_ON(ret);
  241. } else if (!pt) {
  242. e->val64 = read_pte64(vgpu->gvt->dev_priv, index);
  243. } else {
  244. e->val64 = *((u64 *)pt + index);
  245. }
  246. return e;
  247. }
  248. static inline struct intel_gvt_gtt_entry *gtt_set_entry64(void *pt,
  249. struct intel_gvt_gtt_entry *e,
  250. unsigned long index, bool hypervisor_access, unsigned long gpa,
  251. struct intel_vgpu *vgpu)
  252. {
  253. const struct intel_gvt_device_info *info = &vgpu->gvt->device_info;
  254. int ret;
  255. if (WARN_ON(info->gtt_entry_size != 8))
  256. return e;
  257. if (hypervisor_access) {
  258. ret = intel_gvt_hypervisor_write_gpa(vgpu, gpa +
  259. (index << info->gtt_entry_size_shift),
  260. &e->val64, 8);
  261. WARN_ON(ret);
  262. } else if (!pt) {
  263. write_pte64(vgpu->gvt->dev_priv, index, e->val64);
  264. } else {
  265. *((u64 *)pt + index) = e->val64;
  266. }
  267. return e;
  268. }
  269. #define GTT_HAW 46
  270. #define ADDR_1G_MASK (((1UL << (GTT_HAW - 30 + 1)) - 1) << 30)
  271. #define ADDR_2M_MASK (((1UL << (GTT_HAW - 21 + 1)) - 1) << 21)
  272. #define ADDR_4K_MASK (((1UL << (GTT_HAW - 12 + 1)) - 1) << 12)
  273. static unsigned long gen8_gtt_get_pfn(struct intel_gvt_gtt_entry *e)
  274. {
  275. unsigned long pfn;
  276. if (e->type == GTT_TYPE_PPGTT_PTE_1G_ENTRY)
  277. pfn = (e->val64 & ADDR_1G_MASK) >> 12;
  278. else if (e->type == GTT_TYPE_PPGTT_PTE_2M_ENTRY)
  279. pfn = (e->val64 & ADDR_2M_MASK) >> 12;
  280. else
  281. pfn = (e->val64 & ADDR_4K_MASK) >> 12;
  282. return pfn;
  283. }
  284. static void gen8_gtt_set_pfn(struct intel_gvt_gtt_entry *e, unsigned long pfn)
  285. {
  286. if (e->type == GTT_TYPE_PPGTT_PTE_1G_ENTRY) {
  287. e->val64 &= ~ADDR_1G_MASK;
  288. pfn &= (ADDR_1G_MASK >> 12);
  289. } else if (e->type == GTT_TYPE_PPGTT_PTE_2M_ENTRY) {
  290. e->val64 &= ~ADDR_2M_MASK;
  291. pfn &= (ADDR_2M_MASK >> 12);
  292. } else {
  293. e->val64 &= ~ADDR_4K_MASK;
  294. pfn &= (ADDR_4K_MASK >> 12);
  295. }
  296. e->val64 |= (pfn << 12);
  297. }
  298. static bool gen8_gtt_test_pse(struct intel_gvt_gtt_entry *e)
  299. {
  300. /* Entry doesn't have PSE bit. */
  301. if (get_pse_type(e->type) == GTT_TYPE_INVALID)
  302. return false;
  303. e->type = get_entry_type(e->type);
  304. if (!(e->val64 & (1 << 7)))
  305. return false;
  306. e->type = get_pse_type(e->type);
  307. return true;
  308. }
  309. static bool gen8_gtt_test_present(struct intel_gvt_gtt_entry *e)
  310. {
  311. /*
  312. * i915 writes PDP root pointer registers without present bit,
  313. * it also works, so we need to treat root pointer entry
  314. * specifically.
  315. */
  316. if (e->type == GTT_TYPE_PPGTT_ROOT_L3_ENTRY
  317. || e->type == GTT_TYPE_PPGTT_ROOT_L4_ENTRY)
  318. return (e->val64 != 0);
  319. else
  320. return (e->val64 & (1 << 0));
  321. }
  322. static void gtt_entry_clear_present(struct intel_gvt_gtt_entry *e)
  323. {
  324. e->val64 &= ~(1 << 0);
  325. }
  326. /*
  327. * Per-platform GMA routines.
  328. */
  329. static unsigned long gma_to_ggtt_pte_index(unsigned long gma)
  330. {
  331. unsigned long x = (gma >> GTT_PAGE_SHIFT);
  332. trace_gma_index(__func__, gma, x);
  333. return x;
  334. }
  335. #define DEFINE_PPGTT_GMA_TO_INDEX(prefix, ename, exp) \
  336. static unsigned long prefix##_gma_to_##ename##_index(unsigned long gma) \
  337. { \
  338. unsigned long x = (exp); \
  339. trace_gma_index(__func__, gma, x); \
  340. return x; \
  341. }
  342. DEFINE_PPGTT_GMA_TO_INDEX(gen8, pte, (gma >> 12 & 0x1ff));
  343. DEFINE_PPGTT_GMA_TO_INDEX(gen8, pde, (gma >> 21 & 0x1ff));
  344. DEFINE_PPGTT_GMA_TO_INDEX(gen8, l3_pdp, (gma >> 30 & 0x3));
  345. DEFINE_PPGTT_GMA_TO_INDEX(gen8, l4_pdp, (gma >> 30 & 0x1ff));
  346. DEFINE_PPGTT_GMA_TO_INDEX(gen8, pml4, (gma >> 39 & 0x1ff));
  347. static struct intel_gvt_gtt_pte_ops gen8_gtt_pte_ops = {
  348. .get_entry = gtt_get_entry64,
  349. .set_entry = gtt_set_entry64,
  350. .clear_present = gtt_entry_clear_present,
  351. .test_present = gen8_gtt_test_present,
  352. .test_pse = gen8_gtt_test_pse,
  353. .get_pfn = gen8_gtt_get_pfn,
  354. .set_pfn = gen8_gtt_set_pfn,
  355. };
  356. static struct intel_gvt_gtt_gma_ops gen8_gtt_gma_ops = {
  357. .gma_to_ggtt_pte_index = gma_to_ggtt_pte_index,
  358. .gma_to_pte_index = gen8_gma_to_pte_index,
  359. .gma_to_pde_index = gen8_gma_to_pde_index,
  360. .gma_to_l3_pdp_index = gen8_gma_to_l3_pdp_index,
  361. .gma_to_l4_pdp_index = gen8_gma_to_l4_pdp_index,
  362. .gma_to_pml4_index = gen8_gma_to_pml4_index,
  363. };
  364. static int gtt_entry_p2m(struct intel_vgpu *vgpu, struct intel_gvt_gtt_entry *p,
  365. struct intel_gvt_gtt_entry *m)
  366. {
  367. struct intel_gvt_gtt_pte_ops *ops = vgpu->gvt->gtt.pte_ops;
  368. unsigned long gfn, mfn;
  369. *m = *p;
  370. if (!ops->test_present(p))
  371. return 0;
  372. gfn = ops->get_pfn(p);
  373. mfn = intel_gvt_hypervisor_gfn_to_mfn(vgpu, gfn);
  374. if (mfn == INTEL_GVT_INVALID_ADDR) {
  375. gvt_err("fail to translate gfn: 0x%lx\n", gfn);
  376. return -ENXIO;
  377. }
  378. ops->set_pfn(m, mfn);
  379. return 0;
  380. }
  381. /*
  382. * MM helpers.
  383. */
  384. struct intel_gvt_gtt_entry *intel_vgpu_mm_get_entry(struct intel_vgpu_mm *mm,
  385. void *page_table, struct intel_gvt_gtt_entry *e,
  386. unsigned long index)
  387. {
  388. struct intel_gvt *gvt = mm->vgpu->gvt;
  389. struct intel_gvt_gtt_pte_ops *ops = gvt->gtt.pte_ops;
  390. e->type = mm->page_table_entry_type;
  391. ops->get_entry(page_table, e, index, false, 0, mm->vgpu);
  392. ops->test_pse(e);
  393. return e;
  394. }
  395. struct intel_gvt_gtt_entry *intel_vgpu_mm_set_entry(struct intel_vgpu_mm *mm,
  396. void *page_table, struct intel_gvt_gtt_entry *e,
  397. unsigned long index)
  398. {
  399. struct intel_gvt *gvt = mm->vgpu->gvt;
  400. struct intel_gvt_gtt_pte_ops *ops = gvt->gtt.pte_ops;
  401. return ops->set_entry(page_table, e, index, false, 0, mm->vgpu);
  402. }
  403. /*
  404. * PPGTT shadow page table helpers.
  405. */
  406. static inline struct intel_gvt_gtt_entry *ppgtt_spt_get_entry(
  407. struct intel_vgpu_ppgtt_spt *spt,
  408. void *page_table, int type,
  409. struct intel_gvt_gtt_entry *e, unsigned long index,
  410. bool guest)
  411. {
  412. struct intel_gvt *gvt = spt->vgpu->gvt;
  413. struct intel_gvt_gtt_pte_ops *ops = gvt->gtt.pte_ops;
  414. e->type = get_entry_type(type);
  415. if (WARN(!gtt_type_is_entry(e->type), "invalid entry type\n"))
  416. return e;
  417. ops->get_entry(page_table, e, index, guest,
  418. spt->guest_page.gfn << GTT_PAGE_SHIFT,
  419. spt->vgpu);
  420. ops->test_pse(e);
  421. return e;
  422. }
  423. static inline struct intel_gvt_gtt_entry *ppgtt_spt_set_entry(
  424. struct intel_vgpu_ppgtt_spt *spt,
  425. void *page_table, int type,
  426. struct intel_gvt_gtt_entry *e, unsigned long index,
  427. bool guest)
  428. {
  429. struct intel_gvt *gvt = spt->vgpu->gvt;
  430. struct intel_gvt_gtt_pte_ops *ops = gvt->gtt.pte_ops;
  431. if (WARN(!gtt_type_is_entry(e->type), "invalid entry type\n"))
  432. return e;
  433. return ops->set_entry(page_table, e, index, guest,
  434. spt->guest_page.gfn << GTT_PAGE_SHIFT,
  435. spt->vgpu);
  436. }
  437. #define ppgtt_get_guest_entry(spt, e, index) \
  438. ppgtt_spt_get_entry(spt, NULL, \
  439. spt->guest_page_type, e, index, true)
  440. #define ppgtt_set_guest_entry(spt, e, index) \
  441. ppgtt_spt_set_entry(spt, NULL, \
  442. spt->guest_page_type, e, index, true)
  443. #define ppgtt_get_shadow_entry(spt, e, index) \
  444. ppgtt_spt_get_entry(spt, spt->shadow_page.vaddr, \
  445. spt->shadow_page.type, e, index, false)
  446. #define ppgtt_set_shadow_entry(spt, e, index) \
  447. ppgtt_spt_set_entry(spt, spt->shadow_page.vaddr, \
  448. spt->shadow_page.type, e, index, false)
  449. /**
  450. * intel_vgpu_init_guest_page - init a guest page data structure
  451. * @vgpu: a vGPU
  452. * @p: a guest page data structure
  453. * @gfn: guest memory page frame number
  454. * @handler: function will be called when target guest memory page has
  455. * been modified.
  456. *
  457. * This function is called when user wants to track a guest memory page.
  458. *
  459. * Returns:
  460. * Zero on success, negative error code if failed.
  461. */
  462. int intel_vgpu_init_guest_page(struct intel_vgpu *vgpu,
  463. struct intel_vgpu_guest_page *p,
  464. unsigned long gfn,
  465. int (*handler)(void *, u64, void *, int),
  466. void *data)
  467. {
  468. INIT_HLIST_NODE(&p->node);
  469. p->writeprotection = false;
  470. p->gfn = gfn;
  471. p->handler = handler;
  472. p->data = data;
  473. p->oos_page = NULL;
  474. p->write_cnt = 0;
  475. hash_add(vgpu->gtt.guest_page_hash_table, &p->node, p->gfn);
  476. return 0;
  477. }
  478. static int detach_oos_page(struct intel_vgpu *vgpu,
  479. struct intel_vgpu_oos_page *oos_page);
  480. /**
  481. * intel_vgpu_clean_guest_page - release the resource owned by guest page data
  482. * structure
  483. * @vgpu: a vGPU
  484. * @p: a tracked guest page
  485. *
  486. * This function is called when user tries to stop tracking a guest memory
  487. * page.
  488. */
  489. void intel_vgpu_clean_guest_page(struct intel_vgpu *vgpu,
  490. struct intel_vgpu_guest_page *p)
  491. {
  492. if (!hlist_unhashed(&p->node))
  493. hash_del(&p->node);
  494. if (p->oos_page)
  495. detach_oos_page(vgpu, p->oos_page);
  496. if (p->writeprotection)
  497. intel_gvt_hypervisor_unset_wp_page(vgpu, p);
  498. }
  499. /**
  500. * intel_vgpu_find_guest_page - find a guest page data structure by GFN.
  501. * @vgpu: a vGPU
  502. * @gfn: guest memory page frame number
  503. *
  504. * This function is called when emulation logic wants to know if a trapped GFN
  505. * is a tracked guest page.
  506. *
  507. * Returns:
  508. * Pointer to guest page data structure, NULL if failed.
  509. */
  510. struct intel_vgpu_guest_page *intel_vgpu_find_guest_page(
  511. struct intel_vgpu *vgpu, unsigned long gfn)
  512. {
  513. struct intel_vgpu_guest_page *p;
  514. hash_for_each_possible(vgpu->gtt.guest_page_hash_table,
  515. p, node, gfn) {
  516. if (p->gfn == gfn)
  517. return p;
  518. }
  519. return NULL;
  520. }
  521. static inline int init_shadow_page(struct intel_vgpu *vgpu,
  522. struct intel_vgpu_shadow_page *p, int type)
  523. {
  524. p->vaddr = page_address(p->page);
  525. p->type = type;
  526. INIT_HLIST_NODE(&p->node);
  527. p->mfn = intel_gvt_hypervisor_virt_to_mfn(p->vaddr);
  528. if (p->mfn == INTEL_GVT_INVALID_ADDR)
  529. return -EFAULT;
  530. hash_add(vgpu->gtt.shadow_page_hash_table, &p->node, p->mfn);
  531. return 0;
  532. }
  533. static inline void clean_shadow_page(struct intel_vgpu_shadow_page *p)
  534. {
  535. if (!hlist_unhashed(&p->node))
  536. hash_del(&p->node);
  537. }
  538. static inline struct intel_vgpu_shadow_page *find_shadow_page(
  539. struct intel_vgpu *vgpu, unsigned long mfn)
  540. {
  541. struct intel_vgpu_shadow_page *p;
  542. hash_for_each_possible(vgpu->gtt.shadow_page_hash_table,
  543. p, node, mfn) {
  544. if (p->mfn == mfn)
  545. return p;
  546. }
  547. return NULL;
  548. }
  549. #define guest_page_to_ppgtt_spt(ptr) \
  550. container_of(ptr, struct intel_vgpu_ppgtt_spt, guest_page)
  551. #define shadow_page_to_ppgtt_spt(ptr) \
  552. container_of(ptr, struct intel_vgpu_ppgtt_spt, shadow_page)
  553. static void *alloc_spt(gfp_t gfp_mask)
  554. {
  555. struct intel_vgpu_ppgtt_spt *spt;
  556. spt = kzalloc(sizeof(*spt), gfp_mask);
  557. if (!spt)
  558. return NULL;
  559. spt->shadow_page.page = alloc_page(gfp_mask);
  560. if (!spt->shadow_page.page) {
  561. kfree(spt);
  562. return NULL;
  563. }
  564. return spt;
  565. }
  566. static void free_spt(struct intel_vgpu_ppgtt_spt *spt)
  567. {
  568. __free_page(spt->shadow_page.page);
  569. kfree(spt);
  570. }
  571. static void ppgtt_free_shadow_page(struct intel_vgpu_ppgtt_spt *spt)
  572. {
  573. trace_spt_free(spt->vgpu->id, spt, spt->shadow_page.type);
  574. clean_shadow_page(&spt->shadow_page);
  575. intel_vgpu_clean_guest_page(spt->vgpu, &spt->guest_page);
  576. list_del_init(&spt->post_shadow_list);
  577. free_spt(spt);
  578. }
  579. static void ppgtt_free_all_shadow_page(struct intel_vgpu *vgpu)
  580. {
  581. struct hlist_node *n;
  582. struct intel_vgpu_shadow_page *sp;
  583. int i;
  584. hash_for_each_safe(vgpu->gtt.shadow_page_hash_table, i, n, sp, node)
  585. ppgtt_free_shadow_page(shadow_page_to_ppgtt_spt(sp));
  586. }
  587. static int ppgtt_handle_guest_write_page_table_bytes(void *gp,
  588. u64 pa, void *p_data, int bytes);
  589. static int ppgtt_write_protection_handler(void *gp, u64 pa,
  590. void *p_data, int bytes)
  591. {
  592. struct intel_vgpu_guest_page *gpt = (struct intel_vgpu_guest_page *)gp;
  593. int ret;
  594. if (bytes != 4 && bytes != 8)
  595. return -EINVAL;
  596. if (!gpt->writeprotection)
  597. return -EINVAL;
  598. ret = ppgtt_handle_guest_write_page_table_bytes(gp,
  599. pa, p_data, bytes);
  600. if (ret)
  601. return ret;
  602. return ret;
  603. }
  604. static int reclaim_one_mm(struct intel_gvt *gvt);
  605. static struct intel_vgpu_ppgtt_spt *ppgtt_alloc_shadow_page(
  606. struct intel_vgpu *vgpu, int type, unsigned long gfn)
  607. {
  608. struct intel_vgpu_ppgtt_spt *spt = NULL;
  609. int ret;
  610. retry:
  611. spt = alloc_spt(GFP_KERNEL | __GFP_ZERO);
  612. if (!spt) {
  613. if (reclaim_one_mm(vgpu->gvt))
  614. goto retry;
  615. gvt_err("fail to allocate ppgtt shadow page\n");
  616. return ERR_PTR(-ENOMEM);
  617. }
  618. spt->vgpu = vgpu;
  619. spt->guest_page_type = type;
  620. atomic_set(&spt->refcount, 1);
  621. INIT_LIST_HEAD(&spt->post_shadow_list);
  622. /*
  623. * TODO: guest page type may be different with shadow page type,
  624. * when we support PSE page in future.
  625. */
  626. ret = init_shadow_page(vgpu, &spt->shadow_page, type);
  627. if (ret) {
  628. gvt_err("fail to initialize shadow page for spt\n");
  629. goto err;
  630. }
  631. ret = intel_vgpu_init_guest_page(vgpu, &spt->guest_page,
  632. gfn, ppgtt_write_protection_handler, NULL);
  633. if (ret) {
  634. gvt_err("fail to initialize guest page for spt\n");
  635. goto err;
  636. }
  637. trace_spt_alloc(vgpu->id, spt, type, spt->shadow_page.mfn, gfn);
  638. return spt;
  639. err:
  640. ppgtt_free_shadow_page(spt);
  641. return ERR_PTR(ret);
  642. }
  643. static struct intel_vgpu_ppgtt_spt *ppgtt_find_shadow_page(
  644. struct intel_vgpu *vgpu, unsigned long mfn)
  645. {
  646. struct intel_vgpu_shadow_page *p = find_shadow_page(vgpu, mfn);
  647. if (p)
  648. return shadow_page_to_ppgtt_spt(p);
  649. gvt_err("vgpu%d: fail to find ppgtt shadow page: 0x%lx\n",
  650. vgpu->id, mfn);
  651. return NULL;
  652. }
  653. #define pt_entry_size_shift(spt) \
  654. ((spt)->vgpu->gvt->device_info.gtt_entry_size_shift)
  655. #define pt_entries(spt) \
  656. (GTT_PAGE_SIZE >> pt_entry_size_shift(spt))
  657. #define for_each_present_guest_entry(spt, e, i) \
  658. for (i = 0; i < pt_entries(spt); i++) \
  659. if (spt->vgpu->gvt->gtt.pte_ops->test_present( \
  660. ppgtt_get_guest_entry(spt, e, i)))
  661. #define for_each_present_shadow_entry(spt, e, i) \
  662. for (i = 0; i < pt_entries(spt); i++) \
  663. if (spt->vgpu->gvt->gtt.pte_ops->test_present( \
  664. ppgtt_get_shadow_entry(spt, e, i)))
  665. static void ppgtt_get_shadow_page(struct intel_vgpu_ppgtt_spt *spt)
  666. {
  667. int v = atomic_read(&spt->refcount);
  668. trace_spt_refcount(spt->vgpu->id, "inc", spt, v, (v + 1));
  669. atomic_inc(&spt->refcount);
  670. }
  671. static int ppgtt_invalidate_shadow_page(struct intel_vgpu_ppgtt_spt *spt);
  672. static int ppgtt_invalidate_shadow_page_by_shadow_entry(struct intel_vgpu *vgpu,
  673. struct intel_gvt_gtt_entry *e)
  674. {
  675. struct intel_gvt_gtt_pte_ops *ops = vgpu->gvt->gtt.pte_ops;
  676. struct intel_vgpu_ppgtt_spt *s;
  677. intel_gvt_gtt_type_t cur_pt_type;
  678. if (WARN_ON(!gtt_type_is_pt(get_next_pt_type(e->type))))
  679. return -EINVAL;
  680. if (e->type != GTT_TYPE_PPGTT_ROOT_L3_ENTRY
  681. && e->type != GTT_TYPE_PPGTT_ROOT_L4_ENTRY) {
  682. cur_pt_type = get_next_pt_type(e->type) + 1;
  683. if (ops->get_pfn(e) ==
  684. vgpu->gtt.scratch_pt[cur_pt_type].page_mfn)
  685. return 0;
  686. }
  687. s = ppgtt_find_shadow_page(vgpu, ops->get_pfn(e));
  688. if (!s) {
  689. gvt_err("vgpu%d: fail to find shadow page: mfn: 0x%lx\n",
  690. vgpu->id, ops->get_pfn(e));
  691. return -ENXIO;
  692. }
  693. return ppgtt_invalidate_shadow_page(s);
  694. }
  695. static int ppgtt_invalidate_shadow_page(struct intel_vgpu_ppgtt_spt *spt)
  696. {
  697. struct intel_gvt_gtt_entry e;
  698. unsigned long index;
  699. int ret;
  700. int v = atomic_read(&spt->refcount);
  701. trace_spt_change(spt->vgpu->id, "die", spt,
  702. spt->guest_page.gfn, spt->shadow_page.type);
  703. trace_spt_refcount(spt->vgpu->id, "dec", spt, v, (v - 1));
  704. if (atomic_dec_return(&spt->refcount) > 0)
  705. return 0;
  706. if (gtt_type_is_pte_pt(spt->shadow_page.type))
  707. goto release;
  708. for_each_present_shadow_entry(spt, &e, index) {
  709. if (!gtt_type_is_pt(get_next_pt_type(e.type))) {
  710. gvt_err("GVT doesn't support pse bit for now\n");
  711. return -EINVAL;
  712. }
  713. ret = ppgtt_invalidate_shadow_page_by_shadow_entry(
  714. spt->vgpu, &e);
  715. if (ret)
  716. goto fail;
  717. }
  718. release:
  719. trace_spt_change(spt->vgpu->id, "release", spt,
  720. spt->guest_page.gfn, spt->shadow_page.type);
  721. ppgtt_free_shadow_page(spt);
  722. return 0;
  723. fail:
  724. gvt_err("vgpu%d: fail: shadow page %p shadow entry 0x%llx type %d\n",
  725. spt->vgpu->id, spt, e.val64, e.type);
  726. return ret;
  727. }
  728. static int ppgtt_populate_shadow_page(struct intel_vgpu_ppgtt_spt *spt);
  729. static struct intel_vgpu_ppgtt_spt *ppgtt_populate_shadow_page_by_guest_entry(
  730. struct intel_vgpu *vgpu, struct intel_gvt_gtt_entry *we)
  731. {
  732. struct intel_gvt_gtt_pte_ops *ops = vgpu->gvt->gtt.pte_ops;
  733. struct intel_vgpu_ppgtt_spt *s = NULL;
  734. struct intel_vgpu_guest_page *g;
  735. int ret;
  736. if (WARN_ON(!gtt_type_is_pt(get_next_pt_type(we->type)))) {
  737. ret = -EINVAL;
  738. goto fail;
  739. }
  740. g = intel_vgpu_find_guest_page(vgpu, ops->get_pfn(we));
  741. if (g) {
  742. s = guest_page_to_ppgtt_spt(g);
  743. ppgtt_get_shadow_page(s);
  744. } else {
  745. int type = get_next_pt_type(we->type);
  746. s = ppgtt_alloc_shadow_page(vgpu, type, ops->get_pfn(we));
  747. if (IS_ERR(s)) {
  748. ret = PTR_ERR(s);
  749. goto fail;
  750. }
  751. ret = intel_gvt_hypervisor_set_wp_page(vgpu, &s->guest_page);
  752. if (ret)
  753. goto fail;
  754. ret = ppgtt_populate_shadow_page(s);
  755. if (ret)
  756. goto fail;
  757. trace_spt_change(vgpu->id, "new", s, s->guest_page.gfn,
  758. s->shadow_page.type);
  759. }
  760. return s;
  761. fail:
  762. gvt_err("vgpu%d: fail: shadow page %p guest entry 0x%llx type %d\n",
  763. vgpu->id, s, we->val64, we->type);
  764. return ERR_PTR(ret);
  765. }
  766. static inline void ppgtt_generate_shadow_entry(struct intel_gvt_gtt_entry *se,
  767. struct intel_vgpu_ppgtt_spt *s, struct intel_gvt_gtt_entry *ge)
  768. {
  769. struct intel_gvt_gtt_pte_ops *ops = s->vgpu->gvt->gtt.pte_ops;
  770. se->type = ge->type;
  771. se->val64 = ge->val64;
  772. ops->set_pfn(se, s->shadow_page.mfn);
  773. }
  774. static int ppgtt_populate_shadow_page(struct intel_vgpu_ppgtt_spt *spt)
  775. {
  776. struct intel_vgpu *vgpu = spt->vgpu;
  777. struct intel_vgpu_ppgtt_spt *s;
  778. struct intel_gvt_gtt_entry se, ge;
  779. unsigned long i;
  780. int ret;
  781. trace_spt_change(spt->vgpu->id, "born", spt,
  782. spt->guest_page.gfn, spt->shadow_page.type);
  783. if (gtt_type_is_pte_pt(spt->shadow_page.type)) {
  784. for_each_present_guest_entry(spt, &ge, i) {
  785. ret = gtt_entry_p2m(vgpu, &ge, &se);
  786. if (ret)
  787. goto fail;
  788. ppgtt_set_shadow_entry(spt, &se, i);
  789. }
  790. return 0;
  791. }
  792. for_each_present_guest_entry(spt, &ge, i) {
  793. if (!gtt_type_is_pt(get_next_pt_type(ge.type))) {
  794. gvt_err("GVT doesn't support pse bit now\n");
  795. ret = -EINVAL;
  796. goto fail;
  797. }
  798. s = ppgtt_populate_shadow_page_by_guest_entry(vgpu, &ge);
  799. if (IS_ERR(s)) {
  800. ret = PTR_ERR(s);
  801. goto fail;
  802. }
  803. ppgtt_get_shadow_entry(spt, &se, i);
  804. ppgtt_generate_shadow_entry(&se, s, &ge);
  805. ppgtt_set_shadow_entry(spt, &se, i);
  806. }
  807. return 0;
  808. fail:
  809. gvt_err("vgpu%d: fail: shadow page %p guest entry 0x%llx type %d\n",
  810. vgpu->id, spt, ge.val64, ge.type);
  811. return ret;
  812. }
  813. static int ppgtt_handle_guest_entry_removal(struct intel_vgpu_guest_page *gpt,
  814. unsigned long index)
  815. {
  816. struct intel_vgpu_ppgtt_spt *spt = guest_page_to_ppgtt_spt(gpt);
  817. struct intel_vgpu_shadow_page *sp = &spt->shadow_page;
  818. struct intel_vgpu *vgpu = spt->vgpu;
  819. struct intel_gvt_gtt_pte_ops *ops = vgpu->gvt->gtt.pte_ops;
  820. struct intel_gvt_gtt_entry e;
  821. int ret;
  822. ppgtt_get_shadow_entry(spt, &e, index);
  823. trace_gpt_change(spt->vgpu->id, "remove", spt, sp->type, e.val64,
  824. index);
  825. if (!ops->test_present(&e))
  826. return 0;
  827. if (ops->get_pfn(&e) == vgpu->gtt.scratch_pt[sp->type].page_mfn)
  828. return 0;
  829. if (gtt_type_is_pt(get_next_pt_type(e.type))) {
  830. struct intel_vgpu_ppgtt_spt *s =
  831. ppgtt_find_shadow_page(vgpu, ops->get_pfn(&e));
  832. if (!s) {
  833. gvt_err("fail to find guest page\n");
  834. ret = -ENXIO;
  835. goto fail;
  836. }
  837. ret = ppgtt_invalidate_shadow_page(s);
  838. if (ret)
  839. goto fail;
  840. }
  841. ops->set_pfn(&e, vgpu->gtt.scratch_pt[sp->type].page_mfn);
  842. ppgtt_set_shadow_entry(spt, &e, index);
  843. return 0;
  844. fail:
  845. gvt_err("vgpu%d: fail: shadow page %p guest entry 0x%llx type %d\n",
  846. vgpu->id, spt, e.val64, e.type);
  847. return ret;
  848. }
  849. static int ppgtt_handle_guest_entry_add(struct intel_vgpu_guest_page *gpt,
  850. struct intel_gvt_gtt_entry *we, unsigned long index)
  851. {
  852. struct intel_vgpu_ppgtt_spt *spt = guest_page_to_ppgtt_spt(gpt);
  853. struct intel_vgpu_shadow_page *sp = &spt->shadow_page;
  854. struct intel_vgpu *vgpu = spt->vgpu;
  855. struct intel_gvt_gtt_entry m;
  856. struct intel_vgpu_ppgtt_spt *s;
  857. int ret;
  858. trace_gpt_change(spt->vgpu->id, "add", spt, sp->type,
  859. we->val64, index);
  860. if (gtt_type_is_pt(get_next_pt_type(we->type))) {
  861. s = ppgtt_populate_shadow_page_by_guest_entry(vgpu, we);
  862. if (IS_ERR(s)) {
  863. ret = PTR_ERR(s);
  864. goto fail;
  865. }
  866. ppgtt_get_shadow_entry(spt, &m, index);
  867. ppgtt_generate_shadow_entry(&m, s, we);
  868. ppgtt_set_shadow_entry(spt, &m, index);
  869. } else {
  870. ret = gtt_entry_p2m(vgpu, we, &m);
  871. if (ret)
  872. goto fail;
  873. ppgtt_set_shadow_entry(spt, &m, index);
  874. }
  875. return 0;
  876. fail:
  877. gvt_err("vgpu%d: fail: spt %p guest entry 0x%llx type %d\n", vgpu->id,
  878. spt, we->val64, we->type);
  879. return ret;
  880. }
  881. static int sync_oos_page(struct intel_vgpu *vgpu,
  882. struct intel_vgpu_oos_page *oos_page)
  883. {
  884. const struct intel_gvt_device_info *info = &vgpu->gvt->device_info;
  885. struct intel_gvt *gvt = vgpu->gvt;
  886. struct intel_gvt_gtt_pte_ops *ops = gvt->gtt.pte_ops;
  887. struct intel_vgpu_ppgtt_spt *spt =
  888. guest_page_to_ppgtt_spt(oos_page->guest_page);
  889. struct intel_gvt_gtt_entry old, new, m;
  890. int index;
  891. int ret;
  892. trace_oos_change(vgpu->id, "sync", oos_page->id,
  893. oos_page->guest_page, spt->guest_page_type);
  894. old.type = new.type = get_entry_type(spt->guest_page_type);
  895. old.val64 = new.val64 = 0;
  896. for (index = 0; index < (GTT_PAGE_SIZE >> info->gtt_entry_size_shift);
  897. index++) {
  898. ops->get_entry(oos_page->mem, &old, index, false, 0, vgpu);
  899. ops->get_entry(NULL, &new, index, true,
  900. oos_page->guest_page->gfn << PAGE_SHIFT, vgpu);
  901. if (old.val64 == new.val64
  902. && !test_and_clear_bit(index, spt->post_shadow_bitmap))
  903. continue;
  904. trace_oos_sync(vgpu->id, oos_page->id,
  905. oos_page->guest_page, spt->guest_page_type,
  906. new.val64, index);
  907. ret = gtt_entry_p2m(vgpu, &new, &m);
  908. if (ret)
  909. return ret;
  910. ops->set_entry(oos_page->mem, &new, index, false, 0, vgpu);
  911. ppgtt_set_shadow_entry(spt, &m, index);
  912. }
  913. oos_page->guest_page->write_cnt = 0;
  914. list_del_init(&spt->post_shadow_list);
  915. return 0;
  916. }
  917. static int detach_oos_page(struct intel_vgpu *vgpu,
  918. struct intel_vgpu_oos_page *oos_page)
  919. {
  920. struct intel_gvt *gvt = vgpu->gvt;
  921. struct intel_vgpu_ppgtt_spt *spt =
  922. guest_page_to_ppgtt_spt(oos_page->guest_page);
  923. trace_oos_change(vgpu->id, "detach", oos_page->id,
  924. oos_page->guest_page, spt->guest_page_type);
  925. oos_page->guest_page->write_cnt = 0;
  926. oos_page->guest_page->oos_page = NULL;
  927. oos_page->guest_page = NULL;
  928. list_del_init(&oos_page->vm_list);
  929. list_move_tail(&oos_page->list, &gvt->gtt.oos_page_free_list_head);
  930. return 0;
  931. }
  932. static int attach_oos_page(struct intel_vgpu *vgpu,
  933. struct intel_vgpu_oos_page *oos_page,
  934. struct intel_vgpu_guest_page *gpt)
  935. {
  936. struct intel_gvt *gvt = vgpu->gvt;
  937. int ret;
  938. ret = intel_gvt_hypervisor_read_gpa(vgpu, gpt->gfn << GTT_PAGE_SHIFT,
  939. oos_page->mem, GTT_PAGE_SIZE);
  940. if (ret)
  941. return ret;
  942. oos_page->guest_page = gpt;
  943. gpt->oos_page = oos_page;
  944. list_move_tail(&oos_page->list, &gvt->gtt.oos_page_use_list_head);
  945. trace_oos_change(vgpu->id, "attach", gpt->oos_page->id,
  946. gpt, guest_page_to_ppgtt_spt(gpt)->guest_page_type);
  947. return 0;
  948. }
  949. static int ppgtt_set_guest_page_sync(struct intel_vgpu *vgpu,
  950. struct intel_vgpu_guest_page *gpt)
  951. {
  952. int ret;
  953. ret = intel_gvt_hypervisor_set_wp_page(vgpu, gpt);
  954. if (ret)
  955. return ret;
  956. trace_oos_change(vgpu->id, "set page sync", gpt->oos_page->id,
  957. gpt, guest_page_to_ppgtt_spt(gpt)->guest_page_type);
  958. list_del_init(&gpt->oos_page->vm_list);
  959. return sync_oos_page(vgpu, gpt->oos_page);
  960. }
  961. static int ppgtt_allocate_oos_page(struct intel_vgpu *vgpu,
  962. struct intel_vgpu_guest_page *gpt)
  963. {
  964. struct intel_gvt *gvt = vgpu->gvt;
  965. struct intel_gvt_gtt *gtt = &gvt->gtt;
  966. struct intel_vgpu_oos_page *oos_page = gpt->oos_page;
  967. int ret;
  968. WARN(oos_page, "shadow PPGTT page has already has a oos page\n");
  969. if (list_empty(&gtt->oos_page_free_list_head)) {
  970. oos_page = container_of(gtt->oos_page_use_list_head.next,
  971. struct intel_vgpu_oos_page, list);
  972. ret = ppgtt_set_guest_page_sync(vgpu, oos_page->guest_page);
  973. if (ret)
  974. return ret;
  975. ret = detach_oos_page(vgpu, oos_page);
  976. if (ret)
  977. return ret;
  978. } else
  979. oos_page = container_of(gtt->oos_page_free_list_head.next,
  980. struct intel_vgpu_oos_page, list);
  981. return attach_oos_page(vgpu, oos_page, gpt);
  982. }
  983. static int ppgtt_set_guest_page_oos(struct intel_vgpu *vgpu,
  984. struct intel_vgpu_guest_page *gpt)
  985. {
  986. struct intel_vgpu_oos_page *oos_page = gpt->oos_page;
  987. if (WARN(!oos_page, "shadow PPGTT page should have a oos page\n"))
  988. return -EINVAL;
  989. trace_oos_change(vgpu->id, "set page out of sync", gpt->oos_page->id,
  990. gpt, guest_page_to_ppgtt_spt(gpt)->guest_page_type);
  991. list_add_tail(&oos_page->vm_list, &vgpu->gtt.oos_page_list_head);
  992. return intel_gvt_hypervisor_unset_wp_page(vgpu, gpt);
  993. }
  994. /**
  995. * intel_vgpu_sync_oos_pages - sync all the out-of-synced shadow for vGPU
  996. * @vgpu: a vGPU
  997. *
  998. * This function is called before submitting a guest workload to host,
  999. * to sync all the out-of-synced shadow for vGPU
  1000. *
  1001. * Returns:
  1002. * Zero on success, negative error code if failed.
  1003. */
  1004. int intel_vgpu_sync_oos_pages(struct intel_vgpu *vgpu)
  1005. {
  1006. struct list_head *pos, *n;
  1007. struct intel_vgpu_oos_page *oos_page;
  1008. int ret;
  1009. if (!enable_out_of_sync)
  1010. return 0;
  1011. list_for_each_safe(pos, n, &vgpu->gtt.oos_page_list_head) {
  1012. oos_page = container_of(pos,
  1013. struct intel_vgpu_oos_page, vm_list);
  1014. ret = ppgtt_set_guest_page_sync(vgpu, oos_page->guest_page);
  1015. if (ret)
  1016. return ret;
  1017. }
  1018. return 0;
  1019. }
  1020. /*
  1021. * The heart of PPGTT shadow page table.
  1022. */
  1023. static int ppgtt_handle_guest_write_page_table(
  1024. struct intel_vgpu_guest_page *gpt,
  1025. struct intel_gvt_gtt_entry *we, unsigned long index)
  1026. {
  1027. struct intel_vgpu_ppgtt_spt *spt = guest_page_to_ppgtt_spt(gpt);
  1028. struct intel_vgpu *vgpu = spt->vgpu;
  1029. struct intel_gvt_gtt_pte_ops *ops = vgpu->gvt->gtt.pte_ops;
  1030. int ret;
  1031. int new_present;
  1032. new_present = ops->test_present(we);
  1033. ret = ppgtt_handle_guest_entry_removal(gpt, index);
  1034. if (ret)
  1035. goto fail;
  1036. if (new_present) {
  1037. ret = ppgtt_handle_guest_entry_add(gpt, we, index);
  1038. if (ret)
  1039. goto fail;
  1040. }
  1041. return 0;
  1042. fail:
  1043. gvt_err("vgpu%d: fail: shadow page %p guest entry 0x%llx type %d.\n",
  1044. vgpu->id, spt, we->val64, we->type);
  1045. return ret;
  1046. }
  1047. static inline bool can_do_out_of_sync(struct intel_vgpu_guest_page *gpt)
  1048. {
  1049. return enable_out_of_sync
  1050. && gtt_type_is_pte_pt(
  1051. guest_page_to_ppgtt_spt(gpt)->guest_page_type)
  1052. && gpt->write_cnt >= 2;
  1053. }
  1054. static void ppgtt_set_post_shadow(struct intel_vgpu_ppgtt_spt *spt,
  1055. unsigned long index)
  1056. {
  1057. set_bit(index, spt->post_shadow_bitmap);
  1058. if (!list_empty(&spt->post_shadow_list))
  1059. return;
  1060. list_add_tail(&spt->post_shadow_list,
  1061. &spt->vgpu->gtt.post_shadow_list_head);
  1062. }
  1063. /**
  1064. * intel_vgpu_flush_post_shadow - flush the post shadow transactions
  1065. * @vgpu: a vGPU
  1066. *
  1067. * This function is called before submitting a guest workload to host,
  1068. * to flush all the post shadows for a vGPU.
  1069. *
  1070. * Returns:
  1071. * Zero on success, negative error code if failed.
  1072. */
  1073. int intel_vgpu_flush_post_shadow(struct intel_vgpu *vgpu)
  1074. {
  1075. struct list_head *pos, *n;
  1076. struct intel_vgpu_ppgtt_spt *spt;
  1077. struct intel_gvt_gtt_entry ge;
  1078. unsigned long index;
  1079. int ret;
  1080. list_for_each_safe(pos, n, &vgpu->gtt.post_shadow_list_head) {
  1081. spt = container_of(pos, struct intel_vgpu_ppgtt_spt,
  1082. post_shadow_list);
  1083. for_each_set_bit(index, spt->post_shadow_bitmap,
  1084. GTT_ENTRY_NUM_IN_ONE_PAGE) {
  1085. ppgtt_get_guest_entry(spt, &ge, index);
  1086. ret = ppgtt_handle_guest_write_page_table(
  1087. &spt->guest_page, &ge, index);
  1088. if (ret)
  1089. return ret;
  1090. clear_bit(index, spt->post_shadow_bitmap);
  1091. }
  1092. list_del_init(&spt->post_shadow_list);
  1093. }
  1094. return 0;
  1095. }
  1096. static int ppgtt_handle_guest_write_page_table_bytes(void *gp,
  1097. u64 pa, void *p_data, int bytes)
  1098. {
  1099. struct intel_vgpu_guest_page *gpt = (struct intel_vgpu_guest_page *)gp;
  1100. struct intel_vgpu_ppgtt_spt *spt = guest_page_to_ppgtt_spt(gpt);
  1101. struct intel_vgpu *vgpu = spt->vgpu;
  1102. struct intel_gvt_gtt_pte_ops *ops = vgpu->gvt->gtt.pte_ops;
  1103. const struct intel_gvt_device_info *info = &vgpu->gvt->device_info;
  1104. struct intel_gvt_gtt_entry we;
  1105. unsigned long index;
  1106. int ret;
  1107. index = (pa & (PAGE_SIZE - 1)) >> info->gtt_entry_size_shift;
  1108. ppgtt_get_guest_entry(spt, &we, index);
  1109. ops->test_pse(&we);
  1110. if (bytes == info->gtt_entry_size) {
  1111. ret = ppgtt_handle_guest_write_page_table(gpt, &we, index);
  1112. if (ret)
  1113. return ret;
  1114. } else {
  1115. if (!test_bit(index, spt->post_shadow_bitmap)) {
  1116. ret = ppgtt_handle_guest_entry_removal(gpt, index);
  1117. if (ret)
  1118. return ret;
  1119. }
  1120. ppgtt_set_post_shadow(spt, index);
  1121. }
  1122. if (!enable_out_of_sync)
  1123. return 0;
  1124. gpt->write_cnt++;
  1125. if (gpt->oos_page)
  1126. ops->set_entry(gpt->oos_page->mem, &we, index,
  1127. false, 0, vgpu);
  1128. if (can_do_out_of_sync(gpt)) {
  1129. if (!gpt->oos_page)
  1130. ppgtt_allocate_oos_page(vgpu, gpt);
  1131. ret = ppgtt_set_guest_page_oos(vgpu, gpt);
  1132. if (ret < 0)
  1133. return ret;
  1134. }
  1135. return 0;
  1136. }
  1137. /*
  1138. * mm page table allocation policy for bdw+
  1139. * - for ggtt, only virtual page table will be allocated.
  1140. * - for ppgtt, dedicated virtual/shadow page table will be allocated.
  1141. */
  1142. static int gen8_mm_alloc_page_table(struct intel_vgpu_mm *mm)
  1143. {
  1144. struct intel_vgpu *vgpu = mm->vgpu;
  1145. struct intel_gvt *gvt = vgpu->gvt;
  1146. const struct intel_gvt_device_info *info = &gvt->device_info;
  1147. void *mem;
  1148. if (mm->type == INTEL_GVT_MM_PPGTT) {
  1149. mm->page_table_entry_cnt = 4;
  1150. mm->page_table_entry_size = mm->page_table_entry_cnt *
  1151. info->gtt_entry_size;
  1152. mem = kzalloc(mm->has_shadow_page_table ?
  1153. mm->page_table_entry_size * 2
  1154. : mm->page_table_entry_size, GFP_KERNEL);
  1155. if (!mem)
  1156. return -ENOMEM;
  1157. mm->virtual_page_table = mem;
  1158. if (!mm->has_shadow_page_table)
  1159. return 0;
  1160. mm->shadow_page_table = mem + mm->page_table_entry_size;
  1161. } else if (mm->type == INTEL_GVT_MM_GGTT) {
  1162. mm->page_table_entry_cnt =
  1163. (gvt_ggtt_gm_sz(gvt) >> GTT_PAGE_SHIFT);
  1164. mm->page_table_entry_size = mm->page_table_entry_cnt *
  1165. info->gtt_entry_size;
  1166. mem = vzalloc(mm->page_table_entry_size);
  1167. if (!mem)
  1168. return -ENOMEM;
  1169. mm->virtual_page_table = mem;
  1170. }
  1171. return 0;
  1172. }
  1173. static void gen8_mm_free_page_table(struct intel_vgpu_mm *mm)
  1174. {
  1175. if (mm->type == INTEL_GVT_MM_PPGTT) {
  1176. kfree(mm->virtual_page_table);
  1177. } else if (mm->type == INTEL_GVT_MM_GGTT) {
  1178. if (mm->virtual_page_table)
  1179. vfree(mm->virtual_page_table);
  1180. }
  1181. mm->virtual_page_table = mm->shadow_page_table = NULL;
  1182. }
  1183. static void invalidate_mm(struct intel_vgpu_mm *mm)
  1184. {
  1185. struct intel_vgpu *vgpu = mm->vgpu;
  1186. struct intel_gvt *gvt = vgpu->gvt;
  1187. struct intel_gvt_gtt *gtt = &gvt->gtt;
  1188. struct intel_gvt_gtt_pte_ops *ops = gtt->pte_ops;
  1189. struct intel_gvt_gtt_entry se;
  1190. int i;
  1191. if (WARN_ON(!mm->has_shadow_page_table || !mm->shadowed))
  1192. return;
  1193. for (i = 0; i < mm->page_table_entry_cnt; i++) {
  1194. ppgtt_get_shadow_root_entry(mm, &se, i);
  1195. if (!ops->test_present(&se))
  1196. continue;
  1197. ppgtt_invalidate_shadow_page_by_shadow_entry(
  1198. vgpu, &se);
  1199. se.val64 = 0;
  1200. ppgtt_set_shadow_root_entry(mm, &se, i);
  1201. trace_gpt_change(vgpu->id, "destroy root pointer",
  1202. NULL, se.type, se.val64, i);
  1203. }
  1204. mm->shadowed = false;
  1205. }
  1206. /**
  1207. * intel_vgpu_destroy_mm - destroy a mm object
  1208. * @mm: a kref object
  1209. *
  1210. * This function is used to destroy a mm object for vGPU
  1211. *
  1212. */
  1213. void intel_vgpu_destroy_mm(struct kref *mm_ref)
  1214. {
  1215. struct intel_vgpu_mm *mm = container_of(mm_ref, typeof(*mm), ref);
  1216. struct intel_vgpu *vgpu = mm->vgpu;
  1217. struct intel_gvt *gvt = vgpu->gvt;
  1218. struct intel_gvt_gtt *gtt = &gvt->gtt;
  1219. if (!mm->initialized)
  1220. goto out;
  1221. list_del(&mm->list);
  1222. list_del(&mm->lru_list);
  1223. if (mm->has_shadow_page_table)
  1224. invalidate_mm(mm);
  1225. gtt->mm_free_page_table(mm);
  1226. out:
  1227. kfree(mm);
  1228. }
  1229. static int shadow_mm(struct intel_vgpu_mm *mm)
  1230. {
  1231. struct intel_vgpu *vgpu = mm->vgpu;
  1232. struct intel_gvt *gvt = vgpu->gvt;
  1233. struct intel_gvt_gtt *gtt = &gvt->gtt;
  1234. struct intel_gvt_gtt_pte_ops *ops = gtt->pte_ops;
  1235. struct intel_vgpu_ppgtt_spt *spt;
  1236. struct intel_gvt_gtt_entry ge, se;
  1237. int i;
  1238. int ret;
  1239. if (WARN_ON(!mm->has_shadow_page_table || mm->shadowed))
  1240. return 0;
  1241. mm->shadowed = true;
  1242. for (i = 0; i < mm->page_table_entry_cnt; i++) {
  1243. ppgtt_get_guest_root_entry(mm, &ge, i);
  1244. if (!ops->test_present(&ge))
  1245. continue;
  1246. trace_gpt_change(vgpu->id, __func__, NULL,
  1247. ge.type, ge.val64, i);
  1248. spt = ppgtt_populate_shadow_page_by_guest_entry(vgpu, &ge);
  1249. if (IS_ERR(spt)) {
  1250. gvt_err("fail to populate guest root pointer\n");
  1251. ret = PTR_ERR(spt);
  1252. goto fail;
  1253. }
  1254. ppgtt_generate_shadow_entry(&se, spt, &ge);
  1255. ppgtt_set_shadow_root_entry(mm, &se, i);
  1256. trace_gpt_change(vgpu->id, "populate root pointer",
  1257. NULL, se.type, se.val64, i);
  1258. }
  1259. return 0;
  1260. fail:
  1261. invalidate_mm(mm);
  1262. return ret;
  1263. }
  1264. /**
  1265. * intel_vgpu_create_mm - create a mm object for a vGPU
  1266. * @vgpu: a vGPU
  1267. * @mm_type: mm object type, should be PPGTT or GGTT
  1268. * @virtual_page_table: page table root pointers. Could be NULL if user wants
  1269. * to populate shadow later.
  1270. * @page_table_level: describe the page table level of the mm object
  1271. * @pde_base_index: pde root pointer base in GGTT MMIO.
  1272. *
  1273. * This function is used to create a mm object for a vGPU.
  1274. *
  1275. * Returns:
  1276. * Zero on success, negative error code in pointer if failed.
  1277. */
  1278. struct intel_vgpu_mm *intel_vgpu_create_mm(struct intel_vgpu *vgpu,
  1279. int mm_type, void *virtual_page_table, int page_table_level,
  1280. u32 pde_base_index)
  1281. {
  1282. struct intel_gvt *gvt = vgpu->gvt;
  1283. struct intel_gvt_gtt *gtt = &gvt->gtt;
  1284. struct intel_vgpu_mm *mm;
  1285. int ret;
  1286. mm = kzalloc(sizeof(*mm), GFP_KERNEL);
  1287. if (!mm) {
  1288. ret = -ENOMEM;
  1289. goto fail;
  1290. }
  1291. mm->type = mm_type;
  1292. if (page_table_level == 1)
  1293. mm->page_table_entry_type = GTT_TYPE_GGTT_PTE;
  1294. else if (page_table_level == 3)
  1295. mm->page_table_entry_type = GTT_TYPE_PPGTT_ROOT_L3_ENTRY;
  1296. else if (page_table_level == 4)
  1297. mm->page_table_entry_type = GTT_TYPE_PPGTT_ROOT_L4_ENTRY;
  1298. else {
  1299. WARN_ON(1);
  1300. ret = -EINVAL;
  1301. goto fail;
  1302. }
  1303. mm->page_table_level = page_table_level;
  1304. mm->pde_base_index = pde_base_index;
  1305. mm->vgpu = vgpu;
  1306. mm->has_shadow_page_table = !!(mm_type == INTEL_GVT_MM_PPGTT);
  1307. kref_init(&mm->ref);
  1308. atomic_set(&mm->pincount, 0);
  1309. INIT_LIST_HEAD(&mm->list);
  1310. INIT_LIST_HEAD(&mm->lru_list);
  1311. list_add_tail(&mm->list, &vgpu->gtt.mm_list_head);
  1312. ret = gtt->mm_alloc_page_table(mm);
  1313. if (ret) {
  1314. gvt_err("fail to allocate page table for mm\n");
  1315. goto fail;
  1316. }
  1317. mm->initialized = true;
  1318. if (virtual_page_table)
  1319. memcpy(mm->virtual_page_table, virtual_page_table,
  1320. mm->page_table_entry_size);
  1321. if (mm->has_shadow_page_table) {
  1322. ret = shadow_mm(mm);
  1323. if (ret)
  1324. goto fail;
  1325. list_add_tail(&mm->lru_list, &gvt->gtt.mm_lru_list_head);
  1326. }
  1327. return mm;
  1328. fail:
  1329. gvt_err("fail to create mm\n");
  1330. if (mm)
  1331. intel_gvt_mm_unreference(mm);
  1332. return ERR_PTR(ret);
  1333. }
  1334. /**
  1335. * intel_vgpu_unpin_mm - decrease the pin count of a vGPU mm object
  1336. * @mm: a vGPU mm object
  1337. *
  1338. * This function is called when user doesn't want to use a vGPU mm object
  1339. */
  1340. void intel_vgpu_unpin_mm(struct intel_vgpu_mm *mm)
  1341. {
  1342. if (WARN_ON(mm->type != INTEL_GVT_MM_PPGTT))
  1343. return;
  1344. atomic_dec(&mm->pincount);
  1345. }
  1346. /**
  1347. * intel_vgpu_pin_mm - increase the pin count of a vGPU mm object
  1348. * @vgpu: a vGPU
  1349. *
  1350. * This function is called when user wants to use a vGPU mm object. If this
  1351. * mm object hasn't been shadowed yet, the shadow will be populated at this
  1352. * time.
  1353. *
  1354. * Returns:
  1355. * Zero on success, negative error code if failed.
  1356. */
  1357. int intel_vgpu_pin_mm(struct intel_vgpu_mm *mm)
  1358. {
  1359. int ret;
  1360. if (WARN_ON(mm->type != INTEL_GVT_MM_PPGTT))
  1361. return 0;
  1362. atomic_inc(&mm->pincount);
  1363. if (!mm->shadowed) {
  1364. ret = shadow_mm(mm);
  1365. if (ret)
  1366. return ret;
  1367. }
  1368. list_del_init(&mm->lru_list);
  1369. list_add_tail(&mm->lru_list, &mm->vgpu->gvt->gtt.mm_lru_list_head);
  1370. return 0;
  1371. }
  1372. static int reclaim_one_mm(struct intel_gvt *gvt)
  1373. {
  1374. struct intel_vgpu_mm *mm;
  1375. struct list_head *pos, *n;
  1376. list_for_each_safe(pos, n, &gvt->gtt.mm_lru_list_head) {
  1377. mm = container_of(pos, struct intel_vgpu_mm, lru_list);
  1378. if (mm->type != INTEL_GVT_MM_PPGTT)
  1379. continue;
  1380. if (atomic_read(&mm->pincount))
  1381. continue;
  1382. list_del_init(&mm->lru_list);
  1383. invalidate_mm(mm);
  1384. return 1;
  1385. }
  1386. return 0;
  1387. }
  1388. /*
  1389. * GMA translation APIs.
  1390. */
  1391. static inline int ppgtt_get_next_level_entry(struct intel_vgpu_mm *mm,
  1392. struct intel_gvt_gtt_entry *e, unsigned long index, bool guest)
  1393. {
  1394. struct intel_vgpu *vgpu = mm->vgpu;
  1395. struct intel_gvt_gtt_pte_ops *ops = vgpu->gvt->gtt.pte_ops;
  1396. struct intel_vgpu_ppgtt_spt *s;
  1397. if (WARN_ON(!mm->has_shadow_page_table))
  1398. return -EINVAL;
  1399. s = ppgtt_find_shadow_page(vgpu, ops->get_pfn(e));
  1400. if (!s)
  1401. return -ENXIO;
  1402. if (!guest)
  1403. ppgtt_get_shadow_entry(s, e, index);
  1404. else
  1405. ppgtt_get_guest_entry(s, e, index);
  1406. return 0;
  1407. }
  1408. /**
  1409. * intel_vgpu_gma_to_gpa - translate a gma to GPA
  1410. * @mm: mm object. could be a PPGTT or GGTT mm object
  1411. * @gma: graphics memory address in this mm object
  1412. *
  1413. * This function is used to translate a graphics memory address in specific
  1414. * graphics memory space to guest physical address.
  1415. *
  1416. * Returns:
  1417. * Guest physical address on success, INTEL_GVT_INVALID_ADDR if failed.
  1418. */
  1419. unsigned long intel_vgpu_gma_to_gpa(struct intel_vgpu_mm *mm, unsigned long gma)
  1420. {
  1421. struct intel_vgpu *vgpu = mm->vgpu;
  1422. struct intel_gvt *gvt = vgpu->gvt;
  1423. struct intel_gvt_gtt_pte_ops *pte_ops = gvt->gtt.pte_ops;
  1424. struct intel_gvt_gtt_gma_ops *gma_ops = gvt->gtt.gma_ops;
  1425. unsigned long gpa = INTEL_GVT_INVALID_ADDR;
  1426. unsigned long gma_index[4];
  1427. struct intel_gvt_gtt_entry e;
  1428. int i, index;
  1429. int ret;
  1430. if (mm->type != INTEL_GVT_MM_GGTT && mm->type != INTEL_GVT_MM_PPGTT)
  1431. return INTEL_GVT_INVALID_ADDR;
  1432. if (mm->type == INTEL_GVT_MM_GGTT) {
  1433. if (!vgpu_gmadr_is_valid(vgpu, gma))
  1434. goto err;
  1435. ggtt_get_guest_entry(mm, &e,
  1436. gma_ops->gma_to_ggtt_pte_index(gma));
  1437. gpa = (pte_ops->get_pfn(&e) << GTT_PAGE_SHIFT)
  1438. + (gma & ~GTT_PAGE_MASK);
  1439. trace_gma_translate(vgpu->id, "ggtt", 0, 0, gma, gpa);
  1440. return gpa;
  1441. }
  1442. switch (mm->page_table_level) {
  1443. case 4:
  1444. ppgtt_get_shadow_root_entry(mm, &e, 0);
  1445. gma_index[0] = gma_ops->gma_to_pml4_index(gma);
  1446. gma_index[1] = gma_ops->gma_to_l4_pdp_index(gma);
  1447. gma_index[2] = gma_ops->gma_to_pde_index(gma);
  1448. gma_index[3] = gma_ops->gma_to_pte_index(gma);
  1449. index = 4;
  1450. break;
  1451. case 3:
  1452. ppgtt_get_shadow_root_entry(mm, &e,
  1453. gma_ops->gma_to_l3_pdp_index(gma));
  1454. gma_index[0] = gma_ops->gma_to_pde_index(gma);
  1455. gma_index[1] = gma_ops->gma_to_pte_index(gma);
  1456. index = 2;
  1457. break;
  1458. case 2:
  1459. ppgtt_get_shadow_root_entry(mm, &e,
  1460. gma_ops->gma_to_pde_index(gma));
  1461. gma_index[0] = gma_ops->gma_to_pte_index(gma);
  1462. index = 1;
  1463. break;
  1464. default:
  1465. WARN_ON(1);
  1466. goto err;
  1467. }
  1468. /* walk into the shadow page table and get gpa from guest entry */
  1469. for (i = 0; i < index; i++) {
  1470. ret = ppgtt_get_next_level_entry(mm, &e, gma_index[i],
  1471. (i == index - 1));
  1472. if (ret)
  1473. goto err;
  1474. }
  1475. gpa = (pte_ops->get_pfn(&e) << GTT_PAGE_SHIFT)
  1476. + (gma & ~GTT_PAGE_MASK);
  1477. trace_gma_translate(vgpu->id, "ppgtt", 0,
  1478. mm->page_table_level, gma, gpa);
  1479. return gpa;
  1480. err:
  1481. gvt_err("invalid mm type: %d gma %lx\n", mm->type, gma);
  1482. return INTEL_GVT_INVALID_ADDR;
  1483. }
  1484. static int emulate_gtt_mmio_read(struct intel_vgpu *vgpu,
  1485. unsigned int off, void *p_data, unsigned int bytes)
  1486. {
  1487. struct intel_vgpu_mm *ggtt_mm = vgpu->gtt.ggtt_mm;
  1488. const struct intel_gvt_device_info *info = &vgpu->gvt->device_info;
  1489. unsigned long index = off >> info->gtt_entry_size_shift;
  1490. struct intel_gvt_gtt_entry e;
  1491. if (bytes != 4 && bytes != 8)
  1492. return -EINVAL;
  1493. ggtt_get_guest_entry(ggtt_mm, &e, index);
  1494. memcpy(p_data, (void *)&e.val64 + (off & (info->gtt_entry_size - 1)),
  1495. bytes);
  1496. return 0;
  1497. }
  1498. /**
  1499. * intel_vgpu_emulate_gtt_mmio_read - emulate GTT MMIO register read
  1500. * @vgpu: a vGPU
  1501. * @off: register offset
  1502. * @p_data: data will be returned to guest
  1503. * @bytes: data length
  1504. *
  1505. * This function is used to emulate the GTT MMIO register read
  1506. *
  1507. * Returns:
  1508. * Zero on success, error code if failed.
  1509. */
  1510. int intel_vgpu_emulate_gtt_mmio_read(struct intel_vgpu *vgpu, unsigned int off,
  1511. void *p_data, unsigned int bytes)
  1512. {
  1513. const struct intel_gvt_device_info *info = &vgpu->gvt->device_info;
  1514. int ret;
  1515. if (bytes != 4 && bytes != 8)
  1516. return -EINVAL;
  1517. off -= info->gtt_start_offset;
  1518. ret = emulate_gtt_mmio_read(vgpu, off, p_data, bytes);
  1519. return ret;
  1520. }
  1521. static int emulate_gtt_mmio_write(struct intel_vgpu *vgpu, unsigned int off,
  1522. void *p_data, unsigned int bytes)
  1523. {
  1524. struct intel_gvt *gvt = vgpu->gvt;
  1525. const struct intel_gvt_device_info *info = &gvt->device_info;
  1526. struct intel_vgpu_mm *ggtt_mm = vgpu->gtt.ggtt_mm;
  1527. struct intel_gvt_gtt_pte_ops *ops = gvt->gtt.pte_ops;
  1528. unsigned long g_gtt_index = off >> info->gtt_entry_size_shift;
  1529. unsigned long gma;
  1530. struct intel_gvt_gtt_entry e, m;
  1531. int ret;
  1532. if (bytes != 4 && bytes != 8)
  1533. return -EINVAL;
  1534. gma = g_gtt_index << GTT_PAGE_SHIFT;
  1535. /* the VM may configure the whole GM space when ballooning is used */
  1536. if (WARN_ONCE(!vgpu_gmadr_is_valid(vgpu, gma),
  1537. "vgpu%d: found oob ggtt write, offset %x\n",
  1538. vgpu->id, off)) {
  1539. return 0;
  1540. }
  1541. ggtt_get_guest_entry(ggtt_mm, &e, g_gtt_index);
  1542. memcpy((void *)&e.val64 + (off & (info->gtt_entry_size - 1)), p_data,
  1543. bytes);
  1544. if (ops->test_present(&e)) {
  1545. ret = gtt_entry_p2m(vgpu, &e, &m);
  1546. if (ret) {
  1547. gvt_err("vgpu%d: fail to translate guest gtt entry\n",
  1548. vgpu->id);
  1549. return ret;
  1550. }
  1551. } else {
  1552. m = e;
  1553. m.val64 = 0;
  1554. }
  1555. ggtt_set_shadow_entry(ggtt_mm, &m, g_gtt_index);
  1556. ggtt_set_guest_entry(ggtt_mm, &e, g_gtt_index);
  1557. return 0;
  1558. }
  1559. /*
  1560. * intel_vgpu_emulate_gtt_mmio_write - emulate GTT MMIO register write
  1561. * @vgpu: a vGPU
  1562. * @off: register offset
  1563. * @p_data: data from guest write
  1564. * @bytes: data length
  1565. *
  1566. * This function is used to emulate the GTT MMIO register write
  1567. *
  1568. * Returns:
  1569. * Zero on success, error code if failed.
  1570. */
  1571. int intel_vgpu_emulate_gtt_mmio_write(struct intel_vgpu *vgpu, unsigned int off,
  1572. void *p_data, unsigned int bytes)
  1573. {
  1574. const struct intel_gvt_device_info *info = &vgpu->gvt->device_info;
  1575. int ret;
  1576. if (bytes != 4 && bytes != 8)
  1577. return -EINVAL;
  1578. off -= info->gtt_start_offset;
  1579. ret = emulate_gtt_mmio_write(vgpu, off, p_data, bytes);
  1580. return ret;
  1581. }
  1582. static int alloc_scratch_pages(struct intel_vgpu *vgpu,
  1583. intel_gvt_gtt_type_t type)
  1584. {
  1585. struct intel_vgpu_gtt *gtt = &vgpu->gtt;
  1586. struct intel_gvt_gtt_pte_ops *ops = vgpu->gvt->gtt.pte_ops;
  1587. int page_entry_num = GTT_PAGE_SIZE >>
  1588. vgpu->gvt->device_info.gtt_entry_size_shift;
  1589. void *scratch_pt;
  1590. unsigned long mfn;
  1591. int i;
  1592. if (WARN_ON(type < GTT_TYPE_PPGTT_PTE_PT || type >= GTT_TYPE_MAX))
  1593. return -EINVAL;
  1594. scratch_pt = (void *)get_zeroed_page(GFP_KERNEL);
  1595. if (!scratch_pt) {
  1596. gvt_err("fail to allocate scratch page\n");
  1597. return -ENOMEM;
  1598. }
  1599. mfn = intel_gvt_hypervisor_virt_to_mfn(scratch_pt);
  1600. if (mfn == INTEL_GVT_INVALID_ADDR) {
  1601. gvt_err("fail to translate vaddr:0x%lx\n", (unsigned long)scratch_pt);
  1602. free_page((unsigned long)scratch_pt);
  1603. return -EFAULT;
  1604. }
  1605. gtt->scratch_pt[type].page_mfn = mfn;
  1606. gtt->scratch_pt[type].page = virt_to_page(scratch_pt);
  1607. gvt_dbg_mm("vgpu%d create scratch_pt: type %d mfn=0x%lx\n",
  1608. vgpu->id, type, mfn);
  1609. /* Build the tree by full filled the scratch pt with the entries which
  1610. * point to the next level scratch pt or scratch page. The
  1611. * scratch_pt[type] indicate the scratch pt/scratch page used by the
  1612. * 'type' pt.
  1613. * e.g. scratch_pt[GTT_TYPE_PPGTT_PDE_PT] is used by
  1614. * GTT_TYPE_PPGTT_PDE_PT level pt, that means this scratch_pt it self
  1615. * is GTT_TYPE_PPGTT_PTE_PT, and full filled by scratch page mfn.
  1616. */
  1617. if (type > GTT_TYPE_PPGTT_PTE_PT && type < GTT_TYPE_MAX) {
  1618. struct intel_gvt_gtt_entry se;
  1619. memset(&se, 0, sizeof(struct intel_gvt_gtt_entry));
  1620. se.type = get_entry_type(type - 1);
  1621. ops->set_pfn(&se, gtt->scratch_pt[type - 1].page_mfn);
  1622. /* The entry parameters like present/writeable/cache type
  1623. * set to the same as i915's scratch page tree.
  1624. */
  1625. se.val64 |= _PAGE_PRESENT | _PAGE_RW;
  1626. if (type == GTT_TYPE_PPGTT_PDE_PT)
  1627. se.val64 |= PPAT_CACHED_INDEX;
  1628. for (i = 0; i < page_entry_num; i++)
  1629. ops->set_entry(scratch_pt, &se, i, false, 0, vgpu);
  1630. }
  1631. return 0;
  1632. }
  1633. static int release_scratch_page_tree(struct intel_vgpu *vgpu)
  1634. {
  1635. int i;
  1636. for (i = GTT_TYPE_PPGTT_PTE_PT; i < GTT_TYPE_MAX; i++) {
  1637. if (vgpu->gtt.scratch_pt[i].page != NULL) {
  1638. __free_page(vgpu->gtt.scratch_pt[i].page);
  1639. vgpu->gtt.scratch_pt[i].page = NULL;
  1640. vgpu->gtt.scratch_pt[i].page_mfn = 0;
  1641. }
  1642. }
  1643. return 0;
  1644. }
  1645. static int create_scratch_page_tree(struct intel_vgpu *vgpu)
  1646. {
  1647. int i, ret;
  1648. for (i = GTT_TYPE_PPGTT_PTE_PT; i < GTT_TYPE_MAX; i++) {
  1649. ret = alloc_scratch_pages(vgpu, i);
  1650. if (ret)
  1651. goto err;
  1652. }
  1653. return 0;
  1654. err:
  1655. release_scratch_page_tree(vgpu);
  1656. return ret;
  1657. }
  1658. /**
  1659. * intel_vgpu_init_gtt - initialize per-vGPU graphics memory virulization
  1660. * @vgpu: a vGPU
  1661. *
  1662. * This function is used to initialize per-vGPU graphics memory virtualization
  1663. * components.
  1664. *
  1665. * Returns:
  1666. * Zero on success, error code if failed.
  1667. */
  1668. int intel_vgpu_init_gtt(struct intel_vgpu *vgpu)
  1669. {
  1670. struct intel_vgpu_gtt *gtt = &vgpu->gtt;
  1671. struct intel_vgpu_mm *ggtt_mm;
  1672. hash_init(gtt->guest_page_hash_table);
  1673. hash_init(gtt->shadow_page_hash_table);
  1674. INIT_LIST_HEAD(&gtt->mm_list_head);
  1675. INIT_LIST_HEAD(&gtt->oos_page_list_head);
  1676. INIT_LIST_HEAD(&gtt->post_shadow_list_head);
  1677. intel_vgpu_reset_ggtt(vgpu);
  1678. ggtt_mm = intel_vgpu_create_mm(vgpu, INTEL_GVT_MM_GGTT,
  1679. NULL, 1, 0);
  1680. if (IS_ERR(ggtt_mm)) {
  1681. gvt_err("fail to create mm for ggtt.\n");
  1682. return PTR_ERR(ggtt_mm);
  1683. }
  1684. gtt->ggtt_mm = ggtt_mm;
  1685. return create_scratch_page_tree(vgpu);
  1686. }
  1687. /**
  1688. * intel_vgpu_clean_gtt - clean up per-vGPU graphics memory virulization
  1689. * @vgpu: a vGPU
  1690. *
  1691. * This function is used to clean up per-vGPU graphics memory virtualization
  1692. * components.
  1693. *
  1694. * Returns:
  1695. * Zero on success, error code if failed.
  1696. */
  1697. void intel_vgpu_clean_gtt(struct intel_vgpu *vgpu)
  1698. {
  1699. struct list_head *pos, *n;
  1700. struct intel_vgpu_mm *mm;
  1701. ppgtt_free_all_shadow_page(vgpu);
  1702. release_scratch_page_tree(vgpu);
  1703. list_for_each_safe(pos, n, &vgpu->gtt.mm_list_head) {
  1704. mm = container_of(pos, struct intel_vgpu_mm, list);
  1705. vgpu->gvt->gtt.mm_free_page_table(mm);
  1706. list_del(&mm->list);
  1707. list_del(&mm->lru_list);
  1708. kfree(mm);
  1709. }
  1710. }
  1711. static void clean_spt_oos(struct intel_gvt *gvt)
  1712. {
  1713. struct intel_gvt_gtt *gtt = &gvt->gtt;
  1714. struct list_head *pos, *n;
  1715. struct intel_vgpu_oos_page *oos_page;
  1716. WARN(!list_empty(&gtt->oos_page_use_list_head),
  1717. "someone is still using oos page\n");
  1718. list_for_each_safe(pos, n, &gtt->oos_page_free_list_head) {
  1719. oos_page = container_of(pos, struct intel_vgpu_oos_page, list);
  1720. list_del(&oos_page->list);
  1721. kfree(oos_page);
  1722. }
  1723. }
  1724. static int setup_spt_oos(struct intel_gvt *gvt)
  1725. {
  1726. struct intel_gvt_gtt *gtt = &gvt->gtt;
  1727. struct intel_vgpu_oos_page *oos_page;
  1728. int i;
  1729. int ret;
  1730. INIT_LIST_HEAD(&gtt->oos_page_free_list_head);
  1731. INIT_LIST_HEAD(&gtt->oos_page_use_list_head);
  1732. for (i = 0; i < preallocated_oos_pages; i++) {
  1733. oos_page = kzalloc(sizeof(*oos_page), GFP_KERNEL);
  1734. if (!oos_page) {
  1735. gvt_err("fail to pre-allocate oos page\n");
  1736. ret = -ENOMEM;
  1737. goto fail;
  1738. }
  1739. INIT_LIST_HEAD(&oos_page->list);
  1740. INIT_LIST_HEAD(&oos_page->vm_list);
  1741. oos_page->id = i;
  1742. list_add_tail(&oos_page->list, &gtt->oos_page_free_list_head);
  1743. }
  1744. gvt_dbg_mm("%d oos pages preallocated\n", i);
  1745. return 0;
  1746. fail:
  1747. clean_spt_oos(gvt);
  1748. return ret;
  1749. }
  1750. /**
  1751. * intel_vgpu_find_ppgtt_mm - find a PPGTT mm object
  1752. * @vgpu: a vGPU
  1753. * @page_table_level: PPGTT page table level
  1754. * @root_entry: PPGTT page table root pointers
  1755. *
  1756. * This function is used to find a PPGTT mm object from mm object pool
  1757. *
  1758. * Returns:
  1759. * pointer to mm object on success, NULL if failed.
  1760. */
  1761. struct intel_vgpu_mm *intel_vgpu_find_ppgtt_mm(struct intel_vgpu *vgpu,
  1762. int page_table_level, void *root_entry)
  1763. {
  1764. struct list_head *pos;
  1765. struct intel_vgpu_mm *mm;
  1766. u64 *src, *dst;
  1767. list_for_each(pos, &vgpu->gtt.mm_list_head) {
  1768. mm = container_of(pos, struct intel_vgpu_mm, list);
  1769. if (mm->type != INTEL_GVT_MM_PPGTT)
  1770. continue;
  1771. if (mm->page_table_level != page_table_level)
  1772. continue;
  1773. src = root_entry;
  1774. dst = mm->virtual_page_table;
  1775. if (page_table_level == 3) {
  1776. if (src[0] == dst[0]
  1777. && src[1] == dst[1]
  1778. && src[2] == dst[2]
  1779. && src[3] == dst[3])
  1780. return mm;
  1781. } else {
  1782. if (src[0] == dst[0])
  1783. return mm;
  1784. }
  1785. }
  1786. return NULL;
  1787. }
  1788. /**
  1789. * intel_vgpu_g2v_create_ppgtt_mm - create a PPGTT mm object from
  1790. * g2v notification
  1791. * @vgpu: a vGPU
  1792. * @page_table_level: PPGTT page table level
  1793. *
  1794. * This function is used to create a PPGTT mm object from a guest to GVT-g
  1795. * notification.
  1796. *
  1797. * Returns:
  1798. * Zero on success, negative error code if failed.
  1799. */
  1800. int intel_vgpu_g2v_create_ppgtt_mm(struct intel_vgpu *vgpu,
  1801. int page_table_level)
  1802. {
  1803. u64 *pdp = (u64 *)&vgpu_vreg64(vgpu, vgtif_reg(pdp[0]));
  1804. struct intel_vgpu_mm *mm;
  1805. if (WARN_ON((page_table_level != 4) && (page_table_level != 3)))
  1806. return -EINVAL;
  1807. mm = intel_vgpu_find_ppgtt_mm(vgpu, page_table_level, pdp);
  1808. if (mm) {
  1809. intel_gvt_mm_reference(mm);
  1810. } else {
  1811. mm = intel_vgpu_create_mm(vgpu, INTEL_GVT_MM_PPGTT,
  1812. pdp, page_table_level, 0);
  1813. if (IS_ERR(mm)) {
  1814. gvt_err("fail to create mm\n");
  1815. return PTR_ERR(mm);
  1816. }
  1817. }
  1818. return 0;
  1819. }
  1820. /**
  1821. * intel_vgpu_g2v_destroy_ppgtt_mm - destroy a PPGTT mm object from
  1822. * g2v notification
  1823. * @vgpu: a vGPU
  1824. * @page_table_level: PPGTT page table level
  1825. *
  1826. * This function is used to create a PPGTT mm object from a guest to GVT-g
  1827. * notification.
  1828. *
  1829. * Returns:
  1830. * Zero on success, negative error code if failed.
  1831. */
  1832. int intel_vgpu_g2v_destroy_ppgtt_mm(struct intel_vgpu *vgpu,
  1833. int page_table_level)
  1834. {
  1835. u64 *pdp = (u64 *)&vgpu_vreg64(vgpu, vgtif_reg(pdp[0]));
  1836. struct intel_vgpu_mm *mm;
  1837. if (WARN_ON((page_table_level != 4) && (page_table_level != 3)))
  1838. return -EINVAL;
  1839. mm = intel_vgpu_find_ppgtt_mm(vgpu, page_table_level, pdp);
  1840. if (!mm) {
  1841. gvt_err("fail to find ppgtt instance.\n");
  1842. return -EINVAL;
  1843. }
  1844. intel_gvt_mm_unreference(mm);
  1845. return 0;
  1846. }
  1847. /**
  1848. * intel_gvt_init_gtt - initialize mm components of a GVT device
  1849. * @gvt: GVT device
  1850. *
  1851. * This function is called at the initialization stage, to initialize
  1852. * the mm components of a GVT device.
  1853. *
  1854. * Returns:
  1855. * zero on success, negative error code if failed.
  1856. */
  1857. int intel_gvt_init_gtt(struct intel_gvt *gvt)
  1858. {
  1859. int ret;
  1860. void *page;
  1861. gvt_dbg_core("init gtt\n");
  1862. if (IS_BROADWELL(gvt->dev_priv) || IS_SKYLAKE(gvt->dev_priv)) {
  1863. gvt->gtt.pte_ops = &gen8_gtt_pte_ops;
  1864. gvt->gtt.gma_ops = &gen8_gtt_gma_ops;
  1865. gvt->gtt.mm_alloc_page_table = gen8_mm_alloc_page_table;
  1866. gvt->gtt.mm_free_page_table = gen8_mm_free_page_table;
  1867. } else {
  1868. return -ENODEV;
  1869. }
  1870. page = (void *)get_zeroed_page(GFP_KERNEL);
  1871. if (!page) {
  1872. gvt_err("fail to allocate scratch ggtt page\n");
  1873. return -ENOMEM;
  1874. }
  1875. gvt->gtt.scratch_ggtt_page = virt_to_page(page);
  1876. gvt->gtt.scratch_ggtt_mfn = intel_gvt_hypervisor_virt_to_mfn(page);
  1877. if (gvt->gtt.scratch_ggtt_mfn == INTEL_GVT_INVALID_ADDR) {
  1878. gvt_err("fail to translate scratch ggtt page\n");
  1879. __free_page(gvt->gtt.scratch_ggtt_page);
  1880. return -EFAULT;
  1881. }
  1882. if (enable_out_of_sync) {
  1883. ret = setup_spt_oos(gvt);
  1884. if (ret) {
  1885. gvt_err("fail to initialize SPT oos\n");
  1886. return ret;
  1887. }
  1888. }
  1889. INIT_LIST_HEAD(&gvt->gtt.mm_lru_list_head);
  1890. return 0;
  1891. }
  1892. /**
  1893. * intel_gvt_clean_gtt - clean up mm components of a GVT device
  1894. * @gvt: GVT device
  1895. *
  1896. * This function is called at the driver unloading stage, to clean up the
  1897. * the mm components of a GVT device.
  1898. *
  1899. */
  1900. void intel_gvt_clean_gtt(struct intel_gvt *gvt)
  1901. {
  1902. __free_page(gvt->gtt.scratch_ggtt_page);
  1903. if (enable_out_of_sync)
  1904. clean_spt_oos(gvt);
  1905. }
  1906. /**
  1907. * intel_vgpu_reset_ggtt - reset the GGTT entry
  1908. * @vgpu: a vGPU
  1909. *
  1910. * This function is called at the vGPU create stage
  1911. * to reset all the GGTT entries.
  1912. *
  1913. */
  1914. void intel_vgpu_reset_ggtt(struct intel_vgpu *vgpu)
  1915. {
  1916. struct intel_gvt *gvt = vgpu->gvt;
  1917. struct intel_gvt_gtt_pte_ops *ops = vgpu->gvt->gtt.pte_ops;
  1918. u32 index;
  1919. u32 offset;
  1920. u32 num_entries;
  1921. struct intel_gvt_gtt_entry e;
  1922. memset(&e, 0, sizeof(struct intel_gvt_gtt_entry));
  1923. e.type = GTT_TYPE_GGTT_PTE;
  1924. ops->set_pfn(&e, gvt->gtt.scratch_ggtt_mfn);
  1925. e.val64 |= _PAGE_PRESENT;
  1926. index = vgpu_aperture_gmadr_base(vgpu) >> PAGE_SHIFT;
  1927. num_entries = vgpu_aperture_sz(vgpu) >> PAGE_SHIFT;
  1928. for (offset = 0; offset < num_entries; offset++)
  1929. ops->set_entry(NULL, &e, index + offset, false, 0, vgpu);
  1930. index = vgpu_hidden_gmadr_base(vgpu) >> PAGE_SHIFT;
  1931. num_entries = vgpu_hidden_sz(vgpu) >> PAGE_SHIFT;
  1932. for (offset = 0; offset < num_entries; offset++)
  1933. ops->set_entry(NULL, &e, index + offset, false, 0, vgpu);
  1934. }
  1935. /**
  1936. * intel_vgpu_reset_gtt - reset the all GTT related status
  1937. * @vgpu: a vGPU
  1938. * @dmlr: true for vGPU Device Model Level Reset, false for GT Reset
  1939. *
  1940. * This function is called from vfio core to reset reset all
  1941. * GTT related status, including GGTT, PPGTT, scratch page.
  1942. *
  1943. */
  1944. void intel_vgpu_reset_gtt(struct intel_vgpu *vgpu, bool dmlr)
  1945. {
  1946. int i;
  1947. ppgtt_free_all_shadow_page(vgpu);
  1948. if (!dmlr)
  1949. return;
  1950. intel_vgpu_reset_ggtt(vgpu);
  1951. /* clear scratch page for security */
  1952. for (i = GTT_TYPE_PPGTT_PTE_PT; i < GTT_TYPE_MAX; i++) {
  1953. if (vgpu->gtt.scratch_pt[i].page != NULL)
  1954. memset(page_address(vgpu->gtt.scratch_pt[i].page),
  1955. 0, PAGE_SIZE);
  1956. }
  1957. }