intel_lrc.c 12 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592
  1. /*
  2. * SPDX-License-Identifier: MIT
  3. *
  4. * Copyright © 2018 Intel Corporation
  5. */
  6. #include "../i915_selftest.h"
  7. #include "igt_flush_test.h"
  8. #include "mock_context.h"
  9. struct spinner {
  10. struct drm_i915_private *i915;
  11. struct drm_i915_gem_object *hws;
  12. struct drm_i915_gem_object *obj;
  13. u32 *batch;
  14. void *seqno;
  15. };
  16. static int spinner_init(struct spinner *spin, struct drm_i915_private *i915)
  17. {
  18. unsigned int mode;
  19. void *vaddr;
  20. int err;
  21. GEM_BUG_ON(INTEL_GEN(i915) < 8);
  22. memset(spin, 0, sizeof(*spin));
  23. spin->i915 = i915;
  24. spin->hws = i915_gem_object_create_internal(i915, PAGE_SIZE);
  25. if (IS_ERR(spin->hws)) {
  26. err = PTR_ERR(spin->hws);
  27. goto err;
  28. }
  29. spin->obj = i915_gem_object_create_internal(i915, PAGE_SIZE);
  30. if (IS_ERR(spin->obj)) {
  31. err = PTR_ERR(spin->obj);
  32. goto err_hws;
  33. }
  34. i915_gem_object_set_cache_level(spin->hws, I915_CACHE_LLC);
  35. vaddr = i915_gem_object_pin_map(spin->hws, I915_MAP_WB);
  36. if (IS_ERR(vaddr)) {
  37. err = PTR_ERR(vaddr);
  38. goto err_obj;
  39. }
  40. spin->seqno = memset(vaddr, 0xff, PAGE_SIZE);
  41. mode = HAS_LLC(i915) ? I915_MAP_WB : I915_MAP_WC;
  42. vaddr = i915_gem_object_pin_map(spin->obj, mode);
  43. if (IS_ERR(vaddr)) {
  44. err = PTR_ERR(vaddr);
  45. goto err_unpin_hws;
  46. }
  47. spin->batch = vaddr;
  48. return 0;
  49. err_unpin_hws:
  50. i915_gem_object_unpin_map(spin->hws);
  51. err_obj:
  52. i915_gem_object_put(spin->obj);
  53. err_hws:
  54. i915_gem_object_put(spin->hws);
  55. err:
  56. return err;
  57. }
  58. static unsigned int seqno_offset(u64 fence)
  59. {
  60. return offset_in_page(sizeof(u32) * fence);
  61. }
  62. static u64 hws_address(const struct i915_vma *hws,
  63. const struct i915_request *rq)
  64. {
  65. return hws->node.start + seqno_offset(rq->fence.context);
  66. }
  67. static int emit_recurse_batch(struct spinner *spin,
  68. struct i915_request *rq,
  69. u32 arbitration_command)
  70. {
  71. struct i915_address_space *vm = &rq->gem_context->ppgtt->vm;
  72. struct i915_vma *hws, *vma;
  73. u32 *batch;
  74. int err;
  75. vma = i915_vma_instance(spin->obj, vm, NULL);
  76. if (IS_ERR(vma))
  77. return PTR_ERR(vma);
  78. hws = i915_vma_instance(spin->hws, vm, NULL);
  79. if (IS_ERR(hws))
  80. return PTR_ERR(hws);
  81. err = i915_vma_pin(vma, 0, 0, PIN_USER);
  82. if (err)
  83. return err;
  84. err = i915_vma_pin(hws, 0, 0, PIN_USER);
  85. if (err)
  86. goto unpin_vma;
  87. err = i915_vma_move_to_active(vma, rq, 0);
  88. if (err)
  89. goto unpin_hws;
  90. if (!i915_gem_object_has_active_reference(vma->obj)) {
  91. i915_gem_object_get(vma->obj);
  92. i915_gem_object_set_active_reference(vma->obj);
  93. }
  94. err = i915_vma_move_to_active(hws, rq, 0);
  95. if (err)
  96. goto unpin_hws;
  97. if (!i915_gem_object_has_active_reference(hws->obj)) {
  98. i915_gem_object_get(hws->obj);
  99. i915_gem_object_set_active_reference(hws->obj);
  100. }
  101. batch = spin->batch;
  102. *batch++ = MI_STORE_DWORD_IMM_GEN4;
  103. *batch++ = lower_32_bits(hws_address(hws, rq));
  104. *batch++ = upper_32_bits(hws_address(hws, rq));
  105. *batch++ = rq->fence.seqno;
  106. *batch++ = arbitration_command;
  107. *batch++ = MI_BATCH_BUFFER_START | 1 << 8 | 1;
  108. *batch++ = lower_32_bits(vma->node.start);
  109. *batch++ = upper_32_bits(vma->node.start);
  110. *batch++ = MI_BATCH_BUFFER_END; /* not reached */
  111. i915_gem_chipset_flush(spin->i915);
  112. err = rq->engine->emit_bb_start(rq, vma->node.start, PAGE_SIZE, 0);
  113. unpin_hws:
  114. i915_vma_unpin(hws);
  115. unpin_vma:
  116. i915_vma_unpin(vma);
  117. return err;
  118. }
  119. static struct i915_request *
  120. spinner_create_request(struct spinner *spin,
  121. struct i915_gem_context *ctx,
  122. struct intel_engine_cs *engine,
  123. u32 arbitration_command)
  124. {
  125. struct i915_request *rq;
  126. int err;
  127. rq = i915_request_alloc(engine, ctx);
  128. if (IS_ERR(rq))
  129. return rq;
  130. err = emit_recurse_batch(spin, rq, arbitration_command);
  131. if (err) {
  132. i915_request_add(rq);
  133. return ERR_PTR(err);
  134. }
  135. return rq;
  136. }
  137. static u32 hws_seqno(const struct spinner *spin, const struct i915_request *rq)
  138. {
  139. u32 *seqno = spin->seqno + seqno_offset(rq->fence.context);
  140. return READ_ONCE(*seqno);
  141. }
  142. static void spinner_end(struct spinner *spin)
  143. {
  144. *spin->batch = MI_BATCH_BUFFER_END;
  145. i915_gem_chipset_flush(spin->i915);
  146. }
  147. static void spinner_fini(struct spinner *spin)
  148. {
  149. spinner_end(spin);
  150. i915_gem_object_unpin_map(spin->obj);
  151. i915_gem_object_put(spin->obj);
  152. i915_gem_object_unpin_map(spin->hws);
  153. i915_gem_object_put(spin->hws);
  154. }
  155. static bool wait_for_spinner(struct spinner *spin, struct i915_request *rq)
  156. {
  157. if (!wait_event_timeout(rq->execute,
  158. READ_ONCE(rq->global_seqno),
  159. msecs_to_jiffies(10)))
  160. return false;
  161. return !(wait_for_us(i915_seqno_passed(hws_seqno(spin, rq),
  162. rq->fence.seqno),
  163. 10) &&
  164. wait_for(i915_seqno_passed(hws_seqno(spin, rq),
  165. rq->fence.seqno),
  166. 1000));
  167. }
  168. static int live_sanitycheck(void *arg)
  169. {
  170. struct drm_i915_private *i915 = arg;
  171. struct intel_engine_cs *engine;
  172. struct i915_gem_context *ctx;
  173. enum intel_engine_id id;
  174. struct spinner spin;
  175. int err = -ENOMEM;
  176. if (!HAS_LOGICAL_RING_CONTEXTS(i915))
  177. return 0;
  178. mutex_lock(&i915->drm.struct_mutex);
  179. intel_runtime_pm_get(i915);
  180. if (spinner_init(&spin, i915))
  181. goto err_unlock;
  182. ctx = kernel_context(i915);
  183. if (!ctx)
  184. goto err_spin;
  185. for_each_engine(engine, i915, id) {
  186. struct i915_request *rq;
  187. rq = spinner_create_request(&spin, ctx, engine, MI_NOOP);
  188. if (IS_ERR(rq)) {
  189. err = PTR_ERR(rq);
  190. goto err_ctx;
  191. }
  192. i915_request_add(rq);
  193. if (!wait_for_spinner(&spin, rq)) {
  194. GEM_TRACE("spinner failed to start\n");
  195. GEM_TRACE_DUMP();
  196. i915_gem_set_wedged(i915);
  197. err = -EIO;
  198. goto err_ctx;
  199. }
  200. spinner_end(&spin);
  201. if (igt_flush_test(i915, I915_WAIT_LOCKED)) {
  202. err = -EIO;
  203. goto err_ctx;
  204. }
  205. }
  206. err = 0;
  207. err_ctx:
  208. kernel_context_close(ctx);
  209. err_spin:
  210. spinner_fini(&spin);
  211. err_unlock:
  212. igt_flush_test(i915, I915_WAIT_LOCKED);
  213. intel_runtime_pm_put(i915);
  214. mutex_unlock(&i915->drm.struct_mutex);
  215. return err;
  216. }
  217. static int live_preempt(void *arg)
  218. {
  219. struct drm_i915_private *i915 = arg;
  220. struct i915_gem_context *ctx_hi, *ctx_lo;
  221. struct spinner spin_hi, spin_lo;
  222. struct intel_engine_cs *engine;
  223. enum intel_engine_id id;
  224. int err = -ENOMEM;
  225. if (!HAS_LOGICAL_RING_PREEMPTION(i915))
  226. return 0;
  227. mutex_lock(&i915->drm.struct_mutex);
  228. intel_runtime_pm_get(i915);
  229. if (spinner_init(&spin_hi, i915))
  230. goto err_unlock;
  231. if (spinner_init(&spin_lo, i915))
  232. goto err_spin_hi;
  233. ctx_hi = kernel_context(i915);
  234. if (!ctx_hi)
  235. goto err_spin_lo;
  236. ctx_hi->sched.priority = I915_CONTEXT_MAX_USER_PRIORITY;
  237. ctx_lo = kernel_context(i915);
  238. if (!ctx_lo)
  239. goto err_ctx_hi;
  240. ctx_lo->sched.priority = I915_CONTEXT_MIN_USER_PRIORITY;
  241. for_each_engine(engine, i915, id) {
  242. struct i915_request *rq;
  243. rq = spinner_create_request(&spin_lo, ctx_lo, engine,
  244. MI_ARB_CHECK);
  245. if (IS_ERR(rq)) {
  246. err = PTR_ERR(rq);
  247. goto err_ctx_lo;
  248. }
  249. i915_request_add(rq);
  250. if (!wait_for_spinner(&spin_lo, rq)) {
  251. GEM_TRACE("lo spinner failed to start\n");
  252. GEM_TRACE_DUMP();
  253. i915_gem_set_wedged(i915);
  254. err = -EIO;
  255. goto err_ctx_lo;
  256. }
  257. rq = spinner_create_request(&spin_hi, ctx_hi, engine,
  258. MI_ARB_CHECK);
  259. if (IS_ERR(rq)) {
  260. spinner_end(&spin_lo);
  261. err = PTR_ERR(rq);
  262. goto err_ctx_lo;
  263. }
  264. i915_request_add(rq);
  265. if (!wait_for_spinner(&spin_hi, rq)) {
  266. GEM_TRACE("hi spinner failed to start\n");
  267. GEM_TRACE_DUMP();
  268. i915_gem_set_wedged(i915);
  269. err = -EIO;
  270. goto err_ctx_lo;
  271. }
  272. spinner_end(&spin_hi);
  273. spinner_end(&spin_lo);
  274. if (igt_flush_test(i915, I915_WAIT_LOCKED)) {
  275. err = -EIO;
  276. goto err_ctx_lo;
  277. }
  278. }
  279. err = 0;
  280. err_ctx_lo:
  281. kernel_context_close(ctx_lo);
  282. err_ctx_hi:
  283. kernel_context_close(ctx_hi);
  284. err_spin_lo:
  285. spinner_fini(&spin_lo);
  286. err_spin_hi:
  287. spinner_fini(&spin_hi);
  288. err_unlock:
  289. igt_flush_test(i915, I915_WAIT_LOCKED);
  290. intel_runtime_pm_put(i915);
  291. mutex_unlock(&i915->drm.struct_mutex);
  292. return err;
  293. }
  294. static int live_late_preempt(void *arg)
  295. {
  296. struct drm_i915_private *i915 = arg;
  297. struct i915_gem_context *ctx_hi, *ctx_lo;
  298. struct spinner spin_hi, spin_lo;
  299. struct intel_engine_cs *engine;
  300. struct i915_sched_attr attr = {};
  301. enum intel_engine_id id;
  302. int err = -ENOMEM;
  303. if (!HAS_LOGICAL_RING_PREEMPTION(i915))
  304. return 0;
  305. mutex_lock(&i915->drm.struct_mutex);
  306. intel_runtime_pm_get(i915);
  307. if (spinner_init(&spin_hi, i915))
  308. goto err_unlock;
  309. if (spinner_init(&spin_lo, i915))
  310. goto err_spin_hi;
  311. ctx_hi = kernel_context(i915);
  312. if (!ctx_hi)
  313. goto err_spin_lo;
  314. ctx_lo = kernel_context(i915);
  315. if (!ctx_lo)
  316. goto err_ctx_hi;
  317. for_each_engine(engine, i915, id) {
  318. struct i915_request *rq;
  319. rq = spinner_create_request(&spin_lo, ctx_lo, engine,
  320. MI_ARB_CHECK);
  321. if (IS_ERR(rq)) {
  322. err = PTR_ERR(rq);
  323. goto err_ctx_lo;
  324. }
  325. i915_request_add(rq);
  326. if (!wait_for_spinner(&spin_lo, rq)) {
  327. pr_err("First context failed to start\n");
  328. goto err_wedged;
  329. }
  330. rq = spinner_create_request(&spin_hi, ctx_hi, engine, MI_NOOP);
  331. if (IS_ERR(rq)) {
  332. spinner_end(&spin_lo);
  333. err = PTR_ERR(rq);
  334. goto err_ctx_lo;
  335. }
  336. i915_request_add(rq);
  337. if (wait_for_spinner(&spin_hi, rq)) {
  338. pr_err("Second context overtook first?\n");
  339. goto err_wedged;
  340. }
  341. attr.priority = I915_PRIORITY_MAX;
  342. engine->schedule(rq, &attr);
  343. if (!wait_for_spinner(&spin_hi, rq)) {
  344. pr_err("High priority context failed to preempt the low priority context\n");
  345. GEM_TRACE_DUMP();
  346. goto err_wedged;
  347. }
  348. spinner_end(&spin_hi);
  349. spinner_end(&spin_lo);
  350. if (igt_flush_test(i915, I915_WAIT_LOCKED)) {
  351. err = -EIO;
  352. goto err_ctx_lo;
  353. }
  354. }
  355. err = 0;
  356. err_ctx_lo:
  357. kernel_context_close(ctx_lo);
  358. err_ctx_hi:
  359. kernel_context_close(ctx_hi);
  360. err_spin_lo:
  361. spinner_fini(&spin_lo);
  362. err_spin_hi:
  363. spinner_fini(&spin_hi);
  364. err_unlock:
  365. igt_flush_test(i915, I915_WAIT_LOCKED);
  366. intel_runtime_pm_put(i915);
  367. mutex_unlock(&i915->drm.struct_mutex);
  368. return err;
  369. err_wedged:
  370. spinner_end(&spin_hi);
  371. spinner_end(&spin_lo);
  372. i915_gem_set_wedged(i915);
  373. err = -EIO;
  374. goto err_ctx_lo;
  375. }
  376. static int live_preempt_hang(void *arg)
  377. {
  378. struct drm_i915_private *i915 = arg;
  379. struct i915_gem_context *ctx_hi, *ctx_lo;
  380. struct spinner spin_hi, spin_lo;
  381. struct intel_engine_cs *engine;
  382. enum intel_engine_id id;
  383. int err = -ENOMEM;
  384. if (!HAS_LOGICAL_RING_PREEMPTION(i915))
  385. return 0;
  386. if (!intel_has_reset_engine(i915))
  387. return 0;
  388. mutex_lock(&i915->drm.struct_mutex);
  389. intel_runtime_pm_get(i915);
  390. if (spinner_init(&spin_hi, i915))
  391. goto err_unlock;
  392. if (spinner_init(&spin_lo, i915))
  393. goto err_spin_hi;
  394. ctx_hi = kernel_context(i915);
  395. if (!ctx_hi)
  396. goto err_spin_lo;
  397. ctx_hi->sched.priority = I915_CONTEXT_MAX_USER_PRIORITY;
  398. ctx_lo = kernel_context(i915);
  399. if (!ctx_lo)
  400. goto err_ctx_hi;
  401. ctx_lo->sched.priority = I915_CONTEXT_MIN_USER_PRIORITY;
  402. for_each_engine(engine, i915, id) {
  403. struct i915_request *rq;
  404. if (!intel_engine_has_preemption(engine))
  405. continue;
  406. rq = spinner_create_request(&spin_lo, ctx_lo, engine,
  407. MI_ARB_CHECK);
  408. if (IS_ERR(rq)) {
  409. err = PTR_ERR(rq);
  410. goto err_ctx_lo;
  411. }
  412. i915_request_add(rq);
  413. if (!wait_for_spinner(&spin_lo, rq)) {
  414. GEM_TRACE("lo spinner failed to start\n");
  415. GEM_TRACE_DUMP();
  416. i915_gem_set_wedged(i915);
  417. err = -EIO;
  418. goto err_ctx_lo;
  419. }
  420. rq = spinner_create_request(&spin_hi, ctx_hi, engine,
  421. MI_ARB_CHECK);
  422. if (IS_ERR(rq)) {
  423. spinner_end(&spin_lo);
  424. err = PTR_ERR(rq);
  425. goto err_ctx_lo;
  426. }
  427. init_completion(&engine->execlists.preempt_hang.completion);
  428. engine->execlists.preempt_hang.inject_hang = true;
  429. i915_request_add(rq);
  430. if (!wait_for_completion_timeout(&engine->execlists.preempt_hang.completion,
  431. HZ / 10)) {
  432. pr_err("Preemption did not occur within timeout!");
  433. GEM_TRACE_DUMP();
  434. i915_gem_set_wedged(i915);
  435. err = -EIO;
  436. goto err_ctx_lo;
  437. }
  438. set_bit(I915_RESET_ENGINE + id, &i915->gpu_error.flags);
  439. i915_reset_engine(engine, NULL);
  440. clear_bit(I915_RESET_ENGINE + id, &i915->gpu_error.flags);
  441. engine->execlists.preempt_hang.inject_hang = false;
  442. if (!wait_for_spinner(&spin_hi, rq)) {
  443. GEM_TRACE("hi spinner failed to start\n");
  444. GEM_TRACE_DUMP();
  445. i915_gem_set_wedged(i915);
  446. err = -EIO;
  447. goto err_ctx_lo;
  448. }
  449. spinner_end(&spin_hi);
  450. spinner_end(&spin_lo);
  451. if (igt_flush_test(i915, I915_WAIT_LOCKED)) {
  452. err = -EIO;
  453. goto err_ctx_lo;
  454. }
  455. }
  456. err = 0;
  457. err_ctx_lo:
  458. kernel_context_close(ctx_lo);
  459. err_ctx_hi:
  460. kernel_context_close(ctx_hi);
  461. err_spin_lo:
  462. spinner_fini(&spin_lo);
  463. err_spin_hi:
  464. spinner_fini(&spin_hi);
  465. err_unlock:
  466. igt_flush_test(i915, I915_WAIT_LOCKED);
  467. intel_runtime_pm_put(i915);
  468. mutex_unlock(&i915->drm.struct_mutex);
  469. return err;
  470. }
  471. int intel_execlists_live_selftests(struct drm_i915_private *i915)
  472. {
  473. static const struct i915_subtest tests[] = {
  474. SUBTEST(live_sanitycheck),
  475. SUBTEST(live_preempt),
  476. SUBTEST(live_late_preempt),
  477. SUBTEST(live_preempt_hang),
  478. };
  479. if (!HAS_EXECLISTS(i915))
  480. return 0;
  481. if (i915_terminally_wedged(&i915->gpu_error))
  482. return 0;
  483. return i915_subtests(tests, i915);
  484. }