i915_sw_fence.c 12 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525
  1. /*
  2. * (C) Copyright 2016 Intel Corporation
  3. *
  4. * This program is free software; you can redistribute it and/or
  5. * modify it under the terms of the GNU General Public License
  6. * as published by the Free Software Foundation; version 2
  7. * of the License.
  8. */
  9. #include <linux/slab.h>
  10. #include <linux/dma-fence.h>
  11. #include <linux/reservation.h>
  12. #include "i915_sw_fence.h"
  13. #define I915_SW_FENCE_FLAG_ALLOC BIT(3) /* after WQ_FLAG_* for safety */
  14. static DEFINE_SPINLOCK(i915_sw_fence_lock);
  15. enum {
  16. DEBUG_FENCE_IDLE = 0,
  17. DEBUG_FENCE_NOTIFY,
  18. };
  19. #ifdef CONFIG_DRM_I915_SW_FENCE_DEBUG_OBJECTS
  20. static void *i915_sw_fence_debug_hint(void *addr)
  21. {
  22. return (void *)(((struct i915_sw_fence *)addr)->flags & I915_SW_FENCE_MASK);
  23. }
  24. static struct debug_obj_descr i915_sw_fence_debug_descr = {
  25. .name = "i915_sw_fence",
  26. .debug_hint = i915_sw_fence_debug_hint,
  27. };
  28. static inline void debug_fence_init(struct i915_sw_fence *fence)
  29. {
  30. debug_object_init(fence, &i915_sw_fence_debug_descr);
  31. }
  32. static inline void debug_fence_activate(struct i915_sw_fence *fence)
  33. {
  34. debug_object_activate(fence, &i915_sw_fence_debug_descr);
  35. }
  36. static inline void debug_fence_set_state(struct i915_sw_fence *fence,
  37. int old, int new)
  38. {
  39. debug_object_active_state(fence, &i915_sw_fence_debug_descr, old, new);
  40. }
  41. static inline void debug_fence_deactivate(struct i915_sw_fence *fence)
  42. {
  43. debug_object_deactivate(fence, &i915_sw_fence_debug_descr);
  44. }
  45. static inline void debug_fence_destroy(struct i915_sw_fence *fence)
  46. {
  47. debug_object_destroy(fence, &i915_sw_fence_debug_descr);
  48. }
  49. static inline void debug_fence_free(struct i915_sw_fence *fence)
  50. {
  51. debug_object_free(fence, &i915_sw_fence_debug_descr);
  52. smp_wmb(); /* flush the change in state before reallocation */
  53. }
  54. static inline void debug_fence_assert(struct i915_sw_fence *fence)
  55. {
  56. debug_object_assert_init(fence, &i915_sw_fence_debug_descr);
  57. }
  58. #else
  59. static inline void debug_fence_init(struct i915_sw_fence *fence)
  60. {
  61. }
  62. static inline void debug_fence_activate(struct i915_sw_fence *fence)
  63. {
  64. }
  65. static inline void debug_fence_set_state(struct i915_sw_fence *fence,
  66. int old, int new)
  67. {
  68. }
  69. static inline void debug_fence_deactivate(struct i915_sw_fence *fence)
  70. {
  71. }
  72. static inline void debug_fence_destroy(struct i915_sw_fence *fence)
  73. {
  74. }
  75. static inline void debug_fence_free(struct i915_sw_fence *fence)
  76. {
  77. }
  78. static inline void debug_fence_assert(struct i915_sw_fence *fence)
  79. {
  80. }
  81. #endif
  82. static int __i915_sw_fence_notify(struct i915_sw_fence *fence,
  83. enum i915_sw_fence_notify state)
  84. {
  85. i915_sw_fence_notify_t fn;
  86. fn = (i915_sw_fence_notify_t)(fence->flags & I915_SW_FENCE_MASK);
  87. return fn(fence, state);
  88. }
  89. #ifdef CONFIG_DRM_I915_SW_FENCE_DEBUG_OBJECTS
  90. void i915_sw_fence_fini(struct i915_sw_fence *fence)
  91. {
  92. debug_fence_free(fence);
  93. }
  94. #endif
  95. static void i915_sw_fence_release(struct kref *kref)
  96. {
  97. struct i915_sw_fence *fence = container_of(kref, typeof(*fence), kref);
  98. WARN_ON(atomic_read(&fence->pending) > 0);
  99. debug_fence_destroy(fence);
  100. if (fence->flags & I915_SW_FENCE_MASK) {
  101. __i915_sw_fence_notify(fence, FENCE_FREE);
  102. } else {
  103. i915_sw_fence_fini(fence);
  104. kfree(fence);
  105. }
  106. }
  107. static void i915_sw_fence_put(struct i915_sw_fence *fence)
  108. {
  109. debug_fence_assert(fence);
  110. kref_put(&fence->kref, i915_sw_fence_release);
  111. }
  112. static struct i915_sw_fence *i915_sw_fence_get(struct i915_sw_fence *fence)
  113. {
  114. debug_fence_assert(fence);
  115. kref_get(&fence->kref);
  116. return fence;
  117. }
  118. static void __i915_sw_fence_wake_up_all(struct i915_sw_fence *fence,
  119. struct list_head *continuation)
  120. {
  121. wait_queue_head_t *x = &fence->wait;
  122. wait_queue_t *pos, *next;
  123. unsigned long flags;
  124. debug_fence_deactivate(fence);
  125. atomic_set_release(&fence->pending, -1); /* 0 -> -1 [done] */
  126. /*
  127. * To prevent unbounded recursion as we traverse the graph of
  128. * i915_sw_fences, we move the task_list from this, the next ready
  129. * fence, to the tail of the original fence's task_list
  130. * (and so added to the list to be woken).
  131. */
  132. spin_lock_irqsave_nested(&x->lock, flags, 1 + !!continuation);
  133. if (continuation) {
  134. list_for_each_entry_safe(pos, next, &x->task_list, task_list) {
  135. if (pos->func == autoremove_wake_function)
  136. pos->func(pos, TASK_NORMAL, 0, continuation);
  137. else
  138. list_move_tail(&pos->task_list, continuation);
  139. }
  140. } else {
  141. LIST_HEAD(extra);
  142. do {
  143. list_for_each_entry_safe(pos, next,
  144. &x->task_list, task_list)
  145. pos->func(pos, TASK_NORMAL, 0, &extra);
  146. if (list_empty(&extra))
  147. break;
  148. list_splice_tail_init(&extra, &x->task_list);
  149. } while (1);
  150. }
  151. spin_unlock_irqrestore(&x->lock, flags);
  152. debug_fence_assert(fence);
  153. }
  154. static void __i915_sw_fence_complete(struct i915_sw_fence *fence,
  155. struct list_head *continuation)
  156. {
  157. debug_fence_assert(fence);
  158. if (!atomic_dec_and_test(&fence->pending))
  159. return;
  160. debug_fence_set_state(fence, DEBUG_FENCE_IDLE, DEBUG_FENCE_NOTIFY);
  161. if (fence->flags & I915_SW_FENCE_MASK &&
  162. __i915_sw_fence_notify(fence, FENCE_COMPLETE) != NOTIFY_DONE)
  163. return;
  164. debug_fence_set_state(fence, DEBUG_FENCE_NOTIFY, DEBUG_FENCE_IDLE);
  165. __i915_sw_fence_wake_up_all(fence, continuation);
  166. }
  167. static void i915_sw_fence_complete(struct i915_sw_fence *fence)
  168. {
  169. debug_fence_assert(fence);
  170. if (WARN_ON(i915_sw_fence_done(fence)))
  171. return;
  172. __i915_sw_fence_complete(fence, NULL);
  173. }
  174. static void i915_sw_fence_await(struct i915_sw_fence *fence)
  175. {
  176. debug_fence_assert(fence);
  177. WARN_ON(atomic_inc_return(&fence->pending) <= 1);
  178. }
  179. void __i915_sw_fence_init(struct i915_sw_fence *fence,
  180. i915_sw_fence_notify_t fn,
  181. const char *name,
  182. struct lock_class_key *key)
  183. {
  184. BUG_ON((unsigned long)fn & ~I915_SW_FENCE_MASK);
  185. debug_fence_init(fence);
  186. __init_waitqueue_head(&fence->wait, name, key);
  187. kref_init(&fence->kref);
  188. atomic_set(&fence->pending, 1);
  189. fence->flags = (unsigned long)fn;
  190. }
  191. static void __i915_sw_fence_commit(struct i915_sw_fence *fence)
  192. {
  193. i915_sw_fence_complete(fence);
  194. i915_sw_fence_put(fence);
  195. }
  196. void i915_sw_fence_commit(struct i915_sw_fence *fence)
  197. {
  198. debug_fence_activate(fence);
  199. __i915_sw_fence_commit(fence);
  200. }
  201. static int i915_sw_fence_wake(wait_queue_t *wq, unsigned mode, int flags, void *key)
  202. {
  203. list_del(&wq->task_list);
  204. __i915_sw_fence_complete(wq->private, key);
  205. i915_sw_fence_put(wq->private);
  206. if (wq->flags & I915_SW_FENCE_FLAG_ALLOC)
  207. kfree(wq);
  208. return 0;
  209. }
  210. static bool __i915_sw_fence_check_if_after(struct i915_sw_fence *fence,
  211. const struct i915_sw_fence * const signaler)
  212. {
  213. wait_queue_t *wq;
  214. if (__test_and_set_bit(I915_SW_FENCE_CHECKED_BIT, &fence->flags))
  215. return false;
  216. if (fence == signaler)
  217. return true;
  218. list_for_each_entry(wq, &fence->wait.task_list, task_list) {
  219. if (wq->func != i915_sw_fence_wake)
  220. continue;
  221. if (__i915_sw_fence_check_if_after(wq->private, signaler))
  222. return true;
  223. }
  224. return false;
  225. }
  226. static void __i915_sw_fence_clear_checked_bit(struct i915_sw_fence *fence)
  227. {
  228. wait_queue_t *wq;
  229. if (!__test_and_clear_bit(I915_SW_FENCE_CHECKED_BIT, &fence->flags))
  230. return;
  231. list_for_each_entry(wq, &fence->wait.task_list, task_list) {
  232. if (wq->func != i915_sw_fence_wake)
  233. continue;
  234. __i915_sw_fence_clear_checked_bit(wq->private);
  235. }
  236. }
  237. static bool i915_sw_fence_check_if_after(struct i915_sw_fence *fence,
  238. const struct i915_sw_fence * const signaler)
  239. {
  240. unsigned long flags;
  241. bool err;
  242. if (!IS_ENABLED(CONFIG_I915_SW_FENCE_CHECK_DAG))
  243. return false;
  244. spin_lock_irqsave(&i915_sw_fence_lock, flags);
  245. err = __i915_sw_fence_check_if_after(fence, signaler);
  246. __i915_sw_fence_clear_checked_bit(fence);
  247. spin_unlock_irqrestore(&i915_sw_fence_lock, flags);
  248. return err;
  249. }
  250. static int __i915_sw_fence_await_sw_fence(struct i915_sw_fence *fence,
  251. struct i915_sw_fence *signaler,
  252. wait_queue_t *wq, gfp_t gfp)
  253. {
  254. unsigned long flags;
  255. int pending;
  256. debug_fence_assert(fence);
  257. if (i915_sw_fence_done(signaler))
  258. return 0;
  259. debug_fence_assert(signaler);
  260. /* The dependency graph must be acyclic. */
  261. if (unlikely(i915_sw_fence_check_if_after(fence, signaler)))
  262. return -EINVAL;
  263. pending = 0;
  264. if (!wq) {
  265. wq = kmalloc(sizeof(*wq), gfp);
  266. if (!wq) {
  267. if (!gfpflags_allow_blocking(gfp))
  268. return -ENOMEM;
  269. i915_sw_fence_wait(signaler);
  270. return 0;
  271. }
  272. pending |= I915_SW_FENCE_FLAG_ALLOC;
  273. }
  274. INIT_LIST_HEAD(&wq->task_list);
  275. wq->flags = pending;
  276. wq->func = i915_sw_fence_wake;
  277. wq->private = i915_sw_fence_get(fence);
  278. i915_sw_fence_await(fence);
  279. spin_lock_irqsave(&signaler->wait.lock, flags);
  280. if (likely(!i915_sw_fence_done(signaler))) {
  281. __add_wait_queue_tail(&signaler->wait, wq);
  282. pending = 1;
  283. } else {
  284. i915_sw_fence_wake(wq, 0, 0, NULL);
  285. pending = 0;
  286. }
  287. spin_unlock_irqrestore(&signaler->wait.lock, flags);
  288. return pending;
  289. }
  290. int i915_sw_fence_await_sw_fence(struct i915_sw_fence *fence,
  291. struct i915_sw_fence *signaler,
  292. wait_queue_t *wq)
  293. {
  294. return __i915_sw_fence_await_sw_fence(fence, signaler, wq, 0);
  295. }
  296. int i915_sw_fence_await_sw_fence_gfp(struct i915_sw_fence *fence,
  297. struct i915_sw_fence *signaler,
  298. gfp_t gfp)
  299. {
  300. return __i915_sw_fence_await_sw_fence(fence, signaler, NULL, gfp);
  301. }
  302. struct i915_sw_dma_fence_cb {
  303. struct dma_fence_cb base;
  304. struct i915_sw_fence *fence;
  305. struct dma_fence *dma;
  306. struct timer_list timer;
  307. };
  308. static void timer_i915_sw_fence_wake(unsigned long data)
  309. {
  310. struct i915_sw_dma_fence_cb *cb = (struct i915_sw_dma_fence_cb *)data;
  311. printk(KERN_WARNING "asynchronous wait on fence %s:%s:%x timed out\n",
  312. cb->dma->ops->get_driver_name(cb->dma),
  313. cb->dma->ops->get_timeline_name(cb->dma),
  314. cb->dma->seqno);
  315. dma_fence_put(cb->dma);
  316. cb->dma = NULL;
  317. __i915_sw_fence_commit(cb->fence);
  318. cb->timer.function = NULL;
  319. }
  320. static void dma_i915_sw_fence_wake(struct dma_fence *dma,
  321. struct dma_fence_cb *data)
  322. {
  323. struct i915_sw_dma_fence_cb *cb = container_of(data, typeof(*cb), base);
  324. del_timer_sync(&cb->timer);
  325. if (cb->timer.function)
  326. __i915_sw_fence_commit(cb->fence);
  327. dma_fence_put(cb->dma);
  328. kfree(cb);
  329. }
  330. int i915_sw_fence_await_dma_fence(struct i915_sw_fence *fence,
  331. struct dma_fence *dma,
  332. unsigned long timeout,
  333. gfp_t gfp)
  334. {
  335. struct i915_sw_dma_fence_cb *cb;
  336. int ret;
  337. debug_fence_assert(fence);
  338. if (dma_fence_is_signaled(dma))
  339. return 0;
  340. cb = kmalloc(sizeof(*cb), gfp);
  341. if (!cb) {
  342. if (!gfpflags_allow_blocking(gfp))
  343. return -ENOMEM;
  344. return dma_fence_wait(dma, false);
  345. }
  346. cb->fence = i915_sw_fence_get(fence);
  347. i915_sw_fence_await(fence);
  348. cb->dma = NULL;
  349. __setup_timer(&cb->timer,
  350. timer_i915_sw_fence_wake, (unsigned long)cb,
  351. TIMER_IRQSAFE);
  352. if (timeout) {
  353. cb->dma = dma_fence_get(dma);
  354. mod_timer(&cb->timer, round_jiffies_up(jiffies + timeout));
  355. }
  356. ret = dma_fence_add_callback(dma, &cb->base, dma_i915_sw_fence_wake);
  357. if (ret == 0) {
  358. ret = 1;
  359. } else {
  360. dma_i915_sw_fence_wake(dma, &cb->base);
  361. if (ret == -ENOENT) /* fence already signaled */
  362. ret = 0;
  363. }
  364. return ret;
  365. }
  366. int i915_sw_fence_await_reservation(struct i915_sw_fence *fence,
  367. struct reservation_object *resv,
  368. const struct dma_fence_ops *exclude,
  369. bool write,
  370. unsigned long timeout,
  371. gfp_t gfp)
  372. {
  373. struct dma_fence *excl;
  374. int ret = 0, pending;
  375. debug_fence_assert(fence);
  376. if (write) {
  377. struct dma_fence **shared;
  378. unsigned int count, i;
  379. ret = reservation_object_get_fences_rcu(resv,
  380. &excl, &count, &shared);
  381. if (ret)
  382. return ret;
  383. for (i = 0; i < count; i++) {
  384. if (shared[i]->ops == exclude)
  385. continue;
  386. pending = i915_sw_fence_await_dma_fence(fence,
  387. shared[i],
  388. timeout,
  389. gfp);
  390. if (pending < 0) {
  391. ret = pending;
  392. break;
  393. }
  394. ret |= pending;
  395. }
  396. for (i = 0; i < count; i++)
  397. dma_fence_put(shared[i]);
  398. kfree(shared);
  399. } else {
  400. excl = reservation_object_get_excl_rcu(resv);
  401. }
  402. if (ret >= 0 && excl && excl->ops != exclude) {
  403. pending = i915_sw_fence_await_dma_fence(fence,
  404. excl,
  405. timeout,
  406. gfp);
  407. if (pending < 0)
  408. ret = pending;
  409. else
  410. ret |= pending;
  411. }
  412. dma_fence_put(excl);
  413. return ret;
  414. }