virtgpu_drv.h 14 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433
  1. /*
  2. * Copyright (C) 2015 Red Hat, Inc.
  3. * All Rights Reserved.
  4. *
  5. * Permission is hereby granted, free of charge, to any person obtaining
  6. * a copy of this software and associated documentation files (the
  7. * "Software"), to deal in the Software without restriction, including
  8. * without limitation the rights to use, copy, modify, merge, publish,
  9. * distribute, sublicense, and/or sell copies of the Software, and to
  10. * permit persons to whom the Software is furnished to do so, subject to
  11. * the following conditions:
  12. *
  13. * The above copyright notice and this permission notice (including the
  14. * next paragraph) shall be included in all copies or substantial
  15. * portions of the Software.
  16. *
  17. * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
  18. * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
  19. * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
  20. * IN NO EVENT SHALL THE COPYRIGHT OWNER(S) AND/OR ITS SUPPLIERS BE
  21. * LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
  22. * OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
  23. * WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
  24. */
  25. #ifndef VIRTIO_DRV_H
  26. #define VIRTIO_DRV_H
  27. #include <linux/virtio.h>
  28. #include <linux/virtio_ids.h>
  29. #include <linux/virtio_config.h>
  30. #include <linux/virtio_gpu.h>
  31. #include <drm/drmP.h>
  32. #include <drm/drm_gem.h>
  33. #include <drm/drm_atomic.h>
  34. #include <drm/drm_crtc_helper.h>
  35. #include <drm/drm_encoder.h>
  36. #include <drm/drm_fb_helper.h>
  37. #include <drm/ttm/ttm_bo_api.h>
  38. #include <drm/ttm/ttm_bo_driver.h>
  39. #include <drm/ttm/ttm_placement.h>
  40. #include <drm/ttm/ttm_module.h>
  41. #define DRIVER_NAME "virtio_gpu"
  42. #define DRIVER_DESC "virtio GPU"
  43. #define DRIVER_DATE "0"
  44. #define DRIVER_MAJOR 0
  45. #define DRIVER_MINOR 0
  46. #define DRIVER_PATCHLEVEL 1
  47. /* virtgpu_drm_bus.c */
  48. int drm_virtio_init(struct drm_driver *driver, struct virtio_device *vdev);
  49. struct virtio_gpu_object {
  50. struct drm_gem_object gem_base;
  51. uint32_t hw_res_handle;
  52. struct sg_table *pages;
  53. uint32_t mapped;
  54. void *vmap;
  55. bool dumb;
  56. struct ttm_place placement_code;
  57. struct ttm_placement placement;
  58. struct ttm_buffer_object tbo;
  59. struct ttm_bo_kmap_obj kmap;
  60. };
  61. #define gem_to_virtio_gpu_obj(gobj) \
  62. container_of((gobj), struct virtio_gpu_object, gem_base)
  63. struct virtio_gpu_vbuffer;
  64. struct virtio_gpu_device;
  65. typedef void (*virtio_gpu_resp_cb)(struct virtio_gpu_device *vgdev,
  66. struct virtio_gpu_vbuffer *vbuf);
  67. struct virtio_gpu_fence_driver {
  68. atomic64_t last_seq;
  69. uint64_t sync_seq;
  70. uint64_t context;
  71. struct list_head fences;
  72. spinlock_t lock;
  73. };
  74. struct virtio_gpu_fence {
  75. struct dma_fence f;
  76. struct virtio_gpu_fence_driver *drv;
  77. struct list_head node;
  78. uint64_t seq;
  79. };
  80. #define to_virtio_fence(x) \
  81. container_of(x, struct virtio_gpu_fence, f)
  82. struct virtio_gpu_vbuffer {
  83. char *buf;
  84. int size;
  85. void *data_buf;
  86. uint32_t data_size;
  87. char *resp_buf;
  88. int resp_size;
  89. virtio_gpu_resp_cb resp_cb;
  90. struct list_head list;
  91. };
  92. struct virtio_gpu_output {
  93. int index;
  94. struct drm_crtc crtc;
  95. struct drm_connector conn;
  96. struct drm_encoder enc;
  97. struct virtio_gpu_display_one info;
  98. struct virtio_gpu_update_cursor cursor;
  99. int cur_x;
  100. int cur_y;
  101. bool enabled;
  102. };
  103. #define drm_crtc_to_virtio_gpu_output(x) \
  104. container_of(x, struct virtio_gpu_output, crtc)
  105. #define drm_connector_to_virtio_gpu_output(x) \
  106. container_of(x, struct virtio_gpu_output, conn)
  107. #define drm_encoder_to_virtio_gpu_output(x) \
  108. container_of(x, struct virtio_gpu_output, enc)
  109. struct virtio_gpu_framebuffer {
  110. struct drm_framebuffer base;
  111. int x1, y1, x2, y2; /* dirty rect */
  112. spinlock_t dirty_lock;
  113. uint32_t hw_res_handle;
  114. };
  115. #define to_virtio_gpu_framebuffer(x) \
  116. container_of(x, struct virtio_gpu_framebuffer, base)
  117. struct virtio_gpu_fbdev {
  118. struct drm_fb_helper helper;
  119. struct virtio_gpu_framebuffer vgfb;
  120. struct virtio_gpu_device *vgdev;
  121. struct delayed_work work;
  122. };
  123. struct virtio_gpu_mman {
  124. struct ttm_bo_global_ref bo_global_ref;
  125. struct drm_global_reference mem_global_ref;
  126. bool mem_global_referenced;
  127. struct ttm_bo_device bdev;
  128. };
  129. struct virtio_gpu_fbdev;
  130. struct virtio_gpu_queue {
  131. struct virtqueue *vq;
  132. spinlock_t qlock;
  133. wait_queue_head_t ack_queue;
  134. struct work_struct dequeue_work;
  135. };
  136. struct virtio_gpu_drv_capset {
  137. uint32_t id;
  138. uint32_t max_version;
  139. uint32_t max_size;
  140. };
  141. struct virtio_gpu_drv_cap_cache {
  142. struct list_head head;
  143. void *caps_cache;
  144. uint32_t id;
  145. uint32_t version;
  146. uint32_t size;
  147. atomic_t is_valid;
  148. };
  149. struct virtio_gpu_device {
  150. struct device *dev;
  151. struct drm_device *ddev;
  152. struct virtio_device *vdev;
  153. struct virtio_gpu_mman mman;
  154. /* pointer to fbdev info structure */
  155. struct virtio_gpu_fbdev *vgfbdev;
  156. struct virtio_gpu_output outputs[VIRTIO_GPU_MAX_SCANOUTS];
  157. uint32_t num_scanouts;
  158. struct virtio_gpu_queue ctrlq;
  159. struct virtio_gpu_queue cursorq;
  160. struct kmem_cache *vbufs;
  161. bool vqs_ready;
  162. struct idr resource_idr;
  163. spinlock_t resource_idr_lock;
  164. wait_queue_head_t resp_wq;
  165. /* current display info */
  166. spinlock_t display_info_lock;
  167. bool display_info_pending;
  168. struct virtio_gpu_fence_driver fence_drv;
  169. struct idr ctx_id_idr;
  170. spinlock_t ctx_id_idr_lock;
  171. bool has_virgl_3d;
  172. struct work_struct config_changed_work;
  173. struct virtio_gpu_drv_capset *capsets;
  174. uint32_t num_capsets;
  175. struct list_head cap_cache;
  176. };
  177. struct virtio_gpu_fpriv {
  178. uint32_t ctx_id;
  179. };
  180. /* virtio_ioctl.c */
  181. #define DRM_VIRTIO_NUM_IOCTLS 10
  182. extern struct drm_ioctl_desc virtio_gpu_ioctls[DRM_VIRTIO_NUM_IOCTLS];
  183. /* virtio_kms.c */
  184. int virtio_gpu_driver_load(struct drm_device *dev, unsigned long flags);
  185. void virtio_gpu_driver_unload(struct drm_device *dev);
  186. int virtio_gpu_driver_open(struct drm_device *dev, struct drm_file *file);
  187. void virtio_gpu_driver_postclose(struct drm_device *dev, struct drm_file *file);
  188. /* virtio_gem.c */
  189. void virtio_gpu_gem_free_object(struct drm_gem_object *gem_obj);
  190. int virtio_gpu_gem_init(struct virtio_gpu_device *vgdev);
  191. void virtio_gpu_gem_fini(struct virtio_gpu_device *vgdev);
  192. int virtio_gpu_gem_create(struct drm_file *file,
  193. struct drm_device *dev,
  194. uint64_t size,
  195. struct drm_gem_object **obj_p,
  196. uint32_t *handle_p);
  197. int virtio_gpu_gem_object_open(struct drm_gem_object *obj,
  198. struct drm_file *file);
  199. void virtio_gpu_gem_object_close(struct drm_gem_object *obj,
  200. struct drm_file *file);
  201. struct virtio_gpu_object *virtio_gpu_alloc_object(struct drm_device *dev,
  202. size_t size, bool kernel,
  203. bool pinned);
  204. int virtio_gpu_mode_dumb_create(struct drm_file *file_priv,
  205. struct drm_device *dev,
  206. struct drm_mode_create_dumb *args);
  207. int virtio_gpu_mode_dumb_mmap(struct drm_file *file_priv,
  208. struct drm_device *dev,
  209. uint32_t handle, uint64_t *offset_p);
  210. /* virtio_fb */
  211. #define VIRTIO_GPUFB_CONN_LIMIT 1
  212. int virtio_gpu_fbdev_init(struct virtio_gpu_device *vgdev);
  213. void virtio_gpu_fbdev_fini(struct virtio_gpu_device *vgdev);
  214. int virtio_gpu_surface_dirty(struct virtio_gpu_framebuffer *qfb,
  215. struct drm_clip_rect *clips,
  216. unsigned int num_clips);
  217. /* virtio vg */
  218. int virtio_gpu_alloc_vbufs(struct virtio_gpu_device *vgdev);
  219. void virtio_gpu_free_vbufs(struct virtio_gpu_device *vgdev);
  220. void virtio_gpu_resource_id_get(struct virtio_gpu_device *vgdev,
  221. uint32_t *resid);
  222. void virtio_gpu_resource_id_put(struct virtio_gpu_device *vgdev, uint32_t id);
  223. void virtio_gpu_cmd_create_resource(struct virtio_gpu_device *vgdev,
  224. uint32_t resource_id,
  225. uint32_t format,
  226. uint32_t width,
  227. uint32_t height);
  228. void virtio_gpu_cmd_unref_resource(struct virtio_gpu_device *vgdev,
  229. uint32_t resource_id);
  230. void virtio_gpu_cmd_transfer_to_host_2d(struct virtio_gpu_device *vgdev,
  231. struct virtio_gpu_object *bo,
  232. uint64_t offset,
  233. __le32 width, __le32 height,
  234. __le32 x, __le32 y,
  235. struct virtio_gpu_fence **fence);
  236. void virtio_gpu_cmd_resource_flush(struct virtio_gpu_device *vgdev,
  237. uint32_t resource_id,
  238. uint32_t x, uint32_t y,
  239. uint32_t width, uint32_t height);
  240. void virtio_gpu_cmd_set_scanout(struct virtio_gpu_device *vgdev,
  241. uint32_t scanout_id, uint32_t resource_id,
  242. uint32_t width, uint32_t height,
  243. uint32_t x, uint32_t y);
  244. int virtio_gpu_object_attach(struct virtio_gpu_device *vgdev,
  245. struct virtio_gpu_object *obj,
  246. uint32_t resource_id,
  247. struct virtio_gpu_fence **fence);
  248. void virtio_gpu_object_detach(struct virtio_gpu_device *vgdev,
  249. struct virtio_gpu_object *obj);
  250. int virtio_gpu_attach_status_page(struct virtio_gpu_device *vgdev);
  251. int virtio_gpu_detach_status_page(struct virtio_gpu_device *vgdev);
  252. void virtio_gpu_cursor_ping(struct virtio_gpu_device *vgdev,
  253. struct virtio_gpu_output *output);
  254. int virtio_gpu_cmd_get_display_info(struct virtio_gpu_device *vgdev);
  255. int virtio_gpu_cmd_get_capset_info(struct virtio_gpu_device *vgdev, int idx);
  256. int virtio_gpu_cmd_get_capset(struct virtio_gpu_device *vgdev,
  257. int idx, int version,
  258. struct virtio_gpu_drv_cap_cache **cache_p);
  259. void virtio_gpu_cmd_context_create(struct virtio_gpu_device *vgdev, uint32_t id,
  260. uint32_t nlen, const char *name);
  261. void virtio_gpu_cmd_context_destroy(struct virtio_gpu_device *vgdev,
  262. uint32_t id);
  263. void virtio_gpu_cmd_context_attach_resource(struct virtio_gpu_device *vgdev,
  264. uint32_t ctx_id,
  265. uint32_t resource_id);
  266. void virtio_gpu_cmd_context_detach_resource(struct virtio_gpu_device *vgdev,
  267. uint32_t ctx_id,
  268. uint32_t resource_id);
  269. void virtio_gpu_cmd_submit(struct virtio_gpu_device *vgdev,
  270. void *data, uint32_t data_size,
  271. uint32_t ctx_id, struct virtio_gpu_fence **fence);
  272. void virtio_gpu_cmd_transfer_from_host_3d(struct virtio_gpu_device *vgdev,
  273. uint32_t resource_id, uint32_t ctx_id,
  274. uint64_t offset, uint32_t level,
  275. struct virtio_gpu_box *box,
  276. struct virtio_gpu_fence **fence);
  277. void virtio_gpu_cmd_transfer_to_host_3d(struct virtio_gpu_device *vgdev,
  278. struct virtio_gpu_object *bo,
  279. uint32_t ctx_id,
  280. uint64_t offset, uint32_t level,
  281. struct virtio_gpu_box *box,
  282. struct virtio_gpu_fence **fence);
  283. void
  284. virtio_gpu_cmd_resource_create_3d(struct virtio_gpu_device *vgdev,
  285. struct virtio_gpu_resource_create_3d *rc_3d,
  286. struct virtio_gpu_fence **fence);
  287. void virtio_gpu_ctrl_ack(struct virtqueue *vq);
  288. void virtio_gpu_cursor_ack(struct virtqueue *vq);
  289. void virtio_gpu_fence_ack(struct virtqueue *vq);
  290. void virtio_gpu_dequeue_ctrl_func(struct work_struct *work);
  291. void virtio_gpu_dequeue_cursor_func(struct work_struct *work);
  292. void virtio_gpu_dequeue_fence_func(struct work_struct *work);
  293. /* virtio_gpu_display.c */
  294. int virtio_gpu_framebuffer_init(struct drm_device *dev,
  295. struct virtio_gpu_framebuffer *vgfb,
  296. const struct drm_mode_fb_cmd2 *mode_cmd,
  297. struct drm_gem_object *obj);
  298. int virtio_gpu_modeset_init(struct virtio_gpu_device *vgdev);
  299. void virtio_gpu_modeset_fini(struct virtio_gpu_device *vgdev);
  300. /* virtio_gpu_plane.c */
  301. uint32_t virtio_gpu_translate_format(uint32_t drm_fourcc);
  302. struct drm_plane *virtio_gpu_plane_init(struct virtio_gpu_device *vgdev,
  303. enum drm_plane_type type,
  304. int index);
  305. /* virtio_gpu_ttm.c */
  306. int virtio_gpu_ttm_init(struct virtio_gpu_device *vgdev);
  307. void virtio_gpu_ttm_fini(struct virtio_gpu_device *vgdev);
  308. int virtio_gpu_mmap(struct file *filp, struct vm_area_struct *vma);
  309. /* virtio_gpu_fence.c */
  310. int virtio_gpu_fence_emit(struct virtio_gpu_device *vgdev,
  311. struct virtio_gpu_ctrl_hdr *cmd_hdr,
  312. struct virtio_gpu_fence **fence);
  313. void virtio_gpu_fence_event_process(struct virtio_gpu_device *vdev,
  314. u64 last_seq);
  315. /* virtio_gpu_object */
  316. int virtio_gpu_object_create(struct virtio_gpu_device *vgdev,
  317. unsigned long size, bool kernel, bool pinned,
  318. struct virtio_gpu_object **bo_ptr);
  319. void virtio_gpu_object_kunmap(struct virtio_gpu_object *bo);
  320. int virtio_gpu_object_kmap(struct virtio_gpu_object *bo);
  321. int virtio_gpu_object_get_sg_table(struct virtio_gpu_device *qdev,
  322. struct virtio_gpu_object *bo);
  323. void virtio_gpu_object_free_sg_table(struct virtio_gpu_object *bo);
  324. int virtio_gpu_object_wait(struct virtio_gpu_object *bo, bool no_wait);
  325. /* virtgpu_prime.c */
  326. int virtgpu_gem_prime_pin(struct drm_gem_object *obj);
  327. void virtgpu_gem_prime_unpin(struct drm_gem_object *obj);
  328. struct sg_table *virtgpu_gem_prime_get_sg_table(struct drm_gem_object *obj);
  329. struct drm_gem_object *virtgpu_gem_prime_import_sg_table(
  330. struct drm_device *dev, struct dma_buf_attachment *attach,
  331. struct sg_table *sgt);
  332. void *virtgpu_gem_prime_vmap(struct drm_gem_object *obj);
  333. void virtgpu_gem_prime_vunmap(struct drm_gem_object *obj, void *vaddr);
  334. int virtgpu_gem_prime_mmap(struct drm_gem_object *obj,
  335. struct vm_area_struct *vma);
  336. static inline struct virtio_gpu_object*
  337. virtio_gpu_object_ref(struct virtio_gpu_object *bo)
  338. {
  339. ttm_bo_get(&bo->tbo);
  340. return bo;
  341. }
  342. static inline void virtio_gpu_object_unref(struct virtio_gpu_object **bo)
  343. {
  344. struct ttm_buffer_object *tbo;
  345. if ((*bo) == NULL)
  346. return;
  347. tbo = &((*bo)->tbo);
  348. ttm_bo_put(tbo);
  349. *bo = NULL;
  350. }
  351. static inline u64 virtio_gpu_object_mmap_offset(struct virtio_gpu_object *bo)
  352. {
  353. return drm_vma_node_offset_addr(&bo->tbo.vma_node);
  354. }
  355. static inline int virtio_gpu_object_reserve(struct virtio_gpu_object *bo,
  356. bool no_wait)
  357. {
  358. int r;
  359. r = ttm_bo_reserve(&bo->tbo, true, no_wait, NULL);
  360. if (unlikely(r != 0)) {
  361. if (r != -ERESTARTSYS) {
  362. struct virtio_gpu_device *qdev =
  363. bo->gem_base.dev->dev_private;
  364. dev_err(qdev->dev, "%p reserve failed\n", bo);
  365. }
  366. return r;
  367. }
  368. return 0;
  369. }
  370. static inline void virtio_gpu_object_unreserve(struct virtio_gpu_object *bo)
  371. {
  372. ttm_bo_unreserve(&bo->tbo);
  373. }
  374. /* virgl debufs */
  375. int virtio_gpu_debugfs_init(struct drm_minor *minor);
  376. #endif