vgem_drv.c 12 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517
  1. /*
  2. * Copyright 2011 Red Hat, Inc.
  3. * Copyright © 2014 The Chromium OS Authors
  4. *
  5. * Permission is hereby granted, free of charge, to any person obtaining a
  6. * copy of this software and associated documentation files (the "Software")
  7. * to deal in the software without restriction, including without limitation
  8. * on the rights to use, copy, modify, merge, publish, distribute, sub
  9. * license, and/or sell copies of the Software, and to permit persons to whom
  10. * them Software is furnished to do so, subject to the following conditions:
  11. *
  12. * The above copyright notice and this permission notice (including the next
  13. * paragraph) shall be included in all copies or substantial portions of the
  14. * Software.
  15. *
  16. * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
  17. * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTIBILITY,
  18. * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
  19. * THE AUTHORS BE LIABLE FOR ANY CLAIM, DAMAGES, OR OTHER LIABILITY, WHETHER
  20. * IN AN ACTION OF CONTRACT, TORT, OR OTHERWISE, ARISING FROM, OUT OF OR IN
  21. * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
  22. *
  23. * Authors:
  24. * Adam Jackson <ajax@redhat.com>
  25. * Ben Widawsky <ben@bwidawsk.net>
  26. */
  27. /**
  28. * This is vgem, a (non-hardware-backed) GEM service. This is used by Mesa's
  29. * software renderer and the X server for efficient buffer sharing.
  30. */
  31. #include <linux/module.h>
  32. #include <linux/ramfs.h>
  33. #include <linux/shmem_fs.h>
  34. #include <linux/dma-buf.h>
  35. #include "vgem_drv.h"
  36. #define DRIVER_NAME "vgem"
  37. #define DRIVER_DESC "Virtual GEM provider"
  38. #define DRIVER_DATE "20120112"
  39. #define DRIVER_MAJOR 1
  40. #define DRIVER_MINOR 0
  41. static struct vgem_device {
  42. struct drm_device drm;
  43. struct platform_device *platform;
  44. } *vgem_device;
  45. static void vgem_gem_free_object(struct drm_gem_object *obj)
  46. {
  47. struct drm_vgem_gem_object *vgem_obj = to_vgem_bo(obj);
  48. kvfree(vgem_obj->pages);
  49. mutex_destroy(&vgem_obj->pages_lock);
  50. if (obj->import_attach)
  51. drm_prime_gem_destroy(obj, vgem_obj->table);
  52. drm_gem_object_release(obj);
  53. kfree(vgem_obj);
  54. }
  55. static int vgem_gem_fault(struct vm_fault *vmf)
  56. {
  57. struct vm_area_struct *vma = vmf->vma;
  58. struct drm_vgem_gem_object *obj = vma->vm_private_data;
  59. /* We don't use vmf->pgoff since that has the fake offset */
  60. unsigned long vaddr = vmf->address;
  61. int ret;
  62. loff_t num_pages;
  63. pgoff_t page_offset;
  64. page_offset = (vaddr - vma->vm_start) >> PAGE_SHIFT;
  65. num_pages = DIV_ROUND_UP(obj->base.size, PAGE_SIZE);
  66. if (page_offset > num_pages)
  67. return VM_FAULT_SIGBUS;
  68. ret = -ENOENT;
  69. mutex_lock(&obj->pages_lock);
  70. if (obj->pages) {
  71. get_page(obj->pages[page_offset]);
  72. vmf->page = obj->pages[page_offset];
  73. ret = 0;
  74. }
  75. mutex_unlock(&obj->pages_lock);
  76. if (ret) {
  77. struct page *page;
  78. page = shmem_read_mapping_page(
  79. file_inode(obj->base.filp)->i_mapping,
  80. page_offset);
  81. if (!IS_ERR(page)) {
  82. vmf->page = page;
  83. ret = 0;
  84. } else switch (PTR_ERR(page)) {
  85. case -ENOSPC:
  86. case -ENOMEM:
  87. ret = VM_FAULT_OOM;
  88. break;
  89. case -EBUSY:
  90. ret = VM_FAULT_RETRY;
  91. break;
  92. case -EFAULT:
  93. case -EINVAL:
  94. ret = VM_FAULT_SIGBUS;
  95. break;
  96. default:
  97. WARN_ON(PTR_ERR(page));
  98. ret = VM_FAULT_SIGBUS;
  99. break;
  100. }
  101. }
  102. return ret;
  103. }
  104. static const struct vm_operations_struct vgem_gem_vm_ops = {
  105. .fault = vgem_gem_fault,
  106. .open = drm_gem_vm_open,
  107. .close = drm_gem_vm_close,
  108. };
  109. static int vgem_open(struct drm_device *dev, struct drm_file *file)
  110. {
  111. struct vgem_file *vfile;
  112. int ret;
  113. vfile = kzalloc(sizeof(*vfile), GFP_KERNEL);
  114. if (!vfile)
  115. return -ENOMEM;
  116. file->driver_priv = vfile;
  117. ret = vgem_fence_open(vfile);
  118. if (ret) {
  119. kfree(vfile);
  120. return ret;
  121. }
  122. return 0;
  123. }
  124. static void vgem_postclose(struct drm_device *dev, struct drm_file *file)
  125. {
  126. struct vgem_file *vfile = file->driver_priv;
  127. vgem_fence_close(vfile);
  128. kfree(vfile);
  129. }
  130. static struct drm_vgem_gem_object *__vgem_gem_create(struct drm_device *dev,
  131. unsigned long size)
  132. {
  133. struct drm_vgem_gem_object *obj;
  134. int ret;
  135. obj = kzalloc(sizeof(*obj), GFP_KERNEL);
  136. if (!obj)
  137. return ERR_PTR(-ENOMEM);
  138. ret = drm_gem_object_init(dev, &obj->base, roundup(size, PAGE_SIZE));
  139. if (ret) {
  140. kfree(obj);
  141. return ERR_PTR(ret);
  142. }
  143. mutex_init(&obj->pages_lock);
  144. return obj;
  145. }
  146. static void __vgem_gem_destroy(struct drm_vgem_gem_object *obj)
  147. {
  148. drm_gem_object_release(&obj->base);
  149. kfree(obj);
  150. }
  151. static struct drm_gem_object *vgem_gem_create(struct drm_device *dev,
  152. struct drm_file *file,
  153. unsigned int *handle,
  154. unsigned long size)
  155. {
  156. struct drm_vgem_gem_object *obj;
  157. int ret;
  158. obj = __vgem_gem_create(dev, size);
  159. if (IS_ERR(obj))
  160. return ERR_CAST(obj);
  161. ret = drm_gem_handle_create(file, &obj->base, handle);
  162. drm_gem_object_put_unlocked(&obj->base);
  163. if (ret)
  164. goto err;
  165. return &obj->base;
  166. err:
  167. __vgem_gem_destroy(obj);
  168. return ERR_PTR(ret);
  169. }
  170. static int vgem_gem_dumb_create(struct drm_file *file, struct drm_device *dev,
  171. struct drm_mode_create_dumb *args)
  172. {
  173. struct drm_gem_object *gem_object;
  174. u64 pitch, size;
  175. pitch = args->width * DIV_ROUND_UP(args->bpp, 8);
  176. size = args->height * pitch;
  177. if (size == 0)
  178. return -EINVAL;
  179. gem_object = vgem_gem_create(dev, file, &args->handle, size);
  180. if (IS_ERR(gem_object))
  181. return PTR_ERR(gem_object);
  182. args->size = gem_object->size;
  183. args->pitch = pitch;
  184. DRM_DEBUG_DRIVER("Created object of size %lld\n", size);
  185. return 0;
  186. }
  187. static int vgem_gem_dumb_map(struct drm_file *file, struct drm_device *dev,
  188. uint32_t handle, uint64_t *offset)
  189. {
  190. struct drm_gem_object *obj;
  191. int ret;
  192. obj = drm_gem_object_lookup(file, handle);
  193. if (!obj)
  194. return -ENOENT;
  195. if (!obj->filp) {
  196. ret = -EINVAL;
  197. goto unref;
  198. }
  199. ret = drm_gem_create_mmap_offset(obj);
  200. if (ret)
  201. goto unref;
  202. *offset = drm_vma_node_offset_addr(&obj->vma_node);
  203. unref:
  204. drm_gem_object_put_unlocked(obj);
  205. return ret;
  206. }
  207. static struct drm_ioctl_desc vgem_ioctls[] = {
  208. DRM_IOCTL_DEF_DRV(VGEM_FENCE_ATTACH, vgem_fence_attach_ioctl, DRM_AUTH|DRM_RENDER_ALLOW),
  209. DRM_IOCTL_DEF_DRV(VGEM_FENCE_SIGNAL, vgem_fence_signal_ioctl, DRM_AUTH|DRM_RENDER_ALLOW),
  210. };
  211. static int vgem_mmap(struct file *filp, struct vm_area_struct *vma)
  212. {
  213. unsigned long flags = vma->vm_flags;
  214. int ret;
  215. ret = drm_gem_mmap(filp, vma);
  216. if (ret)
  217. return ret;
  218. /* Keep the WC mmaping set by drm_gem_mmap() but our pages
  219. * are ordinary and not special.
  220. */
  221. vma->vm_flags = flags | VM_DONTEXPAND | VM_DONTDUMP;
  222. return 0;
  223. }
  224. static const struct file_operations vgem_driver_fops = {
  225. .owner = THIS_MODULE,
  226. .open = drm_open,
  227. .mmap = vgem_mmap,
  228. .poll = drm_poll,
  229. .read = drm_read,
  230. .unlocked_ioctl = drm_ioctl,
  231. .compat_ioctl = drm_compat_ioctl,
  232. .release = drm_release,
  233. };
  234. static struct page **vgem_pin_pages(struct drm_vgem_gem_object *bo)
  235. {
  236. mutex_lock(&bo->pages_lock);
  237. if (bo->pages_pin_count++ == 0) {
  238. struct page **pages;
  239. pages = drm_gem_get_pages(&bo->base);
  240. if (IS_ERR(pages)) {
  241. bo->pages_pin_count--;
  242. mutex_unlock(&bo->pages_lock);
  243. return pages;
  244. }
  245. bo->pages = pages;
  246. }
  247. mutex_unlock(&bo->pages_lock);
  248. return bo->pages;
  249. }
  250. static void vgem_unpin_pages(struct drm_vgem_gem_object *bo)
  251. {
  252. mutex_lock(&bo->pages_lock);
  253. if (--bo->pages_pin_count == 0) {
  254. drm_gem_put_pages(&bo->base, bo->pages, true, true);
  255. bo->pages = NULL;
  256. }
  257. mutex_unlock(&bo->pages_lock);
  258. }
  259. static int vgem_prime_pin(struct drm_gem_object *obj)
  260. {
  261. struct drm_vgem_gem_object *bo = to_vgem_bo(obj);
  262. long n_pages = obj->size >> PAGE_SHIFT;
  263. struct page **pages;
  264. pages = vgem_pin_pages(bo);
  265. if (IS_ERR(pages))
  266. return PTR_ERR(pages);
  267. /* Flush the object from the CPU cache so that importers can rely
  268. * on coherent indirect access via the exported dma-address.
  269. */
  270. drm_clflush_pages(pages, n_pages);
  271. return 0;
  272. }
  273. static void vgem_prime_unpin(struct drm_gem_object *obj)
  274. {
  275. struct drm_vgem_gem_object *bo = to_vgem_bo(obj);
  276. vgem_unpin_pages(bo);
  277. }
  278. static struct sg_table *vgem_prime_get_sg_table(struct drm_gem_object *obj)
  279. {
  280. struct drm_vgem_gem_object *bo = to_vgem_bo(obj);
  281. return drm_prime_pages_to_sg(bo->pages, bo->base.size >> PAGE_SHIFT);
  282. }
  283. static struct drm_gem_object* vgem_prime_import(struct drm_device *dev,
  284. struct dma_buf *dma_buf)
  285. {
  286. struct vgem_device *vgem = container_of(dev, typeof(*vgem), drm);
  287. return drm_gem_prime_import_dev(dev, dma_buf, &vgem->platform->dev);
  288. }
  289. static struct drm_gem_object *vgem_prime_import_sg_table(struct drm_device *dev,
  290. struct dma_buf_attachment *attach, struct sg_table *sg)
  291. {
  292. struct drm_vgem_gem_object *obj;
  293. int npages;
  294. obj = __vgem_gem_create(dev, attach->dmabuf->size);
  295. if (IS_ERR(obj))
  296. return ERR_CAST(obj);
  297. npages = PAGE_ALIGN(attach->dmabuf->size) / PAGE_SIZE;
  298. obj->table = sg;
  299. obj->pages = kvmalloc_array(npages, sizeof(struct page *), GFP_KERNEL);
  300. if (!obj->pages) {
  301. __vgem_gem_destroy(obj);
  302. return ERR_PTR(-ENOMEM);
  303. }
  304. obj->pages_pin_count++; /* perma-pinned */
  305. drm_prime_sg_to_page_addr_arrays(obj->table, obj->pages, NULL,
  306. npages);
  307. return &obj->base;
  308. }
  309. static void *vgem_prime_vmap(struct drm_gem_object *obj)
  310. {
  311. struct drm_vgem_gem_object *bo = to_vgem_bo(obj);
  312. long n_pages = obj->size >> PAGE_SHIFT;
  313. struct page **pages;
  314. pages = vgem_pin_pages(bo);
  315. if (IS_ERR(pages))
  316. return NULL;
  317. return vmap(pages, n_pages, 0, pgprot_writecombine(PAGE_KERNEL));
  318. }
  319. static void vgem_prime_vunmap(struct drm_gem_object *obj, void *vaddr)
  320. {
  321. struct drm_vgem_gem_object *bo = to_vgem_bo(obj);
  322. vunmap(vaddr);
  323. vgem_unpin_pages(bo);
  324. }
  325. static int vgem_prime_mmap(struct drm_gem_object *obj,
  326. struct vm_area_struct *vma)
  327. {
  328. int ret;
  329. if (obj->size < vma->vm_end - vma->vm_start)
  330. return -EINVAL;
  331. if (!obj->filp)
  332. return -ENODEV;
  333. ret = call_mmap(obj->filp, vma);
  334. if (ret)
  335. return ret;
  336. fput(vma->vm_file);
  337. vma->vm_file = get_file(obj->filp);
  338. vma->vm_flags |= VM_DONTEXPAND | VM_DONTDUMP;
  339. vma->vm_page_prot = pgprot_writecombine(vm_get_page_prot(vma->vm_flags));
  340. return 0;
  341. }
  342. static void vgem_release(struct drm_device *dev)
  343. {
  344. struct vgem_device *vgem = container_of(dev, typeof(*vgem), drm);
  345. platform_device_unregister(vgem->platform);
  346. drm_dev_fini(&vgem->drm);
  347. kfree(vgem);
  348. }
  349. static struct drm_driver vgem_driver = {
  350. .driver_features = DRIVER_GEM | DRIVER_PRIME,
  351. .release = vgem_release,
  352. .open = vgem_open,
  353. .postclose = vgem_postclose,
  354. .gem_free_object_unlocked = vgem_gem_free_object,
  355. .gem_vm_ops = &vgem_gem_vm_ops,
  356. .ioctls = vgem_ioctls,
  357. .num_ioctls = ARRAY_SIZE(vgem_ioctls),
  358. .fops = &vgem_driver_fops,
  359. .dumb_create = vgem_gem_dumb_create,
  360. .dumb_map_offset = vgem_gem_dumb_map,
  361. .prime_handle_to_fd = drm_gem_prime_handle_to_fd,
  362. .prime_fd_to_handle = drm_gem_prime_fd_to_handle,
  363. .gem_prime_pin = vgem_prime_pin,
  364. .gem_prime_unpin = vgem_prime_unpin,
  365. .gem_prime_import = vgem_prime_import,
  366. .gem_prime_export = drm_gem_prime_export,
  367. .gem_prime_import_sg_table = vgem_prime_import_sg_table,
  368. .gem_prime_get_sg_table = vgem_prime_get_sg_table,
  369. .gem_prime_vmap = vgem_prime_vmap,
  370. .gem_prime_vunmap = vgem_prime_vunmap,
  371. .gem_prime_mmap = vgem_prime_mmap,
  372. .name = DRIVER_NAME,
  373. .desc = DRIVER_DESC,
  374. .date = DRIVER_DATE,
  375. .major = DRIVER_MAJOR,
  376. .minor = DRIVER_MINOR,
  377. };
  378. static int __init vgem_init(void)
  379. {
  380. int ret;
  381. vgem_device = kzalloc(sizeof(*vgem_device), GFP_KERNEL);
  382. if (!vgem_device)
  383. return -ENOMEM;
  384. ret = drm_dev_init(&vgem_device->drm, &vgem_driver, NULL);
  385. if (ret)
  386. goto out_free;
  387. vgem_device->platform =
  388. platform_device_register_simple("vgem", -1, NULL, 0);
  389. if (IS_ERR(vgem_device->platform)) {
  390. ret = PTR_ERR(vgem_device->platform);
  391. goto out_fini;
  392. }
  393. dma_coerce_mask_and_coherent(&vgem_device->platform->dev,
  394. DMA_BIT_MASK(64));
  395. /* Final step: expose the device/driver to userspace */
  396. ret = drm_dev_register(&vgem_device->drm, 0);
  397. if (ret)
  398. goto out_unregister;
  399. return 0;
  400. out_unregister:
  401. platform_device_unregister(vgem_device->platform);
  402. out_fini:
  403. drm_dev_fini(&vgem_device->drm);
  404. out_free:
  405. kfree(vgem_device);
  406. return ret;
  407. }
  408. static void __exit vgem_exit(void)
  409. {
  410. drm_dev_unregister(&vgem_device->drm);
  411. drm_dev_unref(&vgem_device->drm);
  412. }
  413. module_init(vgem_init);
  414. module_exit(vgem_exit);
  415. MODULE_AUTHOR("Red Hat, Inc.");
  416. MODULE_AUTHOR("Intel Corporation");
  417. MODULE_DESCRIPTION(DRIVER_DESC);
  418. MODULE_LICENSE("GPL and additional rights");