tee_core.c 22 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898
  1. /*
  2. * Copyright (c) 2015-2016, Linaro Limited
  3. *
  4. * This software is licensed under the terms of the GNU General Public
  5. * License version 2, as published by the Free Software Foundation, and
  6. * may be copied, distributed, and modified under those terms.
  7. *
  8. * This program is distributed in the hope that it will be useful,
  9. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  10. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
  11. * GNU General Public License for more details.
  12. *
  13. */
  14. #define pr_fmt(fmt) "%s: " fmt, __func__
  15. #include <linux/cdev.h>
  16. #include <linux/device.h>
  17. #include <linux/fs.h>
  18. #include <linux/idr.h>
  19. #include <linux/module.h>
  20. #include <linux/slab.h>
  21. #include <linux/tee_drv.h>
  22. #include <linux/uaccess.h>
  23. #include "tee_private.h"
  24. #define TEE_NUM_DEVICES 32
  25. #define TEE_IOCTL_PARAM_SIZE(x) (sizeof(struct tee_param) * (x))
  26. /*
  27. * Unprivileged devices in the lower half range and privileged devices in
  28. * the upper half range.
  29. */
  30. static DECLARE_BITMAP(dev_mask, TEE_NUM_DEVICES);
  31. static DEFINE_SPINLOCK(driver_lock);
  32. static struct class *tee_class;
  33. static dev_t tee_devt;
  34. static int tee_open(struct inode *inode, struct file *filp)
  35. {
  36. int rc;
  37. struct tee_device *teedev;
  38. struct tee_context *ctx;
  39. teedev = container_of(inode->i_cdev, struct tee_device, cdev);
  40. if (!tee_device_get(teedev))
  41. return -EINVAL;
  42. ctx = kzalloc(sizeof(*ctx), GFP_KERNEL);
  43. if (!ctx) {
  44. rc = -ENOMEM;
  45. goto err;
  46. }
  47. ctx->teedev = teedev;
  48. INIT_LIST_HEAD(&ctx->list_shm);
  49. filp->private_data = ctx;
  50. rc = teedev->desc->ops->open(ctx);
  51. if (rc)
  52. goto err;
  53. return 0;
  54. err:
  55. kfree(ctx);
  56. tee_device_put(teedev);
  57. return rc;
  58. }
  59. static int tee_release(struct inode *inode, struct file *filp)
  60. {
  61. struct tee_context *ctx = filp->private_data;
  62. struct tee_device *teedev = ctx->teedev;
  63. struct tee_shm *shm;
  64. ctx->teedev->desc->ops->release(ctx);
  65. mutex_lock(&ctx->teedev->mutex);
  66. list_for_each_entry(shm, &ctx->list_shm, link)
  67. shm->ctx = NULL;
  68. mutex_unlock(&ctx->teedev->mutex);
  69. kfree(ctx);
  70. tee_device_put(teedev);
  71. return 0;
  72. }
  73. static int tee_ioctl_version(struct tee_context *ctx,
  74. struct tee_ioctl_version_data __user *uvers)
  75. {
  76. struct tee_ioctl_version_data vers;
  77. ctx->teedev->desc->ops->get_version(ctx->teedev, &vers);
  78. if (ctx->teedev->desc->flags & TEE_DESC_PRIVILEGED)
  79. vers.gen_caps |= TEE_GEN_CAP_PRIVILEGED;
  80. if (copy_to_user(uvers, &vers, sizeof(vers)))
  81. return -EFAULT;
  82. return 0;
  83. }
  84. static int tee_ioctl_shm_alloc(struct tee_context *ctx,
  85. struct tee_ioctl_shm_alloc_data __user *udata)
  86. {
  87. long ret;
  88. struct tee_ioctl_shm_alloc_data data;
  89. struct tee_shm *shm;
  90. if (copy_from_user(&data, udata, sizeof(data)))
  91. return -EFAULT;
  92. /* Currently no input flags are supported */
  93. if (data.flags)
  94. return -EINVAL;
  95. data.id = -1;
  96. shm = tee_shm_alloc(ctx, data.size, TEE_SHM_MAPPED | TEE_SHM_DMA_BUF);
  97. if (IS_ERR(shm))
  98. return PTR_ERR(shm);
  99. data.id = shm->id;
  100. data.flags = shm->flags;
  101. data.size = shm->size;
  102. if (copy_to_user(udata, &data, sizeof(data)))
  103. ret = -EFAULT;
  104. else
  105. ret = tee_shm_get_fd(shm);
  106. /*
  107. * When user space closes the file descriptor the shared memory
  108. * should be freed or if tee_shm_get_fd() failed then it will
  109. * be freed immediately.
  110. */
  111. tee_shm_put(shm);
  112. return ret;
  113. }
  114. static int params_from_user(struct tee_context *ctx, struct tee_param *params,
  115. size_t num_params,
  116. struct tee_ioctl_param __user *uparams)
  117. {
  118. size_t n;
  119. for (n = 0; n < num_params; n++) {
  120. struct tee_shm *shm;
  121. struct tee_ioctl_param ip;
  122. if (copy_from_user(&ip, uparams + n, sizeof(ip)))
  123. return -EFAULT;
  124. /* All unused attribute bits has to be zero */
  125. if (ip.attr & ~TEE_IOCTL_PARAM_ATTR_TYPE_MASK)
  126. return -EINVAL;
  127. params[n].attr = ip.attr;
  128. switch (ip.attr) {
  129. case TEE_IOCTL_PARAM_ATTR_TYPE_NONE:
  130. case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_OUTPUT:
  131. break;
  132. case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INPUT:
  133. case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INOUT:
  134. params[n].u.value.a = ip.a;
  135. params[n].u.value.b = ip.b;
  136. params[n].u.value.c = ip.c;
  137. break;
  138. case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INPUT:
  139. case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_OUTPUT:
  140. case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INOUT:
  141. /*
  142. * If we fail to get a pointer to a shared memory
  143. * object (and increase the ref count) from an
  144. * identifier we return an error. All pointers that
  145. * has been added in params have an increased ref
  146. * count. It's the callers responibility to do
  147. * tee_shm_put() on all resolved pointers.
  148. */
  149. shm = tee_shm_get_from_id(ctx, ip.c);
  150. if (IS_ERR(shm))
  151. return PTR_ERR(shm);
  152. params[n].u.memref.shm_offs = ip.a;
  153. params[n].u.memref.size = ip.b;
  154. params[n].u.memref.shm = shm;
  155. break;
  156. default:
  157. /* Unknown attribute */
  158. return -EINVAL;
  159. }
  160. }
  161. return 0;
  162. }
  163. static int params_to_user(struct tee_ioctl_param __user *uparams,
  164. size_t num_params, struct tee_param *params)
  165. {
  166. size_t n;
  167. for (n = 0; n < num_params; n++) {
  168. struct tee_ioctl_param __user *up = uparams + n;
  169. struct tee_param *p = params + n;
  170. switch (p->attr) {
  171. case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_OUTPUT:
  172. case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INOUT:
  173. if (put_user(p->u.value.a, &up->a) ||
  174. put_user(p->u.value.b, &up->b) ||
  175. put_user(p->u.value.c, &up->c))
  176. return -EFAULT;
  177. break;
  178. case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_OUTPUT:
  179. case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INOUT:
  180. if (put_user((u64)p->u.memref.size, &up->b))
  181. return -EFAULT;
  182. default:
  183. break;
  184. }
  185. }
  186. return 0;
  187. }
  188. static bool param_is_memref(struct tee_param *param)
  189. {
  190. switch (param->attr & TEE_IOCTL_PARAM_ATTR_TYPE_MASK) {
  191. case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INPUT:
  192. case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_OUTPUT:
  193. case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INOUT:
  194. return true;
  195. default:
  196. return false;
  197. }
  198. }
  199. static int tee_ioctl_open_session(struct tee_context *ctx,
  200. struct tee_ioctl_buf_data __user *ubuf)
  201. {
  202. int rc;
  203. size_t n;
  204. struct tee_ioctl_buf_data buf;
  205. struct tee_ioctl_open_session_arg __user *uarg;
  206. struct tee_ioctl_open_session_arg arg;
  207. struct tee_ioctl_param __user *uparams = NULL;
  208. struct tee_param *params = NULL;
  209. bool have_session = false;
  210. if (!ctx->teedev->desc->ops->open_session)
  211. return -EINVAL;
  212. if (copy_from_user(&buf, ubuf, sizeof(buf)))
  213. return -EFAULT;
  214. if (buf.buf_len > TEE_MAX_ARG_SIZE ||
  215. buf.buf_len < sizeof(struct tee_ioctl_open_session_arg))
  216. return -EINVAL;
  217. uarg = u64_to_user_ptr(buf.buf_ptr);
  218. if (copy_from_user(&arg, uarg, sizeof(arg)))
  219. return -EFAULT;
  220. if (sizeof(arg) + TEE_IOCTL_PARAM_SIZE(arg.num_params) != buf.buf_len)
  221. return -EINVAL;
  222. if (arg.num_params) {
  223. params = kcalloc(arg.num_params, sizeof(struct tee_param),
  224. GFP_KERNEL);
  225. if (!params)
  226. return -ENOMEM;
  227. uparams = uarg->params;
  228. rc = params_from_user(ctx, params, arg.num_params, uparams);
  229. if (rc)
  230. goto out;
  231. }
  232. rc = ctx->teedev->desc->ops->open_session(ctx, &arg, params);
  233. if (rc)
  234. goto out;
  235. have_session = true;
  236. if (put_user(arg.session, &uarg->session) ||
  237. put_user(arg.ret, &uarg->ret) ||
  238. put_user(arg.ret_origin, &uarg->ret_origin)) {
  239. rc = -EFAULT;
  240. goto out;
  241. }
  242. rc = params_to_user(uparams, arg.num_params, params);
  243. out:
  244. /*
  245. * If we've succeeded to open the session but failed to communicate
  246. * it back to user space, close the session again to avoid leakage.
  247. */
  248. if (rc && have_session && ctx->teedev->desc->ops->close_session)
  249. ctx->teedev->desc->ops->close_session(ctx, arg.session);
  250. if (params) {
  251. /* Decrease ref count for all valid shared memory pointers */
  252. for (n = 0; n < arg.num_params; n++)
  253. if (param_is_memref(params + n) &&
  254. params[n].u.memref.shm)
  255. tee_shm_put(params[n].u.memref.shm);
  256. kfree(params);
  257. }
  258. return rc;
  259. }
  260. static int tee_ioctl_invoke(struct tee_context *ctx,
  261. struct tee_ioctl_buf_data __user *ubuf)
  262. {
  263. int rc;
  264. size_t n;
  265. struct tee_ioctl_buf_data buf;
  266. struct tee_ioctl_invoke_arg __user *uarg;
  267. struct tee_ioctl_invoke_arg arg;
  268. struct tee_ioctl_param __user *uparams = NULL;
  269. struct tee_param *params = NULL;
  270. if (!ctx->teedev->desc->ops->invoke_func)
  271. return -EINVAL;
  272. if (copy_from_user(&buf, ubuf, sizeof(buf)))
  273. return -EFAULT;
  274. if (buf.buf_len > TEE_MAX_ARG_SIZE ||
  275. buf.buf_len < sizeof(struct tee_ioctl_invoke_arg))
  276. return -EINVAL;
  277. uarg = u64_to_user_ptr(buf.buf_ptr);
  278. if (copy_from_user(&arg, uarg, sizeof(arg)))
  279. return -EFAULT;
  280. if (sizeof(arg) + TEE_IOCTL_PARAM_SIZE(arg.num_params) != buf.buf_len)
  281. return -EINVAL;
  282. if (arg.num_params) {
  283. params = kcalloc(arg.num_params, sizeof(struct tee_param),
  284. GFP_KERNEL);
  285. if (!params)
  286. return -ENOMEM;
  287. uparams = uarg->params;
  288. rc = params_from_user(ctx, params, arg.num_params, uparams);
  289. if (rc)
  290. goto out;
  291. }
  292. rc = ctx->teedev->desc->ops->invoke_func(ctx, &arg, params);
  293. if (rc)
  294. goto out;
  295. if (put_user(arg.ret, &uarg->ret) ||
  296. put_user(arg.ret_origin, &uarg->ret_origin)) {
  297. rc = -EFAULT;
  298. goto out;
  299. }
  300. rc = params_to_user(uparams, arg.num_params, params);
  301. out:
  302. if (params) {
  303. /* Decrease ref count for all valid shared memory pointers */
  304. for (n = 0; n < arg.num_params; n++)
  305. if (param_is_memref(params + n) &&
  306. params[n].u.memref.shm)
  307. tee_shm_put(params[n].u.memref.shm);
  308. kfree(params);
  309. }
  310. return rc;
  311. }
  312. static int tee_ioctl_cancel(struct tee_context *ctx,
  313. struct tee_ioctl_cancel_arg __user *uarg)
  314. {
  315. struct tee_ioctl_cancel_arg arg;
  316. if (!ctx->teedev->desc->ops->cancel_req)
  317. return -EINVAL;
  318. if (copy_from_user(&arg, uarg, sizeof(arg)))
  319. return -EFAULT;
  320. return ctx->teedev->desc->ops->cancel_req(ctx, arg.cancel_id,
  321. arg.session);
  322. }
  323. static int
  324. tee_ioctl_close_session(struct tee_context *ctx,
  325. struct tee_ioctl_close_session_arg __user *uarg)
  326. {
  327. struct tee_ioctl_close_session_arg arg;
  328. if (!ctx->teedev->desc->ops->close_session)
  329. return -EINVAL;
  330. if (copy_from_user(&arg, uarg, sizeof(arg)))
  331. return -EFAULT;
  332. return ctx->teedev->desc->ops->close_session(ctx, arg.session);
  333. }
  334. static int params_to_supp(struct tee_context *ctx,
  335. struct tee_ioctl_param __user *uparams,
  336. size_t num_params, struct tee_param *params)
  337. {
  338. size_t n;
  339. for (n = 0; n < num_params; n++) {
  340. struct tee_ioctl_param ip;
  341. struct tee_param *p = params + n;
  342. ip.attr = p->attr & TEE_IOCTL_PARAM_ATTR_TYPE_MASK;
  343. switch (p->attr) {
  344. case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INPUT:
  345. case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INOUT:
  346. ip.a = p->u.value.a;
  347. ip.b = p->u.value.b;
  348. ip.c = p->u.value.c;
  349. break;
  350. case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INPUT:
  351. case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_OUTPUT:
  352. case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INOUT:
  353. ip.b = p->u.memref.size;
  354. if (!p->u.memref.shm) {
  355. ip.a = 0;
  356. ip.c = (u64)-1; /* invalid shm id */
  357. break;
  358. }
  359. ip.a = p->u.memref.shm_offs;
  360. ip.c = p->u.memref.shm->id;
  361. break;
  362. default:
  363. ip.a = 0;
  364. ip.b = 0;
  365. ip.c = 0;
  366. break;
  367. }
  368. if (copy_to_user(uparams + n, &ip, sizeof(ip)))
  369. return -EFAULT;
  370. }
  371. return 0;
  372. }
  373. static int tee_ioctl_supp_recv(struct tee_context *ctx,
  374. struct tee_ioctl_buf_data __user *ubuf)
  375. {
  376. int rc;
  377. struct tee_ioctl_buf_data buf;
  378. struct tee_iocl_supp_recv_arg __user *uarg;
  379. struct tee_param *params;
  380. u32 num_params;
  381. u32 func;
  382. if (!ctx->teedev->desc->ops->supp_recv)
  383. return -EINVAL;
  384. if (copy_from_user(&buf, ubuf, sizeof(buf)))
  385. return -EFAULT;
  386. if (buf.buf_len > TEE_MAX_ARG_SIZE ||
  387. buf.buf_len < sizeof(struct tee_iocl_supp_recv_arg))
  388. return -EINVAL;
  389. uarg = u64_to_user_ptr(buf.buf_ptr);
  390. if (get_user(num_params, &uarg->num_params))
  391. return -EFAULT;
  392. if (sizeof(*uarg) + TEE_IOCTL_PARAM_SIZE(num_params) != buf.buf_len)
  393. return -EINVAL;
  394. params = kcalloc(num_params, sizeof(struct tee_param), GFP_KERNEL);
  395. if (!params)
  396. return -ENOMEM;
  397. rc = ctx->teedev->desc->ops->supp_recv(ctx, &func, &num_params, params);
  398. if (rc)
  399. goto out;
  400. if (put_user(func, &uarg->func) ||
  401. put_user(num_params, &uarg->num_params)) {
  402. rc = -EFAULT;
  403. goto out;
  404. }
  405. rc = params_to_supp(ctx, uarg->params, num_params, params);
  406. out:
  407. kfree(params);
  408. return rc;
  409. }
  410. static int params_from_supp(struct tee_param *params, size_t num_params,
  411. struct tee_ioctl_param __user *uparams)
  412. {
  413. size_t n;
  414. for (n = 0; n < num_params; n++) {
  415. struct tee_param *p = params + n;
  416. struct tee_ioctl_param ip;
  417. if (copy_from_user(&ip, uparams + n, sizeof(ip)))
  418. return -EFAULT;
  419. /* All unused attribute bits has to be zero */
  420. if (ip.attr & ~TEE_IOCTL_PARAM_ATTR_TYPE_MASK)
  421. return -EINVAL;
  422. p->attr = ip.attr;
  423. switch (ip.attr) {
  424. case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_OUTPUT:
  425. case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INOUT:
  426. /* Only out and in/out values can be updated */
  427. p->u.value.a = ip.a;
  428. p->u.value.b = ip.b;
  429. p->u.value.c = ip.c;
  430. break;
  431. case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_OUTPUT:
  432. case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INOUT:
  433. /*
  434. * Only the size of the memref can be updated.
  435. * Since we don't have access to the original
  436. * parameters here, only store the supplied size.
  437. * The driver will copy the updated size into the
  438. * original parameters.
  439. */
  440. p->u.memref.shm = NULL;
  441. p->u.memref.shm_offs = 0;
  442. p->u.memref.size = ip.b;
  443. break;
  444. default:
  445. memset(&p->u, 0, sizeof(p->u));
  446. break;
  447. }
  448. }
  449. return 0;
  450. }
  451. static int tee_ioctl_supp_send(struct tee_context *ctx,
  452. struct tee_ioctl_buf_data __user *ubuf)
  453. {
  454. long rc;
  455. struct tee_ioctl_buf_data buf;
  456. struct tee_iocl_supp_send_arg __user *uarg;
  457. struct tee_param *params;
  458. u32 num_params;
  459. u32 ret;
  460. /* Not valid for this driver */
  461. if (!ctx->teedev->desc->ops->supp_send)
  462. return -EINVAL;
  463. if (copy_from_user(&buf, ubuf, sizeof(buf)))
  464. return -EFAULT;
  465. if (buf.buf_len > TEE_MAX_ARG_SIZE ||
  466. buf.buf_len < sizeof(struct tee_iocl_supp_send_arg))
  467. return -EINVAL;
  468. uarg = u64_to_user_ptr(buf.buf_ptr);
  469. if (get_user(ret, &uarg->ret) ||
  470. get_user(num_params, &uarg->num_params))
  471. return -EFAULT;
  472. if (sizeof(*uarg) + TEE_IOCTL_PARAM_SIZE(num_params) > buf.buf_len)
  473. return -EINVAL;
  474. params = kcalloc(num_params, sizeof(struct tee_param), GFP_KERNEL);
  475. if (!params)
  476. return -ENOMEM;
  477. rc = params_from_supp(params, num_params, uarg->params);
  478. if (rc)
  479. goto out;
  480. rc = ctx->teedev->desc->ops->supp_send(ctx, ret, num_params, params);
  481. out:
  482. kfree(params);
  483. return rc;
  484. }
  485. static long tee_ioctl(struct file *filp, unsigned int cmd, unsigned long arg)
  486. {
  487. struct tee_context *ctx = filp->private_data;
  488. void __user *uarg = (void __user *)arg;
  489. switch (cmd) {
  490. case TEE_IOC_VERSION:
  491. return tee_ioctl_version(ctx, uarg);
  492. case TEE_IOC_SHM_ALLOC:
  493. return tee_ioctl_shm_alloc(ctx, uarg);
  494. case TEE_IOC_OPEN_SESSION:
  495. return tee_ioctl_open_session(ctx, uarg);
  496. case TEE_IOC_INVOKE:
  497. return tee_ioctl_invoke(ctx, uarg);
  498. case TEE_IOC_CANCEL:
  499. return tee_ioctl_cancel(ctx, uarg);
  500. case TEE_IOC_CLOSE_SESSION:
  501. return tee_ioctl_close_session(ctx, uarg);
  502. case TEE_IOC_SUPPL_RECV:
  503. return tee_ioctl_supp_recv(ctx, uarg);
  504. case TEE_IOC_SUPPL_SEND:
  505. return tee_ioctl_supp_send(ctx, uarg);
  506. default:
  507. return -EINVAL;
  508. }
  509. }
  510. static const struct file_operations tee_fops = {
  511. .owner = THIS_MODULE,
  512. .open = tee_open,
  513. .release = tee_release,
  514. .unlocked_ioctl = tee_ioctl,
  515. .compat_ioctl = tee_ioctl,
  516. };
  517. static void tee_release_device(struct device *dev)
  518. {
  519. struct tee_device *teedev = container_of(dev, struct tee_device, dev);
  520. spin_lock(&driver_lock);
  521. clear_bit(teedev->id, dev_mask);
  522. spin_unlock(&driver_lock);
  523. mutex_destroy(&teedev->mutex);
  524. idr_destroy(&teedev->idr);
  525. kfree(teedev);
  526. }
  527. /**
  528. * tee_device_alloc() - Allocate a new struct tee_device instance
  529. * @teedesc: Descriptor for this driver
  530. * @dev: Parent device for this device
  531. * @pool: Shared memory pool, NULL if not used
  532. * @driver_data: Private driver data for this device
  533. *
  534. * Allocates a new struct tee_device instance. The device is
  535. * removed by tee_device_unregister().
  536. *
  537. * @returns a pointer to a 'struct tee_device' or an ERR_PTR on failure
  538. */
  539. struct tee_device *tee_device_alloc(const struct tee_desc *teedesc,
  540. struct device *dev,
  541. struct tee_shm_pool *pool,
  542. void *driver_data)
  543. {
  544. struct tee_device *teedev;
  545. void *ret;
  546. int rc;
  547. int offs = 0;
  548. if (!teedesc || !teedesc->name || !teedesc->ops ||
  549. !teedesc->ops->get_version || !teedesc->ops->open ||
  550. !teedesc->ops->release || !pool)
  551. return ERR_PTR(-EINVAL);
  552. teedev = kzalloc(sizeof(*teedev), GFP_KERNEL);
  553. if (!teedev) {
  554. ret = ERR_PTR(-ENOMEM);
  555. goto err;
  556. }
  557. if (teedesc->flags & TEE_DESC_PRIVILEGED)
  558. offs = TEE_NUM_DEVICES / 2;
  559. spin_lock(&driver_lock);
  560. teedev->id = find_next_zero_bit(dev_mask, TEE_NUM_DEVICES, offs);
  561. if (teedev->id < TEE_NUM_DEVICES)
  562. set_bit(teedev->id, dev_mask);
  563. spin_unlock(&driver_lock);
  564. if (teedev->id >= TEE_NUM_DEVICES) {
  565. ret = ERR_PTR(-ENOMEM);
  566. goto err;
  567. }
  568. snprintf(teedev->name, sizeof(teedev->name), "tee%s%d",
  569. teedesc->flags & TEE_DESC_PRIVILEGED ? "priv" : "",
  570. teedev->id - offs);
  571. teedev->dev.class = tee_class;
  572. teedev->dev.release = tee_release_device;
  573. teedev->dev.parent = dev;
  574. teedev->dev.devt = MKDEV(MAJOR(tee_devt), teedev->id);
  575. rc = dev_set_name(&teedev->dev, "%s", teedev->name);
  576. if (rc) {
  577. ret = ERR_PTR(rc);
  578. goto err_devt;
  579. }
  580. cdev_init(&teedev->cdev, &tee_fops);
  581. teedev->cdev.owner = teedesc->owner;
  582. teedev->cdev.kobj.parent = &teedev->dev.kobj;
  583. dev_set_drvdata(&teedev->dev, driver_data);
  584. device_initialize(&teedev->dev);
  585. /* 1 as tee_device_unregister() does one final tee_device_put() */
  586. teedev->num_users = 1;
  587. init_completion(&teedev->c_no_users);
  588. mutex_init(&teedev->mutex);
  589. idr_init(&teedev->idr);
  590. teedev->desc = teedesc;
  591. teedev->pool = pool;
  592. return teedev;
  593. err_devt:
  594. unregister_chrdev_region(teedev->dev.devt, 1);
  595. err:
  596. pr_err("could not register %s driver\n",
  597. teedesc->flags & TEE_DESC_PRIVILEGED ? "privileged" : "client");
  598. if (teedev && teedev->id < TEE_NUM_DEVICES) {
  599. spin_lock(&driver_lock);
  600. clear_bit(teedev->id, dev_mask);
  601. spin_unlock(&driver_lock);
  602. }
  603. kfree(teedev);
  604. return ret;
  605. }
  606. EXPORT_SYMBOL_GPL(tee_device_alloc);
  607. static ssize_t implementation_id_show(struct device *dev,
  608. struct device_attribute *attr, char *buf)
  609. {
  610. struct tee_device *teedev = container_of(dev, struct tee_device, dev);
  611. struct tee_ioctl_version_data vers;
  612. teedev->desc->ops->get_version(teedev, &vers);
  613. return scnprintf(buf, PAGE_SIZE, "%d\n", vers.impl_id);
  614. }
  615. static DEVICE_ATTR_RO(implementation_id);
  616. static struct attribute *tee_dev_attrs[] = {
  617. &dev_attr_implementation_id.attr,
  618. NULL
  619. };
  620. static const struct attribute_group tee_dev_group = {
  621. .attrs = tee_dev_attrs,
  622. };
  623. /**
  624. * tee_device_register() - Registers a TEE device
  625. * @teedev: Device to register
  626. *
  627. * tee_device_unregister() need to be called to remove the @teedev if
  628. * this function fails.
  629. *
  630. * @returns < 0 on failure
  631. */
  632. int tee_device_register(struct tee_device *teedev)
  633. {
  634. int rc;
  635. if (teedev->flags & TEE_DEVICE_FLAG_REGISTERED) {
  636. dev_err(&teedev->dev, "attempt to register twice\n");
  637. return -EINVAL;
  638. }
  639. rc = cdev_add(&teedev->cdev, teedev->dev.devt, 1);
  640. if (rc) {
  641. dev_err(&teedev->dev,
  642. "unable to cdev_add() %s, major %d, minor %d, err=%d\n",
  643. teedev->name, MAJOR(teedev->dev.devt),
  644. MINOR(teedev->dev.devt), rc);
  645. return rc;
  646. }
  647. rc = device_add(&teedev->dev);
  648. if (rc) {
  649. dev_err(&teedev->dev,
  650. "unable to device_add() %s, major %d, minor %d, err=%d\n",
  651. teedev->name, MAJOR(teedev->dev.devt),
  652. MINOR(teedev->dev.devt), rc);
  653. goto err_device_add;
  654. }
  655. rc = sysfs_create_group(&teedev->dev.kobj, &tee_dev_group);
  656. if (rc) {
  657. dev_err(&teedev->dev,
  658. "failed to create sysfs attributes, err=%d\n", rc);
  659. goto err_sysfs_create_group;
  660. }
  661. teedev->flags |= TEE_DEVICE_FLAG_REGISTERED;
  662. return 0;
  663. err_sysfs_create_group:
  664. device_del(&teedev->dev);
  665. err_device_add:
  666. cdev_del(&teedev->cdev);
  667. return rc;
  668. }
  669. EXPORT_SYMBOL_GPL(tee_device_register);
  670. void tee_device_put(struct tee_device *teedev)
  671. {
  672. mutex_lock(&teedev->mutex);
  673. /* Shouldn't put in this state */
  674. if (!WARN_ON(!teedev->desc)) {
  675. teedev->num_users--;
  676. if (!teedev->num_users) {
  677. teedev->desc = NULL;
  678. complete(&teedev->c_no_users);
  679. }
  680. }
  681. mutex_unlock(&teedev->mutex);
  682. }
  683. bool tee_device_get(struct tee_device *teedev)
  684. {
  685. mutex_lock(&teedev->mutex);
  686. if (!teedev->desc) {
  687. mutex_unlock(&teedev->mutex);
  688. return false;
  689. }
  690. teedev->num_users++;
  691. mutex_unlock(&teedev->mutex);
  692. return true;
  693. }
  694. /**
  695. * tee_device_unregister() - Removes a TEE device
  696. * @teedev: Device to unregister
  697. *
  698. * This function should be called to remove the @teedev even if
  699. * tee_device_register() hasn't been called yet. Does nothing if
  700. * @teedev is NULL.
  701. */
  702. void tee_device_unregister(struct tee_device *teedev)
  703. {
  704. if (!teedev)
  705. return;
  706. if (teedev->flags & TEE_DEVICE_FLAG_REGISTERED) {
  707. sysfs_remove_group(&teedev->dev.kobj, &tee_dev_group);
  708. cdev_del(&teedev->cdev);
  709. device_del(&teedev->dev);
  710. }
  711. tee_device_put(teedev);
  712. wait_for_completion(&teedev->c_no_users);
  713. /*
  714. * No need to take a mutex any longer now since teedev->desc was
  715. * set to NULL before teedev->c_no_users was completed.
  716. */
  717. teedev->pool = NULL;
  718. put_device(&teedev->dev);
  719. }
  720. EXPORT_SYMBOL_GPL(tee_device_unregister);
  721. /**
  722. * tee_get_drvdata() - Return driver_data pointer
  723. * @teedev: Device containing the driver_data pointer
  724. * @returns the driver_data pointer supplied to tee_register().
  725. */
  726. void *tee_get_drvdata(struct tee_device *teedev)
  727. {
  728. return dev_get_drvdata(&teedev->dev);
  729. }
  730. EXPORT_SYMBOL_GPL(tee_get_drvdata);
  731. static int __init tee_init(void)
  732. {
  733. int rc;
  734. tee_class = class_create(THIS_MODULE, "tee");
  735. if (IS_ERR(tee_class)) {
  736. pr_err("couldn't create class\n");
  737. return PTR_ERR(tee_class);
  738. }
  739. rc = alloc_chrdev_region(&tee_devt, 0, TEE_NUM_DEVICES, "tee");
  740. if (rc) {
  741. pr_err("failed to allocate char dev region\n");
  742. class_destroy(tee_class);
  743. tee_class = NULL;
  744. }
  745. return rc;
  746. }
  747. static void __exit tee_exit(void)
  748. {
  749. class_destroy(tee_class);
  750. tee_class = NULL;
  751. unregister_chrdev_region(tee_devt, TEE_NUM_DEVICES);
  752. }
  753. subsys_initcall(tee_init);
  754. module_exit(tee_exit);
  755. MODULE_AUTHOR("Linaro");
  756. MODULE_DESCRIPTION("TEE Driver");
  757. MODULE_VERSION("1.0");
  758. MODULE_LICENSE("GPL v2");