Loading...
1// SPDX-License-Identifier: GPL-2.0-only
2/*
3 * Copyright (c) 2015-2016, Linaro Limited
4 */
5
6#define pr_fmt(fmt) "%s: " fmt, __func__
7
8#include <linux/cdev.h>
9#include <linux/cred.h>
10#include <linux/fs.h>
11#include <linux/idr.h>
12#include <linux/module.h>
13#include <linux/slab.h>
14#include <linux/tee_drv.h>
15#include <linux/uaccess.h>
16#include <crypto/hash.h>
17#include <crypto/sha.h>
18#include "tee_private.h"
19
20#define TEE_NUM_DEVICES 32
21
22#define TEE_IOCTL_PARAM_SIZE(x) (sizeof(struct tee_param) * (x))
23
24#define TEE_UUID_NS_NAME_SIZE 128
25
26/*
27 * TEE Client UUID name space identifier (UUIDv4)
28 *
29 * Value here is random UUID that is allocated as name space identifier for
30 * forming Client UUID's for TEE environment using UUIDv5 scheme.
31 */
32static const uuid_t tee_client_uuid_ns = UUID_INIT(0x58ac9ca0, 0x2086, 0x4683,
33 0xa1, 0xb8, 0xec, 0x4b,
34 0xc0, 0x8e, 0x01, 0xb6);
35
36/*
37 * Unprivileged devices in the lower half range and privileged devices in
38 * the upper half range.
39 */
40static DECLARE_BITMAP(dev_mask, TEE_NUM_DEVICES);
41static DEFINE_SPINLOCK(driver_lock);
42
43static struct class *tee_class;
44static dev_t tee_devt;
45
46static struct tee_context *teedev_open(struct tee_device *teedev)
47{
48 int rc;
49 struct tee_context *ctx;
50
51 if (!tee_device_get(teedev))
52 return ERR_PTR(-EINVAL);
53
54 ctx = kzalloc(sizeof(*ctx), GFP_KERNEL);
55 if (!ctx) {
56 rc = -ENOMEM;
57 goto err;
58 }
59
60 kref_init(&ctx->refcount);
61 ctx->teedev = teedev;
62 rc = teedev->desc->ops->open(ctx);
63 if (rc)
64 goto err;
65
66 return ctx;
67err:
68 kfree(ctx);
69 tee_device_put(teedev);
70 return ERR_PTR(rc);
71
72}
73
74void teedev_ctx_get(struct tee_context *ctx)
75{
76 if (ctx->releasing)
77 return;
78
79 kref_get(&ctx->refcount);
80}
81
82static void teedev_ctx_release(struct kref *ref)
83{
84 struct tee_context *ctx = container_of(ref, struct tee_context,
85 refcount);
86 ctx->releasing = true;
87 ctx->teedev->desc->ops->release(ctx);
88 kfree(ctx);
89}
90
91void teedev_ctx_put(struct tee_context *ctx)
92{
93 if (ctx->releasing)
94 return;
95
96 kref_put(&ctx->refcount, teedev_ctx_release);
97}
98
99static void teedev_close_context(struct tee_context *ctx)
100{
101 tee_device_put(ctx->teedev);
102 teedev_ctx_put(ctx);
103}
104
105static int tee_open(struct inode *inode, struct file *filp)
106{
107 struct tee_context *ctx;
108
109 ctx = teedev_open(container_of(inode->i_cdev, struct tee_device, cdev));
110 if (IS_ERR(ctx))
111 return PTR_ERR(ctx);
112
113 /*
114 * Default user-space behaviour is to wait for tee-supplicant
115 * if not present for any requests in this context.
116 */
117 ctx->supp_nowait = false;
118 filp->private_data = ctx;
119 return 0;
120}
121
122static int tee_release(struct inode *inode, struct file *filp)
123{
124 teedev_close_context(filp->private_data);
125 return 0;
126}
127
128/**
129 * uuid_v5() - Calculate UUIDv5
130 * @uuid: Resulting UUID
131 * @ns: Name space ID for UUIDv5 function
132 * @name: Name for UUIDv5 function
133 * @size: Size of name
134 *
135 * UUIDv5 is specific in RFC 4122.
136 *
137 * This implements section (for SHA-1):
138 * 4.3. Algorithm for Creating a Name-Based UUID
139 */
140static int uuid_v5(uuid_t *uuid, const uuid_t *ns, const void *name,
141 size_t size)
142{
143 unsigned char hash[SHA1_DIGEST_SIZE];
144 struct crypto_shash *shash = NULL;
145 struct shash_desc *desc = NULL;
146 int rc;
147
148 shash = crypto_alloc_shash("sha1", 0, 0);
149 if (IS_ERR(shash)) {
150 rc = PTR_ERR(shash);
151 pr_err("shash(sha1) allocation failed\n");
152 return rc;
153 }
154
155 desc = kzalloc(sizeof(*desc) + crypto_shash_descsize(shash),
156 GFP_KERNEL);
157 if (!desc) {
158 rc = -ENOMEM;
159 goto out_free_shash;
160 }
161
162 desc->tfm = shash;
163
164 rc = crypto_shash_init(desc);
165 if (rc < 0)
166 goto out_free_desc;
167
168 rc = crypto_shash_update(desc, (const u8 *)ns, sizeof(*ns));
169 if (rc < 0)
170 goto out_free_desc;
171
172 rc = crypto_shash_update(desc, (const u8 *)name, size);
173 if (rc < 0)
174 goto out_free_desc;
175
176 rc = crypto_shash_final(desc, hash);
177 if (rc < 0)
178 goto out_free_desc;
179
180 memcpy(uuid->b, hash, UUID_SIZE);
181
182 /* Tag for version 5 */
183 uuid->b[6] = (hash[6] & 0x0F) | 0x50;
184 uuid->b[8] = (hash[8] & 0x3F) | 0x80;
185
186out_free_desc:
187 kfree(desc);
188
189out_free_shash:
190 crypto_free_shash(shash);
191 return rc;
192}
193
194int tee_session_calc_client_uuid(uuid_t *uuid, u32 connection_method,
195 const u8 connection_data[TEE_IOCTL_UUID_LEN])
196{
197 gid_t ns_grp = (gid_t)-1;
198 kgid_t grp = INVALID_GID;
199 char *name = NULL;
200 int name_len;
201 int rc;
202
203 if (connection_method == TEE_IOCTL_LOGIN_PUBLIC) {
204 /* Nil UUID to be passed to TEE environment */
205 uuid_copy(uuid, &uuid_null);
206 return 0;
207 }
208
209 /*
210 * In Linux environment client UUID is based on UUIDv5.
211 *
212 * Determine client UUID with following semantics for 'name':
213 *
214 * For TEEC_LOGIN_USER:
215 * uid=<uid>
216 *
217 * For TEEC_LOGIN_GROUP:
218 * gid=<gid>
219 *
220 */
221
222 name = kzalloc(TEE_UUID_NS_NAME_SIZE, GFP_KERNEL);
223 if (!name)
224 return -ENOMEM;
225
226 switch (connection_method) {
227 case TEE_IOCTL_LOGIN_USER:
228 name_len = snprintf(name, TEE_UUID_NS_NAME_SIZE, "uid=%x",
229 current_euid().val);
230 if (name_len >= TEE_UUID_NS_NAME_SIZE) {
231 rc = -E2BIG;
232 goto out_free_name;
233 }
234 break;
235
236 case TEE_IOCTL_LOGIN_GROUP:
237 memcpy(&ns_grp, connection_data, sizeof(gid_t));
238 grp = make_kgid(current_user_ns(), ns_grp);
239 if (!gid_valid(grp) || !in_egroup_p(grp)) {
240 rc = -EPERM;
241 goto out_free_name;
242 }
243
244 name_len = snprintf(name, TEE_UUID_NS_NAME_SIZE, "gid=%x",
245 grp.val);
246 if (name_len >= TEE_UUID_NS_NAME_SIZE) {
247 rc = -E2BIG;
248 goto out_free_name;
249 }
250 break;
251
252 default:
253 rc = -EINVAL;
254 goto out_free_name;
255 }
256
257 rc = uuid_v5(uuid, &tee_client_uuid_ns, name, name_len);
258out_free_name:
259 kfree(name);
260
261 return rc;
262}
263EXPORT_SYMBOL_GPL(tee_session_calc_client_uuid);
264
265static int tee_ioctl_version(struct tee_context *ctx,
266 struct tee_ioctl_version_data __user *uvers)
267{
268 struct tee_ioctl_version_data vers;
269
270 ctx->teedev->desc->ops->get_version(ctx->teedev, &vers);
271
272 if (ctx->teedev->desc->flags & TEE_DESC_PRIVILEGED)
273 vers.gen_caps |= TEE_GEN_CAP_PRIVILEGED;
274
275 if (copy_to_user(uvers, &vers, sizeof(vers)))
276 return -EFAULT;
277
278 return 0;
279}
280
281static int tee_ioctl_shm_alloc(struct tee_context *ctx,
282 struct tee_ioctl_shm_alloc_data __user *udata)
283{
284 long ret;
285 struct tee_ioctl_shm_alloc_data data;
286 struct tee_shm *shm;
287
288 if (copy_from_user(&data, udata, sizeof(data)))
289 return -EFAULT;
290
291 /* Currently no input flags are supported */
292 if (data.flags)
293 return -EINVAL;
294
295 shm = tee_shm_alloc(ctx, data.size, TEE_SHM_MAPPED | TEE_SHM_DMA_BUF);
296 if (IS_ERR(shm))
297 return PTR_ERR(shm);
298
299 data.id = shm->id;
300 data.flags = shm->flags;
301 data.size = shm->size;
302
303 if (copy_to_user(udata, &data, sizeof(data)))
304 ret = -EFAULT;
305 else
306 ret = tee_shm_get_fd(shm);
307
308 /*
309 * When user space closes the file descriptor the shared memory
310 * should be freed or if tee_shm_get_fd() failed then it will
311 * be freed immediately.
312 */
313 tee_shm_put(shm);
314 return ret;
315}
316
317static int
318tee_ioctl_shm_register(struct tee_context *ctx,
319 struct tee_ioctl_shm_register_data __user *udata)
320{
321 long ret;
322 struct tee_ioctl_shm_register_data data;
323 struct tee_shm *shm;
324
325 if (copy_from_user(&data, udata, sizeof(data)))
326 return -EFAULT;
327
328 /* Currently no input flags are supported */
329 if (data.flags)
330 return -EINVAL;
331
332 shm = tee_shm_register(ctx, data.addr, data.length,
333 TEE_SHM_DMA_BUF | TEE_SHM_USER_MAPPED);
334 if (IS_ERR(shm))
335 return PTR_ERR(shm);
336
337 data.id = shm->id;
338 data.flags = shm->flags;
339 data.length = shm->size;
340
341 if (copy_to_user(udata, &data, sizeof(data)))
342 ret = -EFAULT;
343 else
344 ret = tee_shm_get_fd(shm);
345 /*
346 * When user space closes the file descriptor the shared memory
347 * should be freed or if tee_shm_get_fd() failed then it will
348 * be freed immediately.
349 */
350 tee_shm_put(shm);
351 return ret;
352}
353
354static int params_from_user(struct tee_context *ctx, struct tee_param *params,
355 size_t num_params,
356 struct tee_ioctl_param __user *uparams)
357{
358 size_t n;
359
360 for (n = 0; n < num_params; n++) {
361 struct tee_shm *shm;
362 struct tee_ioctl_param ip;
363
364 if (copy_from_user(&ip, uparams + n, sizeof(ip)))
365 return -EFAULT;
366
367 /* All unused attribute bits has to be zero */
368 if (ip.attr & ~TEE_IOCTL_PARAM_ATTR_MASK)
369 return -EINVAL;
370
371 params[n].attr = ip.attr;
372 switch (ip.attr & TEE_IOCTL_PARAM_ATTR_TYPE_MASK) {
373 case TEE_IOCTL_PARAM_ATTR_TYPE_NONE:
374 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_OUTPUT:
375 break;
376 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INPUT:
377 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INOUT:
378 params[n].u.value.a = ip.a;
379 params[n].u.value.b = ip.b;
380 params[n].u.value.c = ip.c;
381 break;
382 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INPUT:
383 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_OUTPUT:
384 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INOUT:
385 /*
386 * If we fail to get a pointer to a shared memory
387 * object (and increase the ref count) from an
388 * identifier we return an error. All pointers that
389 * has been added in params have an increased ref
390 * count. It's the callers responibility to do
391 * tee_shm_put() on all resolved pointers.
392 */
393 shm = tee_shm_get_from_id(ctx, ip.c);
394 if (IS_ERR(shm))
395 return PTR_ERR(shm);
396
397 /*
398 * Ensure offset + size does not overflow offset
399 * and does not overflow the size of the referred
400 * shared memory object.
401 */
402 if ((ip.a + ip.b) < ip.a ||
403 (ip.a + ip.b) > shm->size) {
404 tee_shm_put(shm);
405 return -EINVAL;
406 }
407
408 params[n].u.memref.shm_offs = ip.a;
409 params[n].u.memref.size = ip.b;
410 params[n].u.memref.shm = shm;
411 break;
412 default:
413 /* Unknown attribute */
414 return -EINVAL;
415 }
416 }
417 return 0;
418}
419
420static int params_to_user(struct tee_ioctl_param __user *uparams,
421 size_t num_params, struct tee_param *params)
422{
423 size_t n;
424
425 for (n = 0; n < num_params; n++) {
426 struct tee_ioctl_param __user *up = uparams + n;
427 struct tee_param *p = params + n;
428
429 switch (p->attr) {
430 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_OUTPUT:
431 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INOUT:
432 if (put_user(p->u.value.a, &up->a) ||
433 put_user(p->u.value.b, &up->b) ||
434 put_user(p->u.value.c, &up->c))
435 return -EFAULT;
436 break;
437 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_OUTPUT:
438 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INOUT:
439 if (put_user((u64)p->u.memref.size, &up->b))
440 return -EFAULT;
441 default:
442 break;
443 }
444 }
445 return 0;
446}
447
448static int tee_ioctl_open_session(struct tee_context *ctx,
449 struct tee_ioctl_buf_data __user *ubuf)
450{
451 int rc;
452 size_t n;
453 struct tee_ioctl_buf_data buf;
454 struct tee_ioctl_open_session_arg __user *uarg;
455 struct tee_ioctl_open_session_arg arg;
456 struct tee_ioctl_param __user *uparams = NULL;
457 struct tee_param *params = NULL;
458 bool have_session = false;
459
460 if (!ctx->teedev->desc->ops->open_session)
461 return -EINVAL;
462
463 if (copy_from_user(&buf, ubuf, sizeof(buf)))
464 return -EFAULT;
465
466 if (buf.buf_len > TEE_MAX_ARG_SIZE ||
467 buf.buf_len < sizeof(struct tee_ioctl_open_session_arg))
468 return -EINVAL;
469
470 uarg = u64_to_user_ptr(buf.buf_ptr);
471 if (copy_from_user(&arg, uarg, sizeof(arg)))
472 return -EFAULT;
473
474 if (sizeof(arg) + TEE_IOCTL_PARAM_SIZE(arg.num_params) != buf.buf_len)
475 return -EINVAL;
476
477 if (arg.num_params) {
478 params = kcalloc(arg.num_params, sizeof(struct tee_param),
479 GFP_KERNEL);
480 if (!params)
481 return -ENOMEM;
482 uparams = uarg->params;
483 rc = params_from_user(ctx, params, arg.num_params, uparams);
484 if (rc)
485 goto out;
486 }
487
488 if (arg.clnt_login >= TEE_IOCTL_LOGIN_REE_KERNEL_MIN &&
489 arg.clnt_login <= TEE_IOCTL_LOGIN_REE_KERNEL_MAX) {
490 pr_debug("login method not allowed for user-space client\n");
491 rc = -EPERM;
492 goto out;
493 }
494
495 rc = ctx->teedev->desc->ops->open_session(ctx, &arg, params);
496 if (rc)
497 goto out;
498 have_session = true;
499
500 if (put_user(arg.session, &uarg->session) ||
501 put_user(arg.ret, &uarg->ret) ||
502 put_user(arg.ret_origin, &uarg->ret_origin)) {
503 rc = -EFAULT;
504 goto out;
505 }
506 rc = params_to_user(uparams, arg.num_params, params);
507out:
508 /*
509 * If we've succeeded to open the session but failed to communicate
510 * it back to user space, close the session again to avoid leakage.
511 */
512 if (rc && have_session && ctx->teedev->desc->ops->close_session)
513 ctx->teedev->desc->ops->close_session(ctx, arg.session);
514
515 if (params) {
516 /* Decrease ref count for all valid shared memory pointers */
517 for (n = 0; n < arg.num_params; n++)
518 if (tee_param_is_memref(params + n) &&
519 params[n].u.memref.shm)
520 tee_shm_put(params[n].u.memref.shm);
521 kfree(params);
522 }
523
524 return rc;
525}
526
527static int tee_ioctl_invoke(struct tee_context *ctx,
528 struct tee_ioctl_buf_data __user *ubuf)
529{
530 int rc;
531 size_t n;
532 struct tee_ioctl_buf_data buf;
533 struct tee_ioctl_invoke_arg __user *uarg;
534 struct tee_ioctl_invoke_arg arg;
535 struct tee_ioctl_param __user *uparams = NULL;
536 struct tee_param *params = NULL;
537
538 if (!ctx->teedev->desc->ops->invoke_func)
539 return -EINVAL;
540
541 if (copy_from_user(&buf, ubuf, sizeof(buf)))
542 return -EFAULT;
543
544 if (buf.buf_len > TEE_MAX_ARG_SIZE ||
545 buf.buf_len < sizeof(struct tee_ioctl_invoke_arg))
546 return -EINVAL;
547
548 uarg = u64_to_user_ptr(buf.buf_ptr);
549 if (copy_from_user(&arg, uarg, sizeof(arg)))
550 return -EFAULT;
551
552 if (sizeof(arg) + TEE_IOCTL_PARAM_SIZE(arg.num_params) != buf.buf_len)
553 return -EINVAL;
554
555 if (arg.num_params) {
556 params = kcalloc(arg.num_params, sizeof(struct tee_param),
557 GFP_KERNEL);
558 if (!params)
559 return -ENOMEM;
560 uparams = uarg->params;
561 rc = params_from_user(ctx, params, arg.num_params, uparams);
562 if (rc)
563 goto out;
564 }
565
566 rc = ctx->teedev->desc->ops->invoke_func(ctx, &arg, params);
567 if (rc)
568 goto out;
569
570 if (put_user(arg.ret, &uarg->ret) ||
571 put_user(arg.ret_origin, &uarg->ret_origin)) {
572 rc = -EFAULT;
573 goto out;
574 }
575 rc = params_to_user(uparams, arg.num_params, params);
576out:
577 if (params) {
578 /* Decrease ref count for all valid shared memory pointers */
579 for (n = 0; n < arg.num_params; n++)
580 if (tee_param_is_memref(params + n) &&
581 params[n].u.memref.shm)
582 tee_shm_put(params[n].u.memref.shm);
583 kfree(params);
584 }
585 return rc;
586}
587
588static int tee_ioctl_cancel(struct tee_context *ctx,
589 struct tee_ioctl_cancel_arg __user *uarg)
590{
591 struct tee_ioctl_cancel_arg arg;
592
593 if (!ctx->teedev->desc->ops->cancel_req)
594 return -EINVAL;
595
596 if (copy_from_user(&arg, uarg, sizeof(arg)))
597 return -EFAULT;
598
599 return ctx->teedev->desc->ops->cancel_req(ctx, arg.cancel_id,
600 arg.session);
601}
602
603static int
604tee_ioctl_close_session(struct tee_context *ctx,
605 struct tee_ioctl_close_session_arg __user *uarg)
606{
607 struct tee_ioctl_close_session_arg arg;
608
609 if (!ctx->teedev->desc->ops->close_session)
610 return -EINVAL;
611
612 if (copy_from_user(&arg, uarg, sizeof(arg)))
613 return -EFAULT;
614
615 return ctx->teedev->desc->ops->close_session(ctx, arg.session);
616}
617
618static int params_to_supp(struct tee_context *ctx,
619 struct tee_ioctl_param __user *uparams,
620 size_t num_params, struct tee_param *params)
621{
622 size_t n;
623
624 for (n = 0; n < num_params; n++) {
625 struct tee_ioctl_param ip;
626 struct tee_param *p = params + n;
627
628 ip.attr = p->attr;
629 switch (p->attr & TEE_IOCTL_PARAM_ATTR_TYPE_MASK) {
630 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INPUT:
631 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INOUT:
632 ip.a = p->u.value.a;
633 ip.b = p->u.value.b;
634 ip.c = p->u.value.c;
635 break;
636 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INPUT:
637 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_OUTPUT:
638 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INOUT:
639 ip.b = p->u.memref.size;
640 if (!p->u.memref.shm) {
641 ip.a = 0;
642 ip.c = (u64)-1; /* invalid shm id */
643 break;
644 }
645 ip.a = p->u.memref.shm_offs;
646 ip.c = p->u.memref.shm->id;
647 break;
648 default:
649 ip.a = 0;
650 ip.b = 0;
651 ip.c = 0;
652 break;
653 }
654
655 if (copy_to_user(uparams + n, &ip, sizeof(ip)))
656 return -EFAULT;
657 }
658
659 return 0;
660}
661
662static int tee_ioctl_supp_recv(struct tee_context *ctx,
663 struct tee_ioctl_buf_data __user *ubuf)
664{
665 int rc;
666 struct tee_ioctl_buf_data buf;
667 struct tee_iocl_supp_recv_arg __user *uarg;
668 struct tee_param *params;
669 u32 num_params;
670 u32 func;
671
672 if (!ctx->teedev->desc->ops->supp_recv)
673 return -EINVAL;
674
675 if (copy_from_user(&buf, ubuf, sizeof(buf)))
676 return -EFAULT;
677
678 if (buf.buf_len > TEE_MAX_ARG_SIZE ||
679 buf.buf_len < sizeof(struct tee_iocl_supp_recv_arg))
680 return -EINVAL;
681
682 uarg = u64_to_user_ptr(buf.buf_ptr);
683 if (get_user(num_params, &uarg->num_params))
684 return -EFAULT;
685
686 if (sizeof(*uarg) + TEE_IOCTL_PARAM_SIZE(num_params) != buf.buf_len)
687 return -EINVAL;
688
689 params = kcalloc(num_params, sizeof(struct tee_param), GFP_KERNEL);
690 if (!params)
691 return -ENOMEM;
692
693 rc = params_from_user(ctx, params, num_params, uarg->params);
694 if (rc)
695 goto out;
696
697 rc = ctx->teedev->desc->ops->supp_recv(ctx, &func, &num_params, params);
698 if (rc)
699 goto out;
700
701 if (put_user(func, &uarg->func) ||
702 put_user(num_params, &uarg->num_params)) {
703 rc = -EFAULT;
704 goto out;
705 }
706
707 rc = params_to_supp(ctx, uarg->params, num_params, params);
708out:
709 kfree(params);
710 return rc;
711}
712
713static int params_from_supp(struct tee_param *params, size_t num_params,
714 struct tee_ioctl_param __user *uparams)
715{
716 size_t n;
717
718 for (n = 0; n < num_params; n++) {
719 struct tee_param *p = params + n;
720 struct tee_ioctl_param ip;
721
722 if (copy_from_user(&ip, uparams + n, sizeof(ip)))
723 return -EFAULT;
724
725 /* All unused attribute bits has to be zero */
726 if (ip.attr & ~TEE_IOCTL_PARAM_ATTR_MASK)
727 return -EINVAL;
728
729 p->attr = ip.attr;
730 switch (ip.attr & TEE_IOCTL_PARAM_ATTR_TYPE_MASK) {
731 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_OUTPUT:
732 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INOUT:
733 /* Only out and in/out values can be updated */
734 p->u.value.a = ip.a;
735 p->u.value.b = ip.b;
736 p->u.value.c = ip.c;
737 break;
738 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_OUTPUT:
739 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INOUT:
740 /*
741 * Only the size of the memref can be updated.
742 * Since we don't have access to the original
743 * parameters here, only store the supplied size.
744 * The driver will copy the updated size into the
745 * original parameters.
746 */
747 p->u.memref.shm = NULL;
748 p->u.memref.shm_offs = 0;
749 p->u.memref.size = ip.b;
750 break;
751 default:
752 memset(&p->u, 0, sizeof(p->u));
753 break;
754 }
755 }
756 return 0;
757}
758
759static int tee_ioctl_supp_send(struct tee_context *ctx,
760 struct tee_ioctl_buf_data __user *ubuf)
761{
762 long rc;
763 struct tee_ioctl_buf_data buf;
764 struct tee_iocl_supp_send_arg __user *uarg;
765 struct tee_param *params;
766 u32 num_params;
767 u32 ret;
768
769 /* Not valid for this driver */
770 if (!ctx->teedev->desc->ops->supp_send)
771 return -EINVAL;
772
773 if (copy_from_user(&buf, ubuf, sizeof(buf)))
774 return -EFAULT;
775
776 if (buf.buf_len > TEE_MAX_ARG_SIZE ||
777 buf.buf_len < sizeof(struct tee_iocl_supp_send_arg))
778 return -EINVAL;
779
780 uarg = u64_to_user_ptr(buf.buf_ptr);
781 if (get_user(ret, &uarg->ret) ||
782 get_user(num_params, &uarg->num_params))
783 return -EFAULT;
784
785 if (sizeof(*uarg) + TEE_IOCTL_PARAM_SIZE(num_params) > buf.buf_len)
786 return -EINVAL;
787
788 params = kcalloc(num_params, sizeof(struct tee_param), GFP_KERNEL);
789 if (!params)
790 return -ENOMEM;
791
792 rc = params_from_supp(params, num_params, uarg->params);
793 if (rc)
794 goto out;
795
796 rc = ctx->teedev->desc->ops->supp_send(ctx, ret, num_params, params);
797out:
798 kfree(params);
799 return rc;
800}
801
802static long tee_ioctl(struct file *filp, unsigned int cmd, unsigned long arg)
803{
804 struct tee_context *ctx = filp->private_data;
805 void __user *uarg = (void __user *)arg;
806
807 switch (cmd) {
808 case TEE_IOC_VERSION:
809 return tee_ioctl_version(ctx, uarg);
810 case TEE_IOC_SHM_ALLOC:
811 return tee_ioctl_shm_alloc(ctx, uarg);
812 case TEE_IOC_SHM_REGISTER:
813 return tee_ioctl_shm_register(ctx, uarg);
814 case TEE_IOC_OPEN_SESSION:
815 return tee_ioctl_open_session(ctx, uarg);
816 case TEE_IOC_INVOKE:
817 return tee_ioctl_invoke(ctx, uarg);
818 case TEE_IOC_CANCEL:
819 return tee_ioctl_cancel(ctx, uarg);
820 case TEE_IOC_CLOSE_SESSION:
821 return tee_ioctl_close_session(ctx, uarg);
822 case TEE_IOC_SUPPL_RECV:
823 return tee_ioctl_supp_recv(ctx, uarg);
824 case TEE_IOC_SUPPL_SEND:
825 return tee_ioctl_supp_send(ctx, uarg);
826 default:
827 return -EINVAL;
828 }
829}
830
831static const struct file_operations tee_fops = {
832 .owner = THIS_MODULE,
833 .open = tee_open,
834 .release = tee_release,
835 .unlocked_ioctl = tee_ioctl,
836 .compat_ioctl = compat_ptr_ioctl,
837};
838
839static void tee_release_device(struct device *dev)
840{
841 struct tee_device *teedev = container_of(dev, struct tee_device, dev);
842
843 spin_lock(&driver_lock);
844 clear_bit(teedev->id, dev_mask);
845 spin_unlock(&driver_lock);
846 mutex_destroy(&teedev->mutex);
847 idr_destroy(&teedev->idr);
848 kfree(teedev);
849}
850
851/**
852 * tee_device_alloc() - Allocate a new struct tee_device instance
853 * @teedesc: Descriptor for this driver
854 * @dev: Parent device for this device
855 * @pool: Shared memory pool, NULL if not used
856 * @driver_data: Private driver data for this device
857 *
858 * Allocates a new struct tee_device instance. The device is
859 * removed by tee_device_unregister().
860 *
861 * @returns a pointer to a 'struct tee_device' or an ERR_PTR on failure
862 */
863struct tee_device *tee_device_alloc(const struct tee_desc *teedesc,
864 struct device *dev,
865 struct tee_shm_pool *pool,
866 void *driver_data)
867{
868 struct tee_device *teedev;
869 void *ret;
870 int rc, max_id;
871 int offs = 0;
872
873 if (!teedesc || !teedesc->name || !teedesc->ops ||
874 !teedesc->ops->get_version || !teedesc->ops->open ||
875 !teedesc->ops->release || !pool)
876 return ERR_PTR(-EINVAL);
877
878 teedev = kzalloc(sizeof(*teedev), GFP_KERNEL);
879 if (!teedev) {
880 ret = ERR_PTR(-ENOMEM);
881 goto err;
882 }
883
884 max_id = TEE_NUM_DEVICES / 2;
885
886 if (teedesc->flags & TEE_DESC_PRIVILEGED) {
887 offs = TEE_NUM_DEVICES / 2;
888 max_id = TEE_NUM_DEVICES;
889 }
890
891 spin_lock(&driver_lock);
892 teedev->id = find_next_zero_bit(dev_mask, max_id, offs);
893 if (teedev->id < max_id)
894 set_bit(teedev->id, dev_mask);
895 spin_unlock(&driver_lock);
896
897 if (teedev->id >= max_id) {
898 ret = ERR_PTR(-ENOMEM);
899 goto err;
900 }
901
902 snprintf(teedev->name, sizeof(teedev->name), "tee%s%d",
903 teedesc->flags & TEE_DESC_PRIVILEGED ? "priv" : "",
904 teedev->id - offs);
905
906 teedev->dev.class = tee_class;
907 teedev->dev.release = tee_release_device;
908 teedev->dev.parent = dev;
909
910 teedev->dev.devt = MKDEV(MAJOR(tee_devt), teedev->id);
911
912 rc = dev_set_name(&teedev->dev, "%s", teedev->name);
913 if (rc) {
914 ret = ERR_PTR(rc);
915 goto err_devt;
916 }
917
918 cdev_init(&teedev->cdev, &tee_fops);
919 teedev->cdev.owner = teedesc->owner;
920 teedev->cdev.kobj.parent = &teedev->dev.kobj;
921
922 dev_set_drvdata(&teedev->dev, driver_data);
923 device_initialize(&teedev->dev);
924
925 /* 1 as tee_device_unregister() does one final tee_device_put() */
926 teedev->num_users = 1;
927 init_completion(&teedev->c_no_users);
928 mutex_init(&teedev->mutex);
929 idr_init(&teedev->idr);
930
931 teedev->desc = teedesc;
932 teedev->pool = pool;
933
934 return teedev;
935err_devt:
936 unregister_chrdev_region(teedev->dev.devt, 1);
937err:
938 pr_err("could not register %s driver\n",
939 teedesc->flags & TEE_DESC_PRIVILEGED ? "privileged" : "client");
940 if (teedev && teedev->id < TEE_NUM_DEVICES) {
941 spin_lock(&driver_lock);
942 clear_bit(teedev->id, dev_mask);
943 spin_unlock(&driver_lock);
944 }
945 kfree(teedev);
946 return ret;
947}
948EXPORT_SYMBOL_GPL(tee_device_alloc);
949
950static ssize_t implementation_id_show(struct device *dev,
951 struct device_attribute *attr, char *buf)
952{
953 struct tee_device *teedev = container_of(dev, struct tee_device, dev);
954 struct tee_ioctl_version_data vers;
955
956 teedev->desc->ops->get_version(teedev, &vers);
957 return scnprintf(buf, PAGE_SIZE, "%d\n", vers.impl_id);
958}
959static DEVICE_ATTR_RO(implementation_id);
960
961static struct attribute *tee_dev_attrs[] = {
962 &dev_attr_implementation_id.attr,
963 NULL
964};
965
966static const struct attribute_group tee_dev_group = {
967 .attrs = tee_dev_attrs,
968};
969
970/**
971 * tee_device_register() - Registers a TEE device
972 * @teedev: Device to register
973 *
974 * tee_device_unregister() need to be called to remove the @teedev if
975 * this function fails.
976 *
977 * @returns < 0 on failure
978 */
979int tee_device_register(struct tee_device *teedev)
980{
981 int rc;
982
983 if (teedev->flags & TEE_DEVICE_FLAG_REGISTERED) {
984 dev_err(&teedev->dev, "attempt to register twice\n");
985 return -EINVAL;
986 }
987
988 rc = cdev_add(&teedev->cdev, teedev->dev.devt, 1);
989 if (rc) {
990 dev_err(&teedev->dev,
991 "unable to cdev_add() %s, major %d, minor %d, err=%d\n",
992 teedev->name, MAJOR(teedev->dev.devt),
993 MINOR(teedev->dev.devt), rc);
994 return rc;
995 }
996
997 rc = device_add(&teedev->dev);
998 if (rc) {
999 dev_err(&teedev->dev,
1000 "unable to device_add() %s, major %d, minor %d, err=%d\n",
1001 teedev->name, MAJOR(teedev->dev.devt),
1002 MINOR(teedev->dev.devt), rc);
1003 goto err_device_add;
1004 }
1005
1006 rc = sysfs_create_group(&teedev->dev.kobj, &tee_dev_group);
1007 if (rc) {
1008 dev_err(&teedev->dev,
1009 "failed to create sysfs attributes, err=%d\n", rc);
1010 goto err_sysfs_create_group;
1011 }
1012
1013 teedev->flags |= TEE_DEVICE_FLAG_REGISTERED;
1014 return 0;
1015
1016err_sysfs_create_group:
1017 device_del(&teedev->dev);
1018err_device_add:
1019 cdev_del(&teedev->cdev);
1020 return rc;
1021}
1022EXPORT_SYMBOL_GPL(tee_device_register);
1023
1024void tee_device_put(struct tee_device *teedev)
1025{
1026 mutex_lock(&teedev->mutex);
1027 /* Shouldn't put in this state */
1028 if (!WARN_ON(!teedev->desc)) {
1029 teedev->num_users--;
1030 if (!teedev->num_users) {
1031 teedev->desc = NULL;
1032 complete(&teedev->c_no_users);
1033 }
1034 }
1035 mutex_unlock(&teedev->mutex);
1036}
1037
1038bool tee_device_get(struct tee_device *teedev)
1039{
1040 mutex_lock(&teedev->mutex);
1041 if (!teedev->desc) {
1042 mutex_unlock(&teedev->mutex);
1043 return false;
1044 }
1045 teedev->num_users++;
1046 mutex_unlock(&teedev->mutex);
1047 return true;
1048}
1049
1050/**
1051 * tee_device_unregister() - Removes a TEE device
1052 * @teedev: Device to unregister
1053 *
1054 * This function should be called to remove the @teedev even if
1055 * tee_device_register() hasn't been called yet. Does nothing if
1056 * @teedev is NULL.
1057 */
1058void tee_device_unregister(struct tee_device *teedev)
1059{
1060 if (!teedev)
1061 return;
1062
1063 if (teedev->flags & TEE_DEVICE_FLAG_REGISTERED) {
1064 sysfs_remove_group(&teedev->dev.kobj, &tee_dev_group);
1065 cdev_del(&teedev->cdev);
1066 device_del(&teedev->dev);
1067 }
1068
1069 tee_device_put(teedev);
1070 wait_for_completion(&teedev->c_no_users);
1071
1072 /*
1073 * No need to take a mutex any longer now since teedev->desc was
1074 * set to NULL before teedev->c_no_users was completed.
1075 */
1076
1077 teedev->pool = NULL;
1078
1079 put_device(&teedev->dev);
1080}
1081EXPORT_SYMBOL_GPL(tee_device_unregister);
1082
1083/**
1084 * tee_get_drvdata() - Return driver_data pointer
1085 * @teedev: Device containing the driver_data pointer
1086 * @returns the driver_data pointer supplied to tee_register().
1087 */
1088void *tee_get_drvdata(struct tee_device *teedev)
1089{
1090 return dev_get_drvdata(&teedev->dev);
1091}
1092EXPORT_SYMBOL_GPL(tee_get_drvdata);
1093
1094struct match_dev_data {
1095 struct tee_ioctl_version_data *vers;
1096 const void *data;
1097 int (*match)(struct tee_ioctl_version_data *, const void *);
1098};
1099
1100static int match_dev(struct device *dev, const void *data)
1101{
1102 const struct match_dev_data *match_data = data;
1103 struct tee_device *teedev = container_of(dev, struct tee_device, dev);
1104
1105 teedev->desc->ops->get_version(teedev, match_data->vers);
1106 return match_data->match(match_data->vers, match_data->data);
1107}
1108
1109struct tee_context *
1110tee_client_open_context(struct tee_context *start,
1111 int (*match)(struct tee_ioctl_version_data *,
1112 const void *),
1113 const void *data, struct tee_ioctl_version_data *vers)
1114{
1115 struct device *dev = NULL;
1116 struct device *put_dev = NULL;
1117 struct tee_context *ctx = NULL;
1118 struct tee_ioctl_version_data v;
1119 struct match_dev_data match_data = { vers ? vers : &v, data, match };
1120
1121 if (start)
1122 dev = &start->teedev->dev;
1123
1124 do {
1125 dev = class_find_device(tee_class, dev, &match_data, match_dev);
1126 if (!dev) {
1127 ctx = ERR_PTR(-ENOENT);
1128 break;
1129 }
1130
1131 put_device(put_dev);
1132 put_dev = dev;
1133
1134 ctx = teedev_open(container_of(dev, struct tee_device, dev));
1135 } while (IS_ERR(ctx) && PTR_ERR(ctx) != -ENOMEM);
1136
1137 put_device(put_dev);
1138 /*
1139 * Default behaviour for in kernel client is to not wait for
1140 * tee-supplicant if not present for any requests in this context.
1141 * Also this flag could be configured again before call to
1142 * tee_client_open_session() if any in kernel client requires
1143 * different behaviour.
1144 */
1145 if (!IS_ERR(ctx))
1146 ctx->supp_nowait = true;
1147
1148 return ctx;
1149}
1150EXPORT_SYMBOL_GPL(tee_client_open_context);
1151
1152void tee_client_close_context(struct tee_context *ctx)
1153{
1154 teedev_close_context(ctx);
1155}
1156EXPORT_SYMBOL_GPL(tee_client_close_context);
1157
1158void tee_client_get_version(struct tee_context *ctx,
1159 struct tee_ioctl_version_data *vers)
1160{
1161 ctx->teedev->desc->ops->get_version(ctx->teedev, vers);
1162}
1163EXPORT_SYMBOL_GPL(tee_client_get_version);
1164
1165int tee_client_open_session(struct tee_context *ctx,
1166 struct tee_ioctl_open_session_arg *arg,
1167 struct tee_param *param)
1168{
1169 if (!ctx->teedev->desc->ops->open_session)
1170 return -EINVAL;
1171 return ctx->teedev->desc->ops->open_session(ctx, arg, param);
1172}
1173EXPORT_SYMBOL_GPL(tee_client_open_session);
1174
1175int tee_client_close_session(struct tee_context *ctx, u32 session)
1176{
1177 if (!ctx->teedev->desc->ops->close_session)
1178 return -EINVAL;
1179 return ctx->teedev->desc->ops->close_session(ctx, session);
1180}
1181EXPORT_SYMBOL_GPL(tee_client_close_session);
1182
1183int tee_client_invoke_func(struct tee_context *ctx,
1184 struct tee_ioctl_invoke_arg *arg,
1185 struct tee_param *param)
1186{
1187 if (!ctx->teedev->desc->ops->invoke_func)
1188 return -EINVAL;
1189 return ctx->teedev->desc->ops->invoke_func(ctx, arg, param);
1190}
1191EXPORT_SYMBOL_GPL(tee_client_invoke_func);
1192
1193int tee_client_cancel_req(struct tee_context *ctx,
1194 struct tee_ioctl_cancel_arg *arg)
1195{
1196 if (!ctx->teedev->desc->ops->cancel_req)
1197 return -EINVAL;
1198 return ctx->teedev->desc->ops->cancel_req(ctx, arg->cancel_id,
1199 arg->session);
1200}
1201
1202static int tee_client_device_match(struct device *dev,
1203 struct device_driver *drv)
1204{
1205 const struct tee_client_device_id *id_table;
1206 struct tee_client_device *tee_device;
1207
1208 id_table = to_tee_client_driver(drv)->id_table;
1209 tee_device = to_tee_client_device(dev);
1210
1211 while (!uuid_is_null(&id_table->uuid)) {
1212 if (uuid_equal(&tee_device->id.uuid, &id_table->uuid))
1213 return 1;
1214 id_table++;
1215 }
1216
1217 return 0;
1218}
1219
1220static int tee_client_device_uevent(struct device *dev,
1221 struct kobj_uevent_env *env)
1222{
1223 uuid_t *dev_id = &to_tee_client_device(dev)->id.uuid;
1224
1225 return add_uevent_var(env, "MODALIAS=tee:%pUb", dev_id);
1226}
1227
1228struct bus_type tee_bus_type = {
1229 .name = "tee",
1230 .match = tee_client_device_match,
1231 .uevent = tee_client_device_uevent,
1232};
1233EXPORT_SYMBOL_GPL(tee_bus_type);
1234
1235static int __init tee_init(void)
1236{
1237 int rc;
1238
1239 tee_class = class_create(THIS_MODULE, "tee");
1240 if (IS_ERR(tee_class)) {
1241 pr_err("couldn't create class\n");
1242 return PTR_ERR(tee_class);
1243 }
1244
1245 rc = alloc_chrdev_region(&tee_devt, 0, TEE_NUM_DEVICES, "tee");
1246 if (rc) {
1247 pr_err("failed to allocate char dev region\n");
1248 goto out_unreg_class;
1249 }
1250
1251 rc = bus_register(&tee_bus_type);
1252 if (rc) {
1253 pr_err("failed to register tee bus\n");
1254 goto out_unreg_chrdev;
1255 }
1256
1257 return 0;
1258
1259out_unreg_chrdev:
1260 unregister_chrdev_region(tee_devt, TEE_NUM_DEVICES);
1261out_unreg_class:
1262 class_destroy(tee_class);
1263 tee_class = NULL;
1264
1265 return rc;
1266}
1267
1268static void __exit tee_exit(void)
1269{
1270 bus_unregister(&tee_bus_type);
1271 unregister_chrdev_region(tee_devt, TEE_NUM_DEVICES);
1272 class_destroy(tee_class);
1273 tee_class = NULL;
1274}
1275
1276subsys_initcall(tee_init);
1277module_exit(tee_exit);
1278
1279MODULE_AUTHOR("Linaro");
1280MODULE_DESCRIPTION("TEE Driver");
1281MODULE_VERSION("1.0");
1282MODULE_LICENSE("GPL v2");
1// SPDX-License-Identifier: GPL-2.0-only
2/*
3 * Copyright (c) 2015-2016, Linaro Limited
4 */
5
6#define pr_fmt(fmt) "%s: " fmt, __func__
7
8#include <linux/cdev.h>
9#include <linux/fs.h>
10#include <linux/idr.h>
11#include <linux/module.h>
12#include <linux/slab.h>
13#include <linux/tee_drv.h>
14#include <linux/uaccess.h>
15#include "tee_private.h"
16
17#define TEE_NUM_DEVICES 32
18
19#define TEE_IOCTL_PARAM_SIZE(x) (sizeof(struct tee_param) * (x))
20
21/*
22 * Unprivileged devices in the lower half range and privileged devices in
23 * the upper half range.
24 */
25static DECLARE_BITMAP(dev_mask, TEE_NUM_DEVICES);
26static DEFINE_SPINLOCK(driver_lock);
27
28static struct class *tee_class;
29static dev_t tee_devt;
30
31static struct tee_context *teedev_open(struct tee_device *teedev)
32{
33 int rc;
34 struct tee_context *ctx;
35
36 if (!tee_device_get(teedev))
37 return ERR_PTR(-EINVAL);
38
39 ctx = kzalloc(sizeof(*ctx), GFP_KERNEL);
40 if (!ctx) {
41 rc = -ENOMEM;
42 goto err;
43 }
44
45 kref_init(&ctx->refcount);
46 ctx->teedev = teedev;
47 INIT_LIST_HEAD(&ctx->list_shm);
48 rc = teedev->desc->ops->open(ctx);
49 if (rc)
50 goto err;
51
52 return ctx;
53err:
54 kfree(ctx);
55 tee_device_put(teedev);
56 return ERR_PTR(rc);
57
58}
59
60void teedev_ctx_get(struct tee_context *ctx)
61{
62 if (ctx->releasing)
63 return;
64
65 kref_get(&ctx->refcount);
66}
67
68static void teedev_ctx_release(struct kref *ref)
69{
70 struct tee_context *ctx = container_of(ref, struct tee_context,
71 refcount);
72 ctx->releasing = true;
73 ctx->teedev->desc->ops->release(ctx);
74 kfree(ctx);
75}
76
77void teedev_ctx_put(struct tee_context *ctx)
78{
79 if (ctx->releasing)
80 return;
81
82 kref_put(&ctx->refcount, teedev_ctx_release);
83}
84
85static void teedev_close_context(struct tee_context *ctx)
86{
87 tee_device_put(ctx->teedev);
88 teedev_ctx_put(ctx);
89}
90
91static int tee_open(struct inode *inode, struct file *filp)
92{
93 struct tee_context *ctx;
94
95 ctx = teedev_open(container_of(inode->i_cdev, struct tee_device, cdev));
96 if (IS_ERR(ctx))
97 return PTR_ERR(ctx);
98
99 /*
100 * Default user-space behaviour is to wait for tee-supplicant
101 * if not present for any requests in this context.
102 */
103 ctx->supp_nowait = false;
104 filp->private_data = ctx;
105 return 0;
106}
107
108static int tee_release(struct inode *inode, struct file *filp)
109{
110 teedev_close_context(filp->private_data);
111 return 0;
112}
113
114static int tee_ioctl_version(struct tee_context *ctx,
115 struct tee_ioctl_version_data __user *uvers)
116{
117 struct tee_ioctl_version_data vers;
118
119 ctx->teedev->desc->ops->get_version(ctx->teedev, &vers);
120
121 if (ctx->teedev->desc->flags & TEE_DESC_PRIVILEGED)
122 vers.gen_caps |= TEE_GEN_CAP_PRIVILEGED;
123
124 if (copy_to_user(uvers, &vers, sizeof(vers)))
125 return -EFAULT;
126
127 return 0;
128}
129
130static int tee_ioctl_shm_alloc(struct tee_context *ctx,
131 struct tee_ioctl_shm_alloc_data __user *udata)
132{
133 long ret;
134 struct tee_ioctl_shm_alloc_data data;
135 struct tee_shm *shm;
136
137 if (copy_from_user(&data, udata, sizeof(data)))
138 return -EFAULT;
139
140 /* Currently no input flags are supported */
141 if (data.flags)
142 return -EINVAL;
143
144 shm = tee_shm_alloc(ctx, data.size, TEE_SHM_MAPPED | TEE_SHM_DMA_BUF);
145 if (IS_ERR(shm))
146 return PTR_ERR(shm);
147
148 data.id = shm->id;
149 data.flags = shm->flags;
150 data.size = shm->size;
151
152 if (copy_to_user(udata, &data, sizeof(data)))
153 ret = -EFAULT;
154 else
155 ret = tee_shm_get_fd(shm);
156
157 /*
158 * When user space closes the file descriptor the shared memory
159 * should be freed or if tee_shm_get_fd() failed then it will
160 * be freed immediately.
161 */
162 tee_shm_put(shm);
163 return ret;
164}
165
166static int
167tee_ioctl_shm_register(struct tee_context *ctx,
168 struct tee_ioctl_shm_register_data __user *udata)
169{
170 long ret;
171 struct tee_ioctl_shm_register_data data;
172 struct tee_shm *shm;
173
174 if (copy_from_user(&data, udata, sizeof(data)))
175 return -EFAULT;
176
177 /* Currently no input flags are supported */
178 if (data.flags)
179 return -EINVAL;
180
181 shm = tee_shm_register(ctx, data.addr, data.length,
182 TEE_SHM_DMA_BUF | TEE_SHM_USER_MAPPED);
183 if (IS_ERR(shm))
184 return PTR_ERR(shm);
185
186 data.id = shm->id;
187 data.flags = shm->flags;
188 data.length = shm->size;
189
190 if (copy_to_user(udata, &data, sizeof(data)))
191 ret = -EFAULT;
192 else
193 ret = tee_shm_get_fd(shm);
194 /*
195 * When user space closes the file descriptor the shared memory
196 * should be freed or if tee_shm_get_fd() failed then it will
197 * be freed immediately.
198 */
199 tee_shm_put(shm);
200 return ret;
201}
202
203static int params_from_user(struct tee_context *ctx, struct tee_param *params,
204 size_t num_params,
205 struct tee_ioctl_param __user *uparams)
206{
207 size_t n;
208
209 for (n = 0; n < num_params; n++) {
210 struct tee_shm *shm;
211 struct tee_ioctl_param ip;
212
213 if (copy_from_user(&ip, uparams + n, sizeof(ip)))
214 return -EFAULT;
215
216 /* All unused attribute bits has to be zero */
217 if (ip.attr & ~TEE_IOCTL_PARAM_ATTR_MASK)
218 return -EINVAL;
219
220 params[n].attr = ip.attr;
221 switch (ip.attr & TEE_IOCTL_PARAM_ATTR_TYPE_MASK) {
222 case TEE_IOCTL_PARAM_ATTR_TYPE_NONE:
223 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_OUTPUT:
224 break;
225 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INPUT:
226 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INOUT:
227 params[n].u.value.a = ip.a;
228 params[n].u.value.b = ip.b;
229 params[n].u.value.c = ip.c;
230 break;
231 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INPUT:
232 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_OUTPUT:
233 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INOUT:
234 /*
235 * If we fail to get a pointer to a shared memory
236 * object (and increase the ref count) from an
237 * identifier we return an error. All pointers that
238 * has been added in params have an increased ref
239 * count. It's the callers responibility to do
240 * tee_shm_put() on all resolved pointers.
241 */
242 shm = tee_shm_get_from_id(ctx, ip.c);
243 if (IS_ERR(shm))
244 return PTR_ERR(shm);
245
246 /*
247 * Ensure offset + size does not overflow offset
248 * and does not overflow the size of the referred
249 * shared memory object.
250 */
251 if ((ip.a + ip.b) < ip.a ||
252 (ip.a + ip.b) > shm->size) {
253 tee_shm_put(shm);
254 return -EINVAL;
255 }
256
257 params[n].u.memref.shm_offs = ip.a;
258 params[n].u.memref.size = ip.b;
259 params[n].u.memref.shm = shm;
260 break;
261 default:
262 /* Unknown attribute */
263 return -EINVAL;
264 }
265 }
266 return 0;
267}
268
269static int params_to_user(struct tee_ioctl_param __user *uparams,
270 size_t num_params, struct tee_param *params)
271{
272 size_t n;
273
274 for (n = 0; n < num_params; n++) {
275 struct tee_ioctl_param __user *up = uparams + n;
276 struct tee_param *p = params + n;
277
278 switch (p->attr) {
279 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_OUTPUT:
280 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INOUT:
281 if (put_user(p->u.value.a, &up->a) ||
282 put_user(p->u.value.b, &up->b) ||
283 put_user(p->u.value.c, &up->c))
284 return -EFAULT;
285 break;
286 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_OUTPUT:
287 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INOUT:
288 if (put_user((u64)p->u.memref.size, &up->b))
289 return -EFAULT;
290 default:
291 break;
292 }
293 }
294 return 0;
295}
296
297static int tee_ioctl_open_session(struct tee_context *ctx,
298 struct tee_ioctl_buf_data __user *ubuf)
299{
300 int rc;
301 size_t n;
302 struct tee_ioctl_buf_data buf;
303 struct tee_ioctl_open_session_arg __user *uarg;
304 struct tee_ioctl_open_session_arg arg;
305 struct tee_ioctl_param __user *uparams = NULL;
306 struct tee_param *params = NULL;
307 bool have_session = false;
308
309 if (!ctx->teedev->desc->ops->open_session)
310 return -EINVAL;
311
312 if (copy_from_user(&buf, ubuf, sizeof(buf)))
313 return -EFAULT;
314
315 if (buf.buf_len > TEE_MAX_ARG_SIZE ||
316 buf.buf_len < sizeof(struct tee_ioctl_open_session_arg))
317 return -EINVAL;
318
319 uarg = u64_to_user_ptr(buf.buf_ptr);
320 if (copy_from_user(&arg, uarg, sizeof(arg)))
321 return -EFAULT;
322
323 if (sizeof(arg) + TEE_IOCTL_PARAM_SIZE(arg.num_params) != buf.buf_len)
324 return -EINVAL;
325
326 if (arg.num_params) {
327 params = kcalloc(arg.num_params, sizeof(struct tee_param),
328 GFP_KERNEL);
329 if (!params)
330 return -ENOMEM;
331 uparams = uarg->params;
332 rc = params_from_user(ctx, params, arg.num_params, uparams);
333 if (rc)
334 goto out;
335 }
336
337 rc = ctx->teedev->desc->ops->open_session(ctx, &arg, params);
338 if (rc)
339 goto out;
340 have_session = true;
341
342 if (put_user(arg.session, &uarg->session) ||
343 put_user(arg.ret, &uarg->ret) ||
344 put_user(arg.ret_origin, &uarg->ret_origin)) {
345 rc = -EFAULT;
346 goto out;
347 }
348 rc = params_to_user(uparams, arg.num_params, params);
349out:
350 /*
351 * If we've succeeded to open the session but failed to communicate
352 * it back to user space, close the session again to avoid leakage.
353 */
354 if (rc && have_session && ctx->teedev->desc->ops->close_session)
355 ctx->teedev->desc->ops->close_session(ctx, arg.session);
356
357 if (params) {
358 /* Decrease ref count for all valid shared memory pointers */
359 for (n = 0; n < arg.num_params; n++)
360 if (tee_param_is_memref(params + n) &&
361 params[n].u.memref.shm)
362 tee_shm_put(params[n].u.memref.shm);
363 kfree(params);
364 }
365
366 return rc;
367}
368
369static int tee_ioctl_invoke(struct tee_context *ctx,
370 struct tee_ioctl_buf_data __user *ubuf)
371{
372 int rc;
373 size_t n;
374 struct tee_ioctl_buf_data buf;
375 struct tee_ioctl_invoke_arg __user *uarg;
376 struct tee_ioctl_invoke_arg arg;
377 struct tee_ioctl_param __user *uparams = NULL;
378 struct tee_param *params = NULL;
379
380 if (!ctx->teedev->desc->ops->invoke_func)
381 return -EINVAL;
382
383 if (copy_from_user(&buf, ubuf, sizeof(buf)))
384 return -EFAULT;
385
386 if (buf.buf_len > TEE_MAX_ARG_SIZE ||
387 buf.buf_len < sizeof(struct tee_ioctl_invoke_arg))
388 return -EINVAL;
389
390 uarg = u64_to_user_ptr(buf.buf_ptr);
391 if (copy_from_user(&arg, uarg, sizeof(arg)))
392 return -EFAULT;
393
394 if (sizeof(arg) + TEE_IOCTL_PARAM_SIZE(arg.num_params) != buf.buf_len)
395 return -EINVAL;
396
397 if (arg.num_params) {
398 params = kcalloc(arg.num_params, sizeof(struct tee_param),
399 GFP_KERNEL);
400 if (!params)
401 return -ENOMEM;
402 uparams = uarg->params;
403 rc = params_from_user(ctx, params, arg.num_params, uparams);
404 if (rc)
405 goto out;
406 }
407
408 rc = ctx->teedev->desc->ops->invoke_func(ctx, &arg, params);
409 if (rc)
410 goto out;
411
412 if (put_user(arg.ret, &uarg->ret) ||
413 put_user(arg.ret_origin, &uarg->ret_origin)) {
414 rc = -EFAULT;
415 goto out;
416 }
417 rc = params_to_user(uparams, arg.num_params, params);
418out:
419 if (params) {
420 /* Decrease ref count for all valid shared memory pointers */
421 for (n = 0; n < arg.num_params; n++)
422 if (tee_param_is_memref(params + n) &&
423 params[n].u.memref.shm)
424 tee_shm_put(params[n].u.memref.shm);
425 kfree(params);
426 }
427 return rc;
428}
429
430static int tee_ioctl_cancel(struct tee_context *ctx,
431 struct tee_ioctl_cancel_arg __user *uarg)
432{
433 struct tee_ioctl_cancel_arg arg;
434
435 if (!ctx->teedev->desc->ops->cancel_req)
436 return -EINVAL;
437
438 if (copy_from_user(&arg, uarg, sizeof(arg)))
439 return -EFAULT;
440
441 return ctx->teedev->desc->ops->cancel_req(ctx, arg.cancel_id,
442 arg.session);
443}
444
445static int
446tee_ioctl_close_session(struct tee_context *ctx,
447 struct tee_ioctl_close_session_arg __user *uarg)
448{
449 struct tee_ioctl_close_session_arg arg;
450
451 if (!ctx->teedev->desc->ops->close_session)
452 return -EINVAL;
453
454 if (copy_from_user(&arg, uarg, sizeof(arg)))
455 return -EFAULT;
456
457 return ctx->teedev->desc->ops->close_session(ctx, arg.session);
458}
459
460static int params_to_supp(struct tee_context *ctx,
461 struct tee_ioctl_param __user *uparams,
462 size_t num_params, struct tee_param *params)
463{
464 size_t n;
465
466 for (n = 0; n < num_params; n++) {
467 struct tee_ioctl_param ip;
468 struct tee_param *p = params + n;
469
470 ip.attr = p->attr;
471 switch (p->attr & TEE_IOCTL_PARAM_ATTR_TYPE_MASK) {
472 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INPUT:
473 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INOUT:
474 ip.a = p->u.value.a;
475 ip.b = p->u.value.b;
476 ip.c = p->u.value.c;
477 break;
478 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INPUT:
479 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_OUTPUT:
480 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INOUT:
481 ip.b = p->u.memref.size;
482 if (!p->u.memref.shm) {
483 ip.a = 0;
484 ip.c = (u64)-1; /* invalid shm id */
485 break;
486 }
487 ip.a = p->u.memref.shm_offs;
488 ip.c = p->u.memref.shm->id;
489 break;
490 default:
491 ip.a = 0;
492 ip.b = 0;
493 ip.c = 0;
494 break;
495 }
496
497 if (copy_to_user(uparams + n, &ip, sizeof(ip)))
498 return -EFAULT;
499 }
500
501 return 0;
502}
503
504static int tee_ioctl_supp_recv(struct tee_context *ctx,
505 struct tee_ioctl_buf_data __user *ubuf)
506{
507 int rc;
508 struct tee_ioctl_buf_data buf;
509 struct tee_iocl_supp_recv_arg __user *uarg;
510 struct tee_param *params;
511 u32 num_params;
512 u32 func;
513
514 if (!ctx->teedev->desc->ops->supp_recv)
515 return -EINVAL;
516
517 if (copy_from_user(&buf, ubuf, sizeof(buf)))
518 return -EFAULT;
519
520 if (buf.buf_len > TEE_MAX_ARG_SIZE ||
521 buf.buf_len < sizeof(struct tee_iocl_supp_recv_arg))
522 return -EINVAL;
523
524 uarg = u64_to_user_ptr(buf.buf_ptr);
525 if (get_user(num_params, &uarg->num_params))
526 return -EFAULT;
527
528 if (sizeof(*uarg) + TEE_IOCTL_PARAM_SIZE(num_params) != buf.buf_len)
529 return -EINVAL;
530
531 params = kcalloc(num_params, sizeof(struct tee_param), GFP_KERNEL);
532 if (!params)
533 return -ENOMEM;
534
535 rc = params_from_user(ctx, params, num_params, uarg->params);
536 if (rc)
537 goto out;
538
539 rc = ctx->teedev->desc->ops->supp_recv(ctx, &func, &num_params, params);
540 if (rc)
541 goto out;
542
543 if (put_user(func, &uarg->func) ||
544 put_user(num_params, &uarg->num_params)) {
545 rc = -EFAULT;
546 goto out;
547 }
548
549 rc = params_to_supp(ctx, uarg->params, num_params, params);
550out:
551 kfree(params);
552 return rc;
553}
554
555static int params_from_supp(struct tee_param *params, size_t num_params,
556 struct tee_ioctl_param __user *uparams)
557{
558 size_t n;
559
560 for (n = 0; n < num_params; n++) {
561 struct tee_param *p = params + n;
562 struct tee_ioctl_param ip;
563
564 if (copy_from_user(&ip, uparams + n, sizeof(ip)))
565 return -EFAULT;
566
567 /* All unused attribute bits has to be zero */
568 if (ip.attr & ~TEE_IOCTL_PARAM_ATTR_MASK)
569 return -EINVAL;
570
571 p->attr = ip.attr;
572 switch (ip.attr & TEE_IOCTL_PARAM_ATTR_TYPE_MASK) {
573 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_OUTPUT:
574 case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INOUT:
575 /* Only out and in/out values can be updated */
576 p->u.value.a = ip.a;
577 p->u.value.b = ip.b;
578 p->u.value.c = ip.c;
579 break;
580 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_OUTPUT:
581 case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INOUT:
582 /*
583 * Only the size of the memref can be updated.
584 * Since we don't have access to the original
585 * parameters here, only store the supplied size.
586 * The driver will copy the updated size into the
587 * original parameters.
588 */
589 p->u.memref.shm = NULL;
590 p->u.memref.shm_offs = 0;
591 p->u.memref.size = ip.b;
592 break;
593 default:
594 memset(&p->u, 0, sizeof(p->u));
595 break;
596 }
597 }
598 return 0;
599}
600
601static int tee_ioctl_supp_send(struct tee_context *ctx,
602 struct tee_ioctl_buf_data __user *ubuf)
603{
604 long rc;
605 struct tee_ioctl_buf_data buf;
606 struct tee_iocl_supp_send_arg __user *uarg;
607 struct tee_param *params;
608 u32 num_params;
609 u32 ret;
610
611 /* Not valid for this driver */
612 if (!ctx->teedev->desc->ops->supp_send)
613 return -EINVAL;
614
615 if (copy_from_user(&buf, ubuf, sizeof(buf)))
616 return -EFAULT;
617
618 if (buf.buf_len > TEE_MAX_ARG_SIZE ||
619 buf.buf_len < sizeof(struct tee_iocl_supp_send_arg))
620 return -EINVAL;
621
622 uarg = u64_to_user_ptr(buf.buf_ptr);
623 if (get_user(ret, &uarg->ret) ||
624 get_user(num_params, &uarg->num_params))
625 return -EFAULT;
626
627 if (sizeof(*uarg) + TEE_IOCTL_PARAM_SIZE(num_params) > buf.buf_len)
628 return -EINVAL;
629
630 params = kcalloc(num_params, sizeof(struct tee_param), GFP_KERNEL);
631 if (!params)
632 return -ENOMEM;
633
634 rc = params_from_supp(params, num_params, uarg->params);
635 if (rc)
636 goto out;
637
638 rc = ctx->teedev->desc->ops->supp_send(ctx, ret, num_params, params);
639out:
640 kfree(params);
641 return rc;
642}
643
644static long tee_ioctl(struct file *filp, unsigned int cmd, unsigned long arg)
645{
646 struct tee_context *ctx = filp->private_data;
647 void __user *uarg = (void __user *)arg;
648
649 switch (cmd) {
650 case TEE_IOC_VERSION:
651 return tee_ioctl_version(ctx, uarg);
652 case TEE_IOC_SHM_ALLOC:
653 return tee_ioctl_shm_alloc(ctx, uarg);
654 case TEE_IOC_SHM_REGISTER:
655 return tee_ioctl_shm_register(ctx, uarg);
656 case TEE_IOC_OPEN_SESSION:
657 return tee_ioctl_open_session(ctx, uarg);
658 case TEE_IOC_INVOKE:
659 return tee_ioctl_invoke(ctx, uarg);
660 case TEE_IOC_CANCEL:
661 return tee_ioctl_cancel(ctx, uarg);
662 case TEE_IOC_CLOSE_SESSION:
663 return tee_ioctl_close_session(ctx, uarg);
664 case TEE_IOC_SUPPL_RECV:
665 return tee_ioctl_supp_recv(ctx, uarg);
666 case TEE_IOC_SUPPL_SEND:
667 return tee_ioctl_supp_send(ctx, uarg);
668 default:
669 return -EINVAL;
670 }
671}
672
673static const struct file_operations tee_fops = {
674 .owner = THIS_MODULE,
675 .open = tee_open,
676 .release = tee_release,
677 .unlocked_ioctl = tee_ioctl,
678 .compat_ioctl = tee_ioctl,
679};
680
681static void tee_release_device(struct device *dev)
682{
683 struct tee_device *teedev = container_of(dev, struct tee_device, dev);
684
685 spin_lock(&driver_lock);
686 clear_bit(teedev->id, dev_mask);
687 spin_unlock(&driver_lock);
688 mutex_destroy(&teedev->mutex);
689 idr_destroy(&teedev->idr);
690 kfree(teedev);
691}
692
693/**
694 * tee_device_alloc() - Allocate a new struct tee_device instance
695 * @teedesc: Descriptor for this driver
696 * @dev: Parent device for this device
697 * @pool: Shared memory pool, NULL if not used
698 * @driver_data: Private driver data for this device
699 *
700 * Allocates a new struct tee_device instance. The device is
701 * removed by tee_device_unregister().
702 *
703 * @returns a pointer to a 'struct tee_device' or an ERR_PTR on failure
704 */
705struct tee_device *tee_device_alloc(const struct tee_desc *teedesc,
706 struct device *dev,
707 struct tee_shm_pool *pool,
708 void *driver_data)
709{
710 struct tee_device *teedev;
711 void *ret;
712 int rc, max_id;
713 int offs = 0;
714
715 if (!teedesc || !teedesc->name || !teedesc->ops ||
716 !teedesc->ops->get_version || !teedesc->ops->open ||
717 !teedesc->ops->release || !pool)
718 return ERR_PTR(-EINVAL);
719
720 teedev = kzalloc(sizeof(*teedev), GFP_KERNEL);
721 if (!teedev) {
722 ret = ERR_PTR(-ENOMEM);
723 goto err;
724 }
725
726 max_id = TEE_NUM_DEVICES / 2;
727
728 if (teedesc->flags & TEE_DESC_PRIVILEGED) {
729 offs = TEE_NUM_DEVICES / 2;
730 max_id = TEE_NUM_DEVICES;
731 }
732
733 spin_lock(&driver_lock);
734 teedev->id = find_next_zero_bit(dev_mask, max_id, offs);
735 if (teedev->id < max_id)
736 set_bit(teedev->id, dev_mask);
737 spin_unlock(&driver_lock);
738
739 if (teedev->id >= max_id) {
740 ret = ERR_PTR(-ENOMEM);
741 goto err;
742 }
743
744 snprintf(teedev->name, sizeof(teedev->name), "tee%s%d",
745 teedesc->flags & TEE_DESC_PRIVILEGED ? "priv" : "",
746 teedev->id - offs);
747
748 teedev->dev.class = tee_class;
749 teedev->dev.release = tee_release_device;
750 teedev->dev.parent = dev;
751
752 teedev->dev.devt = MKDEV(MAJOR(tee_devt), teedev->id);
753
754 rc = dev_set_name(&teedev->dev, "%s", teedev->name);
755 if (rc) {
756 ret = ERR_PTR(rc);
757 goto err_devt;
758 }
759
760 cdev_init(&teedev->cdev, &tee_fops);
761 teedev->cdev.owner = teedesc->owner;
762 teedev->cdev.kobj.parent = &teedev->dev.kobj;
763
764 dev_set_drvdata(&teedev->dev, driver_data);
765 device_initialize(&teedev->dev);
766
767 /* 1 as tee_device_unregister() does one final tee_device_put() */
768 teedev->num_users = 1;
769 init_completion(&teedev->c_no_users);
770 mutex_init(&teedev->mutex);
771 idr_init(&teedev->idr);
772
773 teedev->desc = teedesc;
774 teedev->pool = pool;
775
776 return teedev;
777err_devt:
778 unregister_chrdev_region(teedev->dev.devt, 1);
779err:
780 pr_err("could not register %s driver\n",
781 teedesc->flags & TEE_DESC_PRIVILEGED ? "privileged" : "client");
782 if (teedev && teedev->id < TEE_NUM_DEVICES) {
783 spin_lock(&driver_lock);
784 clear_bit(teedev->id, dev_mask);
785 spin_unlock(&driver_lock);
786 }
787 kfree(teedev);
788 return ret;
789}
790EXPORT_SYMBOL_GPL(tee_device_alloc);
791
792static ssize_t implementation_id_show(struct device *dev,
793 struct device_attribute *attr, char *buf)
794{
795 struct tee_device *teedev = container_of(dev, struct tee_device, dev);
796 struct tee_ioctl_version_data vers;
797
798 teedev->desc->ops->get_version(teedev, &vers);
799 return scnprintf(buf, PAGE_SIZE, "%d\n", vers.impl_id);
800}
801static DEVICE_ATTR_RO(implementation_id);
802
803static struct attribute *tee_dev_attrs[] = {
804 &dev_attr_implementation_id.attr,
805 NULL
806};
807
808static const struct attribute_group tee_dev_group = {
809 .attrs = tee_dev_attrs,
810};
811
812/**
813 * tee_device_register() - Registers a TEE device
814 * @teedev: Device to register
815 *
816 * tee_device_unregister() need to be called to remove the @teedev if
817 * this function fails.
818 *
819 * @returns < 0 on failure
820 */
821int tee_device_register(struct tee_device *teedev)
822{
823 int rc;
824
825 if (teedev->flags & TEE_DEVICE_FLAG_REGISTERED) {
826 dev_err(&teedev->dev, "attempt to register twice\n");
827 return -EINVAL;
828 }
829
830 rc = cdev_add(&teedev->cdev, teedev->dev.devt, 1);
831 if (rc) {
832 dev_err(&teedev->dev,
833 "unable to cdev_add() %s, major %d, minor %d, err=%d\n",
834 teedev->name, MAJOR(teedev->dev.devt),
835 MINOR(teedev->dev.devt), rc);
836 return rc;
837 }
838
839 rc = device_add(&teedev->dev);
840 if (rc) {
841 dev_err(&teedev->dev,
842 "unable to device_add() %s, major %d, minor %d, err=%d\n",
843 teedev->name, MAJOR(teedev->dev.devt),
844 MINOR(teedev->dev.devt), rc);
845 goto err_device_add;
846 }
847
848 rc = sysfs_create_group(&teedev->dev.kobj, &tee_dev_group);
849 if (rc) {
850 dev_err(&teedev->dev,
851 "failed to create sysfs attributes, err=%d\n", rc);
852 goto err_sysfs_create_group;
853 }
854
855 teedev->flags |= TEE_DEVICE_FLAG_REGISTERED;
856 return 0;
857
858err_sysfs_create_group:
859 device_del(&teedev->dev);
860err_device_add:
861 cdev_del(&teedev->cdev);
862 return rc;
863}
864EXPORT_SYMBOL_GPL(tee_device_register);
865
866void tee_device_put(struct tee_device *teedev)
867{
868 mutex_lock(&teedev->mutex);
869 /* Shouldn't put in this state */
870 if (!WARN_ON(!teedev->desc)) {
871 teedev->num_users--;
872 if (!teedev->num_users) {
873 teedev->desc = NULL;
874 complete(&teedev->c_no_users);
875 }
876 }
877 mutex_unlock(&teedev->mutex);
878}
879
880bool tee_device_get(struct tee_device *teedev)
881{
882 mutex_lock(&teedev->mutex);
883 if (!teedev->desc) {
884 mutex_unlock(&teedev->mutex);
885 return false;
886 }
887 teedev->num_users++;
888 mutex_unlock(&teedev->mutex);
889 return true;
890}
891
892/**
893 * tee_device_unregister() - Removes a TEE device
894 * @teedev: Device to unregister
895 *
896 * This function should be called to remove the @teedev even if
897 * tee_device_register() hasn't been called yet. Does nothing if
898 * @teedev is NULL.
899 */
900void tee_device_unregister(struct tee_device *teedev)
901{
902 if (!teedev)
903 return;
904
905 if (teedev->flags & TEE_DEVICE_FLAG_REGISTERED) {
906 sysfs_remove_group(&teedev->dev.kobj, &tee_dev_group);
907 cdev_del(&teedev->cdev);
908 device_del(&teedev->dev);
909 }
910
911 tee_device_put(teedev);
912 wait_for_completion(&teedev->c_no_users);
913
914 /*
915 * No need to take a mutex any longer now since teedev->desc was
916 * set to NULL before teedev->c_no_users was completed.
917 */
918
919 teedev->pool = NULL;
920
921 put_device(&teedev->dev);
922}
923EXPORT_SYMBOL_GPL(tee_device_unregister);
924
925/**
926 * tee_get_drvdata() - Return driver_data pointer
927 * @teedev: Device containing the driver_data pointer
928 * @returns the driver_data pointer supplied to tee_register().
929 */
930void *tee_get_drvdata(struct tee_device *teedev)
931{
932 return dev_get_drvdata(&teedev->dev);
933}
934EXPORT_SYMBOL_GPL(tee_get_drvdata);
935
936struct match_dev_data {
937 struct tee_ioctl_version_data *vers;
938 const void *data;
939 int (*match)(struct tee_ioctl_version_data *, const void *);
940};
941
942static int match_dev(struct device *dev, const void *data)
943{
944 const struct match_dev_data *match_data = data;
945 struct tee_device *teedev = container_of(dev, struct tee_device, dev);
946
947 teedev->desc->ops->get_version(teedev, match_data->vers);
948 return match_data->match(match_data->vers, match_data->data);
949}
950
951struct tee_context *
952tee_client_open_context(struct tee_context *start,
953 int (*match)(struct tee_ioctl_version_data *,
954 const void *),
955 const void *data, struct tee_ioctl_version_data *vers)
956{
957 struct device *dev = NULL;
958 struct device *put_dev = NULL;
959 struct tee_context *ctx = NULL;
960 struct tee_ioctl_version_data v;
961 struct match_dev_data match_data = { vers ? vers : &v, data, match };
962
963 if (start)
964 dev = &start->teedev->dev;
965
966 do {
967 dev = class_find_device(tee_class, dev, &match_data, match_dev);
968 if (!dev) {
969 ctx = ERR_PTR(-ENOENT);
970 break;
971 }
972
973 put_device(put_dev);
974 put_dev = dev;
975
976 ctx = teedev_open(container_of(dev, struct tee_device, dev));
977 } while (IS_ERR(ctx) && PTR_ERR(ctx) != -ENOMEM);
978
979 put_device(put_dev);
980 /*
981 * Default behaviour for in kernel client is to not wait for
982 * tee-supplicant if not present for any requests in this context.
983 * Also this flag could be configured again before call to
984 * tee_client_open_session() if any in kernel client requires
985 * different behaviour.
986 */
987 if (!IS_ERR(ctx))
988 ctx->supp_nowait = true;
989
990 return ctx;
991}
992EXPORT_SYMBOL_GPL(tee_client_open_context);
993
994void tee_client_close_context(struct tee_context *ctx)
995{
996 teedev_close_context(ctx);
997}
998EXPORT_SYMBOL_GPL(tee_client_close_context);
999
1000void tee_client_get_version(struct tee_context *ctx,
1001 struct tee_ioctl_version_data *vers)
1002{
1003 ctx->teedev->desc->ops->get_version(ctx->teedev, vers);
1004}
1005EXPORT_SYMBOL_GPL(tee_client_get_version);
1006
1007int tee_client_open_session(struct tee_context *ctx,
1008 struct tee_ioctl_open_session_arg *arg,
1009 struct tee_param *param)
1010{
1011 if (!ctx->teedev->desc->ops->open_session)
1012 return -EINVAL;
1013 return ctx->teedev->desc->ops->open_session(ctx, arg, param);
1014}
1015EXPORT_SYMBOL_GPL(tee_client_open_session);
1016
1017int tee_client_close_session(struct tee_context *ctx, u32 session)
1018{
1019 if (!ctx->teedev->desc->ops->close_session)
1020 return -EINVAL;
1021 return ctx->teedev->desc->ops->close_session(ctx, session);
1022}
1023EXPORT_SYMBOL_GPL(tee_client_close_session);
1024
1025int tee_client_invoke_func(struct tee_context *ctx,
1026 struct tee_ioctl_invoke_arg *arg,
1027 struct tee_param *param)
1028{
1029 if (!ctx->teedev->desc->ops->invoke_func)
1030 return -EINVAL;
1031 return ctx->teedev->desc->ops->invoke_func(ctx, arg, param);
1032}
1033EXPORT_SYMBOL_GPL(tee_client_invoke_func);
1034
1035int tee_client_cancel_req(struct tee_context *ctx,
1036 struct tee_ioctl_cancel_arg *arg)
1037{
1038 if (!ctx->teedev->desc->ops->cancel_req)
1039 return -EINVAL;
1040 return ctx->teedev->desc->ops->cancel_req(ctx, arg->cancel_id,
1041 arg->session);
1042}
1043
1044static int tee_client_device_match(struct device *dev,
1045 struct device_driver *drv)
1046{
1047 const struct tee_client_device_id *id_table;
1048 struct tee_client_device *tee_device;
1049
1050 id_table = to_tee_client_driver(drv)->id_table;
1051 tee_device = to_tee_client_device(dev);
1052
1053 while (!uuid_is_null(&id_table->uuid)) {
1054 if (uuid_equal(&tee_device->id.uuid, &id_table->uuid))
1055 return 1;
1056 id_table++;
1057 }
1058
1059 return 0;
1060}
1061
1062static int tee_client_device_uevent(struct device *dev,
1063 struct kobj_uevent_env *env)
1064{
1065 uuid_t *dev_id = &to_tee_client_device(dev)->id.uuid;
1066
1067 return add_uevent_var(env, "MODALIAS=tee:%pUb", dev_id);
1068}
1069
1070struct bus_type tee_bus_type = {
1071 .name = "tee",
1072 .match = tee_client_device_match,
1073 .uevent = tee_client_device_uevent,
1074};
1075EXPORT_SYMBOL_GPL(tee_bus_type);
1076
1077static int __init tee_init(void)
1078{
1079 int rc;
1080
1081 tee_class = class_create(THIS_MODULE, "tee");
1082 if (IS_ERR(tee_class)) {
1083 pr_err("couldn't create class\n");
1084 return PTR_ERR(tee_class);
1085 }
1086
1087 rc = alloc_chrdev_region(&tee_devt, 0, TEE_NUM_DEVICES, "tee");
1088 if (rc) {
1089 pr_err("failed to allocate char dev region\n");
1090 goto out_unreg_class;
1091 }
1092
1093 rc = bus_register(&tee_bus_type);
1094 if (rc) {
1095 pr_err("failed to register tee bus\n");
1096 goto out_unreg_chrdev;
1097 }
1098
1099 return 0;
1100
1101out_unreg_chrdev:
1102 unregister_chrdev_region(tee_devt, TEE_NUM_DEVICES);
1103out_unreg_class:
1104 class_destroy(tee_class);
1105 tee_class = NULL;
1106
1107 return rc;
1108}
1109
1110static void __exit tee_exit(void)
1111{
1112 bus_unregister(&tee_bus_type);
1113 unregister_chrdev_region(tee_devt, TEE_NUM_DEVICES);
1114 class_destroy(tee_class);
1115 tee_class = NULL;
1116}
1117
1118subsys_initcall(tee_init);
1119module_exit(tee_exit);
1120
1121MODULE_AUTHOR("Linaro");
1122MODULE_DESCRIPTION("TEE Driver");
1123MODULE_VERSION("1.0");
1124MODULE_LICENSE("GPL v2");