|  | /* | 
|  | * Copyright (c) 2015, Linaro Limited | 
|  | * | 
|  | * This software is licensed under the terms of the GNU General Public | 
|  | * License version 2, as published by the Free Software Foundation, and | 
|  | * may be copied, distributed, and modified under those terms. | 
|  | * | 
|  | * This program is distributed in the hope that it will be useful, | 
|  | * but WITHOUT ANY WARRANTY; without even the implied warranty of | 
|  | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the | 
|  | * GNU General Public License for more details. | 
|  | * | 
|  | */ | 
|  | #include <linux/device.h> | 
|  | #include <linux/slab.h> | 
|  | #include <linux/uaccess.h> | 
|  | #include "optee_private.h" | 
|  |  | 
|  | struct optee_supp_req { | 
|  | struct list_head link; | 
|  |  | 
|  | bool in_queue; | 
|  | u32 func; | 
|  | u32 ret; | 
|  | size_t num_params; | 
|  | struct tee_param *param; | 
|  |  | 
|  | struct completion c; | 
|  | }; | 
|  |  | 
|  | void optee_supp_init(struct optee_supp *supp) | 
|  | { | 
|  | memset(supp, 0, sizeof(*supp)); | 
|  | mutex_init(&supp->mutex); | 
|  | init_completion(&supp->reqs_c); | 
|  | idr_init(&supp->idr); | 
|  | INIT_LIST_HEAD(&supp->reqs); | 
|  | supp->req_id = -1; | 
|  | } | 
|  |  | 
|  | void optee_supp_uninit(struct optee_supp *supp) | 
|  | { | 
|  | mutex_destroy(&supp->mutex); | 
|  | idr_destroy(&supp->idr); | 
|  | } | 
|  |  | 
|  | void optee_supp_release(struct optee_supp *supp) | 
|  | { | 
|  | int id; | 
|  | struct optee_supp_req *req; | 
|  | struct optee_supp_req *req_tmp; | 
|  |  | 
|  | mutex_lock(&supp->mutex); | 
|  |  | 
|  | /* Abort all request retrieved by supplicant */ | 
|  | idr_for_each_entry(&supp->idr, req, id) { | 
|  | idr_remove(&supp->idr, id); | 
|  | req->ret = TEEC_ERROR_COMMUNICATION; | 
|  | complete(&req->c); | 
|  | } | 
|  |  | 
|  | /* Abort all queued requests */ | 
|  | list_for_each_entry_safe(req, req_tmp, &supp->reqs, link) { | 
|  | list_del(&req->link); | 
|  | req->in_queue = false; | 
|  | req->ret = TEEC_ERROR_COMMUNICATION; | 
|  | complete(&req->c); | 
|  | } | 
|  |  | 
|  | supp->ctx = NULL; | 
|  | supp->req_id = -1; | 
|  |  | 
|  | mutex_unlock(&supp->mutex); | 
|  | } | 
|  |  | 
|  | /** | 
|  | * optee_supp_thrd_req() - request service from supplicant | 
|  | * @ctx:	context doing the request | 
|  | * @func:	function requested | 
|  | * @num_params:	number of elements in @param array | 
|  | * @param:	parameters for function | 
|  | * | 
|  | * Returns result of operation to be passed to secure world | 
|  | */ | 
|  | u32 optee_supp_thrd_req(struct tee_context *ctx, u32 func, size_t num_params, | 
|  | struct tee_param *param) | 
|  |  | 
|  | { | 
|  | struct optee *optee = tee_get_drvdata(ctx->teedev); | 
|  | struct optee_supp *supp = &optee->supp; | 
|  | struct optee_supp_req *req = kzalloc(sizeof(*req), GFP_KERNEL); | 
|  | bool interruptable; | 
|  | u32 ret; | 
|  |  | 
|  | if (!req) | 
|  | return TEEC_ERROR_OUT_OF_MEMORY; | 
|  |  | 
|  | init_completion(&req->c); | 
|  | req->func = func; | 
|  | req->num_params = num_params; | 
|  | req->param = param; | 
|  |  | 
|  | /* Insert the request in the request list */ | 
|  | mutex_lock(&supp->mutex); | 
|  | list_add_tail(&req->link, &supp->reqs); | 
|  | req->in_queue = true; | 
|  | mutex_unlock(&supp->mutex); | 
|  |  | 
|  | /* Tell an eventual waiter there's a new request */ | 
|  | complete(&supp->reqs_c); | 
|  |  | 
|  | /* | 
|  | * Wait for supplicant to process and return result, once we've | 
|  | * returned from wait_for_completion(&req->c) successfully we have | 
|  | * exclusive access again. | 
|  | */ | 
|  | while (wait_for_completion_interruptible(&req->c)) { | 
|  | mutex_lock(&supp->mutex); | 
|  | interruptable = !supp->ctx; | 
|  | if (interruptable) { | 
|  | /* | 
|  | * There's no supplicant available and since the | 
|  | * supp->mutex currently is held none can | 
|  | * become available until the mutex released | 
|  | * again. | 
|  | * | 
|  | * Interrupting an RPC to supplicant is only | 
|  | * allowed as a way of slightly improving the user | 
|  | * experience in case the supplicant hasn't been | 
|  | * started yet. During normal operation the supplicant | 
|  | * will serve all requests in a timely manner and | 
|  | * interrupting then wouldn't make sense. | 
|  | */ | 
|  | if (req->in_queue) { | 
|  | list_del(&req->link); | 
|  | req->in_queue = false; | 
|  | } | 
|  | } | 
|  | mutex_unlock(&supp->mutex); | 
|  |  | 
|  | if (interruptable) { | 
|  | req->ret = TEEC_ERROR_COMMUNICATION; | 
|  | break; | 
|  | } | 
|  | } | 
|  |  | 
|  | ret = req->ret; | 
|  | kfree(req); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static struct optee_supp_req  *supp_pop_entry(struct optee_supp *supp, | 
|  | int num_params, int *id) | 
|  | { | 
|  | struct optee_supp_req *req; | 
|  |  | 
|  | if (supp->req_id != -1) { | 
|  | /* | 
|  | * Supplicant should not mix synchronous and asnynchronous | 
|  | * requests. | 
|  | */ | 
|  | return ERR_PTR(-EINVAL); | 
|  | } | 
|  |  | 
|  | if (list_empty(&supp->reqs)) | 
|  | return NULL; | 
|  |  | 
|  | req = list_first_entry(&supp->reqs, struct optee_supp_req, link); | 
|  |  | 
|  | if (num_params < req->num_params) { | 
|  | /* Not enough room for parameters */ | 
|  | return ERR_PTR(-EINVAL); | 
|  | } | 
|  |  | 
|  | *id = idr_alloc(&supp->idr, req, 1, 0, GFP_KERNEL); | 
|  | if (*id < 0) | 
|  | return ERR_PTR(-ENOMEM); | 
|  |  | 
|  | list_del(&req->link); | 
|  | req->in_queue = false; | 
|  |  | 
|  | return req; | 
|  | } | 
|  |  | 
|  | static int supp_check_recv_params(size_t num_params, struct tee_param *params, | 
|  | size_t *num_meta) | 
|  | { | 
|  | size_t n; | 
|  |  | 
|  | if (!num_params) | 
|  | return -EINVAL; | 
|  |  | 
|  | /* | 
|  | * If there's memrefs we need to decrease those as they where | 
|  | * increased earlier and we'll even refuse to accept any below. | 
|  | */ | 
|  | for (n = 0; n < num_params; n++) | 
|  | if (tee_param_is_memref(params + n) && params[n].u.memref.shm) | 
|  | tee_shm_put(params[n].u.memref.shm); | 
|  |  | 
|  | /* | 
|  | * We only expect parameters as TEE_IOCTL_PARAM_ATTR_TYPE_NONE with | 
|  | * or without the TEE_IOCTL_PARAM_ATTR_META bit set. | 
|  | */ | 
|  | for (n = 0; n < num_params; n++) | 
|  | if (params[n].attr && | 
|  | params[n].attr != TEE_IOCTL_PARAM_ATTR_META) | 
|  | return -EINVAL; | 
|  |  | 
|  | /* At most we'll need one meta parameter so no need to check for more */ | 
|  | if (params->attr == TEE_IOCTL_PARAM_ATTR_META) | 
|  | *num_meta = 1; | 
|  | else | 
|  | *num_meta = 0; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /** | 
|  | * optee_supp_recv() - receive request for supplicant | 
|  | * @ctx:	context receiving the request | 
|  | * @func:	requested function in supplicant | 
|  | * @num_params:	number of elements allocated in @param, updated with number | 
|  | *		used elements | 
|  | * @param:	space for parameters for @func | 
|  | * | 
|  | * Returns 0 on success or <0 on failure | 
|  | */ | 
|  | int optee_supp_recv(struct tee_context *ctx, u32 *func, u32 *num_params, | 
|  | struct tee_param *param) | 
|  | { | 
|  | struct tee_device *teedev = ctx->teedev; | 
|  | struct optee *optee = tee_get_drvdata(teedev); | 
|  | struct optee_supp *supp = &optee->supp; | 
|  | struct optee_supp_req *req = NULL; | 
|  | int id; | 
|  | size_t num_meta; | 
|  | int rc; | 
|  |  | 
|  | rc = supp_check_recv_params(*num_params, param, &num_meta); | 
|  | if (rc) | 
|  | return rc; | 
|  |  | 
|  | while (true) { | 
|  | mutex_lock(&supp->mutex); | 
|  | req = supp_pop_entry(supp, *num_params - num_meta, &id); | 
|  | mutex_unlock(&supp->mutex); | 
|  |  | 
|  | if (req) { | 
|  | if (IS_ERR(req)) | 
|  | return PTR_ERR(req); | 
|  | break; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * If we didn't get a request we'll block in | 
|  | * wait_for_completion() to avoid needless spinning. | 
|  | * | 
|  | * This is where supplicant will be hanging most of | 
|  | * the time, let's make this interruptable so we | 
|  | * can easily restart supplicant if needed. | 
|  | */ | 
|  | if (wait_for_completion_interruptible(&supp->reqs_c)) | 
|  | return -ERESTARTSYS; | 
|  | } | 
|  |  | 
|  | if (num_meta) { | 
|  | /* | 
|  | * tee-supplicant support meta parameters -> requsts can be | 
|  | * processed asynchronously. | 
|  | */ | 
|  | param->attr = TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INOUT | | 
|  | TEE_IOCTL_PARAM_ATTR_META; | 
|  | param->u.value.a = id; | 
|  | param->u.value.b = 0; | 
|  | param->u.value.c = 0; | 
|  | } else { | 
|  | mutex_lock(&supp->mutex); | 
|  | supp->req_id = id; | 
|  | mutex_unlock(&supp->mutex); | 
|  | } | 
|  |  | 
|  | *func = req->func; | 
|  | *num_params = req->num_params + num_meta; | 
|  | memcpy(param + num_meta, req->param, | 
|  | sizeof(struct tee_param) * req->num_params); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static struct optee_supp_req *supp_pop_req(struct optee_supp *supp, | 
|  | size_t num_params, | 
|  | struct tee_param *param, | 
|  | size_t *num_meta) | 
|  | { | 
|  | struct optee_supp_req *req; | 
|  | int id; | 
|  | size_t nm; | 
|  | const u32 attr = TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INOUT | | 
|  | TEE_IOCTL_PARAM_ATTR_META; | 
|  |  | 
|  | if (!num_params) | 
|  | return ERR_PTR(-EINVAL); | 
|  |  | 
|  | if (supp->req_id == -1) { | 
|  | if (param->attr != attr) | 
|  | return ERR_PTR(-EINVAL); | 
|  | id = param->u.value.a; | 
|  | nm = 1; | 
|  | } else { | 
|  | id = supp->req_id; | 
|  | nm = 0; | 
|  | } | 
|  |  | 
|  | req = idr_find(&supp->idr, id); | 
|  | if (!req) | 
|  | return ERR_PTR(-ENOENT); | 
|  |  | 
|  | if ((num_params - nm) != req->num_params) | 
|  | return ERR_PTR(-EINVAL); | 
|  |  | 
|  | idr_remove(&supp->idr, id); | 
|  | supp->req_id = -1; | 
|  | *num_meta = nm; | 
|  |  | 
|  | return req; | 
|  | } | 
|  |  | 
|  | /** | 
|  | * optee_supp_send() - send result of request from supplicant | 
|  | * @ctx:	context sending result | 
|  | * @ret:	return value of request | 
|  | * @num_params:	number of parameters returned | 
|  | * @param:	returned parameters | 
|  | * | 
|  | * Returns 0 on success or <0 on failure. | 
|  | */ | 
|  | int optee_supp_send(struct tee_context *ctx, u32 ret, u32 num_params, | 
|  | struct tee_param *param) | 
|  | { | 
|  | struct tee_device *teedev = ctx->teedev; | 
|  | struct optee *optee = tee_get_drvdata(teedev); | 
|  | struct optee_supp *supp = &optee->supp; | 
|  | struct optee_supp_req *req; | 
|  | size_t n; | 
|  | size_t num_meta; | 
|  |  | 
|  | mutex_lock(&supp->mutex); | 
|  | req = supp_pop_req(supp, num_params, param, &num_meta); | 
|  | mutex_unlock(&supp->mutex); | 
|  |  | 
|  | if (IS_ERR(req)) { | 
|  | /* Something is wrong, let supplicant restart. */ | 
|  | return PTR_ERR(req); | 
|  | } | 
|  |  | 
|  | /* Update out and in/out parameters */ | 
|  | for (n = 0; n < req->num_params; n++) { | 
|  | struct tee_param *p = req->param + n; | 
|  |  | 
|  | switch (p->attr & TEE_IOCTL_PARAM_ATTR_TYPE_MASK) { | 
|  | case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_OUTPUT: | 
|  | case TEE_IOCTL_PARAM_ATTR_TYPE_VALUE_INOUT: | 
|  | p->u.value.a = param[n + num_meta].u.value.a; | 
|  | p->u.value.b = param[n + num_meta].u.value.b; | 
|  | p->u.value.c = param[n + num_meta].u.value.c; | 
|  | break; | 
|  | case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_OUTPUT: | 
|  | case TEE_IOCTL_PARAM_ATTR_TYPE_MEMREF_INOUT: | 
|  | p->u.memref.size = param[n + num_meta].u.memref.size; | 
|  | break; | 
|  | default: | 
|  | break; | 
|  | } | 
|  | } | 
|  | req->ret = ret; | 
|  |  | 
|  | /* Let the requesting thread continue */ | 
|  | complete(&req->c); | 
|  |  | 
|  | return 0; | 
|  | } |