Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 18244C433FE for ; Tue, 14 Dec 2021 15:00:05 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S235126AbhLNPAD (ORCPT ); Tue, 14 Dec 2021 10:00:03 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:39990 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S235119AbhLNPAB (ORCPT ); Tue, 14 Dec 2021 10:00:01 -0500 Received: from mail-lf1-x136.google.com (mail-lf1-x136.google.com [IPv6:2a00:1450:4864:20::136]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 158D0C061574 for ; Tue, 14 Dec 2021 07:00:01 -0800 (PST) Received: by mail-lf1-x136.google.com with SMTP id bi37so37350821lfb.5 for ; Tue, 14 Dec 2021 07:00:00 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=date:from:to:cc:subject:message-id:references:mime-version :content-disposition:in-reply-to; bh=q4vsNe2rRZx3//QPk8gs0Wb4qTQqXBzPfbXzdJ9Scyg=; b=DPSZmCU+4BTZCYR/yfULepHU1Vmgclf0ZS4JPafHdJ1A1hCG9arKanLZwmS8sXkjhl SPaocj8Bt580recCktsqcT4tADnRpwoPEanyluIEG1WvJt49xO/xkA58JgLUpZ9mwdgL ilyEPSaSWMu87WFAvDx3w11O45SwXZgbObDtjJilln2AB4g/cfEYcmcL46yVU6P4svN6 96SrXOGVLnhwaPaRO7WvrFXjj5p96eEsPzg/WiYCSvkIeuwpIan/MVL8h7WhDfOF2WyM dfC3Sh/GpM31PtB93pRfDunN1c6e99viOfN8/NzTDKjCcqDXT9sP8TUZbTmLmHEAqeTJ zt6g== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=x-gm-message-state:date:from:to:cc:subject:message-id:references :mime-version:content-disposition:in-reply-to; bh=q4vsNe2rRZx3//QPk8gs0Wb4qTQqXBzPfbXzdJ9Scyg=; b=7tdgDdbcFLZ1CGyA1vtvi3O0R0WqiuBH5qpzKZr1SNKAKnAqQYU8sP1sjPEIoNLS9x v05GtgJA/r1LztJbWmHhE6t4W30FdsZA34TtGUnkMEzFJGMmz/s3hN6Y83+le7by8mlz 9jleqqq6XKTI/6+FOpWqg+9w/MH4LmPTzQC3gMNvX3nBtCyrtyzi1aJCepx01qxL4Dp+ uZL0NMd8pD8VYz3fpVw8WyCSXz7mR2NHVnHM4C+0FQ2u8pRSFKKW2BQDfZXjHC44ajwL QJUkO+FMIyDWzm9YU0pQOnP1dqS71LVtIcf7YpP5GyspVQ6o7imTlk57+VjXRj6A3I8z 16yQ== X-Gm-Message-State: AOAM532mmtiAgih3MfJ9Tgn0c+8Zf10ipjtP7pUvCBV7df1obBDysbOu khAal16x3HYuycm3g8CwrBtzoQ== X-Google-Smtp-Source: ABdhPJxnGg5e+j1pNoh2sgUzjENLGN7aFj+Fwj3TpSZBL1JsD4pJ4QzS8Gh4Uva46sJiuABUa18Xxw== X-Received: by 2002:a05:6512:3c8f:: with SMTP id h15mr5168026lfv.411.1639493999240; Tue, 14 Dec 2021 06:59:59 -0800 (PST) Received: from jade (h-94-254-48-165.A175.priv.bahnhof.se. [94.254.48.165]) by smtp.gmail.com with ESMTPSA id b6sm2250ljr.103.2021.12.14.06.59.58 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Tue, 14 Dec 2021 06:59:58 -0800 (PST) Date: Tue, 14 Dec 2021 15:59:57 +0100 From: Jens Wiklander To: Greg KH Cc: linux-kernel@vger.kernel.org, op-tee@lists.trustedfirmware.org, Sumit Garg , Christian =?utf-8?B?S8O2bmln?= , Rijo Thomas , Devaraj Rangasamy , stable@vger.kernel.org, Lars Persson , Patrik Lantz Subject: Re: [PATCH] tee: handle lookup of shm with reference count 0 Message-ID: <20211214145957.GA1800868@jade> References: <20211214123540.1789434-1-jens.wiklander@linaro.org> MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Disposition: inline In-Reply-To: Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Tue, Dec 14, 2021 at 02:44:30PM +0100, Greg KH wrote: > On Tue, Dec 14, 2021 at 01:35:40PM +0100, Jens Wiklander wrote: > > Since the tee subsystem does not keep a strong reference to its idle > > shared memory buffers, it races with other threads that try to destroy a > > shared memory through a close of its dma-buf fd or by unmapping the > > memory. > > > > In tee_shm_get_from_id() when a lookup in teedev->idr has been > > successful, it is possible that the tee_shm is in the dma-buf teardown > > path, but that path is blocked by the teedev mutex. Since we don't have > > an API to tell if the tee_shm is in the dma-buf teardown path or not we > > must find another way of detecting this condition. > > > > Fix this by doing the reference counting directly on the tee_shm using a > > new refcount_t refcount field. dma-buf is replaced by using > > anon_inode_getfd() instead, this separates the life-cycle of the > > underlying file from the tee_shm. tee_shm_put() is updated to hold the > > mutex when decreasing the refcount to 0 and then remove the tee_shm from > > teedev->idr before releasing the mutex. This means that the tee_shm can > > never be found unless it has a refcount larger than 0. > > So you are dropping dma-buf support entirely? And anon_inode_getfd() > works instead? Why do more people not do this as well? I don't know, but it should be noted that we're not doing very much with this file descriptor. We're only using it with mmap() and close(). > > > > > Fixes: 967c9cca2cc5 ("tee: generic TEE subsystem") > > Cc: stable@vger.kernel.org > > Reviewed-by: Lars Persson > > Reviewed-by: Sumit Garg > > Reported-by: Patrik Lantz > > Signed-off-by: Jens Wiklander > > --- > > drivers/tee/tee_shm.c | 174 +++++++++++++++------------------------- > > include/linux/tee_drv.h | 2 +- > > 2 files changed, 67 insertions(+), 109 deletions(-) > > > > diff --git a/drivers/tee/tee_shm.c b/drivers/tee/tee_shm.c > > index 8a8deb95e918..0c82cf981c46 100644 > > --- a/drivers/tee/tee_shm.c > > +++ b/drivers/tee/tee_shm.c > > @@ -1,20 +1,17 @@ > > // SPDX-License-Identifier: GPL-2.0-only > > /* > > - * Copyright (c) 2015-2016, Linaro Limited > > + * Copyright (c) 2015-2021, Linaro Limited > > Nit, did Linaro really make a copyrightable change in 2017, 2018, 2019 > and 2020 as well? If not, please do not claim it. Fair enough, I was a bit lazy 2018 shouldn't be there now that I've checked the log. I'll fix. > > > */ > > +#include > > #include > > -#include > > -#include > > #include > > +#include > > #include > > #include > > #include > > #include > > -#include > > #include "tee_private.h" > > > > -MODULE_IMPORT_NS(DMA_BUF); > > - > > static void release_registered_pages(struct tee_shm *shm) > > { > > if (shm->pages) { > > @@ -31,16 +28,8 @@ static void release_registered_pages(struct tee_shm *shm) > > } > > } > > > > -static void tee_shm_release(struct tee_shm *shm) > > +static void tee_shm_release(struct tee_device *teedev, struct tee_shm *shm) > > { > > - struct tee_device *teedev = shm->ctx->teedev; > > - > > - if (shm->flags & TEE_SHM_DMA_BUF) { > > - mutex_lock(&teedev->mutex); > > - idr_remove(&teedev->idr, shm->id); > > - mutex_unlock(&teedev->mutex); > > - } > > - > > if (shm->flags & TEE_SHM_POOL) { > > struct tee_shm_pool_mgr *poolm; > > > > @@ -67,45 +56,6 @@ static void tee_shm_release(struct tee_shm *shm) > > tee_device_put(teedev); > > } > > > > -static struct sg_table *tee_shm_op_map_dma_buf(struct dma_buf_attachment > > - *attach, enum dma_data_direction dir) > > -{ > > - return NULL; > > -} > > - > > -static void tee_shm_op_unmap_dma_buf(struct dma_buf_attachment *attach, > > - struct sg_table *table, > > - enum dma_data_direction dir) > > -{ > > -} > > - > > -static void tee_shm_op_release(struct dma_buf *dmabuf) > > -{ > > - struct tee_shm *shm = dmabuf->priv; > > - > > - tee_shm_release(shm); > > -} > > - > > -static int tee_shm_op_mmap(struct dma_buf *dmabuf, struct vm_area_struct *vma) > > -{ > > - struct tee_shm *shm = dmabuf->priv; > > - size_t size = vma->vm_end - vma->vm_start; > > - > > - /* Refuse sharing shared memory provided by application */ > > - if (shm->flags & TEE_SHM_USER_MAPPED) > > - return -EINVAL; > > - > > - return remap_pfn_range(vma, vma->vm_start, shm->paddr >> PAGE_SHIFT, > > - size, vma->vm_page_prot); > > -} > > - > > -static const struct dma_buf_ops tee_shm_dma_buf_ops = { > > - .map_dma_buf = tee_shm_op_map_dma_buf, > > - .unmap_dma_buf = tee_shm_op_unmap_dma_buf, > > - .release = tee_shm_op_release, > > - .mmap = tee_shm_op_mmap, > > -}; > > - > > struct tee_shm *tee_shm_alloc(struct tee_context *ctx, size_t size, u32 flags) > > { > > struct tee_device *teedev = ctx->teedev; > > @@ -140,6 +90,7 @@ struct tee_shm *tee_shm_alloc(struct tee_context *ctx, size_t size, u32 flags) > > goto err_dev_put; > > } > > > > + refcount_set(&shm->refcount, 1); > > shm->flags = flags | TEE_SHM_POOL; > > shm->ctx = ctx; > > if (flags & TEE_SHM_DMA_BUF) > > @@ -153,10 +104,7 @@ struct tee_shm *tee_shm_alloc(struct tee_context *ctx, size_t size, u32 flags) > > goto err_kfree; > > } > > > > - > > if (flags & TEE_SHM_DMA_BUF) { > > - DEFINE_DMA_BUF_EXPORT_INFO(exp_info); > > - > > mutex_lock(&teedev->mutex); > > shm->id = idr_alloc(&teedev->idr, shm, 1, 0, GFP_KERNEL); > > mutex_unlock(&teedev->mutex); > > @@ -164,28 +112,11 @@ struct tee_shm *tee_shm_alloc(struct tee_context *ctx, size_t size, u32 flags) > > ret = ERR_PTR(shm->id); > > goto err_pool_free; > > } > > - > > - exp_info.ops = &tee_shm_dma_buf_ops; > > - exp_info.size = shm->size; > > - exp_info.flags = O_RDWR; > > - exp_info.priv = shm; > > - > > - shm->dmabuf = dma_buf_export(&exp_info); > > - if (IS_ERR(shm->dmabuf)) { > > - ret = ERR_CAST(shm->dmabuf); > > - goto err_rem; > > - } > > } > > > > teedev_ctx_get(ctx); > > > > return shm; > > -err_rem: > > - if (flags & TEE_SHM_DMA_BUF) { > > - mutex_lock(&teedev->mutex); > > - idr_remove(&teedev->idr, shm->id); > > - mutex_unlock(&teedev->mutex); > > - } > > err_pool_free: > > poolm->ops->free(poolm, shm); > > err_kfree: > > @@ -246,6 +177,7 @@ struct tee_shm *tee_shm_register(struct tee_context *ctx, unsigned long addr, > > goto err; > > } > > > > + refcount_set(&shm->refcount, 1); > > shm->flags = flags | TEE_SHM_REGISTER; > > shm->ctx = ctx; > > shm->id = -1; > > @@ -306,22 +238,6 @@ struct tee_shm *tee_shm_register(struct tee_context *ctx, unsigned long addr, > > goto err; > > } > > > > - if (flags & TEE_SHM_DMA_BUF) { > > - DEFINE_DMA_BUF_EXPORT_INFO(exp_info); > > - > > - exp_info.ops = &tee_shm_dma_buf_ops; > > - exp_info.size = shm->size; > > - exp_info.flags = O_RDWR; > > - exp_info.priv = shm; > > - > > - shm->dmabuf = dma_buf_export(&exp_info); > > - if (IS_ERR(shm->dmabuf)) { > > - ret = ERR_CAST(shm->dmabuf); > > - teedev->desc->ops->shm_unregister(ctx, shm); > > - goto err; > > - } > > - } > > - > > return shm; > > err: > > if (shm) { > > @@ -339,6 +255,35 @@ struct tee_shm *tee_shm_register(struct tee_context *ctx, unsigned long addr, > > } > > EXPORT_SYMBOL_GPL(tee_shm_register); > > > > +static int tee_shm_fop_release(struct inode *inode, struct file *filp) > > +{ > > + tee_shm_put(filp->private_data); > > + return 0; > > +} > > + > > +static int tee_shm_fop_mmap(struct file *filp, struct vm_area_struct *vma) > > +{ > > + struct tee_shm *shm = filp->private_data; > > + size_t size = vma->vm_end - vma->vm_start; > > + > > + /* Refuse sharing shared memory provided by application */ > > + if (shm->flags & TEE_SHM_USER_MAPPED) > > + return -EINVAL; > > + > > + /* check for overflowing the buffer's size */ > > + if (vma->vm_pgoff + vma_pages(vma) > shm->size >> PAGE_SHIFT) > > + return -EINVAL; > > + > > + return remap_pfn_range(vma, vma->vm_start, shm->paddr >> PAGE_SHIFT, > > + size, vma->vm_page_prot); > > +} > > + > > +static const struct file_operations tee_shm_fops = { > > + .owner = THIS_MODULE, > > + .release = tee_shm_fop_release, > > + .mmap = tee_shm_fop_mmap, > > +}; > > + > > /** > > * tee_shm_get_fd() - Increase reference count and return file descriptor > > * @shm: Shared memory handle > > @@ -351,10 +296,11 @@ int tee_shm_get_fd(struct tee_shm *shm) > > if (!(shm->flags & TEE_SHM_DMA_BUF)) > > return -EINVAL; > > > > - get_dma_buf(shm->dmabuf); > > - fd = dma_buf_fd(shm->dmabuf, O_CLOEXEC); > > + /* matched by tee_shm_put() in tee_shm_op_release() */ > > + refcount_inc(&shm->refcount); > > + fd = anon_inode_getfd("tee_shm", &tee_shm_fops, shm, O_RDWR); > > if (fd < 0) > > - dma_buf_put(shm->dmabuf); > > + tee_shm_put(shm); > > return fd; > > } > > > > @@ -364,17 +310,7 @@ int tee_shm_get_fd(struct tee_shm *shm) > > */ > > void tee_shm_free(struct tee_shm *shm) > > { > > - /* > > - * dma_buf_put() decreases the dmabuf reference counter and will > > - * call tee_shm_release() when the last reference is gone. > > - * > > - * In the case of driver private memory we call tee_shm_release > > - * directly instead as it doesn't have a reference counter. > > - */ > > - if (shm->flags & TEE_SHM_DMA_BUF) > > - dma_buf_put(shm->dmabuf); > > - else > > - tee_shm_release(shm); > > + tee_shm_put(shm); > > } > > EXPORT_SYMBOL_GPL(tee_shm_free); > > > > @@ -481,10 +417,15 @@ struct tee_shm *tee_shm_get_from_id(struct tee_context *ctx, int id) > > teedev = ctx->teedev; > > mutex_lock(&teedev->mutex); > > shm = idr_find(&teedev->idr, id); > > + /* > > + * If the tee_shm was found in the IDR it must have a refcount > > + * larger than 0 due to the guarantee in tee_shm_put() below. So > > + * it's safe to use refcount_inc(). > > + */ > > if (!shm || shm->ctx != ctx) > > shm = ERR_PTR(-EINVAL); > > - else if (shm->flags & TEE_SHM_DMA_BUF) > > - get_dma_buf(shm->dmabuf); > > + else > > + refcount_inc(&shm->refcount); > > mutex_unlock(&teedev->mutex); > > return shm; > > } > > @@ -496,7 +437,24 @@ EXPORT_SYMBOL_GPL(tee_shm_get_from_id); > > */ > > void tee_shm_put(struct tee_shm *shm) > > { > > - if (shm->flags & TEE_SHM_DMA_BUF) > > - dma_buf_put(shm->dmabuf); > > + struct tee_device *teedev = shm->ctx->teedev; > > + bool do_release = false; > > + > > + mutex_lock(&teedev->mutex); > > + if (refcount_dec_and_test(&shm->refcount)) { > > + /* > > + * refcount has reached 0, we must now remove it from the > > + * IDR before releasing the mutex. This will guarantee that > > + * the refcount_inc() in tee_shm_get_from_id() never starts > > + * from 0. > > + */ > > + if (shm->flags & TEE_SHM_DMA_BUF) > > + idr_remove(&teedev->idr, shm->id); > > + do_release = true; > > As you are using a refcount in the "traditional" way, why not just use a > kref instead? That solves your "do_release" mess here. Yes, but it adds another problem. I don't want to hold the mutex when calling tee_shm_release() so that would mean moving idr_remove() to tee_shm_release() again and then use kref_get_unless_zero() in tee_shm_get_from_id() instead. With this approach the tee_shm is removed from the IDR so it cannot be seen any longer when the refcount is 0. I tried implementing it in both ways before and in my opinion this turned out better. Thanks, Jens > > thanks, > > greg k-h