Received: by 2002:a05:6a10:22f:0:0:0:0 with SMTP id 15csp3352205pxk; Mon, 7 Sep 2020 10:21:21 -0700 (PDT) X-Google-Smtp-Source: ABdhPJyolk2lai8tDMWkt3JCYggEKnPING5XTvMbFxlyvmeO3VewgGAs3Gf1YrxkJaSxvIyUhRKQ X-Received: by 2002:a17:906:6b95:: with SMTP id l21mr21459077ejr.317.1599499281677; Mon, 07 Sep 2020 10:21:21 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1599499281; cv=none; d=google.com; s=arc-20160816; b=hNp4EjTUQ7GJXUod5NeswZZY+WLfQSnBOmtCZ1a7Yd1xiKZz4JF54Trb6hZiocVchi gcFTmE9p5xPXxk8qk8MKxHUXnAfR8IcH+m7Uo8/EdYtnsjQMpaAsid/UfAJcjF5n3i3T E6m7ouE3ivyY3vXeXAFJdvLMbCgGZyJguS4jDnS9AkCnfNO073NgtyIJ/ZgLgF3cSTXK FAPt56yaULpFd3oWG3/mfvxwTQ5TwDXRijlb2cjjtQVLqslY7bV+Ss/yZ7AcmPr8M5sT NW2NIEZMWHCQxdSLqMhYmp7SU5TaH+ptsfT42aBqghZkFi1AZoTExtxUnaDmnn/KIuTz ZeHw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:dkim-signature; bh=6ycAA+DdU1xPs+4BPRjymvkLu6X0RXJu9Y30n9A+nhw=; b=FCL2TnvN/7m/NYKCTFF2m1f4kQWJXHUv3CWG6C8sP9Y8aVjv6tQ/gW8owuTp8INx5p K9CE9ZwaMrcvvmwjxlU5rhmsVgXwmJDv8zVYq7gWzyuI8Dyxe2e0C56P/Uy1/hDFURCU faOLbo8mJRNxxJYJuKS4vgYvP9iz0uenyCKw3PQYAjiACT3/maE4G75RdFcCUIBLsRLs 9QV3TPHjaB+NBN+HIpmQTB+VWL3w7fuawfUd6nfXrlfPY7+TPEyvCRS813Fe4MU0N4jr KoYV6EAGdlk+QIw/XjxcIcNRNsy3ZiyrbsN0g30uNCd33DIGreBfxO9yE190aY/GtAJF iDgA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@ffwll.ch header.s=google header.b="iEuCxz9/"; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id z3si10119148ede.201.2020.09.07.10.20.59; Mon, 07 Sep 2020 10:21:21 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@ffwll.ch header.s=google header.b="iEuCxz9/"; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1730795AbgIGRUP (ORCPT + 99 others); Mon, 7 Sep 2020 13:20:15 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:37526 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1729612AbgIGNxU (ORCPT ); Mon, 7 Sep 2020 09:53:20 -0400 Received: from mail-ot1-x343.google.com (mail-ot1-x343.google.com [IPv6:2607:f8b0:4864:20::343]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 7CB7AC061757 for ; Mon, 7 Sep 2020 06:53:15 -0700 (PDT) Received: by mail-ot1-x343.google.com with SMTP id 37so12330299oto.4 for ; Mon, 07 Sep 2020 06:53:15 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=ffwll.ch; s=google; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=6ycAA+DdU1xPs+4BPRjymvkLu6X0RXJu9Y30n9A+nhw=; b=iEuCxz9/GZ6l217tQFZNQnUWpR9k7WJUgYHDT3dvu6LGpJUTO3ZHU8XY45NhdzKDQO 1fKDlDTQJUBRg69fGPFS9NGm88rxiua5V6kLDIHzvXJQl2XTqcbaxMJ2A3jRMSNE3KpM gljXkMSlOUveuv/7r6jtHMbsJfGshsRGGdmd0= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=6ycAA+DdU1xPs+4BPRjymvkLu6X0RXJu9Y30n9A+nhw=; b=cpgoOR/CKOwGvahWPaKLr4Q3jN1eE2FKBD3YFnHUuVCc1IKd6bhoPOn1xouTZIsEwc MLWXOPZ0Fn82YhM2f4X/GcE2dBf6HazmzQps/YTcYn5ZXQPBWvoF6upvvnbezBAD1Kwu 0oGfLMm3zj9I/FNDxEBvJInHS6JZiTKrMlqv1VnQBglKFwCXIx8fUbZ+bAKRmLranecf zO/179q3sq1SAGhbbDXFTHnxcVISxSQQWb3Ok8ZJ+3asknt3qV9ASDnAGDxpDdxnTEpX r2HOG1uS1FjErIDRAD/KzCyTR/Oi9gLkIwyGVP7xLoQYysXs1+0bkSj64fFK1tJGar2l cyQQ== X-Gm-Message-State: AOAM530EeaG0IBXfdWiNLUPrRw1HBTkCpF+eMPddUcnpl5+cpXxBt5jW UCNTEVG2oEXGHqYBA2V3DmeKNduD0VJ8DQkDELTqjA== X-Received: by 2002:a05:6830:14d9:: with SMTP id t25mr14697170otq.188.1599486793337; Mon, 07 Sep 2020 06:53:13 -0700 (PDT) MIME-Version: 1.0 References: <20200907112425.15610-1-kraxel@redhat.com> <20200907112425.15610-2-kraxel@redhat.com> In-Reply-To: <20200907112425.15610-2-kraxel@redhat.com> From: Daniel Vetter Date: Mon, 7 Sep 2020 15:53:02 +0200 Message-ID: Subject: Re: [PATCH v4 1/1] drm: allow limiting the scatter list size. To: Gerd Hoffmann Cc: dri-devel , =?UTF-8?Q?Christian_K=C3=B6nig?= , Alex Deucher , David Airlie , Maarten Lankhorst , Maxime Ripard , Thomas Zimmermann , Lucas Stach , Russell King , Christian Gmeiner , Rob Clark , Sean Paul , Ben Skeggs , Sandy Huang , =?UTF-8?Q?Heiko_St=C3=BCbner?= , Thierry Reding , Jonathan Hunter , Oleksandr Andrushchenko , "open list:RADEON and AMDGPU DRM DRIVERS" , open list , "moderated list:DRM DRIVERS FOR VIVANTE GPU IP" , "open list:DRM DRIVER FOR MSM ADRENO GPU" , "open list:DRM DRIVER FOR MSM ADRENO GPU" , "open list:DRM DRIVER FOR NVIDIA GEFORCE/QUADRO GPUS" , "moderated list:ARM/Rockchip SoC support" , "open list:ARM/Rockchip SoC support" , "open list:DRM DRIVERS FOR NVIDIA TEGRA" , "moderated list:DRM DRIVERS FOR XEN" Content-Type: text/plain; charset="UTF-8" Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon, Sep 7, 2020 at 1:24 PM Gerd Hoffmann wrote: > > Add drm_device argument to drm_prime_pages_to_sg(), so we can > call dma_max_mapping_size() to figure the segment size limit > and call into __sg_alloc_table_from_pages() with the correct > limit. > > This fixes virtio-gpu with sev. Possibly it'll fix other bugs > too given that drm seems to totaly ignore segment size limits > so far ... > > v2: place max_segment in drm driver not gem object. > v3: move max_segment next to the other gem fields. > v4: just use dma_max_mapping_size(). > > Signed-off-by: Gerd Hoffmann Uh, are you sure this works in all cases for virtio? The comments I've found suggest very much not ... Or is that all very old stuff only that no one cares about anymore? -Daniel > --- > include/drm/drm_prime.h | 3 ++- > drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c | 3 ++- > drivers/gpu/drm/drm_gem_shmem_helper.c | 2 +- > drivers/gpu/drm/drm_prime.c | 13 ++++++++++--- > drivers/gpu/drm/etnaviv/etnaviv_gem.c | 3 ++- > drivers/gpu/drm/etnaviv/etnaviv_gem_prime.c | 2 +- > drivers/gpu/drm/msm/msm_gem.c | 2 +- > drivers/gpu/drm/msm/msm_gem_prime.c | 2 +- > drivers/gpu/drm/nouveau/nouveau_prime.c | 2 +- > drivers/gpu/drm/radeon/radeon_prime.c | 2 +- > drivers/gpu/drm/rockchip/rockchip_drm_gem.c | 5 +++-- > drivers/gpu/drm/tegra/gem.c | 2 +- > drivers/gpu/drm/vgem/vgem_drv.c | 2 +- > drivers/gpu/drm/xen/xen_drm_front_gem.c | 3 ++- > 14 files changed, 29 insertions(+), 17 deletions(-) > > diff --git a/include/drm/drm_prime.h b/include/drm/drm_prime.h > index 9af7422b44cf..bf141e74a1c2 100644 > --- a/include/drm/drm_prime.h > +++ b/include/drm/drm_prime.h > @@ -88,7 +88,8 @@ void drm_gem_dmabuf_vunmap(struct dma_buf *dma_buf, void *vaddr); > int drm_gem_prime_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma); > int drm_gem_dmabuf_mmap(struct dma_buf *dma_buf, struct vm_area_struct *vma); > > -struct sg_table *drm_prime_pages_to_sg(struct page **pages, unsigned int nr_pages); > +struct sg_table *drm_prime_pages_to_sg(struct drm_device *dev, > + struct page **pages, unsigned int nr_pages); > struct dma_buf *drm_gem_prime_export(struct drm_gem_object *obj, > int flags); > > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c > index 519ce4427fce..d7050ab95946 100644 > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c > @@ -302,7 +302,8 @@ static struct sg_table *amdgpu_dma_buf_map(struct dma_buf_attachment *attach, > > switch (bo->tbo.mem.mem_type) { > case TTM_PL_TT: > - sgt = drm_prime_pages_to_sg(bo->tbo.ttm->pages, > + sgt = drm_prime_pages_to_sg(obj->dev, > + bo->tbo.ttm->pages, > bo->tbo.num_pages); > if (IS_ERR(sgt)) > return sgt; > diff --git a/drivers/gpu/drm/drm_gem_shmem_helper.c b/drivers/gpu/drm/drm_gem_shmem_helper.c > index 4b7cfbac4daa..0a952f27c184 100644 > --- a/drivers/gpu/drm/drm_gem_shmem_helper.c > +++ b/drivers/gpu/drm/drm_gem_shmem_helper.c > @@ -656,7 +656,7 @@ struct sg_table *drm_gem_shmem_get_sg_table(struct drm_gem_object *obj) > > WARN_ON(shmem->base.import_attach); > > - return drm_prime_pages_to_sg(shmem->pages, obj->size >> PAGE_SHIFT); > + return drm_prime_pages_to_sg(obj->dev, shmem->pages, obj->size >> PAGE_SHIFT); > } > EXPORT_SYMBOL_GPL(drm_gem_shmem_get_sg_table); > > diff --git a/drivers/gpu/drm/drm_prime.c b/drivers/gpu/drm/drm_prime.c > index 1693aa7c14b5..8a6a3c99b7d8 100644 > --- a/drivers/gpu/drm/drm_prime.c > +++ b/drivers/gpu/drm/drm_prime.c > @@ -802,9 +802,11 @@ static const struct dma_buf_ops drm_gem_prime_dmabuf_ops = { > * > * This is useful for implementing &drm_gem_object_funcs.get_sg_table. > */ > -struct sg_table *drm_prime_pages_to_sg(struct page **pages, unsigned int nr_pages) > +struct sg_table *drm_prime_pages_to_sg(struct drm_device *dev, > + struct page **pages, unsigned int nr_pages) > { > struct sg_table *sg = NULL; > + size_t max_segment = 0; > int ret; > > sg = kmalloc(sizeof(struct sg_table), GFP_KERNEL); > @@ -813,8 +815,13 @@ struct sg_table *drm_prime_pages_to_sg(struct page **pages, unsigned int nr_page > goto out; > } > > - ret = sg_alloc_table_from_pages(sg, pages, nr_pages, 0, > - nr_pages << PAGE_SHIFT, GFP_KERNEL); > + if (dev) > + max_segment = dma_max_mapping_size(dev->dev); > + if (max_segment == 0 || max_segment > SCATTERLIST_MAX_SEGMENT) > + max_segment = SCATTERLIST_MAX_SEGMENT; > + ret = __sg_alloc_table_from_pages(sg, pages, nr_pages, 0, > + nr_pages << PAGE_SHIFT, > + max_segment, GFP_KERNEL); > if (ret) > goto out; > > diff --git a/drivers/gpu/drm/etnaviv/etnaviv_gem.c b/drivers/gpu/drm/etnaviv/etnaviv_gem.c > index f06e19e7be04..ea19f1d27275 100644 > --- a/drivers/gpu/drm/etnaviv/etnaviv_gem.c > +++ b/drivers/gpu/drm/etnaviv/etnaviv_gem.c > @@ -103,7 +103,8 @@ struct page **etnaviv_gem_get_pages(struct etnaviv_gem_object *etnaviv_obj) > int npages = etnaviv_obj->base.size >> PAGE_SHIFT; > struct sg_table *sgt; > > - sgt = drm_prime_pages_to_sg(etnaviv_obj->pages, npages); > + sgt = drm_prime_pages_to_sg(etnaviv_obj->base.dev, > + etnaviv_obj->pages, npages); > if (IS_ERR(sgt)) { > dev_err(dev->dev, "failed to allocate sgt: %ld\n", > PTR_ERR(sgt)); > diff --git a/drivers/gpu/drm/etnaviv/etnaviv_gem_prime.c b/drivers/gpu/drm/etnaviv/etnaviv_gem_prime.c > index 6d9e5c3c4dd5..4aa3426a9ba4 100644 > --- a/drivers/gpu/drm/etnaviv/etnaviv_gem_prime.c > +++ b/drivers/gpu/drm/etnaviv/etnaviv_gem_prime.c > @@ -19,7 +19,7 @@ struct sg_table *etnaviv_gem_prime_get_sg_table(struct drm_gem_object *obj) > if (WARN_ON(!etnaviv_obj->pages)) /* should have already pinned! */ > return ERR_PTR(-EINVAL); > > - return drm_prime_pages_to_sg(etnaviv_obj->pages, npages); > + return drm_prime_pages_to_sg(obj->dev, etnaviv_obj->pages, npages); > } > > void *etnaviv_gem_prime_vmap(struct drm_gem_object *obj) > diff --git a/drivers/gpu/drm/msm/msm_gem.c b/drivers/gpu/drm/msm/msm_gem.c > index b2f49152b4d4..b4553caaa196 100644 > --- a/drivers/gpu/drm/msm/msm_gem.c > +++ b/drivers/gpu/drm/msm/msm_gem.c > @@ -126,7 +126,7 @@ static struct page **get_pages(struct drm_gem_object *obj) > > msm_obj->pages = p; > > - msm_obj->sgt = drm_prime_pages_to_sg(p, npages); > + msm_obj->sgt = drm_prime_pages_to_sg(obj->dev, p, npages); > if (IS_ERR(msm_obj->sgt)) { > void *ptr = ERR_CAST(msm_obj->sgt); > > diff --git a/drivers/gpu/drm/msm/msm_gem_prime.c b/drivers/gpu/drm/msm/msm_gem_prime.c > index d7c8948427fe..515ef80816a0 100644 > --- a/drivers/gpu/drm/msm/msm_gem_prime.c > +++ b/drivers/gpu/drm/msm/msm_gem_prime.c > @@ -19,7 +19,7 @@ struct sg_table *msm_gem_prime_get_sg_table(struct drm_gem_object *obj) > if (WARN_ON(!msm_obj->pages)) /* should have already pinned! */ > return NULL; > > - return drm_prime_pages_to_sg(msm_obj->pages, npages); > + return drm_prime_pages_to_sg(obj->dev, msm_obj->pages, npages); > } > > void *msm_gem_prime_vmap(struct drm_gem_object *obj) > diff --git a/drivers/gpu/drm/nouveau/nouveau_prime.c b/drivers/gpu/drm/nouveau/nouveau_prime.c > index bae6a3eccee0..7766b810653f 100644 > --- a/drivers/gpu/drm/nouveau/nouveau_prime.c > +++ b/drivers/gpu/drm/nouveau/nouveau_prime.c > @@ -32,7 +32,7 @@ struct sg_table *nouveau_gem_prime_get_sg_table(struct drm_gem_object *obj) > struct nouveau_bo *nvbo = nouveau_gem_object(obj); > int npages = nvbo->bo.num_pages; > > - return drm_prime_pages_to_sg(nvbo->bo.ttm->pages, npages); > + return drm_prime_pages_to_sg(obj->dev, nvbo->bo.ttm->pages, npages); > } > > void *nouveau_gem_prime_vmap(struct drm_gem_object *obj) > diff --git a/drivers/gpu/drm/radeon/radeon_prime.c b/drivers/gpu/drm/radeon/radeon_prime.c > index b906e8fbd5f3..ea4c900e7c41 100644 > --- a/drivers/gpu/drm/radeon/radeon_prime.c > +++ b/drivers/gpu/drm/radeon/radeon_prime.c > @@ -36,7 +36,7 @@ struct sg_table *radeon_gem_prime_get_sg_table(struct drm_gem_object *obj) > struct radeon_bo *bo = gem_to_radeon_bo(obj); > int npages = bo->tbo.num_pages; > > - return drm_prime_pages_to_sg(bo->tbo.ttm->pages, npages); > + return drm_prime_pages_to_sg(obj->dev, bo->tbo.ttm->pages, npages); > } > > void *radeon_gem_prime_vmap(struct drm_gem_object *obj) > diff --git a/drivers/gpu/drm/rockchip/rockchip_drm_gem.c b/drivers/gpu/drm/rockchip/rockchip_drm_gem.c > index b9275ba7c5a5..77eeaf3439f6 100644 > --- a/drivers/gpu/drm/rockchip/rockchip_drm_gem.c > +++ b/drivers/gpu/drm/rockchip/rockchip_drm_gem.c > @@ -85,7 +85,8 @@ static int rockchip_gem_get_pages(struct rockchip_gem_object *rk_obj) > > rk_obj->num_pages = rk_obj->base.size >> PAGE_SHIFT; > > - rk_obj->sgt = drm_prime_pages_to_sg(rk_obj->pages, rk_obj->num_pages); > + rk_obj->sgt = drm_prime_pages_to_sg(rk_obj->base.dev, > + rk_obj->pages, rk_obj->num_pages); > if (IS_ERR(rk_obj->sgt)) { > ret = PTR_ERR(rk_obj->sgt); > goto err_put_pages; > @@ -442,7 +443,7 @@ struct sg_table *rockchip_gem_prime_get_sg_table(struct drm_gem_object *obj) > int ret; > > if (rk_obj->pages) > - return drm_prime_pages_to_sg(rk_obj->pages, rk_obj->num_pages); > + return drm_prime_pages_to_sg(obj->dev,rk_obj->pages, rk_obj->num_pages); > > sgt = kzalloc(sizeof(*sgt), GFP_KERNEL); > if (!sgt) > diff --git a/drivers/gpu/drm/tegra/gem.c b/drivers/gpu/drm/tegra/gem.c > index 723df142a981..47e2935b8c68 100644 > --- a/drivers/gpu/drm/tegra/gem.c > +++ b/drivers/gpu/drm/tegra/gem.c > @@ -284,7 +284,7 @@ static int tegra_bo_get_pages(struct drm_device *drm, struct tegra_bo *bo) > > bo->num_pages = bo->gem.size >> PAGE_SHIFT; > > - bo->sgt = drm_prime_pages_to_sg(bo->pages, bo->num_pages); > + bo->sgt = drm_prime_pages_to_sg(bo->gem.dev, bo->pages, bo->num_pages); > if (IS_ERR(bo->sgt)) { > err = PTR_ERR(bo->sgt); > goto put_pages; > diff --git a/drivers/gpu/drm/vgem/vgem_drv.c b/drivers/gpu/drm/vgem/vgem_drv.c > index 313339bbff90..15dd41e67de3 100644 > --- a/drivers/gpu/drm/vgem/vgem_drv.c > +++ b/drivers/gpu/drm/vgem/vgem_drv.c > @@ -321,7 +321,7 @@ static struct sg_table *vgem_prime_get_sg_table(struct drm_gem_object *obj) > { > struct drm_vgem_gem_object *bo = to_vgem_bo(obj); > > - return drm_prime_pages_to_sg(bo->pages, bo->base.size >> PAGE_SHIFT); > + return drm_prime_pages_to_sg(obj->dev, bo->pages, bo->base.size >> PAGE_SHIFT); > } > > static struct drm_gem_object* vgem_prime_import(struct drm_device *dev, > diff --git a/drivers/gpu/drm/xen/xen_drm_front_gem.c b/drivers/gpu/drm/xen/xen_drm_front_gem.c > index 39ff95b75357..aed7510e2710 100644 > --- a/drivers/gpu/drm/xen/xen_drm_front_gem.c > +++ b/drivers/gpu/drm/xen/xen_drm_front_gem.c > @@ -179,7 +179,8 @@ struct sg_table *xen_drm_front_gem_get_sg_table(struct drm_gem_object *gem_obj) > if (!xen_obj->pages) > return ERR_PTR(-ENOMEM); > > - return drm_prime_pages_to_sg(xen_obj->pages, xen_obj->num_pages); > + return drm_prime_pages_to_sg(gem_obj->dev, > + xen_obj->pages, xen_obj->num_pages); > } > > struct drm_gem_object * > -- > 2.27.0 > -- Daniel Vetter Software Engineer, Intel Corporation http://blog.ffwll.ch