Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 855ECC636D4 for ; Tue, 31 Jan 2023 03:07:22 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S230346AbjAaDHV (ORCPT ); Mon, 30 Jan 2023 22:07:21 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:37628 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229742AbjAaDHU (ORCPT ); Mon, 30 Jan 2023 22:07:20 -0500 Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.129.124]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id C602429E08 for ; Mon, 30 Jan 2023 19:06:32 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1675134392; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: in-reply-to:in-reply-to:references:references; bh=KdMSX2PGXe45qBgqP8RMLgntuCYyndOvWlGnSWr1I1w=; b=Lrortqv7icKfhEub19HsWOVMh8X/pZjigrG+ootNdXQaHreHpsRKbYpRK/+4lBQSkH84FZ M7bz0VP6U54Bfb0LiWfbgRRQqtkaLI5dQsmZXSyBKtnC+8MIcMXBytuDzHitD4BFSMFE/E 4zAhmEzKBG5sFzwUV9j1Ukvt00JO1Ls= Received: from mail-oi1-f197.google.com (mail-oi1-f197.google.com [209.85.167.197]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.3, cipher=TLS_AES_128_GCM_SHA256) id us-mta-649-oop9K43INPuetYbYdFO9dg-1; Mon, 30 Jan 2023 22:06:29 -0500 X-MC-Unique: oop9K43INPuetYbYdFO9dg-1 Received: by mail-oi1-f197.google.com with SMTP id u9-20020a056808150900b0036f087d05fdso5873838oiw.0 for ; Mon, 30 Jan 2023 19:06:29 -0800 (PST) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=cc:to:subject:message-id:date:from:in-reply-to:references :mime-version:x-gm-message-state:from:to:cc:subject:date:message-id :reply-to; bh=KdMSX2PGXe45qBgqP8RMLgntuCYyndOvWlGnSWr1I1w=; b=qc7Hixc9Ra04Mk/lqQkuoSDqvzEOWliKC2l7a01Vm6oNUErEc8ncBS3l5fyrJiKnu2 Da01sdZ5Zq+XQOCOUJNqmbSbSEPjtjn+e7HPGH69C0c/zoif08rXnT3kD8s0FeeiIvZr r0we8WChD0ghICScIpcdxvTWl9VuAL9tJYooNmBY94MByQ47DZgZr9hImhPSmEwpqf/+ CI2wulE4gio777h9QnjjKgsN9XubQETmSbv7VPWxwFRAC/LUpn3s016urXeLtCntF1Uq X4eF6vtf/bYAg5mJTCvaY6cHPNBoeC1JUIKqOLhKhUjOFOkPOflOhr17WQUjBMWB6M+d XWyw== X-Gm-Message-State: AO0yUKU/Jb0JhPMrhsofTZ5mbS12JNYWWjBu5zKm9ntqcBzaQv5zvZ/l AUJ+NbGH5c+uysf0fK2L3jxNqVp//rSA9y9WjcL/9Os8CPLWtU/zU5dGjTB5zZ638Tbo7Ov7gKL kqSQDv5kWBGXP2KcXC8vcI4DOI2Vml+W85n52kviQ X-Received: by 2002:a05:6870:959e:b0:163:9cea:eea7 with SMTP id k30-20020a056870959e00b001639ceaeea7mr565253oao.35.1675134389122; Mon, 30 Jan 2023 19:06:29 -0800 (PST) X-Google-Smtp-Source: AK7set/gy6zeCHS7dDqTzLgm0l0HH281Gpnu1QcOaLODISHVTb7n1+4QzN8Fzmkkj1rEtYbW+m8iASC8yFpzqnKRkFM= X-Received: by 2002:a05:6870:959e:b0:163:9cea:eea7 with SMTP id k30-20020a056870959e00b001639ceaeea7mr565247oao.35.1675134388839; Mon, 30 Jan 2023 19:06:28 -0800 (PST) MIME-Version: 1.0 References: <20230128031740.166743-1-sunnanyong@huawei.com> In-Reply-To: From: Jason Wang Date: Tue, 31 Jan 2023 11:06:17 +0800 Message-ID: Subject: Re: [PATCH] vhost/vdpa: Add MSI translation tables to iommu for software-managed MSI To: Nanyong Sun Cc: joro@8bytes.org, will@kernel.org, robin.murphy@arm.com, mst@redhat.com, iommu@lists.linux.dev, linux-kernel@vger.kernel.org, kvm@vger.kernel.org, virtualization@lists.linux-foundation.org, netdev@vger.kernel.org, wangrong68@huawei.com Content-Type: text/plain; charset="UTF-8" Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Tue, Jan 31, 2023 at 9:32 AM Nanyong Sun wrote: > > On 2023/1/29 14:02, Jason Wang wrote: > > On Sat, Jan 28, 2023 at 10:25 AM Nanyong Sun wrote: > >> From: Rong Wang > >> > >> Once enable iommu domain for one device, the MSI > >> translation tables have to be there for software-managed MSI. > >> Otherwise, platform with software-managed MSI without an > >> irq bypass function, can not get a correct memory write event > >> from pcie, will not get irqs. > >> The solution is to obtain the MSI phy base address from > >> iommu reserved region, and set it to iommu MSI cookie, > >> then translation tables will be created while request irq. > >> > >> Signed-off-by: Rong Wang > >> Signed-off-by: Nanyong Sun > >> --- > >> drivers/iommu/iommu.c | 1 + > >> drivers/vhost/vdpa.c | 53 ++++++++++++++++++++++++++++++++++++++++--- > >> 2 files changed, 51 insertions(+), 3 deletions(-) > >> > >> diff --git a/drivers/iommu/iommu.c b/drivers/iommu/iommu.c > >> index de91dd88705b..f6c65d5d8e2b 100644 > >> --- a/drivers/iommu/iommu.c > >> +++ b/drivers/iommu/iommu.c > >> @@ -2623,6 +2623,7 @@ void iommu_get_resv_regions(struct device *dev, struct list_head *list) > >> if (ops->get_resv_regions) > >> ops->get_resv_regions(dev, list); > >> } > >> +EXPORT_SYMBOL_GPL(iommu_get_resv_regions); > >> > >> /** > >> * iommu_put_resv_regions - release resered regions > >> diff --git a/drivers/vhost/vdpa.c b/drivers/vhost/vdpa.c > >> index ec32f785dfde..31d3e9ed4cfa 100644 > >> --- a/drivers/vhost/vdpa.c > >> +++ b/drivers/vhost/vdpa.c > >> @@ -1103,6 +1103,48 @@ static ssize_t vhost_vdpa_chr_write_iter(struct kiocb *iocb, > >> return vhost_chr_write_iter(dev, from); > >> } > >> > >> +static bool vhost_vdpa_check_sw_msi(struct list_head *dev_resv_regions, phys_addr_t *base) > >> +{ > >> + struct iommu_resv_region *region; > >> + bool ret = false; > >> + > >> + list_for_each_entry(region, dev_resv_regions, list) { > >> + /* > >> + * The presence of any 'real' MSI regions should take > >> + * precedence over the software-managed one if the > >> + * IOMMU driver happens to advertise both types. > >> + */ > >> + if (region->type == IOMMU_RESV_MSI) { > >> + ret = false; > >> + break; > >> + } > >> + > >> + if (region->type == IOMMU_RESV_SW_MSI) { > >> + *base = region->start; > >> + ret = true; > >> + } > >> + } > >> + > >> + return ret; > >> +} > > Can we unify this with what VFIO had? > Yes, these two functions are just the same. > Do you think move this function to iommu.c, and export from iommu is a > good choice? Probably, we can try and see. > > > >> + > >> +static int vhost_vdpa_get_msi_cookie(struct iommu_domain *domain, struct device *dma_dev) > >> +{ > >> + struct list_head dev_resv_regions; > >> + phys_addr_t resv_msi_base = 0; > >> + int ret = 0; > >> + > >> + INIT_LIST_HEAD(&dev_resv_regions); > >> + iommu_get_resv_regions(dma_dev, &dev_resv_regions); > >> + > >> + if (vhost_vdpa_check_sw_msi(&dev_resv_regions, &resv_msi_base)) > >> + ret = iommu_get_msi_cookie(domain, resv_msi_base); > >> + > >> + iommu_put_resv_regions(dma_dev, &dev_resv_regions); > >> + > >> + return ret; > >> +} > >> + > >> static int vhost_vdpa_alloc_domain(struct vhost_vdpa *v) > >> { > >> struct vdpa_device *vdpa = v->vdpa; > >> @@ -1128,11 +1170,16 @@ static int vhost_vdpa_alloc_domain(struct vhost_vdpa *v) > >> > >> ret = iommu_attach_device(v->domain, dma_dev); > >> if (ret) > >> - goto err_attach; > >> + goto err_alloc_domain; > >> > >> - return 0; > >> + ret = vhost_vdpa_get_msi_cookie(v->domain, dma_dev); > > Do we need to check the overlap mapping and record it in the interval > > tree (as what VFIO did)? > > > > Thanks > Yes, we need to care about this part, I will handle this recently. > Thanks a lot. I think for parents that requires vendor specific mapping logic we probably also need this. But this could be added on top (via a new config ops probably). Thanks > >> + if (ret) > >> + goto err_attach_device; > >> > >> -err_attach: > >> + return 0; > >> +err_attach_device: > >> + iommu_detach_device(v->domain, dma_dev); > >> +err_alloc_domain: > >> iommu_domain_free(v->domain); > >> return ret; > >> } > >> -- > >> 2.25.1 > >> > > . >