Received: by 2002:a05:6a10:f347:0:0:0:0 with SMTP id d7csp3317762pxu; Sun, 29 Nov 2020 23:26:35 -0800 (PST) X-Google-Smtp-Source: ABdhPJxyN200mqoC44RJ1XNqC8z8jRmgmCKiRo39LoGL9RbeYkWyQsKuw0BLhW+lB0VYr/VmfLzX X-Received: by 2002:a17:906:ee2:: with SMTP id x2mr19412299eji.326.1606721195441; Sun, 29 Nov 2020 23:26:35 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1606721195; cv=none; d=google.com; s=arc-20160816; b=T60KE+5OdxpN5QfTPSKGs2O+lQCHMagOwccV9Hg9HBv6hORgK0eMJbfAUKGbSSWdin DLoNGxFlRaF9eEhStDyfAv3v9O0gPDCTJQiRtpNASKEIgzZuBYzRb0v1B356R/EbHEQ0 xgeYoGC/aqnTaVke4CCufZOrwL5IV9TIynp+yscjriVloo0oQ51FmDfGjlrLqtdoaajv Z5m0HrmrCHQnVfasHMye7IVVdfv2dRURQDaH5pjMWqRTq+I0gyE3dvgdxoixtNH+yHl4 IGjlVytCk5cD00GpDaeIYXxr9+OJaPUOac+2WNKeBfBWbeG4Rkgr/hu7RUTyWl2V+CZK MgOg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:content-language :in-reply-to:mime-version:user-agent:date:message-id:references:cc :to:from:subject; bh=cY66um5cshhy9zyRVVAodtLG9q/NoD6gm5+O7hiClDY=; b=xad2zcU+WFlPCgJq1Px9vgkxGGAqoUPN55oTRLbOqZrMS9KDynIvUgV7G+GiXsK/ed Rzx7se8o6P0XZkTNaucqR0CV9ZbAFdybtv2ynM1oQX73ocQh0ONmNIW2pX2Sui70xIAC iy6XQp8BxMBBer3DPuaVZ8YGn6Hourjrz/shmcfYAyIfJP1V3micnkhoVFcCcxa3Gpzo 0Tm/ygpDXCk5xhHcUkLWy+IcE+sF2/7MhQQ9zkGzQiFhFM6J1WXUpZj+x9nvL1WvhOKy rjPnmVrT8vBe/JmQeskXu80qXHmVRlmkxVsfvObA5sybZ4EJ94BmarOG2td7To1JJdBM widw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id mj7si1431404ejb.575.2020.11.29.23.26.11; Sun, 29 Nov 2020 23:26:35 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726299AbgK3HYQ (ORCPT + 99 others); Mon, 30 Nov 2020 02:24:16 -0500 Received: from szxga05-in.huawei.com ([45.249.212.191]:9070 "EHLO szxga05-in.huawei.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1725902AbgK3HYP (ORCPT ); Mon, 30 Nov 2020 02:24:15 -0500 Received: from DGGEMS403-HUB.china.huawei.com (unknown [172.30.72.58]) by szxga05-in.huawei.com (SkyGuard) with ESMTP id 4CkxXN6vK8zLxYq; Mon, 30 Nov 2020 15:23:00 +0800 (CST) Received: from [10.174.184.228] (10.174.184.228) by DGGEMS403-HUB.china.huawei.com (10.3.19.203) with Microsoft SMTP Server id 14.3.487.0; Mon, 30 Nov 2020 15:23:18 +0800 Subject: Re: [RFC PATCH v1 3/4] KVM: arm64: GICv4.1: Restore VLPI's pending state to physical side From: Shenming Lu To: Marc Zyngier CC: James Morse , Julien Thierry , Suzuki K Poulose , Eric Auger , , , , , Christoffer Dall , Alex Williamson , Kirti Wankhede , Cornelia Huck , Neo Jia , , References: <20201123065410.1915-1-lushenming@huawei.com> <20201123065410.1915-4-lushenming@huawei.com> <5c724bb83730cdd5dcf7add9a812fa92@kernel.org> <2d2bcae4f871d239a1af50362f5c11a4@kernel.org> <49610291-cf57-ff78-d0ac-063af24efbb4@huawei.com> Message-ID: <48c10467-30f3-9b5c-bbcb-533a51516dc5@huawei.com> Date: Mon, 30 Nov 2020 15:23:18 +0800 User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:78.0) Gecko/20100101 Thunderbird/78.2.2 MIME-Version: 1.0 In-Reply-To: <49610291-cf57-ff78-d0ac-063af24efbb4@huawei.com> Content-Type: text/plain; charset="utf-8" Content-Language: en-US Content-Transfer-Encoding: 8bit X-Originating-IP: [10.174.184.228] X-CFilter-Loop: Reflected Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 2020/11/24 21:12, Shenming Lu wrote: > On 2020/11/24 16:44, Marc Zyngier wrote: >> On 2020-11-24 08:10, Shenming Lu wrote: >>> On 2020/11/23 17:27, Marc Zyngier wrote: >>>> On 2020-11-23 06:54, Shenming Lu wrote: >>>>> From: Zenghui Yu >>>>> >>>>> When setting the forwarding path of a VLPI, it is more consistent to >>>> >>>> I'm not sure it is more consistent. It is a *new* behaviour, because it only >>>> matters for migration, which has been so far unsupported. >>> >>> Alright, consistent may not be accurate... >>> But I have doubt that whether there is really no need to transfer the >>> pending states >>> from kvm'vgic to VPT in set_forwarding regardless of migration, and the similar >>> for unset_forwarding. >> >> If you have to transfer that state outside of the a save/restore, it means that >> you have missed the programming of the PCI endpoint. This is an established >> restriction that the MSI programming must occur *after* the translation has >> been established using MAPI/MAPTI (see the large comment at the beginning of >> vgic-v4.c). >> >> If you want to revisit this, fair enough. But you will need a lot more than >> just opportunistically transfer the pending state. > > Thanks, I will look at what you mentioned. > >> >>> >>>> >>>>> also transfer the pending state from irq->pending_latch to VPT (especially >>>>> in migration, the pending states of VLPIs are restored into kvm’s vgic >>>>> first). And we currently send "INT+VSYNC" to trigger a VLPI to pending. >>>>> >>>>> Signed-off-by: Zenghui Yu >>>>> Signed-off-by: Shenming Lu >>>>> --- >>>>>  arch/arm64/kvm/vgic/vgic-v4.c | 12 ++++++++++++ >>>>>  1 file changed, 12 insertions(+) >>>>> >>>>> diff --git a/arch/arm64/kvm/vgic/vgic-v4.c b/arch/arm64/kvm/vgic/vgic-v4.c >>>>> index b5fa73c9fd35..cc3ab9cea182 100644 >>>>> --- a/arch/arm64/kvm/vgic/vgic-v4.c >>>>> +++ b/arch/arm64/kvm/vgic/vgic-v4.c >>>>> @@ -418,6 +418,18 @@ int kvm_vgic_v4_set_forwarding(struct kvm *kvm, int virq, >>>>>      irq->host_irq    = virq; >>>>>      atomic_inc(&map.vpe->vlpi_count); >>>>> >>>>> +    /* Transfer pending state */ >>>>> +    ret = irq_set_irqchip_state(irq->host_irq, >>>>> +                    IRQCHIP_STATE_PENDING, >>>>> +                    irq->pending_latch); >>>>> +    WARN_RATELIMIT(ret, "IRQ %d", irq->host_irq); >>>>> + >>>>> +    /* >>>>> +     * Let it be pruned from ap_list later and don't bother >>>>> +     * the List Register. >>>>> +     */ >>>>> +    irq->pending_latch = false; >>>> >>>> It occurs to me that calling into irq_set_irqchip_state() for a large >>>> number of interrupts can take a significant amount of time. It is also >>>> odd that you dump the VPT with the VPE unmapped, but rely on the VPE >>>> being mapped for the opposite operation. >>>> >>>> Shouldn't these be symmetric, all performed while the VPE is unmapped? >>>> It would also save a lot of ITS traffic. >>>> >>> >>> My thought was to use the existing interface directly without unmapping... >>> >>> If you want to unmap the vPE and poke the VPT here, as I said in the cover >>> letter, set/unset_forwarding might also be called when all devices are running >>> at normal run time, in which case the unmapping of the vPE is not allowed... >> >> No, I'm suggesting that you don't do anything here, but instead as a by-product >> of restoring the ITS tables. What goes wrong if you use the >> KVM_DEV_ARM_ITS_RESTORE_TABLE backend instead? > > There is an issue if we do it in the restoring of the ITS tables: the transferring > of the pending state needs the irq to be marked as hw before, which is done by the > pass-through device, but the configuring of the forwarding path of the VLPI depends > on the restoring of the vgic first... It is a circular dependency. > Hi Marc, We are pondering over this problem these days, but still don't get a good solution... Could you give us some advice on this? Or could we move the restoring of the pending states (include the sync from guest RAM and the transfer to HW) to the GIC VM state change handler, which is completely corresponding to save_pending_tables (more symmetric?) and don't expose GICv4... Thanks, Shenming >> >>> Another possible solution is to add a new dedicated interface to QEMU >>> to transfer >>> these pending states to HW in GIC VM state change handler corresponding to >>> save_pending_tables? >> >> Userspace has no way to know we use GICv4, and I intend to keep it >> completely out of the loop. The API is already pretty tortuous, and >> I really don't want to add any extra complexity to it. >> >> Thanks, >> >>         M.