Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1753012AbdFTJQq (ORCPT ); Tue, 20 Jun 2017 05:16:46 -0400 Received: from mail-pf0-f193.google.com ([209.85.192.193]:33122 "EHLO mail-pf0-f193.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751989AbdFTJP7 (ORCPT ); Tue, 20 Jun 2017 05:15:59 -0400 From: guangrong.xiao@gmail.com X-Google-Original-From: xiaoguangrong@tencent.com To: pbonzini@redhat.com, mtosatti@redhat.com, avi.kivity@gmail.com, rkrcmar@redhat.com Cc: kvm@vger.kernel.org, linux-kernel@vger.kernel.org, qemu-devel@nongnu.org, Xiao Guangrong Subject: [PATCH v2 6/7] KVM: MMU: clarify fast_pf_fix_direct_spte Date: Tue, 20 Jun 2017 17:15:25 +0800 Message-Id: <20170620091526.4287-7-xiaoguangrong@tencent.com> X-Mailer: git-send-email 2.9.4 In-Reply-To: <20170620091526.4287-1-xiaoguangrong@tencent.com> References: <20170620091526.4287-1-xiaoguangrong@tencent.com> Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 1595 Lines: 43 From: Xiao Guangrong The writable spte can not be locklessly fixed and add a WARN_ON() to trigger the warning if something out of our mind happens, that is good for us to track if the log for writable spte is missed on the fast path Signed-off-by: Xiao Guangrong --- arch/x86/kvm/mmu.c | 11 ++++++++++- 1 file changed, 10 insertions(+), 1 deletion(-) diff --git a/arch/x86/kvm/mmu.c b/arch/x86/kvm/mmu.c index c011ab0..7711953 100644 --- a/arch/x86/kvm/mmu.c +++ b/arch/x86/kvm/mmu.c @@ -3240,6 +3240,15 @@ fast_pf_fix_direct_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, WARN_ON(!sp->role.direct); /* + * the original spte can not be writable as only the spte which + * fulfills is_access_track_spte() or + * spte_can_locklessly_be_made_writable() can be locklessly fixed, + * for the former, the W bit is always cleared, for the latter, + * there is nothing to do if it is already writable. + */ + WARN_ON(is_writable_pte(old_spte)); + + /* * Theoretically we could also set dirty bit (and flush TLB) here in * order to eliminate unnecessary PML logging. See comments in * set_spte. But fast_page_fault is very unlikely to happen with PML @@ -3254,7 +3263,7 @@ fast_pf_fix_direct_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, if (cmpxchg64(sptep, old_spte, new_spte) != old_spte) return false; - if (is_writable_pte(new_spte) && !is_writable_pte(old_spte)) { + if (is_writable_pte(new_spte)) { /* * The gfn of direct spte is stable since it is * calculated by sp->gfn. -- 2.9.4