Received: by 2002:a25:ef43:0:0:0:0:0 with SMTP id w3csp405407ybm; Tue, 26 May 2020 21:09:41 -0700 (PDT) X-Google-Smtp-Source: ABdhPJxDYhipMbgNCh+9eXRCzuF6L+dz9fn4ljdWYnuQetnVApoyy0qO4MGaurMO9mP8boUqCwmw X-Received: by 2002:a17:906:9493:: with SMTP id t19mr3875591ejx.461.1590552581726; Tue, 26 May 2020 21:09:41 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1590552581; cv=none; d=google.com; s=arc-20160816; b=aaWZBNJsFOzf1iMiUzJFR/qULwPpMU5VwIoE2z9YWJP4sdGhpchpc7JF+Htyq6RLT3 JF2te2h8ZMSfF36duCRAhy5B4F1EzqLPy/hGn7/bhDCflc6ml6+ucxSJPUu+Jozeu947 iNajcgZM9qkTymTE2h90OnK1PpeapaTvhWk2JPDS2Ulu3HHTwUzSL2tNa+ocOuws9Sgq UqLXkoV4XJ/3Q2wJAm1EK/xSKHOJ/AfLAmhoDx5sBw9j6JOLE+P9zi8x8+gmBUaeqOED 98BFs4Rg4BTECY1pjg8/UGzAOiGYodcaK/XFW/miLLjE2FNEteKZniAwDSa7OXfCc7e2 avIA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:in-reply-to:content-disposition :mime-version:references:message-id:subject:cc:to:from:date :dkim-signature; bh=S8Sh+DjtPEAcfjFCLlA/An7vGsa0j6xUtlJx+u0SHCo=; b=CGg3KvpVEjETmNRtSHUomKNYWOBh++Tewl4DvLlti+rroGKBrBI3SbUj4caNgnshHF DaJIoT82EAkZmBgTcOTJEJ8RkB6FUjjLe0arySMvOA7DSIugGDUuhyR4a0jxcZcLa95Z rxQE5PEn7atGE8r7VEQoDnvapf81BUNI9lLtlRHmDxBOAAFgGTLASbvLiWp3cvdOBMv7 xkvFFzQ2X4+AXQWSXSB17W1B5PKnzwdmL3zouz0M5u8zSksx7Fj6PdaiBHuwBdh/n/jI ySb/P2HgpxeZEcVotTJij/SqiBav+xt+8DVZONFaJ5HdijfhUxNGTlNzL1yXziHWgC36 jjBQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@shutemov-name.20150623.gappssmtp.com header.s=20150623 header.b=iaJc4Q0d; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id se24si1104427ejb.652.2020.05.26.21.09.18; Tue, 26 May 2020 21:09:41 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@shutemov-name.20150623.gappssmtp.com header.s=20150623 header.b=iaJc4Q0d; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S2391908AbgEZWKb (ORCPT + 99 others); Tue, 26 May 2020 18:10:31 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:34268 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S2391894AbgEZWKa (ORCPT ); Tue, 26 May 2020 18:10:30 -0400 Received: from mail-lf1-x143.google.com (mail-lf1-x143.google.com [IPv6:2a00:1450:4864:20::143]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 63119C03E96D for ; Tue, 26 May 2020 15:10:29 -0700 (PDT) Received: by mail-lf1-x143.google.com with SMTP id e125so13267479lfd.1 for ; Tue, 26 May 2020 15:10:29 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=shutemov-name.20150623.gappssmtp.com; s=20150623; h=date:from:to:cc:subject:message-id:references:mime-version :content-disposition:in-reply-to; bh=S8Sh+DjtPEAcfjFCLlA/An7vGsa0j6xUtlJx+u0SHCo=; b=iaJc4Q0dVmQB2Nw7TDPjM5G/drSQCJ1I0yTkAzs35+wOBGWa8Y7ejBHHXXwVKMTMQ+ NkNdap9tqvkiIP9m0q7O1axUE2lvJgI92jZ0IOXLF2XRcaUzTSo/ftuzFdVXnVzObmWy 0CMdIgFTfIsBa7I+hhkQf7Fal4GZRNCYXpv14mQOa8XieWI923D6i4MsqVe56fxN2ZF6 rHv8DGtKKWb9J8y+3Z1O0+vbmnh2k4VRBjNiHod5ZDOy54y/27IGfws3tAvpafBWD1L+ Yi4zJuYg5k5tLPgIEIsylRkz7gp6zK4NNvnxczcZ7BEQ7VnR3CUUIWchfSRdBsju+uqK jSdA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:date:from:to:cc:subject:message-id:references :mime-version:content-disposition:in-reply-to; bh=S8Sh+DjtPEAcfjFCLlA/An7vGsa0j6xUtlJx+u0SHCo=; b=HEWSV6qnN5Pt7I0+x8a/eYrx4pR57PY8J1sJgs/68/bHJOR0Zyr9fIQAIZSueL364g CBCI5waUdrye3MjyH/FChSb/N15FFNsSDVeLx7AVZEjpTxMr17gLZUKxkeMJxYNVMP8J +WPoe8RxAX2Q8t26um8zZTQ0bqizR9jwv8hj/6bH0nnTUE9KJVw4RDuJ4wps6ln2a6R1 horyPMH35IMGZUxUKOmJC8rCO0m3DaD+pkLGmJKF4l5r2n3qaPHRFHsGWQoA2zKIGPGK wtxNqP3UV2DMhYfWHi94NLe4o6hrXyhMa4jYtZyGgVYjs93ZpJk8EUah6x1vn9wD4bYa 2DRQ== X-Gm-Message-State: AOAM5333tUchbAlgc3jamTPd/Yzs9m8DsAhr4+WGCQqiSauvNcszCE58 gKhucoiAih4ziM7aBhe/7gNW0g== X-Received: by 2002:a19:5f4e:: with SMTP id a14mr1482768lfj.57.1590531027702; Tue, 26 May 2020 15:10:27 -0700 (PDT) Received: from box.localdomain ([86.57.175.117]) by smtp.gmail.com with ESMTPSA id n2sm278570lfl.53.2020.05.26.15.10.26 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Tue, 26 May 2020 15:10:27 -0700 (PDT) Received: by box.localdomain (Postfix, from userid 1000) id D850A10230F; Wed, 27 May 2020 01:10:27 +0300 (+03) Date: Wed, 27 May 2020 01:10:27 +0300 From: "Kirill A. Shutemov" To: Mike Rapoport Cc: Dave Hansen , Andy Lutomirski , Peter Zijlstra , Paolo Bonzini , Sean Christopherson , Vitaly Kuznetsov , Wanpeng Li , Jim Mattson , Joerg Roedel , David Rientjes , Andrea Arcangeli , Kees Cook , Will Drewry , "Edgecombe, Rick P" , "Kleen, Andi" , x86@kernel.org, kvm@vger.kernel.org, linux-mm@kvack.org, linux-kernel@vger.kernel.org, "Kirill A. Shutemov" Subject: Re: [RFC 16/16] KVM: Unmap protected pages from direct mapping Message-ID: <20200526221027.ixxahg6ya2z5fppy@box> References: <20200522125214.31348-1-kirill.shutemov@linux.intel.com> <20200522125214.31348-17-kirill.shutemov@linux.intel.com> <20200526061638.GA48741@kernel.org> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20200526061638.GA48741@kernel.org> Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Tue, May 26, 2020 at 09:16:38AM +0300, Mike Rapoport wrote: > On Fri, May 22, 2020 at 03:52:14PM +0300, Kirill A. Shutemov wrote: > > If the protected memory feature enabled, unmap guest memory from > > kernel's direct mappings. > > > > Migration and KSM is disabled for protected memory as it would require a > > special treatment. > > > > Signed-off-by: Kirill A. Shutemov > > --- > > arch/x86/mm/pat/set_memory.c | 1 + > > include/linux/kvm_host.h | 3 ++ > > mm/huge_memory.c | 9 +++++ > > mm/ksm.c | 3 ++ > > mm/memory.c | 13 +++++++ > > mm/rmap.c | 4 ++ > > virt/kvm/kvm_main.c | 74 ++++++++++++++++++++++++++++++++++++ > > 7 files changed, 107 insertions(+) > > > > diff --git a/arch/x86/mm/pat/set_memory.c b/arch/x86/mm/pat/set_memory.c > > index 6f075766bb94..13988413af40 100644 > > --- a/arch/x86/mm/pat/set_memory.c > > +++ b/arch/x86/mm/pat/set_memory.c > > @@ -2227,6 +2227,7 @@ void __kernel_map_pages(struct page *page, int numpages, int enable) > > > > arch_flush_lazy_mmu_mode(); > > } > > +EXPORT_SYMBOL_GPL(__kernel_map_pages); > > > > #ifdef CONFIG_HIBERNATION > > bool kernel_page_present(struct page *page) > > diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h > > index b6944f88033d..e1d7762b615c 100644 > > --- a/include/linux/kvm_host.h > > +++ b/include/linux/kvm_host.h > > @@ -705,6 +705,9 @@ int kvm_protect_all_memory(struct kvm *kvm); > > int kvm_protect_memory(struct kvm *kvm, > > unsigned long gfn, unsigned long npages, bool protect); > > > > +void kvm_map_page(struct page *page, int nr_pages); > > +void kvm_unmap_page(struct page *page, int nr_pages); > > + > > int gfn_to_page_many_atomic(struct kvm_memory_slot *slot, gfn_t gfn, > > struct page **pages, int nr_pages); > > > > diff --git a/mm/huge_memory.c b/mm/huge_memory.c > > index c3562648a4ef..d8a444a401cc 100644 > > --- a/mm/huge_memory.c > > +++ b/mm/huge_memory.c > > @@ -33,6 +33,7 @@ > > #include > > #include > > #include > > +#include > > This does not seem right... I agree. I try to find a more clean way to deal with it. > > #include > > #include > > @@ -650,6 +651,10 @@ static vm_fault_t __do_huge_pmd_anonymous_page(struct vm_fault *vmf, > > spin_unlock(vmf->ptl); > > count_vm_event(THP_FAULT_ALLOC); > > count_memcg_events(memcg, THP_FAULT_ALLOC, 1); > > + > > + /* Unmap page from direct mapping */ > > + if (vma_is_kvm_protected(vma)) > > + kvm_unmap_page(page, HPAGE_PMD_NR); > > ... and neither does this. > > I think the map/unmap primitives shoud be a part of the generic mm and > not burried inside KVM. Well, yes. Except, kvm_map_page() also clears the page before bringing it back to direct mappings. Not sure yet how to deal with it. > > return 0; > > @@ -1886,6 +1891,10 @@ int zap_huge_pmd(struct mmu_gather *tlb, struct vm_area_struct *vma, > > page_remove_rmap(page, true); > > VM_BUG_ON_PAGE(page_mapcount(page) < 0, page); > > VM_BUG_ON_PAGE(!PageHead(page), page); > > + > > + /* Map the page back to the direct mapping */ > > + if (vma_is_kvm_protected(vma)) > > + kvm_map_page(page, HPAGE_PMD_NR); > > } else if (thp_migration_supported()) { > > swp_entry_t entry; > > > > diff --git a/mm/ksm.c b/mm/ksm.c > > index 281c00129a2e..942b88782ac2 100644 > > --- a/mm/ksm.c > > +++ b/mm/ksm.c > > @@ -527,6 +527,9 @@ static struct vm_area_struct *find_mergeable_vma(struct mm_struct *mm, > > return NULL; > > if (!(vma->vm_flags & VM_MERGEABLE) || !vma->anon_vma) > > return NULL; > > + /* TODO */ > > Probably this is not something that should be done. For a security > sensitive environment that wants protected memory, KSM woudn't be > relevant anyway... Hm. True. > > diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c > > index 71aac117357f..defc33d3a124 100644 > > --- a/virt/kvm/kvm_main.c > > +++ b/virt/kvm/kvm_main.c > > @@ -51,6 +51,7 @@ > > #include > > #include > > #include > > +#include > > > > #include > > #include > > @@ -2718,6 +2719,72 @@ void kvm_vcpu_mark_page_dirty(struct kvm_vcpu *vcpu, gfn_t gfn) > > } > > EXPORT_SYMBOL_GPL(kvm_vcpu_mark_page_dirty); > > > > +void kvm_map_page(struct page *page, int nr_pages) > > +{ > > + int i; > > + > > + /* Clear page before returning it to the direct mapping */ > > + for (i = 0; i < nr_pages; i++) { > > + void *p = map_page_atomic(page + i); > > + memset(p, 0, PAGE_SIZE); > > + unmap_page_atomic(p); > > + } > > + > > + kernel_map_pages(page, nr_pages, 1); > > +} > > +EXPORT_SYMBOL_GPL(kvm_map_page); > > + > > +void kvm_unmap_page(struct page *page, int nr_pages) > > +{ > > + kernel_map_pages(page, nr_pages, 0); > > +} > > +EXPORT_SYMBOL_GPL(kvm_unmap_page); > > + > > +static int adjust_direct_mapping_pte_range(pmd_t *pmd, unsigned long addr, > > + unsigned long end, > > + struct mm_walk *walk) > > +{ > > + bool protect = (bool)walk->private; > > + pte_t *pte; > > + struct page *page; > > + > > + if (pmd_trans_huge(*pmd)) { > > + page = pmd_page(*pmd); > > + if (is_huge_zero_page(page)) > > + return 0; > > + VM_BUG_ON_PAGE(total_mapcount(page) != 1, page); > > + /* XXX: Would it fail with direct device assignment? */ > > + VM_BUG_ON_PAGE(page_count(page) != 1, page); > > + kernel_map_pages(page, HPAGE_PMD_NR, !protect); > > + return 0; > > + } > > + > > + pte = pte_offset_map(pmd, addr); > > + for (; addr != end; pte++, addr += PAGE_SIZE) { > > + pte_t entry = *pte; > > + > > + if (!pte_present(entry)) > > + continue; > > + > > + if (is_zero_pfn(pte_pfn(entry))) > > + continue; > > + > > + page = pte_page(entry); > > + > > + VM_BUG_ON_PAGE(page_mapcount(page) != 1, page); > > + /* XXX: Would it fail with direct device assignment? */ > > + VM_BUG_ON_PAGE(page_count(page) != > > + total_mapcount(compound_head(page)), page); > > + kernel_map_pages(page, 1, !protect); > > + } > > + > > + return 0; > > +} > > + > > +static const struct mm_walk_ops adjust_direct_mapping_ops = { > > + .pmd_entry = adjust_direct_mapping_pte_range, > > +}; > > + > > All this seem to me an addition to set_memory APIs rather then KVM. Emm?.. I don't think walking userspace mapping is set_memory thing. And kernel_map_pages() is VMM interface already. -- Kirill A. Shutemov