Received: by 2002:ac0:a5a7:0:0:0:0:0 with SMTP id m36-v6csp525401imm; Wed, 11 Jul 2018 06:43:08 -0700 (PDT) X-Google-Smtp-Source: AAOMgpf4JNeE6hz3wkk4kTuNNqy2/E5J6/jXJIdOTzPjt5DP0h8IeOs+bEf3ahcSeMuwteWsRASU X-Received: by 2002:a63:5624:: with SMTP id k36-v6mr26703081pgb.146.1531316588275; Wed, 11 Jul 2018 06:43:08 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1531316588; cv=none; d=google.com; s=arc-20160816; b=0DczhsXg+PIY+fvLdg0NUdZK3rTxEqUdVrTc0M85QkUsgUvJ/0XzySC2jqwEfp4n+C WmAveYUlArGA7QetwU5fG/9YAEviJWZfE2kdlTwepKKZZ+VGrYaleTgTfYCCTC5cZN93 v1hLhRRdq1WCJ0pARRD3yHNg7PxK0cin1mcZUF5ia8A9L0vUb+73Eph7km0z4dHW1OLn uziogVBCFdm+LLudQOx4a3zJdXByIEsPz36KVRuW13tAInDcx0PIZn5IPuAUtTlIbo2S SaR2PiXqkcv9yCJuGx2FbzPCJjBNlvMJmvzOzv8k3uXbeZEKxAJe+SDubNQjnUq/MPZ8 JlpA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding :content-language:in-reply-to:mime-version:user-agent:date :message-id:from:references:cc:to:subject:arc-authentication-results; bh=rYrThbgZaz5m2qN5sHE15+goOpWjceytJ6PWU7HUjxI=; b=Qc9liv/CSIXXhvef3Rs8ceKBv/6r4jR9BSL5Sj0arw1hBv8oY/lZDMkVuNyKjo2Dt3 BWpY0tHbZUkC+FR0FQ9bFO4CNjo4jJuDY79I/QKh4DCUpplP6BsjqtWs1jQLtzEJroxp UW4TX8rFNefdT3EpPndZPwt6hGFFkwwmp9AhVIgP/vYZPsxesAna6luPpYRFqiIGTbLn SHzsD/2r0ys5/aZ48VrDvT50p5OjE5gjZGyHqd/weC/hkru8XZVuQFJ5JqK2OS3sybkA xlpG2O+pKOgIKzynhk9nwNG1unBOAahSVCVOIDNlvMPU4CfCUTZ8quH3gPzLRp3pFoBg XHtA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id j8-v6si18602565plk.261.2018.07.11.06.42.52; Wed, 11 Jul 2018 06:43:08 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S2388140AbeGKNp6 (ORCPT + 99 others); Wed, 11 Jul 2018 09:45:58 -0400 Received: from foss.arm.com ([217.140.101.70]:36090 "EHLO foss.arm.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S2388116AbeGKNp6 (ORCPT ); Wed, 11 Jul 2018 09:45:58 -0400 Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.72.51.249]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id 0730215AD; Wed, 11 Jul 2018 06:41:34 -0700 (PDT) Received: from [10.1.206.73] (en101.cambridge.arm.com [10.1.206.73]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id 761503F5AD; Wed, 11 Jul 2018 06:41:32 -0700 (PDT) Subject: Re: [PATCH v5 4/7] KVM: arm64: Support PUD hugepage in stage2_is_exec() To: Punit Agrawal , kvmarm@lists.cs.columbia.edu Cc: linux-arm-kernel@lists.infradead.org, marc.zyngier@arm.com, christoffer.dall@arm.com, linux-kernel@vger.kernel.org, will.deacon@arm.com, Russell King , Catalin Marinas References: <20180709143835.28971-1-punit.agrawal@arm.com> <20180709144124.29164-1-punit.agrawal@arm.com> <20180709144124.29164-4-punit.agrawal@arm.com> From: Suzuki K Poulose Message-ID: Date: Wed, 11 Jul 2018 14:41:29 +0100 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:52.0) Gecko/20100101 Thunderbird/52.8.0 MIME-Version: 1.0 In-Reply-To: <20180709144124.29164-4-punit.agrawal@arm.com> Content-Type: text/plain; charset=us-ascii; format=flowed Content-Language: en-US Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 09/07/18 15:41, Punit Agrawal wrote: > In preparation for creating PUD hugepages at stage 2, add support for > detecting execute permissions on PUD page table entries. Faults due to > lack of execute permissions on page table entries is used to perform > i-cache invalidation on first execute. > > Provide trivial implementations of arm32 helpers to allow sharing of > code. > > Signed-off-by: Punit Agrawal > Cc: Christoffer Dall > Cc: Marc Zyngier > Cc: Russell King > Cc: Catalin Marinas > Cc: Will Deacon > --- > arch/arm/include/asm/kvm_mmu.h | 6 ++++ > arch/arm64/include/asm/kvm_mmu.h | 5 +++ > arch/arm64/include/asm/pgtable-hwdef.h | 2 ++ > virt/kvm/arm/mmu.c | 49 +++++++++++++++++++++++--- > 4 files changed, 57 insertions(+), 5 deletions(-) > > diff --git a/arch/arm/include/asm/kvm_mmu.h b/arch/arm/include/asm/kvm_mmu.h > index c23722f75d5c..d05c8986e495 100644 > --- a/arch/arm/include/asm/kvm_mmu.h > +++ b/arch/arm/include/asm/kvm_mmu.h > @@ -96,6 +96,12 @@ static inline bool kvm_s2pud_readonly(pud_t *pud) > } > > > +static inline bool kvm_s2pud_exec(pud_t *pud) > +{ > + BUG(); > + return false; > +} > + > static inline void kvm_set_pmd(pmd_t *pmd, pmd_t new_pmd) > { > *pmd = new_pmd; > diff --git a/arch/arm64/include/asm/kvm_mmu.h b/arch/arm64/include/asm/kvm_mmu.h > index 84051930ddfe..15bc1be8f82f 100644 > --- a/arch/arm64/include/asm/kvm_mmu.h > +++ b/arch/arm64/include/asm/kvm_mmu.h > @@ -249,6 +249,11 @@ static inline bool kvm_s2pud_readonly(pud_t *pudp) > return kvm_s2pte_readonly((pte_t *)pudp); > } > > +static inline bool kvm_s2pud_exec(pud_t *pudp) > +{ > + return !(READ_ONCE(pud_val(*pudp)) & PUD_S2_XN); > +} > + > static inline bool kvm_page_empty(void *ptr) > { > struct page *ptr_page = virt_to_page(ptr); > diff --git a/arch/arm64/include/asm/pgtable-hwdef.h b/arch/arm64/include/asm/pgtable-hwdef.h > index fd208eac9f2a..10ae592b78b8 100644 > --- a/arch/arm64/include/asm/pgtable-hwdef.h > +++ b/arch/arm64/include/asm/pgtable-hwdef.h > @@ -193,6 +193,8 @@ > #define PMD_S2_RDWR (_AT(pmdval_t, 3) << 6) /* HAP[2:1] */ > #define PMD_S2_XN (_AT(pmdval_t, 2) << 53) /* XN[1:0] */ > > +#define PUD_S2_XN (_AT(pudval_t, 2) << 53) /* XN[1:0] */ > + > /* > * Memory Attribute override for Stage-2 (MemAttr[3:0]) > */ > diff --git a/virt/kvm/arm/mmu.c b/virt/kvm/arm/mmu.c > index ed8f8271c389..e73909a31e02 100644 > --- a/virt/kvm/arm/mmu.c > +++ b/virt/kvm/arm/mmu.c > @@ -1038,23 +1038,62 @@ static int stage2_set_pmd_huge(struct kvm *kvm, struct kvm_mmu_memory_cache > return 0; > } > > -static bool stage2_is_exec(struct kvm *kvm, phys_addr_t addr) > +/* > + * stage2_get_leaf_entry - walk the stage2 VM page tables and return > + * true if a valid and present leaf-entry is found. A pointer to the > + * leaf-entry is returned in the appropriate level variable - pudpp, > + * pmdpp, ptepp. > + */ > +static bool stage2_get_leaf_entry(struct kvm *kvm, phys_addr_t addr, > + pud_t **pudpp, pmd_t **pmdpp, pte_t **ptepp) > { > + pud_t *pudp; > pmd_t *pmdp; > pte_t *ptep; nit: As mentioned in the other thread, you may initialize the reference pointers to NULL to make sure we start clean and avoid the initialization everywhere this is called. > > - pmdp = stage2_get_pmd(kvm, NULL, addr); > + pudp = stage2_get_pud(kvm, NULL, addr); > + if (!pudp || pud_none(*pudp) || !pud_present(*pudp)) > + return false; > + > + if (pud_huge(*pudp)) { > + *pudpp = pudp; > + return true; > + } > + > + pmdp = stage2_pmd_offset(pudp, addr); > if (!pmdp || pmd_none(*pmdp) || !pmd_present(*pmdp)) > return false; > > - if (pmd_thp_or_huge(*pmdp)) > - return kvm_s2pmd_exec(pmdp); > + if (pmd_thp_or_huge(*pmdp)) { > + *pmdpp = pmdp; > + return true; > + } > > ptep = pte_offset_kernel(pmdp, addr); > if (!ptep || pte_none(*ptep) || !pte_present(*ptep)) > return false; > > - return kvm_s2pte_exec(ptep); > + *ptepp = ptep; > + return true; > +} > + > +static bool stage2_is_exec(struct kvm *kvm, phys_addr_t addr) > +{ > + pud_t *pudp = NULL; > + pmd_t *pmdp = NULL; > + pte_t *ptep = NULL; > + bool found; > + > + found = stage2_get_leaf_entry(kvm, addr, &pudp, &pmdp, &ptep); > + if (!found) > + return false; > + > + if (pudp) > + return kvm_s2pud_exec(pudp); > + else if (pmdp) > + return kvm_s2pmd_exec(pmdp); > + else > + return kvm_s2pte_exec(ptep); > } Reviewed-by: Suzuki K Poulose