Received: by 2002:ad5:474a:0:0:0:0:0 with SMTP id i10csp2980079imu; Sun, 9 Dec 2018 14:11:29 -0800 (PST) X-Google-Smtp-Source: AFSGD/X4UsgAzaR1TxWK7iP9kFQ12xgEUAfv+/7rcOMbL/qrC9JdHBhVTCIIfrQnZ8yXJVJiryZm X-Received: by 2002:a63:4e15:: with SMTP id c21mr8873419pgb.50.1544393489918; Sun, 09 Dec 2018 14:11:29 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1544393489; cv=none; d=google.com; s=arc-20160816; b=a/9TD/tolNx81eoh+kZZ/ZigPjI/9wdVBJzIUHPNTHPG0YG6FWWS/qj6EuzKgPHuuo skBQh+DDP4IsvUIe7ctNv+lB7cL27CROMiYHuwO9L0iWTHYQ0rpb+cygv1LuZlhOCTJC 7BNU2MONI/OwN9sMhARRCdJHszKGIhBFEIodl3Oh68Rm2L6JIRHYP+RXrntf0hcUJAYw kn2sCnmhNi1zCh3HI+2WfmZpUuxNOuXmt7WU+IxvkhTWOxupJkcNeJ7F5etNgOaEMCbp RkVlEvFP1QcYIAR8J1gKzB2WHbluPu01T/U6Vf86ZSzWAIt+3TM6bTXiBv/mdfLv4Lfg ZZ5w== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:in-reply-to:subject:message-id:date:cc:to :from:mime-version:content-transfer-encoding:content-disposition; bh=+Jfg5ieRqz4ZB/Lyq7aa+xTjkqk7wIZGGLfIC6cpRTs=; b=ji8rUQw9Lfv5fDvYLbhqXtzfTU77BXzk8nra6hNlRr/FBDNveJq6M5xPkRk20TtNhM 8Qv92bMz03FGWeVPiYRncr+eNgrcjzL0ThspplHr+YdoZvwUcoYMcantSyGrg/AxZpp5 iIzr8xJ/HjJ0qKar8/rqzMLZkLqvpcn4augnFDpOqAGe02PtGNsYbfnnDYS0dVOBmWRY nZnnrWleLGhfIupj1Z1xFrASXl0yEQDIqJlIWNPNZV4bFf4LG6x9kTIWGuksqwPTifOs fPcPPestESkhh8p6SlZ80KRgE73tgpwNRz+cE+2mH+l3Ee+SRbQbIDCMhpoLixc0Woqy wVpw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id k62si8782862pfc.208.2018.12.09.14.11.14; Sun, 09 Dec 2018 14:11:29 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1728133AbeLIWJy (ORCPT + 99 others); Sun, 9 Dec 2018 17:09:54 -0500 Received: from shadbolt.e.decadent.org.uk ([88.96.1.126]:37482 "EHLO shadbolt.e.decadent.org.uk" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726314AbeLIWJs (ORCPT ); Sun, 9 Dec 2018 17:09:48 -0500 Received: from pub.yeoldevic.com ([81.174.156.145] helo=deadeye) by shadbolt.decadent.org.uk with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.89) (envelope-from ) id 1gW73E-0002pr-Pd; Sun, 09 Dec 2018 21:55:49 +0000 Received: from ben by deadeye with local (Exim 4.91) (envelope-from ) id 1gW72d-0003Nx-B2; Sun, 09 Dec 2018 21:55:11 +0000 Content-Type: text/plain; charset="UTF-8" Content-Disposition: inline Content-Transfer-Encoding: 8bit MIME-Version: 1.0 From: Ben Hutchings To: linux-kernel@vger.kernel.org, stable@vger.kernel.org CC: akpm@linux-foundation.org, "Marc Zyngier" , "Christoffer Dall" , "Punit Agrawal" , "Suzuki Poulose" Date: Sun, 09 Dec 2018 21:50:33 +0000 Message-ID: X-Mailer: LinuxStableQueue (scripts by bwh) X-Patchwork-Hint: ignore Subject: [PATCH 3.16 133/328] KVM: arm/arm64: Skip updating PMD entry if no change In-Reply-To: X-SA-Exim-Connect-IP: 81.174.156.145 X-SA-Exim-Mail-From: ben@decadent.org.uk X-SA-Exim-Scanned: No (on shadbolt.decadent.org.uk); SAEximRunCond expanded to false Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org 3.16.62-rc1 review patch. If anyone has any objections, please let me know. ------------------ From: Punit Agrawal commit 86658b819cd0a9aa584cd84453ed268a6f013770 upstream. Contention on updating a PMD entry by a large number of vcpus can lead to duplicate work when handling stage 2 page faults. As the page table update follows the break-before-make requirement of the architecture, it can lead to repeated refaults due to clearing the entry and flushing the tlbs. This problem is more likely when - * there are large number of vcpus * the mapping is large block mapping such as when using PMD hugepages (512MB) with 64k pages. Fix this by skipping the page table update if there is no change in the entry being updated. Fixes: ad361f093c1e ("KVM: ARM: Support hugetlbfs backed huge pages") Reviewed-by: Suzuki Poulose Acked-by: Christoffer Dall Signed-off-by: Punit Agrawal Signed-off-by: Marc Zyngier [bwh: Backported to 3.16: adjust filename] Signed-off-by: Ben Hutchings --- arch/arm/kvm/mmu.c | 38 +++++++++++++++++++++++++++----------- 1 file changed, 27 insertions(+), 11 deletions(-) --- a/arch/arm/kvm/mmu.c +++ b/arch/arm/kvm/mmu.c @@ -685,19 +685,35 @@ static int stage2_set_pmd_huge(struct kv pmd = stage2_get_pmd(kvm, cache, addr); VM_BUG_ON(!pmd); - /* - * Mapping in huge pages should only happen through a fault. If a - * page is merged into a transparent huge page, the individual - * subpages of that huge page should be unmapped through MMU - * notifiers before we get here. - * - * Merging of CompoundPages is not supported; they should become - * splitting first, unmapped, merged, and mapped back in on-demand. - */ - VM_BUG_ON(pmd_present(*pmd) && pmd_pfn(*pmd) != pmd_pfn(*new_pmd)); - old_pmd = *pmd; if (pmd_present(old_pmd)) { + /* + * Multiple vcpus faulting on the same PMD entry, can + * lead to them sequentially updating the PMD with the + * same value. Following the break-before-make + * (pmd_clear() followed by tlb_flush()) process can + * hinder forward progress due to refaults generated + * on missing translations. + * + * Skip updating the page table if the entry is + * unchanged. + */ + if (pmd_val(old_pmd) == pmd_val(*new_pmd)) + return 0; + + /* + * Mapping in huge pages should only happen through a + * fault. If a page is merged into a transparent huge + * page, the individual subpages of that huge page + * should be unmapped through MMU notifiers before we + * get here. + * + * Merging of CompoundPages is not supported; they + * should become splitting first, unmapped, merged, + * and mapped back in on-demand. + */ + VM_BUG_ON(pmd_pfn(old_pmd) != pmd_pfn(*new_pmd)); + pmd_clear(pmd); kvm_tlb_flush_vmid_ipa(kvm, addr); } else {