Received: by 2002:ac0:a5b6:0:0:0:0:0 with SMTP id m51-v6csp3929857imm; Mon, 4 Jun 2018 11:39:06 -0700 (PDT) X-Google-Smtp-Source: ADUXVKJF0cdpjn2m47FKN0b68Jwe09hQbn+qudVXhprhg54uwTI4NePJmfAIE76u4i/L8ZrUVbd7 X-Received: by 2002:a63:5ec4:: with SMTP id s187-v6mr17798211pgb.72.1528137546825; Mon, 04 Jun 2018 11:39:06 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1528137546; cv=none; d=google.com; s=arc-20160816; b=a+OboYBohSSvPqs0IZ4PrDAIL3bKoHQHAMYNKnZth2lmm/qSZM5XS+a/aXDGo93afB VdRszNHzskvm6lqfuOZfACIVY7KSC7VY48Hul78QbGKg2U8jVUZjBXBB1Te6/cEWWbM/ NkUpXTqqqmYgiGuYNg7hxyb43Bac9dJJ6D2gNeduFel25iFPt2wla91aKJp3lbDPFInn bWmSUOSyFDGjhe1baPdF5TQZH+QfamRxfdj6Oiy+ytYfSb9BcAVcbwnizleq5Jb0Okyf MVwriiCfSadoXsqxhpa25xH6cZQ/ummqX9Ni/Ej+nNk9BBo53P0UDIw/bO0CltLOnbPF NhZg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:arc-authentication-results; bh=fh3o/HYcmk+U8ewuobcn5iKQMlTBVQ+eiuHm9Yrr+Wc=; b=g638+GP5PeN4ntOvKD5naJAaaHu46QIOkPI4rAq6aGuwQrIMi9pj/8RyRcF7im3ggK H0HKgPMjzLCndr5itC1lCu0u5Y/m65btcmYPfasOqRwE7OzZYo0Zs7FmBbQI9PxvEOtD 6LaLzXoQ2TJiu4xduvMEvJohYOH80tBWoQ8vLMbldTXfxYXFEZmX/pz2YvduNcVsPRF0 mvltM1aUowsCrtSGM/BXV6P7J/fabvqMtL9Le/RMIxbBjEDmlynzSWrEPRomOIemFT3K BVBe1S12PF44LDEqfGWpFfM4tHT99kr7+G8F0pldxMpKzK1KuC8r3Fa+WURdKw5NzUlf +3uQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=vmware.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id u8-v6si3796161plh.492.2018.06.04.11.38.30; Mon, 04 Jun 2018 11:39:06 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=vmware.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1751425AbeFDShH (ORCPT + 99 others); Mon, 4 Jun 2018 14:37:07 -0400 Received: from ex13-edg-ou-002.vmware.com ([208.91.0.190]:51805 "EHLO EX13-EDG-OU-002.vmware.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751219AbeFDSga (ORCPT ); Mon, 4 Jun 2018 14:36:30 -0400 Received: from sc9-mailhost2.vmware.com (10.113.161.72) by EX13-EDG-OU-002.vmware.com (10.113.208.156) with Microsoft SMTP Server id 15.0.1156.6; Mon, 4 Jun 2018 11:36:15 -0700 Received: from sc2-haas01-esx0118.eng.vmware.com (sc2-haas01-esx0118.eng.vmware.com [10.172.44.118]) by sc9-mailhost2.vmware.com (Postfix) with ESMTP id 397C7B0A0F; Mon, 4 Jun 2018 11:36:27 -0700 (PDT) From: Nadav Amit To: , CC: Nadav Amit , Juergen Gross , Alok Kataria , Thomas Gleixner , Ingo Molnar , "H. Peter Anvin" , Subject: [PATCH v2 6/9] x86: prevent inline distortion by paravirt ops Date: Mon, 4 Jun 2018 04:21:28 -0700 Message-ID: <20180604112131.59100-7-namit@vmware.com> X-Mailer: git-send-email 2.17.0 In-Reply-To: <20180604112131.59100-1-namit@vmware.com> References: <20180604112131.59100-1-namit@vmware.com> MIME-Version: 1.0 Content-Type: text/plain Received-SPF: None (EX13-EDG-OU-002.vmware.com: namit@vmware.com does not designate permitted sender hosts) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org GCC considers the number of statements in inlined assembly blocks, according to new-lines and semicolons, as an indication to the cost of the block in time and space. This data is distorted by the kernel code, which puts information in alternative sections. As a result, the compiler may perform incorrect inlining and branch optimizations. The solution is to set an assembly macro and call it from the inlined assembly block. As a result GCC considers the inline assembly block as a single instruction. The effect of the patch is a more aggressive inlining, which also causes a size increase of kernel. text data bss dec hex filename 18147336 10226688 2957312 31331336 1de1408 ./vmlinux before 18162555 10226288 2957312 31346155 1de4deb ./vmlinux after (+14819) Static text symbols: Before: 40053 After: 39942 (-111) Cc: Juergen Gross Cc: Alok Kataria Cc: Thomas Gleixner Cc: Ingo Molnar Cc: "H. Peter Anvin" Cc: x86@kernel.org Cc: virtualization@lists.linux-foundation.org Signed-off-by: Nadav Amit --- arch/x86/include/asm/paravirt_types.h | 54 +++++++++++++++------------ arch/x86/kernel/macros.S | 1 + 2 files changed, 31 insertions(+), 24 deletions(-) diff --git a/arch/x86/include/asm/paravirt_types.h b/arch/x86/include/asm/paravirt_types.h index 180bc0bff0fb..2a9c53f64f1a 100644 --- a/arch/x86/include/asm/paravirt_types.h +++ b/arch/x86/include/asm/paravirt_types.h @@ -347,19 +347,15 @@ extern struct pv_lock_ops pv_lock_ops; * Generate some code, and mark it as patchable by the * apply_paravirt() alternate instruction patcher. */ -#define _paravirt_alt(insn_string, type, clobber) \ - "771:\n\t" insn_string "\n" "772:\n" \ - ".pushsection .parainstructions,\"a\"\n" \ - _ASM_ALIGN "\n" \ - _ASM_PTR " 771b\n" \ - " .byte " type "\n" \ - " .byte 772b-771b\n" \ - " .short " clobber "\n" \ - ".popsection\n" +#define _paravirt_alt(type, clobber, pv_opptr) \ + "PARAVIRT_ALT type=" __stringify(type) \ + " clobber=" __stringify(clobber) \ + " pv_opptr=" __stringify(pv_opptr) "\n\t" /* Generate patchable code, with the default asm parameters. */ -#define paravirt_alt(insn_string) \ - _paravirt_alt(insn_string, "%c[paravirt_typenum]", "%c[paravirt_clobber]") +#define paravirt_alt \ + _paravirt_alt("%c[paravirt_typenum]", "%c[paravirt_clobber]", \ + "%c[paravirt_opptr]") /* Simple instruction patching code. */ #define NATIVE_LABEL(a,x,b) "\n\t.globl " a #x "_" #b "\n" a #x "_" #b ":\n\t" @@ -387,16 +383,6 @@ unsigned native_patch(u8 type, u16 clobbers, void *ibuf, int paravirt_disable_iospace(void); -/* - * This generates an indirect call based on the operation type number. - * The type number, computed in PARAVIRT_PATCH, is derived from the - * offset into the paravirt_patch_template structure, and can therefore be - * freely converted back into a structure offset. - */ -#define PARAVIRT_CALL \ - ANNOTATE_RETPOLINE_SAFE \ - "call *%c[paravirt_opptr];" - /* * These macros are intended to wrap calls through one of the paravirt * ops structs, so that they can be later identified and patched at @@ -534,7 +520,7 @@ int paravirt_disable_iospace(void); /* since this condition will never hold */ \ if (sizeof(rettype) > sizeof(unsigned long)) { \ asm volatile(pre \ - paravirt_alt(PARAVIRT_CALL) \ + paravirt_alt \ post \ : call_clbr, ASM_CALL_CONSTRAINT \ : paravirt_type(op), \ @@ -544,7 +530,7 @@ int paravirt_disable_iospace(void); __ret = (rettype)((((u64)__edx) << 32) | __eax); \ } else { \ asm volatile(pre \ - paravirt_alt(PARAVIRT_CALL) \ + paravirt_alt \ post \ : call_clbr, ASM_CALL_CONSTRAINT \ : paravirt_type(op), \ @@ -571,7 +557,7 @@ int paravirt_disable_iospace(void); PVOP_VCALL_ARGS; \ PVOP_TEST_NULL(op); \ asm volatile(pre \ - paravirt_alt(PARAVIRT_CALL) \ + paravirt_alt \ post \ : call_clbr, ASM_CALL_CONSTRAINT \ : paravirt_type(op), \ @@ -691,6 +677,26 @@ struct paravirt_patch_site { extern struct paravirt_patch_site __parainstructions[], __parainstructions_end[]; +#else /* __ASSEMBLY__ */ + +/* + * This generates an indirect call based on the operation type number. + * The type number, computed in PARAVIRT_PATCH, is derived from the + * offset into the paravirt_patch_template structure, and can therefore be + * freely converted back into a structure offset. + */ +.macro PARAVIRT_ALT type:req clobber:req pv_opptr:req +771: ANNOTATE_RETPOLINE_SAFE + call *\pv_opptr +772: .pushsection .parainstructions,"a" + _ASM_ALIGN + _ASM_PTR 771b + .byte \type + .byte 772b-771b + .short \clobber + .popsection +.endm + #endif /* __ASSEMBLY__ */ #endif /* _ASM_X86_PARAVIRT_TYPES_H */ diff --git a/arch/x86/kernel/macros.S b/arch/x86/kernel/macros.S index 66ccb8e823b1..71d8b716b111 100644 --- a/arch/x86/kernel/macros.S +++ b/arch/x86/kernel/macros.S @@ -10,3 +10,4 @@ #include #include #include +#include -- 2.17.0