Received: by 10.192.165.156 with SMTP id m28csp780703imm; Mon, 16 Apr 2018 08:36:33 -0700 (PDT) X-Google-Smtp-Source: AIpwx4/G8jNWppXK7Soh1ffBZmvGet00ENGGbTEhopY2n9RGgh+GbMv8xv1sNiycrsAcaju1kSW6 X-Received: by 2002:a17:902:5381:: with SMTP id c1-v6mr15763486pli.234.1523892993053; Mon, 16 Apr 2018 08:36:33 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1523892993; cv=none; d=google.com; s=arc-20160816; b=yPmIYNFg75Mf7WR40z5EpOtoEowQdCJGtJuKkbaqLfoLcmYSODuyII712SKVgHSQ39 DKc/oAJA3LALbejDhg4FopilkkJ5PhkFmcZt1u9cDqaWgFna5W4yzQm3FC3ZNE73WiIf gwt8m2D9ZqyGlDBTtnwu9yRYEWT97Ynfw76/TNluoG+YNd0LP9jMUSZkUe9Nb4qcI+5y aKlqS1NCn8ELBtVpRtr8XCYIbQCScg409yfXwJhLUCcsrKUJV1eiXyP8jFWXLc0uvj+Y 5l+boZyK3qHzK3mlhuir0prm7/YjsYJAXTAttPP8x9eNdWB7naz16Hp1ZPgJRsBl/mGA WMnw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:references:in-reply-to:message-id:date :subject:cc:to:from:dkim-signature:arc-authentication-results; bh=5MAYDMDeYiDxuXE5s4LEJcEc5BqHNOgZCeQPhRYBVSk=; b=JAfQVicTFtsOVxcq2ryTfL6SGU8v1lqZf7vABGrXzjHwUAhxJJl7thyTARtBfQkWQa LhRh/Are1RJAh1Io/R6Nw6ElntJieKRGRzEsnP30ERMktKvF+O5dhjxdv0o2k3vp9cpn yElHVrIE+6N6fjiVVyELVQAvoR8Fx0TtLPweNk8vzv4YdGJDx7/OpGvAxsaXvbDAX+17 jXW42dlc0+wPQqhVeNeXpTWJYPN250L3qPq5wk8+yw/O0+4nDXYh+cFxis7ud/9ndkof K+2QcKg19RG1J2r2ry9WjC9OAfBzdGI6FXwbaj7Zn7OvtXZHF4bk5r/wEiX7XcAcLZ8O Sb5Q== ARC-Authentication-Results: i=1; mx.google.com; dkim=fail (test mode) header.i=@8bytes.org header.s=mail-1 header.b=fqoYvgX0; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=8bytes.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id l4si10001736pgs.571.2018.04.16.08.36.19; Mon, 16 Apr 2018 08:36:32 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=fail (test mode) header.i=@8bytes.org header.s=mail-1 header.b=fqoYvgX0; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=8bytes.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1753178AbeDPPcn (ORCPT + 99 others); Mon, 16 Apr 2018 11:32:43 -0400 Received: from 8bytes.org ([81.169.241.247]:35932 "EHLO theia.8bytes.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1752892AbeDPPZr (ORCPT ); Mon, 16 Apr 2018 11:25:47 -0400 Received: by theia.8bytes.org (Postfix, from userid 1000) id 5D97C968; Mon, 16 Apr 2018 17:25:38 +0200 (CEST) DKIM-Signature: v=1; a=rsa-sha256; c=simple/simple; d=8bytes.org; s=mail-1; t=1523892339; bh=czZInasadqZYEyKL7gm23F/hesj6ls/BrwELPTXFMEA=; h=From:To:Cc:Subject:Date:In-Reply-To:References:From; b=fqoYvgX0c9Z5EPMmY01pT26qBLUl2eti0AMB3IOo1B9htmFZK574PEAe/0hsbyOqv SyzFf9wWVuPyWqnpDy/yoVQHEULcTRiiqc5WsE2DlwkgCh4aSCcohlHwJjZRDKEBdg FBxUIyN1Ra6Ws0p8lyM11Q78qXq3gh6WLGCX3Efm9DgIWIX5GCE4UTE1J4dtWLhCXC Y2YyslqJs1RPHVrU4JHPwbDnrI+MboqWkWURfbnsDvpx4lQXmGGDbaZK4hofM1EaZ2 YbKnaY8UPplvucWmqGtqMHW1cgxojZDWZA0RY4MxSSoN19tDtPy6CcBkyS8Sf+Y8PC k6RnXnqI4J3+w== From: Joerg Roedel To: Thomas Gleixner , Ingo Molnar , "H . Peter Anvin" Cc: x86@kernel.org, linux-kernel@vger.kernel.org, linux-mm@kvack.org, Linus Torvalds , Andy Lutomirski , Dave Hansen , Josh Poimboeuf , Juergen Gross , Peter Zijlstra , Borislav Petkov , Jiri Kosina , Boris Ostrovsky , Brian Gerst , David Laight , Denys Vlasenko , Eduardo Valentin , Greg KH , Will Deacon , aliguori@amazon.com, daniel.gruss@iaik.tugraz.at, hughd@google.com, keescook@google.com, Andrea Arcangeli , Waiman Long , Pavel Machek , "David H . Gutteridge" , jroedel@suse.de, joro@8bytes.org Subject: [PATCH 13/35] x86/entry/32: Add PTI cr3 switch to non-NMI entry/exit points Date: Mon, 16 Apr 2018 17:25:01 +0200 Message-Id: <1523892323-14741-14-git-send-email-joro@8bytes.org> X-Mailer: git-send-email 2.7.4 In-Reply-To: <1523892323-14741-1-git-send-email-joro@8bytes.org> References: <1523892323-14741-1-git-send-email-joro@8bytes.org> Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org From: Joerg Roedel Add unconditional cr3 switches between user and kernel cr3 to all non-NMI entry and exit points. Signed-off-by: Joerg Roedel --- arch/x86/entry/entry_32.S | 83 ++++++++++++++++++++++++++++++++++++++++++++--- 1 file changed, 79 insertions(+), 4 deletions(-) diff --git a/arch/x86/entry/entry_32.S b/arch/x86/entry/entry_32.S index 71e1cb3..b2b0ecb 100644 --- a/arch/x86/entry/entry_32.S +++ b/arch/x86/entry/entry_32.S @@ -154,6 +154,33 @@ #endif /* CONFIG_X86_32_LAZY_GS */ +/* Unconditionally switch to user cr3 */ +.macro SWITCH_TO_USER_CR3 scratch_reg:req + ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI + + movl %cr3, \scratch_reg + orl $PTI_SWITCH_MASK, \scratch_reg + movl \scratch_reg, %cr3 +.Lend_\@: +.endm + +/* + * Switch to kernel cr3 if not already loaded and return current cr3 in + * \scratch_reg + */ +.macro SWITCH_TO_KERNEL_CR3 scratch_reg:req + ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI + movl %cr3, \scratch_reg + /* Test if we are already on kernel CR3 */ + testl $PTI_SWITCH_MASK, \scratch_reg + jz .Lend_\@ + andl $(~PTI_SWITCH_MASK), \scratch_reg + movl \scratch_reg, %cr3 + /* Return original CR3 in \scratch_reg */ + orl $PTI_SWITCH_MASK, \scratch_reg +.Lend_\@: +.endm + .macro SAVE_ALL pt_regs_ax=%eax switch_stacks=0 cld /* Push segment registers and %eax */ @@ -288,7 +315,6 @@ #endif /* CONFIG_X86_ESPFIX32 */ .endm - /* * Called with pt_regs fully populated and kernel segments loaded, * so we can access PER_CPU and use the integer registers. @@ -301,11 +327,19 @@ */ #define CS_FROM_ENTRY_STACK (1 << 31) +#define CS_FROM_USER_CR3 (1 << 30) .macro SWITCH_TO_KERNEL_STACK ALTERNATIVE "", "jmp .Lend_\@", X86_FEATURE_XENPV + SWITCH_TO_KERNEL_CR3 scratch_reg=%eax + + /* + * %eax now contains the entry cr3 and we carry it forward in + * that register for the time this macro runs + */ + /* Are we on the entry stack? Bail out if not! */ movl PER_CPU_VAR(cpu_entry_area), %edi addl $CPU_ENTRY_AREA_entry_stack, %edi @@ -374,7 +408,8 @@ * but switch back to the entry-stack again when we approach * iret and return to the interrupted code-path. This usually * happens when we hit an exception while restoring user-space - * segment registers on the way back to user-space. + * segment registers on the way back to user-space or when the + * sysenter handler runs with eflags.tf set. * * When we switch to the task-stack here, we can't trust the * contents of the entry-stack anymore, as the exception handler @@ -391,6 +426,7 @@ * * %esi: Entry-Stack pointer (same as %esp) * %edi: Top of the task stack + * %eax: CR3 on kernel entry */ /* Calculate number of bytes on the entry stack in %ecx */ @@ -407,6 +443,14 @@ orl $CS_FROM_ENTRY_STACK, PT_CS(%esp) /* + * Test the cr3 used to enter the kernel and add a marker + * so that we can switch back to it before iret. + */ + testl $PTI_SWITCH_MASK, %eax + jz .Lcopy_pt_regs_\@ + orl $CS_FROM_USER_CR3, PT_CS(%esp) + + /* * %esi and %edi are unchanged, %ecx contains the number of * bytes to copy. The code at .Lcopy_pt_regs_\@ will allocate * the stack-frame on task-stack and copy everything over @@ -472,7 +516,7 @@ /* * This macro handles the case when we return to kernel-mode on the iret - * path and have to switch back to the entry stack. + * path and have to switch back to the entry stack and/or user-cr3 * * See the comments below the .Lentry_from_kernel_\@ label in the * SWITCH_TO_KERNEL_STACK macro for more details. @@ -518,6 +562,18 @@ /* Safe to switch to entry-stack now */ movl %ebx, %esp + /* + * We came from entry-stack and need to check if we also need to + * switch back to user cr3. + */ + testl $CS_FROM_USER_CR3, PT_CS(%esp) + jz .Lend_\@ + + /* Clear marker from stack-frame */ + andl $(~CS_FROM_USER_CR3), PT_CS(%esp) + + SWITCH_TO_USER_CR3 scratch_reg=%eax + .Lend_\@: .endm /* @@ -711,6 +767,18 @@ ENTRY(xen_sysenter_target) * 0(%ebp) arg6 */ ENTRY(entry_SYSENTER_32) + /* + * On entry-stack with all userspace-regs live - save and + * restore eflags and %eax to use it as scratch-reg for the cr3 + * switch. + */ + pushfl + pushl %eax + SWITCH_TO_KERNEL_CR3 scratch_reg=%eax + popl %eax + popfl + + /* Stack empty again, switch to task stack */ movl TSS_entry_stack(%esp), %esp .Lsysenter_past_esp: @@ -791,6 +859,9 @@ ENTRY(entry_SYSENTER_32) /* Switch to entry stack */ movl %eax, %esp + /* Now ready to switch the cr3 */ + SWITCH_TO_USER_CR3 scratch_reg=%eax + /* * Restore all flags except IF. (We restore IF separately because * STI gives a one-instruction window in which we won't be interrupted, @@ -871,7 +942,11 @@ restore_all: .Lrestore_all_notrace: CHECK_AND_APPLY_ESPFIX .Lrestore_nocheck: - RESTORE_REGS 4 # skip orig_eax/error_code + /* Switch back to user CR3 */ + SWITCH_TO_USER_CR3 scratch_reg=%eax + + /* Restore user state */ + RESTORE_REGS pop=4 # skip orig_eax/error_code .Lirq_return: /* * ARCH_HAS_MEMBARRIER_SYNC_CORE rely on IRET core serialization -- 2.7.4