Received: by 10.223.176.5 with SMTP id f5csp3383955wra; Mon, 29 Jan 2018 12:24:48 -0800 (PST) X-Google-Smtp-Source: AH8x226oSGyH1Gzzx661nEVCmG3AT7LS6UuLQk7/3gfVF0soksk6S49XxKRyl/7J9ZUTlwqB0GK2 X-Received: by 2002:a17:902:9a9:: with SMTP id 38-v6mr22600448pln.202.1517257488850; Mon, 29 Jan 2018 12:24:48 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1517257488; cv=none; d=google.com; s=arc-20160816; b=wIFGwEqV0DnRtbIGUm1G0KrTtRxqEGhDmmj0xVbWQliJRnwD4xr+FmQQoNwRHSvDmP VFa5HviTg9qnWzoxSPE0im8OJBZJypzmyY1FU6wdyXYlxsy9eAlTkIW5qf6sdMkTOmii dUTs3ySqUg7Rtr0DbtRWEV1eSYkS8s9rxT+OzL1QI2YbHW6csTcLZynkKtsqf83wKpZh JjnSDybm7H+CZUR4n4uiy1oyWMy6IWwOWIUBuI1Ws9Kve55BM7DPGSqYy5R/ruoy53lZ 0t0AcP5T1WTTJDKTefwt3kEgDN6CmQL4mtgjXnDHZ35Z1OhtZuWSSAXGiIV1FFzOWT8U 2Rgg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:references:in-reply-to:message-id:date :subject:cc:to:from:arc-authentication-results; bh=RyDOMsujwbpUnTep7fzb6eiztgDAa+QC8wLFbuxpum4=; b=xiLSdmozfGZuEQzRon7SNxVNlZTAceCwp1grpAaTMMFA/eMvYkmVooyYI4iHIFbBVd WPSu87R3Moami5Ey67HhkrC21lQTMOvi/zSmZGc3iq6Wfzcgr77eNF4UMqOWnPRib2YE vfy1sUzd+sEUyJ3OwKk9xWHQNPD+6EmznCiGFB429gDipr6NTsa6qa5YXNGPrUv+YIpV YOIyLQlAXAxjxG5YvcxWqFLEdOdfq1fRyCjaOudhN60colfgvX+DkKbgEMto64zdLdOf u0uzxild5gE+KWZEOkW+VQJCVfL4hzxFRAjIghBxO8iUAWyAj/T9nk8zSeZ+qEF4ZLx+ 1VXQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id r12si22937pfh.206.2018.01.29.12.24.34; Mon, 29 Jan 2018 12:24:48 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1754802AbeA2UWd (ORCPT + 99 others); Mon, 29 Jan 2018 15:22:33 -0500 Received: from mail.efficios.com ([167.114.142.141]:53209 "EHLO mail.efficios.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1754604AbeA2UWV (ORCPT ); Mon, 29 Jan 2018 15:22:21 -0500 Received: from localhost (localhost.localdomain [127.0.0.1]) by mail.efficios.com (Postfix) with ESMTP id CD22434015C; Mon, 29 Jan 2018 20:22:52 +0000 (UTC) Received: from mail.efficios.com ([127.0.0.1]) by localhost (evm-mail-1.efficios.com [127.0.0.1]) (amavisd-new, port 10032) with ESMTP id S58ckw_80occ; Mon, 29 Jan 2018 20:22:39 +0000 (UTC) Received: from localhost (localhost.localdomain [127.0.0.1]) by mail.efficios.com (Postfix) with ESMTP id AF30F3402EC; Mon, 29 Jan 2018 20:22:39 +0000 (UTC) X-Virus-Scanned: amavisd-new at efficios.com Received: from mail.efficios.com ([127.0.0.1]) by localhost (evm-mail-1.efficios.com [127.0.0.1]) (amavisd-new, port 10026) with ESMTP id euA4N3b_nPZ7; Mon, 29 Jan 2018 20:22:39 +0000 (UTC) Received: from thinkos.internal.efficios.com (192-222-157-41.qc.cable.ebox.net [192.222.157.41]) by mail.efficios.com (Postfix) with ESMTPSA id 49C453402DE; Mon, 29 Jan 2018 20:22:39 +0000 (UTC) From: Mathieu Desnoyers To: Ingo Molnar , Peter Zijlstra , Thomas Gleixner Cc: linux-kernel@vger.kernel.org, linux-api@vger.kernel.org, Andy Lutomirski , "Paul E . McKenney" , Boqun Feng , Andrew Hunter , Maged Michael , Avi Kivity , Benjamin Herrenschmidt , Paul Mackerras , Michael Ellerman , Dave Watson , "H . Peter Anvin" , Andrea Parri , Russell King , Greg Hackmann , Will Deacon , David Sehr , Linus Torvalds , x86@kernel.org, Mathieu Desnoyers Subject: [PATCH for 4.16 v5 03/11] membarrier: Document scheduler barrier requirements Date: Mon, 29 Jan 2018 15:20:12 -0500 Message-Id: <20180129202020.8515-4-mathieu.desnoyers@efficios.com> X-Mailer: git-send-email 2.11.0 In-Reply-To: <20180129202020.8515-1-mathieu.desnoyers@efficios.com> References: <20180129202020.8515-1-mathieu.desnoyers@efficios.com> Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Document the membarrier requirement on having a full memory barrier in __schedule() after coming from user-space, before storing to rq->curr. It is provided by smp_mb__after_spinlock() in __schedule(). Document that membarrier requires a full barrier on transition from kernel thread to userspace thread. We currently have an implicit barrier from atomic_dec_and_test() in mmdrop() that ensures this. The x86 switch_mm_irqs_off() full barrier is currently provided by many cpumask update operations as well as write_cr3(). Document that write_cr3() provides this barrier. Signed-off-by: Mathieu Desnoyers Acked-by: Peter Zijlstra (Intel) CC: Paul E. McKenney CC: Boqun Feng CC: Andrew Hunter CC: Maged Michael CC: Avi Kivity CC: Benjamin Herrenschmidt CC: Paul Mackerras CC: Michael Ellerman CC: Dave Watson CC: Thomas Gleixner CC: Ingo Molnar CC: "H. Peter Anvin" CC: Andrea Parri CC: x86@kernel.org --- Changes since v1: - Update comments to match reality for code paths which are after storing to rq->curr, before returning to user-space, based on feedback from Andrea Parri. Changes since v2: - Update changelog (smp_mb__before_spinlock -> smp_mb__after_spinlock). Based on feedback from Andrea Parri. Changes since v3: - Clarify comments following feeback from Peter Zijlstra. Changes since v4: - Update comment regarding powerpc barrier. --- arch/x86/mm/tlb.c | 5 +++++ include/linux/sched/mm.h | 5 +++++ kernel/sched/core.c | 37 ++++++++++++++++++++++++++----------- 3 files changed, 36 insertions(+), 11 deletions(-) diff --git a/arch/x86/mm/tlb.c b/arch/x86/mm/tlb.c index 5bfe61a5e8e3..9fa7d2e0e15e 100644 --- a/arch/x86/mm/tlb.c +++ b/arch/x86/mm/tlb.c @@ -228,6 +228,11 @@ void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next, #endif this_cpu_write(cpu_tlbstate.is_lazy, false); + /* + * The membarrier system call requires a full memory barrier + * before returning to user-space, after storing to rq->curr. + * Writing to CR3 provides that full memory barrier. + */ if (real_prev == next) { VM_WARN_ON(this_cpu_read(cpu_tlbstate.ctxs[prev_asid].ctx_id) != next->context.ctx_id); diff --git a/include/linux/sched/mm.h b/include/linux/sched/mm.h index 1754396795f6..28aef7051d73 100644 --- a/include/linux/sched/mm.h +++ b/include/linux/sched/mm.h @@ -39,6 +39,11 @@ static inline void mmgrab(struct mm_struct *mm) extern void __mmdrop(struct mm_struct *); static inline void mmdrop(struct mm_struct *mm) { + /* + * The implicit full barrier implied by atomic_dec_and_test is + * required by the membarrier system call before returning to + * user-space, after storing to rq->curr. + */ if (unlikely(atomic_dec_and_test(&mm->mm_count))) __mmdrop(mm); } diff --git a/kernel/sched/core.c b/kernel/sched/core.c index c7e06dfa804b..f38c4c7e256a 100644 --- a/kernel/sched/core.c +++ b/kernel/sched/core.c @@ -2657,6 +2657,12 @@ static struct rq *finish_task_switch(struct task_struct *prev) finish_arch_post_lock_switch(); fire_sched_in_preempt_notifiers(current); + /* + * When transitioning from a kernel thread to a userspace + * thread, mmdrop()'s implicit full barrier is required by the + * membarrier system call, because the current active_mm can + * become the current mm without going through switch_mm(). + */ if (mm) mmdrop(mm); if (unlikely(prev_state == TASK_DEAD)) { @@ -2762,6 +2768,13 @@ context_switch(struct rq *rq, struct task_struct *prev, */ arch_start_context_switch(prev); + /* + * If mm is non-NULL, we pass through switch_mm(). If mm is + * NULL, we will pass through mmdrop() in finish_task_switch(). + * Both of these contain the full memory barrier required by + * membarrier after storing to rq->curr, before returning to + * user-space. + */ if (!mm) { next->active_mm = oldmm; mmgrab(oldmm); @@ -3298,6 +3311,9 @@ static void __sched notrace __schedule(bool preempt) * Make sure that signal_pending_state()->signal_pending() below * can't be reordered with __set_current_state(TASK_INTERRUPTIBLE) * done by the caller to avoid the race with signal_wake_up(). + * + * The membarrier system call requires a full memory barrier + * after coming from user-space, before storing to rq->curr. */ rq_lock(rq, &rf); smp_mb__after_spinlock(); @@ -3345,17 +3361,16 @@ static void __sched notrace __schedule(bool preempt) /* * The membarrier system call requires each architecture * to have a full memory barrier after updating - * rq->curr, before returning to user-space. For TSO - * (e.g. x86), the architecture must provide its own - * barrier in switch_mm(). For weakly ordered machines - * for which spin_unlock() acts as a full memory - * barrier, finish_lock_switch() in common code takes - * care of this barrier. For weakly ordered machines for - * which spin_unlock() acts as a RELEASE barrier (only - * arm64 and PowerPC), arm64 has a full barrier in - * switch_to(), and PowerPC has - * smp_mb__after_unlock_lock() before - * finish_lock_switch(). + * rq->curr, before returning to user-space. + * + * Here are the schemes providing that barrier on the + * various architectures: + * - mm ? switch_mm() : mmdrop() for x86, s390, sparc, PowerPC. + * switch_mm() rely on membarrier_arch_switch_mm() on PowerPC. + * - finish_lock_switch() for weakly-ordered + * architectures where spin_unlock is a full barrier, + * - switch_to() for arm64 (weakly-ordered, spin_unlock + * is a RELEASE barrier), */ ++*switch_count; -- 2.11.0