Received: by 2002:a25:8b91:0:0:0:0:0 with SMTP id j17csp1758744ybl; Fri, 13 Dec 2019 23:11:21 -0800 (PST) X-Google-Smtp-Source: APXvYqwkxr2Uvfxeg7J73OlgL8nNXaZmueITWhMDFOj+ZDxpE89Qmrt2c4OM2CtJEX8PWyHcEeUI X-Received: by 2002:a05:6830:1b6a:: with SMTP id d10mr19792030ote.52.1576307481579; Fri, 13 Dec 2019 23:11:21 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1576307481; cv=none; d=google.com; s=arc-20160816; b=htZwWWZHhKR9RgkpzkmQKO2wl7tOgPcObJcg0DE7NquE8jMhUHQRqkSdN9HTQYzTeW DuwT/pz7joVPpD82F300zeTnaoU5iyIiCwRbxeyoTrfd2+W/ykBXRCQHPWAeUIcyx1N3 HSxgEp8jYrNGlNzrbcTj6LdKuFvSS3phNyjkI5keVAeskBt0NAQ2qm1i259t0RQ5wns/ ILPQKSM3CSrcLGCueaW+72Ig8VpdrNJh4vAQ3t3NjNwNMs2X5FIz4n0FX9RkYvmV46QD Cxsnv0YBGkaoA97I8/px0o7NIsDulrf+J4nXm6lVSZRwkTbFi6Id4d3WXI8v7d/LE1bi ZiNg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:reply-to:message-id :subject:cc:to:from:date:dkim-signature; bh=mEHpL/68zEg2GZVyYKydIefjzjK8uwYyruXqtUPzGQo=; b=XxL1iZtol0iPFvbUYm4ygoRBcqGj8iLl/phe7mf9m3hXBfK9Utwr8Jv17Jt9JFwLlj Cm6PVIALgbwyamYZGWGvJmKaV3+32lPwE1JaFNGSB+KXGn10hTz+93hS+ngP4CCGuDvV 2Jpgf9EqruDqzvi/huAAecuGDi8v9jjX8+26ObQ/+m/64mI8QUaSYQiAYda6nUMOul9G KA14y/swnUPdBZH+4Kh+Srh8/OsDVdqqvb789RGhtSSRqOmlBXDei7OZ7BVatM0uo2Y/ vGaO+idciDmdVgdfKEIjjxaAoj08tM+xakGmfNpcYGhnFRZU3G5E7Yhsyhcn2+CJxvQT AL0w== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@kernel.org header.s=default header.b=HSQWPjta; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id f190si6522909oig.229.2019.12.13.23.11.09; Fri, 13 Dec 2019 23:11:21 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@kernel.org header.s=default header.b=HSQWPjta; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1725943AbfLNHKY (ORCPT + 99 others); Sat, 14 Dec 2019 02:10:24 -0500 Received: from mail.kernel.org ([198.145.29.99]:36636 "EHLO mail.kernel.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1725871AbfLNHKX (ORCPT ); Sat, 14 Dec 2019 02:10:23 -0500 Received: from paulmck-ThinkPad-P72.home (50-39-105-78.bvtn.or.frontiernet.net [50.39.105.78]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPSA id 403C120724; Sat, 14 Dec 2019 07:10:22 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=default; t=1576307422; bh=hDUvD25Gte+PrDGnam0RqGK0I/9VsEWs4aAMnrFWYtQ=; h=Date:From:To:Cc:Subject:Reply-To:References:In-Reply-To:From; b=HSQWPjtaQxITmfeyByg3fuOEYMgis27/cFEDi8ZsCxqeWqFDMSqyoN8dcqj5Rp9z0 p0hSgWAKNSd+x4Bj7/maYyoHnjiS+1V0iiPlq1jzHA+5IEqjuQmCcshxGOOHLzFvWs qd90FrUwHsvmolceUa0BwrfKA4ZV7Ho48ctzIt2o= Received: by paulmck-ThinkPad-P72.home (Postfix, from userid 1000) id 05C6F352276B; Fri, 13 Dec 2019 23:10:22 -0800 (PST) Date: Fri, 13 Dec 2019 23:10:22 -0800 From: "Paul E. McKenney" To: Marco Elver Cc: kasan-dev@googlegroups.com, linux-kernel@vger.kernel.org, dvyukov@google.com, Randy Dunlap Subject: Re: [PATCH RESEND -rcu/kcsan] kcsan: Prefer __always_inline for fast-path Message-ID: <20191214071021.GJ2889@paulmck-ThinkPad-P72> Reply-To: paulmck@kernel.org References: <20191213204946.251125-1-elver@google.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20191213204946.251125-1-elver@google.com> User-Agent: Mutt/1.9.4 (2018-02-28) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Fri, Dec 13, 2019 at 09:49:46PM +0100, Marco Elver wrote: > Prefer __always_inline for fast-path functions that are called outside > of user_access_save, to avoid generating UACCESS warnings when > optimizing for size (CC_OPTIMIZE_FOR_SIZE). It will also avoid future > surprises with compiler versions that change the inlining heuristic even > when optimizing for performance. > > Report: http://lkml.kernel.org/r/58708908-84a0-0a81-a836-ad97e33dbb62@infradead.org > Reported-by: Randy Dunlap > Acked-by: Randy Dunlap # build-tested > Signed-off-by: Marco Elver Queued for review and testing, thank you! Thanx, Paul > --- > Version rebased on -rcu/kcsan. > > There are 3 locations that would conflict with the style cleanup in > -tip/locking/kcsan: > https://git.kernel.org/pub/scm/linux/kernel/git/tip/tip.git/commit/?h=locking/kcsan&id=5cbaefe9743bf14c9d3106db0cc19f8cb0a3ca22 > > For the conflicting locations the better style is carried over, so that > upon eventual merge the resolution should be trivial. > --- > kernel/kcsan/atomic.h | 2 +- > kernel/kcsan/core.c | 17 ++++++++--------- > kernel/kcsan/encoding.h | 11 +++++------ > 3 files changed, 14 insertions(+), 16 deletions(-) > > diff --git a/kernel/kcsan/atomic.h b/kernel/kcsan/atomic.h > index c9c3fe628011..466e6777533e 100644 > --- a/kernel/kcsan/atomic.h > +++ b/kernel/kcsan/atomic.h > @@ -18,7 +18,7 @@ > * than cast to volatile. Eventually, we hope to be able to remove this > * function. > */ > -static inline bool kcsan_is_atomic(const volatile void *ptr) > +static __always_inline bool kcsan_is_atomic(const volatile void *ptr) > { > /* only jiffies for now */ > return ptr == &jiffies; > diff --git a/kernel/kcsan/core.c b/kernel/kcsan/core.c > index d9410d58c93e..69870645b631 100644 > --- a/kernel/kcsan/core.c > +++ b/kernel/kcsan/core.c > @@ -78,9 +78,8 @@ static atomic_long_t watchpoints[CONFIG_KCSAN_NUM_WATCHPOINTS + NUM_SLOTS - 1]; > */ > static DEFINE_PER_CPU(long, kcsan_skip); > > -static inline atomic_long_t *find_watchpoint(unsigned long addr, size_t size, > - bool expect_write, > - long *encoded_watchpoint) > +static __always_inline atomic_long_t * > +find_watchpoint(unsigned long addr, size_t size, bool expect_write, long *encoded_watchpoint) > { > const int slot = watchpoint_slot(addr); > const unsigned long addr_masked = addr & WATCHPOINT_ADDR_MASK; > @@ -150,8 +149,8 @@ static inline atomic_long_t *insert_watchpoint(unsigned long addr, size_t size, > * 2. the thread that set up the watchpoint already removed it; > * 3. the watchpoint was removed and then re-used. > */ > -static inline bool try_consume_watchpoint(atomic_long_t *watchpoint, > - long encoded_watchpoint) > +static __always_inline bool > +try_consume_watchpoint(atomic_long_t *watchpoint, long encoded_watchpoint) > { > return atomic_long_try_cmpxchg_relaxed(watchpoint, &encoded_watchpoint, > CONSUMED_WATCHPOINT); > @@ -166,7 +165,7 @@ static inline bool remove_watchpoint(atomic_long_t *watchpoint) > CONSUMED_WATCHPOINT; > } > > -static inline struct kcsan_ctx *get_ctx(void) > +static __always_inline struct kcsan_ctx *get_ctx(void) > { > /* > * In interrupt, use raw_cpu_ptr to avoid unnecessary checks, that would > @@ -175,7 +174,7 @@ static inline struct kcsan_ctx *get_ctx(void) > return in_task() ? ¤t->kcsan_ctx : raw_cpu_ptr(&kcsan_cpu_ctx); > } > > -static inline bool is_atomic(const volatile void *ptr) > +static __always_inline bool is_atomic(const volatile void *ptr) > { > struct kcsan_ctx *ctx = get_ctx(); > > @@ -199,7 +198,7 @@ static inline bool is_atomic(const volatile void *ptr) > return kcsan_is_atomic(ptr); > } > > -static inline bool should_watch(const volatile void *ptr, int type) > +static __always_inline bool should_watch(const volatile void *ptr, int type) > { > /* > * Never set up watchpoints when memory operations are atomic. > @@ -232,7 +231,7 @@ static inline void reset_kcsan_skip(void) > this_cpu_write(kcsan_skip, skip_count); > } > > -static inline bool kcsan_is_enabled(void) > +static __always_inline bool kcsan_is_enabled(void) > { > return READ_ONCE(kcsan_enabled) && get_ctx()->disable_count == 0; > } > diff --git a/kernel/kcsan/encoding.h b/kernel/kcsan/encoding.h > index e17bdac0e54b..e527e83ce825 100644 > --- a/kernel/kcsan/encoding.h > +++ b/kernel/kcsan/encoding.h > @@ -58,9 +58,8 @@ static inline long encode_watchpoint(unsigned long addr, size_t size, > (addr & WATCHPOINT_ADDR_MASK)); > } > > -static inline bool decode_watchpoint(long watchpoint, > - unsigned long *addr_masked, size_t *size, > - bool *is_write) > +static __always_inline bool > +decode_watchpoint(long watchpoint, unsigned long *addr_masked, size_t *size, bool *is_write) > { > if (watchpoint == INVALID_WATCHPOINT || > watchpoint == CONSUMED_WATCHPOINT) > @@ -77,13 +76,13 @@ static inline bool decode_watchpoint(long watchpoint, > /* > * Return watchpoint slot for an address. > */ > -static inline int watchpoint_slot(unsigned long addr) > +static __always_inline int watchpoint_slot(unsigned long addr) > { > return (addr / PAGE_SIZE) % CONFIG_KCSAN_NUM_WATCHPOINTS; > } > > -static inline bool matching_access(unsigned long addr1, size_t size1, > - unsigned long addr2, size_t size2) > +static __always_inline bool matching_access(unsigned long addr1, size_t size1, > + unsigned long addr2, size_t size2) > { > unsigned long end_range1 = addr1 + size1 - 1; > unsigned long end_range2 = addr2 + size2 - 1; > -- > 2.24.1.735.g03f4e72817-goog >