Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1752597AbdH3SEh (ORCPT ); Wed, 30 Aug 2017 14:04:37 -0400 Received: from aserp1040.oracle.com ([141.146.126.69]:27933 "EHLO aserp1040.oracle.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751775AbdH3SEe (ORCPT ); Wed, 30 Aug 2017 14:04:34 -0400 From: Pavel Tatashin To: linux@armlinux.org.uk, schwidefsky@de.ibm.com, heiko.carstens@de.ibm.com, john.stultz@linaro.org, sboyd@codeaurora.org, pasha.tatashin@oracle.com, x86@kernel.org, linux-kernel@vger.kernel.org, mingo@redhat.com, peterz@infradead.org, tglx@linutronix.de, hpa@zytor.com, douly.fnst@cn.fujitsu.com Subject: [PATCH v6 1/4] sched/clock: interface to allow timestamps early in boot Date: Wed, 30 Aug 2017 14:03:22 -0400 Message-Id: <1504116205-355281-2-git-send-email-pasha.tatashin@oracle.com> X-Mailer: git-send-email 1.7.1 In-Reply-To: <1504116205-355281-1-git-send-email-pasha.tatashin@oracle.com> References: <1504116205-355281-1-git-send-email-pasha.tatashin@oracle.com> X-Source-IP: aserv0022.oracle.com [141.146.126.234] Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 4481 Lines: 140 In Linux printk() can output timestamps next to every line. This is very useful for tracking regressions, and finding places that can be optimized. However, the timestamps are available only later in boot. On smaller machines it is insignificant amount of time, but on larger it can be many seconds or even minutes into the boot process. This patch adds an interface for platforms with unstable sched clock to show timestamps early in boot. In order to get this functionality a platform must: - Implement u64 sched_clock_early() Clock that returns monotonic time - Call sched_clock_early_init() Tells sched clock that the early clock can be used - Call sched_clock_early_fini() Tells sched clock that the early clock is finished, and sched clock should hand over the operation to permanent clock. Signed-off-by: Pavel Tatashin --- include/linux/sched/clock.h | 4 +++ kernel/sched/clock.c | 63 ++++++++++++++++++++++++++++++++++++++++++++- 2 files changed, 66 insertions(+), 1 deletion(-) diff --git a/include/linux/sched/clock.h b/include/linux/sched/clock.h index a55600ffdf4b..f8291fa28c0c 100644 --- a/include/linux/sched/clock.h +++ b/include/linux/sched/clock.h @@ -63,6 +63,10 @@ extern void sched_clock_tick_stable(void); extern void sched_clock_idle_sleep_event(void); extern void sched_clock_idle_wakeup_event(void); +void sched_clock_early_init(void); +void sched_clock_early_fini(void); +u64 sched_clock_early(void); + /* * As outlined in clock.c, provides a fast, high resolution, nanosecond * time source that is monotonic per cpu argument and has bounded drift diff --git a/kernel/sched/clock.c b/kernel/sched/clock.c index ca0f8fc945c6..2a41791b22fa 100644 --- a/kernel/sched/clock.c +++ b/kernel/sched/clock.c @@ -80,9 +80,17 @@ EXPORT_SYMBOL_GPL(sched_clock); __read_mostly int sched_clock_running; +static bool __read_mostly sched_clock_early_running; + void sched_clock_init(void) { - sched_clock_running = 1; + /* + * We start clock once early clock is finished or if early clock + * was not running. + */ + if (!sched_clock_early_running) + sched_clock_running = 1; + } #ifdef CONFIG_HAVE_UNSTABLE_SCHED_CLOCK @@ -96,6 +104,16 @@ void sched_clock_init(void) static DEFINE_STATIC_KEY_FALSE(__sched_clock_stable); static int __sched_clock_stable_early = 1; +/* + * Because static branches cannot be altered before jump_label_init() is called, + * and early time stamps may be initialized before that, we start with sched + * clock early static branch enabled, and global status disabled. Early in boot + * it is decided whether to enable the global status as well (set + * sched_clock_early_running to true), later when early clock is no longer + * needed, the static branch is disabled to keep hot-path fast. + */ +static DEFINE_STATIC_KEY_TRUE(__use_sched_clock_early); + /* * We want: ktime_get_ns() + __gtod_offset == sched_clock() + __sched_clock_offset */ @@ -362,6 +380,11 @@ u64 sched_clock_cpu(int cpu) if (sched_clock_stable()) return sched_clock() + __sched_clock_offset; + if (static_branch_unlikely(&__use_sched_clock_early)) { + if (sched_clock_early_running) + return sched_clock_early(); + } + if (unlikely(!sched_clock_running)) return 0ull; @@ -444,6 +467,44 @@ void sched_clock_idle_wakeup_event(void) } EXPORT_SYMBOL_GPL(sched_clock_idle_wakeup_event); +u64 __weak sched_clock_early(void) +{ + return 0; +} + +/* + * Is called when sched_clock_early() is about to be finished: notifies sched + * clock that after this call sched_clock_early() cannot be used. + */ +void __init sched_clock_early_fini(void) +{ + struct sched_clock_data *scd = this_scd(); + u64 now_early, now_sched; + + now_early = sched_clock_early(); + now_sched = sched_clock(); + + __gtod_offset = now_early - scd->tick_gtod; + __sched_clock_offset = now_early - now_sched; + + sched_clock_early_running = false; + static_branch_disable(&__use_sched_clock_early); + + /* Now that early clock is finished, start regular sched clock */ + sched_clock_init(); +} + +/* + * Notifies sched clock that early boot clocksource is available, it means that + * the current platform has implemented sched_clock_early(). + * + * The early clock is running until sched_clock_early_fini is called. + */ +void __init sched_clock_early_init(void) +{ + sched_clock_early_running = true; +} + #else /* CONFIG_HAVE_UNSTABLE_SCHED_CLOCK */ u64 sched_clock_cpu(int cpu) -- 2.14.1