2022-04-16 02:38:32

by Thomas Gleixner

[permalink] [raw]
Subject: [patch 06/10] x86/aperfmperf: Restructure arch_scale_freq_tick()

Preparation for sharing code with the CPU frequency portion of the
aperf/mperf code.

No functional change.

Signed-off-by: Thomas Gleixner <[email protected]>
---
arch/x86/kernel/cpu/aperfmperf.c | 36 +++++++++++++++++++++---------------
1 file changed, 21 insertions(+), 15 deletions(-)

--- a/arch/x86/kernel/cpu/aperfmperf.c
+++ b/arch/x86/kernel/cpu/aperfmperf.c
@@ -477,22 +477,9 @@ static DECLARE_WORK(disable_freq_invaria

DEFINE_PER_CPU(unsigned long, arch_freq_scale) = SCHED_CAPACITY_SCALE;

-void arch_scale_freq_tick(void)
+static void scale_freq_tick(u64 acnt, u64 mcnt)
{
- struct aperfmperf *s = this_cpu_ptr(&cpu_samples);
- u64 aperf, mperf, acnt, mcnt, freq_scale;
-
- if (!arch_scale_freq_invariant())
- return;
-
- rdmsrl(MSR_IA32_APERF, aperf);
- rdmsrl(MSR_IA32_MPERF, mperf);
-
- acnt = aperf - s->aperf;
- mcnt = mperf - s->mperf;
-
- s->aperf = aperf;
- s->mperf = mperf;
+ u64 freq_scale;

if (check_shl_overflow(acnt, 2*SCHED_CAPACITY_SHIFT, &acnt))
goto error;
@@ -514,4 +501,23 @@ void arch_scale_freq_tick(void)
pr_warn("Scheduler frequency invariance went wobbly, disabling!\n");
schedule_work(&disable_freq_invariance_work);
}
+
+void arch_scale_freq_tick(void)
+{
+ struct aperfmperf *s = this_cpu_ptr(&cpu_samples);
+ u64 acnt, mcnt, aperf, mperf;
+
+ if (!arch_scale_freq_invariant())
+ return;
+
+ rdmsrl(MSR_IA32_APERF, aperf);
+ rdmsrl(MSR_IA32_MPERF, mperf);
+ acnt = aperf - s->aperf;
+ mcnt = mperf - s->mperf;
+
+ s->aperf = aperf;
+ s->mperf = mperf;
+
+ scale_freq_tick(acnt, mcnt);
+}
#endif /* CONFIG_X86_64 && CONFIG_SMP */


2022-04-22 02:31:43

by Rafael J. Wysocki

[permalink] [raw]
Subject: Re: [patch 06/10] x86/aperfmperf: Restructure arch_scale_freq_tick()

On Fri, Apr 15, 2022 at 9:20 PM Thomas Gleixner <[email protected]> wrote:
>
> Preparation for sharing code with the CPU frequency portion of the
> aperf/mperf code.
>
> No functional change.
>
> Signed-off-by: Thomas Gleixner <[email protected]>

All good AFAICS:

Reviewed-by: Rafael J. Wysocki <[email protected]>

> ---
> arch/x86/kernel/cpu/aperfmperf.c | 36 +++++++++++++++++++++---------------
> 1 file changed, 21 insertions(+), 15 deletions(-)
>
> --- a/arch/x86/kernel/cpu/aperfmperf.c
> +++ b/arch/x86/kernel/cpu/aperfmperf.c
> @@ -477,22 +477,9 @@ static DECLARE_WORK(disable_freq_invaria
>
> DEFINE_PER_CPU(unsigned long, arch_freq_scale) = SCHED_CAPACITY_SCALE;
>
> -void arch_scale_freq_tick(void)
> +static void scale_freq_tick(u64 acnt, u64 mcnt)
> {
> - struct aperfmperf *s = this_cpu_ptr(&cpu_samples);
> - u64 aperf, mperf, acnt, mcnt, freq_scale;
> -
> - if (!arch_scale_freq_invariant())
> - return;
> -
> - rdmsrl(MSR_IA32_APERF, aperf);
> - rdmsrl(MSR_IA32_MPERF, mperf);
> -
> - acnt = aperf - s->aperf;
> - mcnt = mperf - s->mperf;
> -
> - s->aperf = aperf;
> - s->mperf = mperf;
> + u64 freq_scale;
>
> if (check_shl_overflow(acnt, 2*SCHED_CAPACITY_SHIFT, &acnt))
> goto error;
> @@ -514,4 +501,23 @@ void arch_scale_freq_tick(void)
> pr_warn("Scheduler frequency invariance went wobbly, disabling!\n");
> schedule_work(&disable_freq_invariance_work);
> }
> +
> +void arch_scale_freq_tick(void)
> +{
> + struct aperfmperf *s = this_cpu_ptr(&cpu_samples);
> + u64 acnt, mcnt, aperf, mperf;
> +
> + if (!arch_scale_freq_invariant())
> + return;
> +
> + rdmsrl(MSR_IA32_APERF, aperf);
> + rdmsrl(MSR_IA32_MPERF, mperf);
> + acnt = aperf - s->aperf;
> + mcnt = mperf - s->mperf;
> +
> + s->aperf = aperf;
> + s->mperf = mperf;
> +
> + scale_freq_tick(acnt, mcnt);
> +}
> #endif /* CONFIG_X86_64 && CONFIG_SMP */
>

Subject: [tip: x86/cleanups] x86/aperfmperf: Restructure arch_scale_freq_tick()

The following commit has been merged into the x86/cleanups branch of tip:

Commit-ID: 76225ca3666c5b05afb8dcc18a73f002b097af24
Gitweb: https://git.kernel.org/tip/76225ca3666c5b05afb8dcc18a73f002b097af24
Author: Thomas Gleixner <[email protected]>
AuthorDate: Fri, 15 Apr 2022 21:19:57 +02:00
Committer: Thomas Gleixner <[email protected]>
CommitterDate: Wed, 27 Apr 2022 15:51:08 +02:00

x86/aperfmperf: Restructure arch_scale_freq_tick()

Preparation for sharing code with the CPU frequency portion of the
aperf/mperf code.

No functional change.

Signed-off-by: Thomas Gleixner <[email protected]>
Reviewed-by: Rafael J. Wysocki <[email protected]>
Acked-by: Peter Zijlstra (Intel) <[email protected]>
Acked-by: Paul E. McKenney <[email protected]>
Link: https://lore.kernel.org/r/[email protected]

---
arch/x86/kernel/cpu/aperfmperf.c | 36 ++++++++++++++++++-------------
1 file changed, 21 insertions(+), 15 deletions(-)

diff --git a/arch/x86/kernel/cpu/aperfmperf.c b/arch/x86/kernel/cpu/aperfmperf.c
index 6922c77..6220503 100644
--- a/arch/x86/kernel/cpu/aperfmperf.c
+++ b/arch/x86/kernel/cpu/aperfmperf.c
@@ -477,22 +477,9 @@ static DECLARE_WORK(disable_freq_invariance_work,

DEFINE_PER_CPU(unsigned long, arch_freq_scale) = SCHED_CAPACITY_SCALE;

-void arch_scale_freq_tick(void)
+static void scale_freq_tick(u64 acnt, u64 mcnt)
{
- struct aperfmperf *s = this_cpu_ptr(&cpu_samples);
- u64 aperf, mperf, acnt, mcnt, freq_scale;
-
- if (!arch_scale_freq_invariant())
- return;
-
- rdmsrl(MSR_IA32_APERF, aperf);
- rdmsrl(MSR_IA32_MPERF, mperf);
-
- acnt = aperf - s->aperf;
- mcnt = mperf - s->mperf;
-
- s->aperf = aperf;
- s->mperf = mperf;
+ u64 freq_scale;

if (check_shl_overflow(acnt, 2*SCHED_CAPACITY_SHIFT, &acnt))
goto error;
@@ -514,4 +501,23 @@ error:
pr_warn("Scheduler frequency invariance went wobbly, disabling!\n");
schedule_work(&disable_freq_invariance_work);
}
+
+void arch_scale_freq_tick(void)
+{
+ struct aperfmperf *s = this_cpu_ptr(&cpu_samples);
+ u64 acnt, mcnt, aperf, mperf;
+
+ if (!arch_scale_freq_invariant())
+ return;
+
+ rdmsrl(MSR_IA32_APERF, aperf);
+ rdmsrl(MSR_IA32_MPERF, mperf);
+ acnt = aperf - s->aperf;
+ mcnt = mperf - s->mperf;
+
+ s->aperf = aperf;
+ s->mperf = mperf;
+
+ scale_freq_tick(acnt, mcnt);
+}
#endif /* CONFIG_X86_64 && CONFIG_SMP */

Subject: [tip: x86/cleanups] x86/aperfmperf: Restructure arch_scale_freq_tick()

The following commit has been merged into the x86/cleanups branch of tip:

Commit-ID: 73a5fa7d51366a549a9f2e3ee875ae51aa0b5580
Gitweb: https://git.kernel.org/tip/73a5fa7d51366a549a9f2e3ee875ae51aa0b5580
Author: Thomas Gleixner <[email protected]>
AuthorDate: Fri, 15 Apr 2022 21:19:57 +02:00
Committer: Thomas Gleixner <[email protected]>
CommitterDate: Wed, 27 Apr 2022 20:22:19 +02:00

x86/aperfmperf: Restructure arch_scale_freq_tick()

Preparation for sharing code with the CPU frequency portion of the
aperf/mperf code.

No functional change.

Signed-off-by: Thomas Gleixner <[email protected]>
Reviewed-by: Rafael J. Wysocki <[email protected]>
Acked-by: Peter Zijlstra (Intel) <[email protected]>
Acked-by: Paul E. McKenney <[email protected]>
Link: https://lore.kernel.org/r/[email protected]

---
arch/x86/kernel/cpu/aperfmperf.c | 36 ++++++++++++++++++-------------
1 file changed, 21 insertions(+), 15 deletions(-)

diff --git a/arch/x86/kernel/cpu/aperfmperf.c b/arch/x86/kernel/cpu/aperfmperf.c
index 6922c77..6220503 100644
--- a/arch/x86/kernel/cpu/aperfmperf.c
+++ b/arch/x86/kernel/cpu/aperfmperf.c
@@ -477,22 +477,9 @@ static DECLARE_WORK(disable_freq_invariance_work,

DEFINE_PER_CPU(unsigned long, arch_freq_scale) = SCHED_CAPACITY_SCALE;

-void arch_scale_freq_tick(void)
+static void scale_freq_tick(u64 acnt, u64 mcnt)
{
- struct aperfmperf *s = this_cpu_ptr(&cpu_samples);
- u64 aperf, mperf, acnt, mcnt, freq_scale;
-
- if (!arch_scale_freq_invariant())
- return;
-
- rdmsrl(MSR_IA32_APERF, aperf);
- rdmsrl(MSR_IA32_MPERF, mperf);
-
- acnt = aperf - s->aperf;
- mcnt = mperf - s->mperf;
-
- s->aperf = aperf;
- s->mperf = mperf;
+ u64 freq_scale;

if (check_shl_overflow(acnt, 2*SCHED_CAPACITY_SHIFT, &acnt))
goto error;
@@ -514,4 +501,23 @@ error:
pr_warn("Scheduler frequency invariance went wobbly, disabling!\n");
schedule_work(&disable_freq_invariance_work);
}
+
+void arch_scale_freq_tick(void)
+{
+ struct aperfmperf *s = this_cpu_ptr(&cpu_samples);
+ u64 acnt, mcnt, aperf, mperf;
+
+ if (!arch_scale_freq_invariant())
+ return;
+
+ rdmsrl(MSR_IA32_APERF, aperf);
+ rdmsrl(MSR_IA32_MPERF, mperf);
+ acnt = aperf - s->aperf;
+ mcnt = mperf - s->mperf;
+
+ s->aperf = aperf;
+ s->mperf = mperf;
+
+ scale_freq_tick(acnt, mcnt);
+}
#endif /* CONFIG_X86_64 && CONFIG_SMP */