All of lore.kernel.org
 help / color / mirror / Atom feed
From: "Rafael J. Wysocki" <rafael@kernel.org>
To: Thomas Gleixner <tglx@linutronix.de>
Cc: LKML <linux-kernel@vger.kernel.org>,
	"the arch/x86 maintainers" <x86@kernel.org>,
	"Rafael J. Wysocki" <rafael@kernel.org>,
	Linux PM <linux-pm@vger.kernel.org>,
	Eric Dumazet <edumazet@google.com>,
	"Paul E. McKenney" <paulmck@kernel.org>
Subject: Re: [patch 10/10] x86/aperfmperf: Replace arch_freq_get_on_cpu()
Date: Tue, 19 Apr 2022 18:37:27 +0200	[thread overview]
Message-ID: <CAJZ5v0iDJ7k5cyZt8GWp4kBVrk8h74+kY_=dh48m1xnf9G3ksQ@mail.gmail.com> (raw)
In-Reply-To: <20220415161206.934040006@linutronix.de>

On Fri, Apr 15, 2022 at 9:20 PM Thomas Gleixner <tglx@linutronix.de> wrote:
>
> Reading the current CPU frequency from /sys/..../scaling_cur_freq involves
> in the worst case two IPIs due to the ad hoc sampling.
>
> The frequency invariance infrastructure provides the APERF/MPERF samples
> already. Utilize them and consolidate this with the /proc/cpuinfo readout.
>
> The sample is considered valid for 20ms. So for idle or isolated NOHZ full
> CPUs the function returns 0, which is matching the previous behaviour.
>
> The resulting text size vs. the original APERF/MPERF plus the separate
> frequency invariance code:
>
>   text:         2411    ->   723
>   init.text:       0    ->   767
>
> Signed-off-by: Thomas Gleixner <tglx@linutronix.de>

All good AFAICS.

Reviewed-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>

> ---
>  arch/x86/kernel/cpu/aperfmperf.c |   94 ---------------------------------------
>  arch/x86/kernel/cpu/proc.c       |    2
>  2 files changed, 2 insertions(+), 94 deletions(-)
>
> --- a/arch/x86/kernel/cpu/aperfmperf.c
> +++ b/arch/x86/kernel/cpu/aperfmperf.c
> @@ -35,98 +35,6 @@ static DEFINE_PER_CPU_SHARED_ALIGNED(str
>         .seq = SEQCNT_ZERO(cpu_samples.seq)
>  };
>
> -struct aperfmperf_sample {
> -       unsigned int    khz;
> -       atomic_t        scfpending;
> -       ktime_t time;
> -       u64     aperf;
> -       u64     mperf;
> -};
> -
> -static DEFINE_PER_CPU(struct aperfmperf_sample, samples);
> -
> -#define APERFMPERF_CACHE_THRESHOLD_MS  10
> -#define APERFMPERF_REFRESH_DELAY_MS    10
> -#define APERFMPERF_STALE_THRESHOLD_MS  1000
> -
> -/*
> - * aperfmperf_snapshot_khz()
> - * On the current CPU, snapshot APERF, MPERF, and jiffies
> - * unless we already did it within 10ms
> - * calculate kHz, save snapshot
> - */
> -static void aperfmperf_snapshot_khz(void *dummy)
> -{
> -       u64 aperf, aperf_delta;
> -       u64 mperf, mperf_delta;
> -       struct aperfmperf_sample *s = this_cpu_ptr(&samples);
> -       unsigned long flags;
> -
> -       local_irq_save(flags);
> -       rdmsrl(MSR_IA32_APERF, aperf);
> -       rdmsrl(MSR_IA32_MPERF, mperf);
> -       local_irq_restore(flags);
> -
> -       aperf_delta = aperf - s->aperf;
> -       mperf_delta = mperf - s->mperf;
> -
> -       /*
> -        * There is no architectural guarantee that MPERF
> -        * increments faster than we can read it.
> -        */
> -       if (mperf_delta == 0)
> -               return;
> -
> -       s->time = ktime_get();
> -       s->aperf = aperf;
> -       s->mperf = mperf;
> -       s->khz = div64_u64((cpu_khz * aperf_delta), mperf_delta);
> -       atomic_set_release(&s->scfpending, 0);
> -}
> -
> -static bool aperfmperf_snapshot_cpu(int cpu, ktime_t now, bool wait)
> -{
> -       s64 time_delta = ktime_ms_delta(now, per_cpu(samples.time, cpu));
> -       struct aperfmperf_sample *s = per_cpu_ptr(&samples, cpu);
> -
> -       /* Don't bother re-computing within the cache threshold time. */
> -       if (time_delta < APERFMPERF_CACHE_THRESHOLD_MS)
> -               return true;
> -
> -       if (!atomic_xchg(&s->scfpending, 1) || wait)
> -               smp_call_function_single(cpu, aperfmperf_snapshot_khz, NULL, wait);
> -
> -       /* Return false if the previous iteration was too long ago. */
> -       return time_delta <= APERFMPERF_STALE_THRESHOLD_MS;
> -}
> -
> -unsigned int arch_freq_get_on_cpu(int cpu)
> -{
> -       struct aperfmperf_sample *s = per_cpu_ptr(&samples, cpu);
> -
> -       if (!cpu_khz)
> -               return 0;
> -
> -       if (!boot_cpu_has(X86_FEATURE_APERFMPERF))
> -               return 0;
> -
> -       if (!housekeeping_cpu(cpu, HK_TYPE_MISC))
> -               return 0;
> -
> -       if (rcu_is_idle_cpu(cpu))
> -               return 0;
> -
> -       if (aperfmperf_snapshot_cpu(cpu, ktime_get(), true))
> -               return per_cpu(samples.khz, cpu);
> -
> -       msleep(APERFMPERF_REFRESH_DELAY_MS);
> -       atomic_set(&s->scfpending, 1);
> -       smp_mb(); /* ->scfpending before smp_call_function_single(). */
> -       smp_call_function_single(cpu, aperfmperf_snapshot_khz, NULL, 1);
> -
> -       return per_cpu(samples.khz, cpu);
> -}
> -
>  static void init_counter_refs(void)
>  {
>         u64 aperf, mperf;
> @@ -493,7 +401,7 @@ void arch_scale_freq_tick(void)
>   */
>  #define MAX_SAMPLE_AGE ((unsigned long)HZ / 50)
>
> -unsigned int aperfmperf_get_khz(int cpu)
> +unsigned int arch_freq_get_on_cpu(int cpu)
>  {
>         struct aperfmperf *s = per_cpu_ptr(&cpu_samples, cpu);
>         unsigned long last;
> --- a/arch/x86/kernel/cpu/proc.c
> +++ b/arch/x86/kernel/cpu/proc.c
> @@ -84,7 +84,7 @@ static int show_cpuinfo(struct seq_file
>                 seq_printf(m, "microcode\t: 0x%x\n", c->microcode);
>
>         if (cpu_has(c, X86_FEATURE_TSC)) {
> -               unsigned int freq = aperfmperf_get_khz(cpu);
> +               unsigned int freq = arch_freq_get_on_cpu(cpu);
>
>                 if (!freq)
>                         freq = cpufreq_quick_get(cpu);
>

  reply	other threads:[~2022-04-19 16:37 UTC|newest]

Thread overview: 51+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2022-04-15 19:19 [patch 00/10] x86/cpu: Consolidate APERF/MPERF code Thomas Gleixner
2022-04-15 19:19 ` [patch 01/10] x86/aperfmperf: Dont wake idle CPUs in arch_freq_get_on_cpu() Thomas Gleixner
2022-04-19 15:34   ` Rafael J. Wysocki
2022-04-27 13:56   ` [tip: x86/cleanups] " tip-bot2 for Thomas Gleixner
2022-04-15 19:19 ` [patch 02/10] x86/smp: Move APERF/MPERF code where it belongs Thomas Gleixner
2022-04-19 15:40   ` Rafael J. Wysocki
2022-04-27 13:56   ` [tip: x86/cleanups] " tip-bot2 for Thomas Gleixner
2022-04-15 19:19 ` [patch 03/10] x86/aperfmperf: Separate AP/BP frequency invariance init Thomas Gleixner
2022-04-19 16:04   ` Rafael J. Wysocki
2022-04-27 13:56   ` [tip: x86/cleanups] " tip-bot2 for Thomas Gleixner
2022-04-15 19:19 ` [patch 04/10] x86/aperfmperf: Untangle Intel and AMD " Thomas Gleixner
2022-04-19 16:12   ` Rafael J. Wysocki
2022-04-27 13:56   ` [tip: x86/cleanups] " tip-bot2 for Thomas Gleixner
2022-04-27 18:27   ` tip-bot2 for Thomas Gleixner
2022-04-15 19:19 ` [patch 05/10] x86/aperfmperf: Put frequency invariance aperf/mperf data into a struct Thomas Gleixner
2022-04-19 16:15   ` Rafael J. Wysocki
2022-04-27 13:56   ` [tip: x86/cleanups] " tip-bot2 for Thomas Gleixner
2022-04-27 18:27   ` tip-bot2 for Thomas Gleixner
2022-04-15 19:19 ` [patch 06/10] x86/aperfmperf: Restructure arch_scale_freq_tick() Thomas Gleixner
2022-04-19 16:20   ` Rafael J. Wysocki
2022-04-27 13:56   ` [tip: x86/cleanups] " tip-bot2 for Thomas Gleixner
2022-04-27 18:27   ` tip-bot2 for Thomas Gleixner
2022-04-15 19:19 ` [patch 07/10] x86/aperfmperf: Make parts of the frequency invariance code unconditional Thomas Gleixner
2022-04-19 16:27   ` Rafael J. Wysocki
2022-04-27 13:56   ` [tip: x86/cleanups] " tip-bot2 for Thomas Gleixner
2022-04-27 18:27   ` tip-bot2 for Thomas Gleixner
2022-04-15 19:20 ` [patch 08/10] x86/aperfmperf: Store aperf/mperf data for cpu frequency reads Thomas Gleixner
2022-04-19 16:30   ` Rafael J. Wysocki
2022-04-27 13:56   ` [tip: x86/cleanups] " tip-bot2 for Thomas Gleixner
2022-04-27 18:27   ` tip-bot2 for Thomas Gleixner
2022-04-15 19:20 ` [patch 09/10] x86/aperfmperf: Replace aperfmperf_get_khz() Thomas Gleixner
2022-04-19 16:35   ` Rafael J. Wysocki
2022-04-27 13:56   ` [tip: x86/cleanups] " tip-bot2 for Thomas Gleixner
2022-04-27 18:27   ` tip-bot2 for Thomas Gleixner
2022-04-15 19:20 ` [patch 10/10] x86/aperfmperf: Replace arch_freq_get_on_cpu() Thomas Gleixner
2022-04-19 16:37   ` Rafael J. Wysocki [this message]
2022-04-27 13:56   ` [tip: x86/cleanups] " tip-bot2 for Thomas Gleixner
2022-04-27 18:27   ` tip-bot2 for Thomas Gleixner
2022-04-19 15:51 ` [patch 00/10] x86/cpu: Consolidate APERF/MPERF code Eric Dumazet
2022-04-19 20:39   ` Thomas Gleixner
2022-04-19 21:20     ` Eric Dumazet
2022-04-19 16:41 ` Peter Zijlstra
2022-04-19 17:32 ` Doug Smythies
2022-04-19 18:49   ` Rafael J. Wysocki
2022-04-19 21:11     ` Thomas Gleixner
2022-04-20 22:08       ` Doug Smythies
2022-04-25 15:45         ` Thomas Gleixner
2022-04-25 23:20           ` Doug Smythies
2022-04-27 13:56           ` [tip: x86/cleanups] x86/aperfmperf: Integrate the fallback code from show_cpuinfo() tip-bot2 for Thomas Gleixner
2022-04-27 18:27           ` tip-bot2 for Thomas Gleixner
2022-04-19 21:56 ` [patch 00/10] x86/cpu: Consolidate APERF/MPERF code Paul E. McKenney

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to='CAJZ5v0iDJ7k5cyZt8GWp4kBVrk8h74+kY_=dh48m1xnf9G3ksQ@mail.gmail.com' \
    --to=rafael@kernel.org \
    --cc=edumazet@google.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-pm@vger.kernel.org \
    --cc=paulmck@kernel.org \
    --cc=tglx@linutronix.de \
    --cc=x86@kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.