linux-kernel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* [PATCH V2] perf/x86/intel: Apply mid ACK for small core
@ 2021-08-03 13:25 kan.liang
  2021-08-03 14:55 ` Peter Zijlstra
  2021-08-06 12:51 ` [tip: perf/urgent] " tip-bot2 for Kan Liang
  0 siblings, 2 replies; 7+ messages in thread
From: kan.liang @ 2021-08-03 13:25 UTC (permalink / raw)
  To: peterz, mingo, linux-kernel; +Cc: ak, Kan Liang, stable

From: Kan Liang <kan.liang@linux.intel.com>

A warning as below may be occasionally triggered in an ADL machine when
these conditions occur,
- Two perf record commands run one by one. Both record a PEBS event.
- Both runs on small cores.
- They have different adaptive PEBS configuration (PEBS_DATA_CFG).

[  673.663291] WARNING: CPU: 4 PID: 9874 at
arch/x86/events/intel/ds.c:1743
setup_pebs_adaptive_sample_data+0x55e/0x5b0
[  673.663348] RIP: 0010:setup_pebs_adaptive_sample_data+0x55e/0x5b0
[  673.663357] Call Trace:
[  673.663357]  <NMI>
[  673.663357]  intel_pmu_drain_pebs_icl+0x48b/0x810
[  673.663360]  perf_event_nmi_handler+0x41/0x80
[  673.663368]  </NMI>
[  673.663370]  __perf_event_task_sched_in+0x2c2/0x3a0

Different from the big core, the small core requires the ACK right
before re-enabling counters in the NMI handler, otherwise a stale PEBS
record may be dumped into the later NMI handler, which trigger the
warning.

Add a new mid_ack flag to track the case. Add all PMI handler bits in
the struct x86_hybrid_pmu to track the bits for different types of PMUs.
Apply mid ACK for the small cores on an Alder Lake machine.

The existing hybrid() macro has a compile error when taking address of a
bit-field variable. Add a new macro hybrid_bit() to get the bit-field
value of a given PMU.

Fixes: f83d2f91d259 ("perf/x86/intel: Add Alder Lake Hybrid support")
Reported-by: Ammy Yi <ammy.yi@intel.com>
Tested-by: Ammy Yi <ammy.yi@intel.com>
Reviewed-by: Andi Kleen <ak@linux.intel.com>
Signed-off-by: Kan Liang <kan.liang@linux.intel.com>
Cc: stable@vger.kernel.org
---

The V1 patch set can be found at
https://lore.kernel.org/lkml/1625774073-153697-1-git-send-email-kan.liang@linux.intel.com/

Changes since v1:
- Introduce mid ACK. The early ACK in V1 may trigger other issue based
  on the latest test result.
- Add comments regarding early, mid and late ACK.

 arch/x86/events/intel/core.c | 23 +++++++++++++++--------
 arch/x86/events/perf_event.h | 15 +++++++++++++++
 2 files changed, 30 insertions(+), 8 deletions(-)

diff --git a/arch/x86/events/intel/core.c b/arch/x86/events/intel/core.c
index d76be3b..511d1f9 100644
--- a/arch/x86/events/intel/core.c
+++ b/arch/x86/events/intel/core.c
@@ -2904,24 +2904,28 @@ static int handle_pmi_common(struct pt_regs *regs, u64 status)
  */
 static int intel_pmu_handle_irq(struct pt_regs *regs)
 {
-	struct cpu_hw_events *cpuc;
+	struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events);
+	bool late_ack = hybrid_bit(cpuc->pmu, late_ack);
+	bool mid_ack = hybrid_bit(cpuc->pmu, mid_ack);
 	int loops;
 	u64 status;
 	int handled;
 	int pmu_enabled;
 
-	cpuc = this_cpu_ptr(&cpu_hw_events);
-
 	/*
 	 * Save the PMU state.
 	 * It needs to be restored when leaving the handler.
 	 */
 	pmu_enabled = cpuc->enabled;
 	/*
-	 * No known reason to not always do late ACK,
-	 * but just in case do it opt-in.
+	 * In general, the early ACK is only applied for old platforms.
+	 * For the big core starts from Haswell, the late ACK should be
+	 * applied.
+	 * For the small core after Tremont, we have to do the ACK right
+	 * before re-enabling counters, which is in the middle of the
+	 * NMI handler.
 	 */
-	if (!x86_pmu.late_ack)
+	if (!late_ack && !mid_ack)
 		apic_write(APIC_LVTPC, APIC_DM_NMI);
 	intel_bts_disable_local();
 	cpuc->enabled = 0;
@@ -2958,6 +2962,8 @@ static int intel_pmu_handle_irq(struct pt_regs *regs)
 		goto again;
 
 done:
+	if (mid_ack)
+		apic_write(APIC_LVTPC, APIC_DM_NMI);
 	/* Only restore PMU state when it's active. See x86_pmu_disable(). */
 	cpuc->enabled = pmu_enabled;
 	if (pmu_enabled)
@@ -2969,7 +2975,7 @@ static int intel_pmu_handle_irq(struct pt_regs *regs)
 	 * have been reset. This avoids spurious NMIs on
 	 * Haswell CPUs.
 	 */
-	if (x86_pmu.late_ack)
+	if (late_ack)
 		apic_write(APIC_LVTPC, APIC_DM_NMI);
 	return handled;
 }
@@ -6123,7 +6129,6 @@ __init int intel_pmu_init(void)
 		static_branch_enable(&perf_is_hybrid);
 		x86_pmu.num_hybrid_pmus = X86_HYBRID_NUM_PMUS;
 
-		x86_pmu.late_ack = true;
 		x86_pmu.pebs_aliases = NULL;
 		x86_pmu.pebs_prec_dist = true;
 		x86_pmu.pebs_block = true;
@@ -6161,6 +6166,7 @@ __init int intel_pmu_init(void)
 		pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_CORE_IDX];
 		pmu->name = "cpu_core";
 		pmu->cpu_type = hybrid_big;
+		pmu->late_ack = true;
 		if (cpu_feature_enabled(X86_FEATURE_HYBRID_CPU)) {
 			pmu->num_counters = x86_pmu.num_counters + 2;
 			pmu->num_counters_fixed = x86_pmu.num_counters_fixed + 1;
@@ -6186,6 +6192,7 @@ __init int intel_pmu_init(void)
 		pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_ATOM_IDX];
 		pmu->name = "cpu_atom";
 		pmu->cpu_type = hybrid_small;
+		pmu->mid_ack = true;
 		pmu->num_counters = x86_pmu.num_counters;
 		pmu->num_counters_fixed = x86_pmu.num_counters_fixed;
 		pmu->max_pebs_events = x86_pmu.max_pebs_events;
diff --git a/arch/x86/events/perf_event.h b/arch/x86/events/perf_event.h
index ad87cb3..eec7ce8 100644
--- a/arch/x86/events/perf_event.h
+++ b/arch/x86/events/perf_event.h
@@ -655,6 +655,10 @@ struct x86_hybrid_pmu {
 	struct event_constraint		*event_constraints;
 	struct event_constraint		*pebs_constraints;
 	struct extra_reg		*extra_regs;
+
+	unsigned int			late_ack	:1,
+					mid_ack		:1,
+					enabled_ack	:1;
 };
 
 static __always_inline struct x86_hybrid_pmu *hybrid_pmu(struct pmu *pmu)
@@ -685,6 +689,16 @@ extern struct static_key_false perf_is_hybrid;
 	__Fp;						\
 }))
 
+#define hybrid_bit(_pmu, _field)			\
+({							\
+	bool __Fp = x86_pmu._field;			\
+							\
+	if (is_hybrid() && (_pmu))			\
+		__Fp = hybrid_pmu(_pmu)->_field;	\
+							\
+	__Fp;						\
+})
+
 enum hybrid_pmu_type {
 	hybrid_big		= 0x40,
 	hybrid_small		= 0x20,
@@ -754,6 +768,7 @@ struct x86_pmu {
 
 	/* PMI handler bits */
 	unsigned int	late_ack		:1,
+			mid_ack			:1,
 			enabled_ack		:1;
 	/*
 	 * sysfs attrs
-- 
2.7.4


^ permalink raw reply related	[flat|nested] 7+ messages in thread

* Re: [PATCH V2] perf/x86/intel: Apply mid ACK for small core
  2021-08-03 13:25 [PATCH V2] perf/x86/intel: Apply mid ACK for small core kan.liang
@ 2021-08-03 14:55 ` Peter Zijlstra
  2021-08-03 15:20   ` Liang, Kan
  2021-08-06 12:51 ` [tip: perf/urgent] " tip-bot2 for Kan Liang
  1 sibling, 1 reply; 7+ messages in thread
From: Peter Zijlstra @ 2021-08-03 14:55 UTC (permalink / raw)
  To: kan.liang; +Cc: mingo, linux-kernel, ak, stable

On Tue, Aug 03, 2021 at 06:25:28AM -0700, kan.liang@linux.intel.com wrote:
> From: Kan Liang <kan.liang@linux.intel.com>
> 
> A warning as below may be occasionally triggered in an ADL machine when
> these conditions occur,
> - Two perf record commands run one by one. Both record a PEBS event.
> - Both runs on small cores.
> - They have different adaptive PEBS configuration (PEBS_DATA_CFG).
> 
> [  673.663291] WARNING: CPU: 4 PID: 9874 at
> arch/x86/events/intel/ds.c:1743
> setup_pebs_adaptive_sample_data+0x55e/0x5b0
> [  673.663348] RIP: 0010:setup_pebs_adaptive_sample_data+0x55e/0x5b0
> [  673.663357] Call Trace:
> [  673.663357]  <NMI>
> [  673.663357]  intel_pmu_drain_pebs_icl+0x48b/0x810
> [  673.663360]  perf_event_nmi_handler+0x41/0x80
> [  673.663368]  </NMI>
> [  673.663370]  __perf_event_task_sched_in+0x2c2/0x3a0
> 
> Different from the big core, the small core requires the ACK right
> before re-enabling counters in the NMI handler, otherwise a stale PEBS
> record may be dumped into the later NMI handler, which trigger the
> warning.
> 
> Add a new mid_ack flag to track the case. Add all PMI handler bits in
> the struct x86_hybrid_pmu to track the bits for different types of PMUs.
> Apply mid ACK for the small cores on an Alder Lake machine.

Why do we need a new option? Why isn't early (as in not late) good
enough?

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [PATCH V2] perf/x86/intel: Apply mid ACK for small core
  2021-08-03 14:55 ` Peter Zijlstra
@ 2021-08-03 15:20   ` Liang, Kan
  2021-08-03 16:17     ` Peter Zijlstra
  0 siblings, 1 reply; 7+ messages in thread
From: Liang, Kan @ 2021-08-03 15:20 UTC (permalink / raw)
  To: Peter Zijlstra; +Cc: mingo, linux-kernel, ak, stable



On 8/3/2021 10:55 AM, Peter Zijlstra wrote:
> On Tue, Aug 03, 2021 at 06:25:28AM -0700, kan.liang@linux.intel.com wrote:
>> From: Kan Liang <kan.liang@linux.intel.com>
>>
>> A warning as below may be occasionally triggered in an ADL machine when
>> these conditions occur,
>> - Two perf record commands run one by one. Both record a PEBS event.
>> - Both runs on small cores.
>> - They have different adaptive PEBS configuration (PEBS_DATA_CFG).
>>
>> [  673.663291] WARNING: CPU: 4 PID: 9874 at
>> arch/x86/events/intel/ds.c:1743
>> setup_pebs_adaptive_sample_data+0x55e/0x5b0
>> [  673.663348] RIP: 0010:setup_pebs_adaptive_sample_data+0x55e/0x5b0
>> [  673.663357] Call Trace:
>> [  673.663357]  <NMI>
>> [  673.663357]  intel_pmu_drain_pebs_icl+0x48b/0x810
>> [  673.663360]  perf_event_nmi_handler+0x41/0x80
>> [  673.663368]  </NMI>
>> [  673.663370]  __perf_event_task_sched_in+0x2c2/0x3a0
>>
>> Different from the big core, the small core requires the ACK right
>> before re-enabling counters in the NMI handler, otherwise a stale PEBS
>> record may be dumped into the later NMI handler, which trigger the
>> warning.
>>
>> Add a new mid_ack flag to track the case. Add all PMI handler bits in
>> the struct x86_hybrid_pmu to track the bits for different types of PMUs.
>> Apply mid ACK for the small cores on an Alder Lake machine.
> 
> Why do we need a new option? Why isn't early (as in not late) good
> enough?
> 

The early ACK can fix this issue, however it triggers a spurious NMI 
during the stress test. I'm told to do the ACK right before re-enabling 
counters for small cores. That indeed fixes all the issues.

Thanks,
Kan

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [PATCH V2] perf/x86/intel: Apply mid ACK for small core
  2021-08-03 15:20   ` Liang, Kan
@ 2021-08-03 16:17     ` Peter Zijlstra
  2021-08-03 17:00       ` Liang, Kan
  0 siblings, 1 reply; 7+ messages in thread
From: Peter Zijlstra @ 2021-08-03 16:17 UTC (permalink / raw)
  To: Liang, Kan; +Cc: mingo, linux-kernel, ak, stable

On Tue, Aug 03, 2021 at 11:20:20AM -0400, Liang, Kan wrote:
> 
> 
> On 8/3/2021 10:55 AM, Peter Zijlstra wrote:
> > On Tue, Aug 03, 2021 at 06:25:28AM -0700, kan.liang@linux.intel.com wrote:
> > > From: Kan Liang <kan.liang@linux.intel.com>
> > > 
> > > A warning as below may be occasionally triggered in an ADL machine when
> > > these conditions occur,
> > > - Two perf record commands run one by one. Both record a PEBS event.
> > > - Both runs on small cores.
> > > - They have different adaptive PEBS configuration (PEBS_DATA_CFG).
> > > 
> > > [  673.663291] WARNING: CPU: 4 PID: 9874 at
> > > arch/x86/events/intel/ds.c:1743
> > > setup_pebs_adaptive_sample_data+0x55e/0x5b0
> > > [  673.663348] RIP: 0010:setup_pebs_adaptive_sample_data+0x55e/0x5b0
> > > [  673.663357] Call Trace:
> > > [  673.663357]  <NMI>
> > > [  673.663357]  intel_pmu_drain_pebs_icl+0x48b/0x810
> > > [  673.663360]  perf_event_nmi_handler+0x41/0x80
> > > [  673.663368]  </NMI>
> > > [  673.663370]  __perf_event_task_sched_in+0x2c2/0x3a0
> > > 
> > > Different from the big core, the small core requires the ACK right
> > > before re-enabling counters in the NMI handler, otherwise a stale PEBS
> > > record may be dumped into the later NMI handler, which trigger the
> > > warning.
> > > 
> > > Add a new mid_ack flag to track the case. Add all PMI handler bits in
> > > the struct x86_hybrid_pmu to track the bits for different types of PMUs.
> > > Apply mid ACK for the small cores on an Alder Lake machine.
> > 
> > Why do we need a new option? Why isn't early (as in not late) good
> > enough?
> > 
> 
> The early ACK can fix this issue, however it triggers a spurious NMI during
> the stress test. I'm told to do the ACK right before re-enabling counters
> for small cores. That indeed fixes all the issues.

Any chance that would also work for the chips that now use late_ack?

I'm just (desperately) trying to minimize the amount of quirks here ;-)

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [PATCH V2] perf/x86/intel: Apply mid ACK for small core
  2021-08-03 16:17     ` Peter Zijlstra
@ 2021-08-03 17:00       ` Liang, Kan
  2021-08-03 18:00         ` Andi Kleen
  0 siblings, 1 reply; 7+ messages in thread
From: Liang, Kan @ 2021-08-03 17:00 UTC (permalink / raw)
  To: Peter Zijlstra; +Cc: mingo, linux-kernel, ak, stable



On 8/3/2021 12:17 PM, Peter Zijlstra wrote:
> On Tue, Aug 03, 2021 at 11:20:20AM -0400, Liang, Kan wrote:
>>
>>
>> On 8/3/2021 10:55 AM, Peter Zijlstra wrote:
>>> On Tue, Aug 03, 2021 at 06:25:28AM -0700, kan.liang@linux.intel.com wrote:
>>>> From: Kan Liang <kan.liang@linux.intel.com>
>>>>
>>>> A warning as below may be occasionally triggered in an ADL machine when
>>>> these conditions occur,
>>>> - Two perf record commands run one by one. Both record a PEBS event.
>>>> - Both runs on small cores.
>>>> - They have different adaptive PEBS configuration (PEBS_DATA_CFG).
>>>>
>>>> [  673.663291] WARNING: CPU: 4 PID: 9874 at
>>>> arch/x86/events/intel/ds.c:1743
>>>> setup_pebs_adaptive_sample_data+0x55e/0x5b0
>>>> [  673.663348] RIP: 0010:setup_pebs_adaptive_sample_data+0x55e/0x5b0
>>>> [  673.663357] Call Trace:
>>>> [  673.663357]  <NMI>
>>>> [  673.663357]  intel_pmu_drain_pebs_icl+0x48b/0x810
>>>> [  673.663360]  perf_event_nmi_handler+0x41/0x80
>>>> [  673.663368]  </NMI>
>>>> [  673.663370]  __perf_event_task_sched_in+0x2c2/0x3a0
>>>>
>>>> Different from the big core, the small core requires the ACK right
>>>> before re-enabling counters in the NMI handler, otherwise a stale PEBS
>>>> record may be dumped into the later NMI handler, which trigger the
>>>> warning.
>>>>
>>>> Add a new mid_ack flag to track the case. Add all PMI handler bits in
>>>> the struct x86_hybrid_pmu to track the bits for different types of PMUs.
>>>> Apply mid ACK for the small cores on an Alder Lake machine.
>>>
>>> Why do we need a new option? Why isn't early (as in not late) good
>>> enough?
>>>
>>
>> The early ACK can fix this issue, however it triggers a spurious NMI during
>> the stress test. I'm told to do the ACK right before re-enabling counters
>> for small cores. That indeed fixes all the issues.
> 
> Any chance that would also work for the chips that now use late_ack?
>

Let me check and do some tests.

Thanks,
Kan

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [PATCH V2] perf/x86/intel: Apply mid ACK for small core
  2021-08-03 17:00       ` Liang, Kan
@ 2021-08-03 18:00         ` Andi Kleen
  0 siblings, 0 replies; 7+ messages in thread
From: Andi Kleen @ 2021-08-03 18:00 UTC (permalink / raw)
  To: Liang, Kan, Peter Zijlstra; +Cc: mingo, linux-kernel, stable


>> Any chance that would also work for the chips that now use late_ack?
>>
>
> Let me check and do some tests.


I suspect we won't be able to test all, so we'll probably have to keep 
all three options.


-Andi



^ permalink raw reply	[flat|nested] 7+ messages in thread

* [tip: perf/urgent] perf/x86/intel: Apply mid ACK for small core
  2021-08-03 13:25 [PATCH V2] perf/x86/intel: Apply mid ACK for small core kan.liang
  2021-08-03 14:55 ` Peter Zijlstra
@ 2021-08-06 12:51 ` tip-bot2 for Kan Liang
  1 sibling, 0 replies; 7+ messages in thread
From: tip-bot2 for Kan Liang @ 2021-08-06 12:51 UTC (permalink / raw)
  To: linux-tip-commits
  Cc: Ammy Yi, Kan Liang, Peter Zijlstra (Intel),
	Andi Kleen, x86, linux-kernel

The following commit has been merged into the perf/urgent branch of tip:

Commit-ID:     acade6379930dfa7987f4bd9b26d1a701cc1b542
Gitweb:        https://git.kernel.org/tip/acade6379930dfa7987f4bd9b26d1a701cc1b542
Author:        Kan Liang <kan.liang@linux.intel.com>
AuthorDate:    Tue, 03 Aug 2021 06:25:28 -07:00
Committer:     Peter Zijlstra <peterz@infradead.org>
CommitterDate: Fri, 06 Aug 2021 14:25:15 +02:00

perf/x86/intel: Apply mid ACK for small core

A warning as below may be occasionally triggered in an ADL machine when
these conditions occur:

 - Two perf record commands run one by one. Both record a PEBS event.
 - Both runs on small cores.
 - They have different adaptive PEBS configuration (PEBS_DATA_CFG).

  [ ] WARNING: CPU: 4 PID: 9874 at arch/x86/events/intel/ds.c:1743 setup_pebs_adaptive_sample_data+0x55e/0x5b0
  [ ] RIP: 0010:setup_pebs_adaptive_sample_data+0x55e/0x5b0
  [ ] Call Trace:
  [ ]  <NMI>
  [ ]  intel_pmu_drain_pebs_icl+0x48b/0x810
  [ ]  perf_event_nmi_handler+0x41/0x80
  [ ]  </NMI>
  [ ]  __perf_event_task_sched_in+0x2c2/0x3a0

Different from the big core, the small core requires the ACK right
before re-enabling counters in the NMI handler, otherwise a stale PEBS
record may be dumped into the later NMI handler, which trigger the
warning.

Add a new mid_ack flag to track the case. Add all PMI handler bits in
the struct x86_hybrid_pmu to track the bits for different types of
PMUs.  Apply mid ACK for the small cores on an Alder Lake machine.

The existing hybrid() macro has a compile error when taking address of
a bit-field variable. Add a new macro hybrid_bit() to get the
bit-field value of a given PMU.

Fixes: f83d2f91d259 ("perf/x86/intel: Add Alder Lake Hybrid support")
Reported-by: Ammy Yi <ammy.yi@intel.com>
Signed-off-by: Kan Liang <kan.liang@linux.intel.com>
Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Reviewed-by: Andi Kleen <ak@linux.intel.com>
Tested-by: Ammy Yi <ammy.yi@intel.com>
Link: https://lkml.kernel.org/r/1627997128-57891-1-git-send-email-kan.liang@linux.intel.com
---
 arch/x86/events/intel/core.c | 23 +++++++++++++++--------
 arch/x86/events/perf_event.h | 15 +++++++++++++++
 2 files changed, 30 insertions(+), 8 deletions(-)

diff --git a/arch/x86/events/intel/core.c b/arch/x86/events/intel/core.c
index fca7a6e..ac6fd2d 100644
--- a/arch/x86/events/intel/core.c
+++ b/arch/x86/events/intel/core.c
@@ -2904,24 +2904,28 @@ static int handle_pmi_common(struct pt_regs *regs, u64 status)
  */
 static int intel_pmu_handle_irq(struct pt_regs *regs)
 {
-	struct cpu_hw_events *cpuc;
+	struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events);
+	bool late_ack = hybrid_bit(cpuc->pmu, late_ack);
+	bool mid_ack = hybrid_bit(cpuc->pmu, mid_ack);
 	int loops;
 	u64 status;
 	int handled;
 	int pmu_enabled;
 
-	cpuc = this_cpu_ptr(&cpu_hw_events);
-
 	/*
 	 * Save the PMU state.
 	 * It needs to be restored when leaving the handler.
 	 */
 	pmu_enabled = cpuc->enabled;
 	/*
-	 * No known reason to not always do late ACK,
-	 * but just in case do it opt-in.
+	 * In general, the early ACK is only applied for old platforms.
+	 * For the big core starts from Haswell, the late ACK should be
+	 * applied.
+	 * For the small core after Tremont, we have to do the ACK right
+	 * before re-enabling counters, which is in the middle of the
+	 * NMI handler.
 	 */
-	if (!x86_pmu.late_ack)
+	if (!late_ack && !mid_ack)
 		apic_write(APIC_LVTPC, APIC_DM_NMI);
 	intel_bts_disable_local();
 	cpuc->enabled = 0;
@@ -2958,6 +2962,8 @@ again:
 		goto again;
 
 done:
+	if (mid_ack)
+		apic_write(APIC_LVTPC, APIC_DM_NMI);
 	/* Only restore PMU state when it's active. See x86_pmu_disable(). */
 	cpuc->enabled = pmu_enabled;
 	if (pmu_enabled)
@@ -2969,7 +2975,7 @@ done:
 	 * have been reset. This avoids spurious NMIs on
 	 * Haswell CPUs.
 	 */
-	if (x86_pmu.late_ack)
+	if (late_ack)
 		apic_write(APIC_LVTPC, APIC_DM_NMI);
 	return handled;
 }
@@ -6129,7 +6135,6 @@ __init int intel_pmu_init(void)
 		static_branch_enable(&perf_is_hybrid);
 		x86_pmu.num_hybrid_pmus = X86_HYBRID_NUM_PMUS;
 
-		x86_pmu.late_ack = true;
 		x86_pmu.pebs_aliases = NULL;
 		x86_pmu.pebs_prec_dist = true;
 		x86_pmu.pebs_block = true;
@@ -6167,6 +6172,7 @@ __init int intel_pmu_init(void)
 		pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_CORE_IDX];
 		pmu->name = "cpu_core";
 		pmu->cpu_type = hybrid_big;
+		pmu->late_ack = true;
 		if (cpu_feature_enabled(X86_FEATURE_HYBRID_CPU)) {
 			pmu->num_counters = x86_pmu.num_counters + 2;
 			pmu->num_counters_fixed = x86_pmu.num_counters_fixed + 1;
@@ -6192,6 +6198,7 @@ __init int intel_pmu_init(void)
 		pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_ATOM_IDX];
 		pmu->name = "cpu_atom";
 		pmu->cpu_type = hybrid_small;
+		pmu->mid_ack = true;
 		pmu->num_counters = x86_pmu.num_counters;
 		pmu->num_counters_fixed = x86_pmu.num_counters_fixed;
 		pmu->max_pebs_events = x86_pmu.max_pebs_events;
diff --git a/arch/x86/events/perf_event.h b/arch/x86/events/perf_event.h
index 2938c90..e3ac05c 100644
--- a/arch/x86/events/perf_event.h
+++ b/arch/x86/events/perf_event.h
@@ -656,6 +656,10 @@ struct x86_hybrid_pmu {
 	struct event_constraint		*event_constraints;
 	struct event_constraint		*pebs_constraints;
 	struct extra_reg		*extra_regs;
+
+	unsigned int			late_ack	:1,
+					mid_ack		:1,
+					enabled_ack	:1;
 };
 
 static __always_inline struct x86_hybrid_pmu *hybrid_pmu(struct pmu *pmu)
@@ -686,6 +690,16 @@ extern struct static_key_false perf_is_hybrid;
 	__Fp;						\
 }))
 
+#define hybrid_bit(_pmu, _field)			\
+({							\
+	bool __Fp = x86_pmu._field;			\
+							\
+	if (is_hybrid() && (_pmu))			\
+		__Fp = hybrid_pmu(_pmu)->_field;	\
+							\
+	__Fp;						\
+})
+
 enum hybrid_pmu_type {
 	hybrid_big		= 0x40,
 	hybrid_small		= 0x20,
@@ -755,6 +769,7 @@ struct x86_pmu {
 
 	/* PMI handler bits */
 	unsigned int	late_ack		:1,
+			mid_ack			:1,
 			enabled_ack		:1;
 	/*
 	 * sysfs attrs

^ permalink raw reply related	[flat|nested] 7+ messages in thread

end of thread, other threads:[~2021-08-06 12:51 UTC | newest]

Thread overview: 7+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2021-08-03 13:25 [PATCH V2] perf/x86/intel: Apply mid ACK for small core kan.liang
2021-08-03 14:55 ` Peter Zijlstra
2021-08-03 15:20   ` Liang, Kan
2021-08-03 16:17     ` Peter Zijlstra
2021-08-03 17:00       ` Liang, Kan
2021-08-03 18:00         ` Andi Kleen
2021-08-06 12:51 ` [tip: perf/urgent] " tip-bot2 for Kan Liang

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).