All of lore.kernel.org
 help / color / mirror / Atom feed
From: "Michael S. Tsirkin" <mst@redhat.com>
To: qemu-devel@nongnu.org
Cc: Peter Maydell <peter.maydell@linaro.org>,
	Laszlo Ersek <lersek@redhat.com>,
	Igor Mammedov <imammedo@redhat.com>
Subject: [PULL v4 29/48] x68: acpi: trigger SMI before sending hotplug Notify event to OSPM
Date: Tue, 29 Sep 2020 03:22:10 -0400	[thread overview]
Message-ID: <20200929071948.281157-30-mst@redhat.com> (raw)
In-Reply-To: <20200929071948.281157-1-mst@redhat.com>

From: Igor Mammedov <imammedo@redhat.com>

In case firmware has negotiated CPU hotplug SMI feature, generate
AML to describe SMI IO port region and send SMI to firmware
on each CPU hotplug SCI in case new CPUs were hotplugged.

Since new CPUs can be hotplugged while CPU_SCAN_METHOD is running
we can't send SMI before new CPUs are fetched from QEMU as it
could cause sending Notify to a CPU that firmware hasn't seen yet.
So fetch new CPUs into local cache first, then send SMI and
after that send Notify events to cached CPUs. This should ensure
that Notify is sent only to CPUs which were processed by firmware
first.
Any CPUs that were hotplugged after caching will be processed
by the next CPU_SCAN_METHOD, when pending SCI is handled.

Signed-off-by: Igor Mammedov <imammedo@redhat.com>
Reviewed-by: Laszlo Ersek <lersek@redhat.com>
Tested-by: Laszlo Ersek <lersek@redhat.com>
Message-Id: <20200923094650.1301166-10-imammedo@redhat.com>
Reviewed-by: Michael S. Tsirkin <mst@redhat.com>
Signed-off-by: Michael S. Tsirkin <mst@redhat.com>
---
 hw/acpi/cpu.c | 165 +++++++++++++++++++++++++++++++++++++++++---------
 1 file changed, 138 insertions(+), 27 deletions(-)

diff --git a/hw/acpi/cpu.c b/hw/acpi/cpu.c
index 3d6a500fb7..8dd4d8ebbf 100644
--- a/hw/acpi/cpu.c
+++ b/hw/acpi/cpu.c
@@ -14,6 +14,8 @@
 #define ACPI_CPU_CMD_DATA_OFFSET_RW 8
 #define ACPI_CPU_CMD_DATA2_OFFSET_R 0
 
+#define OVMF_CPUHP_SMI_CMD 4
+
 enum {
     CPHP_GET_NEXT_CPU_WITH_EVENT_CMD = 0,
     CPHP_OST_EVENT_CMD = 1,
@@ -321,6 +323,7 @@ const VMStateDescription vmstate_cpu_hotplug = {
 #define CPU_NOTIFY_METHOD "CTFY"
 #define CPU_EJECT_METHOD  "CEJ0"
 #define CPU_OST_METHOD    "COST"
+#define CPU_ADDED_LIST    "CNEW"
 
 #define CPU_ENABLED       "CPEN"
 #define CPU_SELECTOR      "CSEL"
@@ -465,42 +468,150 @@ void build_cpus_aml(Aml *table, MachineState *machine, CPUHotplugFeatures opts,
 
         method = aml_method(CPU_SCAN_METHOD, 0, AML_SERIALIZED);
         {
+            const uint8_t max_cpus_per_pass = 255;
             Aml *else_ctx;
-            Aml *while_ctx;
+            Aml *while_ctx, *while_ctx2;
             Aml *has_event = aml_local(0);
             Aml *dev_chk = aml_int(1);
             Aml *eject_req = aml_int(3);
             Aml *next_cpu_cmd = aml_int(CPHP_GET_NEXT_CPU_WITH_EVENT_CMD);
+            Aml *num_added_cpus = aml_local(1);
+            Aml *cpu_idx = aml_local(2);
+            Aml *uid = aml_local(3);
+            Aml *has_job = aml_local(4);
+            Aml *new_cpus = aml_name(CPU_ADDED_LIST);
 
             aml_append(method, aml_acquire(ctrl_lock, 0xFFFF));
-            aml_append(method, aml_store(one, has_event));
-            while_ctx = aml_while(aml_equal(has_event, one));
+
+            /*
+             * Windows versions newer than XP (including Windows 10/Windows
+             * Server 2019), do support* VarPackageOp but, it is cripled to hold
+             * the same elements number as old PackageOp.
+             * For compatibility with Windows XP (so it won't crash) use ACPI1.0
+             * PackageOp which can hold max 255 elements.
+             *
+             * use named package as old Windows don't support it in local var
+             */
+            aml_append(method, aml_name_decl(CPU_ADDED_LIST,
+                                             aml_package(max_cpus_per_pass)));
+
+            aml_append(method, aml_store(zero, uid));
+            aml_append(method, aml_store(one, has_job));
+            /*
+             * CPU_ADDED_LIST can hold limited number of elements, outer loop
+             * allows to process CPUs in batches which let us to handle more
+             * CPUs than CPU_ADDED_LIST can hold.
+             */
+            while_ctx2 = aml_while(aml_equal(has_job, one));
             {
-                 /* clear loop exit condition, ins_evt/rm_evt checks
-                  * will set it to 1 while next_cpu_cmd returns a CPU
-                  * with events */
-                 aml_append(while_ctx, aml_store(zero, has_event));
-                 aml_append(while_ctx, aml_store(next_cpu_cmd, cpu_cmd));
-                 ifctx = aml_if(aml_equal(ins_evt, one));
-                 {
-                     aml_append(ifctx,
-                         aml_call2(CPU_NOTIFY_METHOD, cpu_data, dev_chk));
-                     aml_append(ifctx, aml_store(one, ins_evt));
-                     aml_append(ifctx, aml_store(one, has_event));
-                 }
-                 aml_append(while_ctx, ifctx);
-                 else_ctx = aml_else();
-                 ifctx = aml_if(aml_equal(rm_evt, one));
-                 {
-                     aml_append(ifctx,
-                         aml_call2(CPU_NOTIFY_METHOD, cpu_data, eject_req));
-                     aml_append(ifctx, aml_store(one, rm_evt));
-                     aml_append(ifctx, aml_store(one, has_event));
-                 }
-                 aml_append(else_ctx, ifctx);
-                 aml_append(while_ctx, else_ctx);
+                aml_append(while_ctx2, aml_store(zero, has_job));
+
+                aml_append(while_ctx2, aml_store(one, has_event));
+                aml_append(while_ctx2, aml_store(zero, num_added_cpus));
+
+                /*
+                 * Scan CPUs, till there are CPUs with events or
+                 * CPU_ADDED_LIST capacity is exhausted
+                 */
+                while_ctx = aml_while(aml_land(aml_equal(has_event, one),
+                                      aml_lless(uid, aml_int(arch_ids->len))));
+                {
+                     /*
+                      * clear loop exit condition, ins_evt/rm_evt checks will
+                      * set it to 1 while next_cpu_cmd returns a CPU with events
+                      */
+                     aml_append(while_ctx, aml_store(zero, has_event));
+
+                     aml_append(while_ctx, aml_store(uid, cpu_selector));
+                     aml_append(while_ctx, aml_store(next_cpu_cmd, cpu_cmd));
+
+                     /*
+                      * wrap around case, scan is complete, exit loop.
+                      * It happens since events are not cleared in scan loop,
+                      * so next_cpu_cmd continues to find already processed CPUs
+                      */
+                     ifctx = aml_if(aml_lless(cpu_data, uid));
+                     {
+                         aml_append(ifctx, aml_break());
+                     }
+                     aml_append(while_ctx, ifctx);
+
+                     /*
+                      * if CPU_ADDED_LIST is full, exit inner loop and process
+                      * collected CPUs
+                      */
+                     ifctx = aml_if(
+                         aml_equal(num_added_cpus, aml_int(max_cpus_per_pass)));
+                     {
+                         aml_append(ifctx, aml_store(one, has_job));
+                         aml_append(ifctx, aml_break());
+                     }
+                     aml_append(while_ctx, ifctx);
+
+                     aml_append(while_ctx, aml_store(cpu_data, uid));
+                     ifctx = aml_if(aml_equal(ins_evt, one));
+                     {
+                         /* cache added CPUs to Notify/Wakeup later */
+                         aml_append(ifctx, aml_store(uid,
+                             aml_index(new_cpus, num_added_cpus)));
+                         aml_append(ifctx, aml_increment(num_added_cpus));
+                         aml_append(ifctx, aml_store(one, has_event));
+                     }
+                     aml_append(while_ctx, ifctx);
+                     else_ctx = aml_else();
+                     ifctx = aml_if(aml_equal(rm_evt, one));
+                     {
+                         aml_append(ifctx,
+                             aml_call2(CPU_NOTIFY_METHOD, uid, eject_req));
+                         aml_append(ifctx, aml_store(one, rm_evt));
+                         aml_append(ifctx, aml_store(one, has_event));
+                     }
+                     aml_append(else_ctx, ifctx);
+                     aml_append(while_ctx, else_ctx);
+                     aml_append(while_ctx, aml_increment(uid));
+                }
+                aml_append(while_ctx2, while_ctx);
+
+                /*
+                 * in case FW negotiated ICH9_LPC_SMI_F_CPU_HOTPLUG_BIT,
+                 * make upcall to FW, so it can pull in new CPUs before
+                 * OS is notified and wakes them up
+                 */
+                if (opts.smi_path) {
+                    ifctx = aml_if(aml_lgreater(num_added_cpus, zero));
+                    {
+                        aml_append(ifctx, aml_store(aml_int(OVMF_CPUHP_SMI_CMD),
+                            aml_name("%s", opts.smi_path)));
+                    }
+                    aml_append(while_ctx2, ifctx);
+                }
+
+                /* Notify OSPM about new CPUs and clear insert events */
+                aml_append(while_ctx2, aml_store(zero, cpu_idx));
+                while_ctx = aml_while(aml_lless(cpu_idx, num_added_cpus));
+                {
+                    aml_append(while_ctx,
+                        aml_store(aml_derefof(aml_index(new_cpus, cpu_idx)),
+                                  uid));
+                    aml_append(while_ctx,
+                        aml_call2(CPU_NOTIFY_METHOD, uid, dev_chk));
+                    aml_append(while_ctx, aml_store(uid, aml_debug()));
+                    aml_append(while_ctx, aml_store(uid, cpu_selector));
+                    aml_append(while_ctx, aml_store(one, ins_evt));
+                    aml_append(while_ctx, aml_increment(cpu_idx));
+                }
+                aml_append(while_ctx2, while_ctx);
+                /*
+                 * If another batch is needed, then it will resume scanning
+                 * exactly at -- and not after -- the last CPU that's currently
+                 * in CPU_ADDED_LIST. In other words, the last CPU in
+                 * CPU_ADDED_LIST is going to be re-checked. That's OK: we've
+                 * just cleared the insert event for *all* CPUs in
+                 * CPU_ADDED_LIST, including the last one. So the scan will
+                 * simply seek past it.
+                 */
             }
-            aml_append(method, while_ctx);
+            aml_append(method, while_ctx2);
             aml_append(method, aml_release(ctrl_lock));
         }
         aml_append(cpus_dev, method);
-- 
MST



  parent reply	other threads:[~2020-09-29  7:41 UTC|newest]

Thread overview: 63+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2020-09-29  7:20 [PULL v4 00/48] virtio,pc,acpi: fixes, tests Michael S. Tsirkin
2020-09-29  7:20 ` [PULL v4 01/48] linux headers: sync to 5.9-rc4 Michael S. Tsirkin
2020-09-29  7:20   ` Michael S. Tsirkin
2020-09-29  7:20 ` [PULL v4 03/48] vhost-vdpa: batch updating IOTLB mappings Michael S. Tsirkin
2020-09-29  7:21 ` [PULL v4 04/48] virtio-mem: detach the element from the virtqueue when error occurs Michael S. Tsirkin
2020-09-29  7:21 ` [PULL v4 05/48] pc: fix auto_enable_numa_with_memhp/auto_enable_numa_with_memdev for the 5.0 machine Michael S. Tsirkin
2020-09-29  7:21 ` [PULL v4 06/48] vhost: recheck dev state in the vhost_migration_log routine Michael S. Tsirkin
2020-09-29  7:21 ` [PULL v4 07/48] vhost: check queue state in the vhost_dev_set_log routine Michael S. Tsirkin
2020-09-29  7:21 ` [PULL v4 08/48] tests/qtest/vhost-user-test: prepare the tests for adding new dev class Michael S. Tsirkin
2020-09-29  7:21 ` [PULL v4 09/48] cphp: remove deprecated cpu-add command(s) Michael S. Tsirkin
2020-09-29  7:21 ` [PULL v4 10/48] virtio-iommu: Check gtrees are non null before destroying them Michael S. Tsirkin
2020-09-29  7:21 ` [PULL v4 11/48] virtio-iommu-pci: force virtio version 1 Michael S. Tsirkin
2020-09-29  7:21 ` [PULL v4 12/48] virtio-pmem-pci: " Michael S. Tsirkin
2020-09-29  7:21 ` [PULL v4 13/48] util/hexdump: introduce qemu_hexdump_line() Michael S. Tsirkin
2020-09-29  7:21 ` [PULL v4 14/48] vhost-vdpa: add trace-events Michael S. Tsirkin
2020-09-29  7:21 ` [PULL v4 15/48] configure: Fix build dependencies with vhost-vdpa Michael S. Tsirkin
2020-09-29  7:21 ` [PULL v4 16/48] virtio: skip legacy support check on machine types less than 5.1 Michael S. Tsirkin
2020-09-29  7:21 ` [PULL v4 17/48] vhost-vsock-pci: force virtio version 1 Michael S. Tsirkin
2020-09-29  7:21 ` [PULL v4 18/48] vhost-user-vsock-pci: " Michael S. Tsirkin
2020-09-29  7:21 ` [PULL v4 19/48] vhost-vsock-ccw: " Michael S. Tsirkin
2020-09-29  7:21 ` [PULL v4 20/48] virtio: update MemoryRegionCaches when guest set bad features Michael S. Tsirkin
2020-09-29  7:21 ` [PULL v4 21/48] x86: lpc9: let firmware negotiate 'CPU hotplug with SMI' features Michael S. Tsirkin
2020-09-29  7:21 ` [PULL v4 22/48] x86: cpuhp: prevent guest crash on CPU hotplug when broadcast SMI is in use Michael S. Tsirkin
2020-09-29  7:21 ` [PULL v4 23/48] x86: cpuhp: refuse cpu hot-unplug request earlier if not supported Michael S. Tsirkin
2020-09-29  7:21 ` [PULL v4 24/48] acpi: add aml_land() and aml_break() primitives Michael S. Tsirkin
2020-09-29  7:21 ` [PULL v4 25/48] tests: acpi: mark to be changed tables in bios-tables-test-allowed-diff Michael S. Tsirkin
2020-09-29  7:22 ` [PULL v4 26/48] x86: ich9: expose "smi_negotiated_features" as a QOM property Michael S. Tsirkin
2020-09-29  7:22 ` [PULL v4 27/48] x86: acpi: introduce AcpiPmInfo::smi_on_cpuhp Michael S. Tsirkin
2020-09-29  7:22 ` [PULL v4 28/48] x86: acpi: introduce the PCI0.SMI0 ACPI device Michael S. Tsirkin
2020-09-29  7:22 ` Michael S. Tsirkin [this message]
2020-09-29  7:22 ` [PULL v4 30/48] tests: acpi: update acpi blobs with new AML Michael S. Tsirkin
2020-09-29  7:22 ` [PULL v4 31/48] hw/smbios: support loading OEM strings values from a file Michael S. Tsirkin
2020-09-29  7:22 ` [PULL v4 32/48] hw/smbios: report error if table size is too large Michael S. Tsirkin
2020-09-29  7:22 ` [PULL v4 33/48] qemu-options: document SMBIOS type 11 settings Michael S. Tsirkin
2020-09-29  7:22 ` [PULL v4 34/48] vhost-user: save features of multiqueues if chardev is closed Michael S. Tsirkin
2021-05-12  7:58   ` Michael S. Tsirkin
2020-09-29  7:22 ` [PULL v4 35/48] tests/acpi: mark addition of table DSDT.roothp for unit testing root pci hotplug Michael S. Tsirkin
2020-09-29  7:22 ` [PULL v4 36/48] tests/acpi: add new unit test to test hotplug off/on feature on the root pci bus Michael S. Tsirkin
2020-09-29  7:22 ` [PULL v4 37/48] tests/acpi: add a new ACPI table in order to test root pci hotplug on/off Michael S. Tsirkin
2020-09-29  7:22 ` [PULL v4 38/48] Fix a gap where acpi_pcihp_find_hotplug_bus() returns a non-hotpluggable bus Michael S. Tsirkin
2020-09-29  7:22 ` [PULL v4 39/48] i440fx/acpi: do not add hotplug related amls for cold plugged bridges Michael S. Tsirkin
2020-09-29  7:22 ` [PULL v4 40/48] tests/acpi: list added acpi table binary file for pci bridge hotplug test Michael S. Tsirkin
2020-09-29  7:22 ` [PULL v4 41/48] tests/acpi: unit test for 'acpi-pci-hotplug-with-bridge-support' bridge flag Michael S. Tsirkin
2020-09-29  7:22 ` [PULL v4 42/48] tests/acpi: add newly added acpi DSDT table blob for pci bridge hotplug flag Michael S. Tsirkin
2020-09-29  7:22 ` [PULL v4 43/48] Add ACPI DSDT tables for q35 that are being updated by the next patch Michael S. Tsirkin
2020-09-29  7:22 ` [PULL v4 44/48] piix4: don't reserve hw resources when hotplug is off globally Michael S. Tsirkin
2020-11-07 10:10   ` Philippe Mathieu-Daudé
2020-11-07 11:14     ` Philippe Mathieu-Daudé
2020-11-07 12:22     ` Ani Sinha
2020-11-07 14:18       ` Philippe Mathieu-Daudé
2020-11-07 14:28         ` Michael S. Tsirkin
2020-09-29  7:22 ` [PULL v4 45/48] tests/acpi: update golden master DSDT binary table blobs for q35 Michael S. Tsirkin
2020-09-29  7:23 ` [PULL v4 46/48] hw: virtio-pmem: detach the element fromt the virtqueue when error occurs Michael S. Tsirkin
2020-09-29  7:23 ` [PULL v4 47/48] libvhost-user: return early on virtqueue errors Michael S. Tsirkin
2020-09-29  7:23 ` [PULL v4 48/48] libvhost-user: return on error in vu_log_queue_fill() Michael S. Tsirkin
2020-09-29  7:25 ` [PULL v4 02/48] vhost: switch to use IOTLB v2 format Michael S. Tsirkin
2020-09-29  8:13 ` [PULL v4 00/48] virtio,pc,acpi: fixes, tests no-reply
2020-09-29  8:50 ` no-reply
2020-09-29 11:02 ` Peter Maydell
2020-09-29 11:04 ` Michael S. Tsirkin
2020-09-29 11:07   ` Peter Maydell
2020-09-29 11:13     ` Michael S. Tsirkin
2020-10-01  9:16       ` Laszlo Ersek

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20200929071948.281157-30-mst@redhat.com \
    --to=mst@redhat.com \
    --cc=imammedo@redhat.com \
    --cc=lersek@redhat.com \
    --cc=peter.maydell@linaro.org \
    --cc=qemu-devel@nongnu.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.