From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-23.3 required=3.0 tests=BAYES_00,DKIMWL_WL_MED, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS, INCLUDES_CR_TRAILER,INCLUDES_PATCH,MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS, USER_IN_DEF_DKIM_WL autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id F15F2C64E7B for ; Tue, 1 Dec 2020 19:01:54 +0000 (UTC) Received: from lists.gnu.org (lists.gnu.org [209.51.188.17]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPS id 02D4A20643 for ; Tue, 1 Dec 2020 19:01:53 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=pass (2048-bit key) header.d=google.com header.i=@google.com header.b="hcCb4xEe" DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 02D4A20643 Authentication-Results: mail.kernel.org; dmarc=fail (p=reject dis=none) header.from=google.com Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=qemu-devel-bounces+qemu-devel=archiver.kernel.org@nongnu.org Received: from localhost ([::1]:56532 helo=lists1p.gnu.org) by lists.gnu.org with esmtp (Exim 4.90_1) (envelope-from ) id 1kkAuO-0008OH-PG for qemu-devel@archiver.kernel.org; Tue, 01 Dec 2020 14:01:52 -0500 Received: from eggs.gnu.org ([2001:470:142:3::10]:34190) by lists.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kkAsh-0007GB-2V for qemu-devel@nongnu.org; Tue, 01 Dec 2020 14:00:07 -0500 Received: from mail-io1-xd41.google.com ([2607:f8b0:4864:20::d41]:46415) by eggs.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_128_GCM_SHA256:128) (Exim 4.90_1) (envelope-from ) id 1kkAsd-0008Cx-Kz for qemu-devel@nongnu.org; Tue, 01 Dec 2020 14:00:06 -0500 Received: by mail-io1-xd41.google.com with SMTP id d8so2698575ioc.13 for ; Tue, 01 Dec 2020 11:00:03 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20161025; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=7RkaQ4hyux8dO/p79wgIdkoG/uhD8I0JO2+wYVtV8Sk=; b=hcCb4xEe1csQol9u2p3Lh8XrVM80dl0xip6phr65arYeGIl76ZfnFqwmbw3oxyGeyp wWbsH2E2D0bVcMW2V2HeJtem61kI9nlh3vpGGGAqcLK2fngo9up5952uKYsYy0NCMv87 jMCTU2CpPsHoqVDPhjE/VSCnKE5aMtTaKaTyUk63JOxoNQHox2ztWUZKbOkpTeo8DL/0 IrTnCw0rz0C/lhqb/zlDrSJycVGeBMfCR+WnV8ATWa2BVnW94M/AtxkHWP9+rpOeTnNV Cmpr8PjB1EHZsx/SkuptLpDigkmTJc++GeuqLvTM3GsZufC5q8WkycoQmffUI8M+O/3i V1Ww== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=7RkaQ4hyux8dO/p79wgIdkoG/uhD8I0JO2+wYVtV8Sk=; b=rrqwpYWYJ58cn+Ek5ZWn5qJwVKX16rizVSXpaOVEzZy5J+kM2KRa22chfWqqKp8RgT H2/mVudDJcYHZQ8cdJKFnIKFfkvSiphPOFDyPe7rn26gGtVu1e00jt69Ua3sWYJ4qF20 qhTss63wO9TYPvJtOkdgYnHdIcYDuiM0x2gadYrITWFGBgS0uHfhBXpG+Fu1KR0zac0E hMIKol2n4sWUx00yYlSAJ6jNjyIWvA5wxRkAllsBFnvyiyqjIdqlSiA+XVmKIvoSaLmk StaBE88iSfr9tVCw4Zrw00t/gIadt+fIwWN0Udn7hd1cAGmsI2o2Cnck/9Nvw5Rg984c CPMQ== X-Gm-Message-State: AOAM531zq6fMEFaxqw8xksCWLqYy6hU/l3Ziycxj/ccDwmD2WTIbuyjT syMmkILjuYUN4e+LXqvWUVjplkLUCLgbSsUDz7l2Tw== X-Google-Smtp-Source: ABdhPJzzjROSmVddyYH306f340ZATrB0TdtL96cnxUBytvNRWY2naziXgU1z7uSw2hTBQC98WxVyjniy43y56t87y3c= X-Received: by 2002:a5e:9706:: with SMTP id w6mr3496889ioj.132.1606849201911; Tue, 01 Dec 2020 11:00:01 -0800 (PST) MIME-Version: 1.0 References: <20201201082142.649007-1-pcc@google.com> In-Reply-To: From: Peter Collingbourne Date: Tue, 1 Dec 2020 10:59:50 -0800 Message-ID: Subject: Re: [PATCH] arm/hvf: Optimize and simplify WFI handling To: Alexander Graf Cc: Frank Yang , Roman Bolshakov , Peter Maydell , Eduardo Habkost , Richard Henderson , qemu-devel , Cameron Esfahani , qemu-arm@nongnu.org, Claudio Fontana , Paolo Bonzini Content-Type: text/plain; charset="UTF-8" Received-SPF: pass client-ip=2607:f8b0:4864:20::d41; envelope-from=pcc@google.com; helo=mail-io1-xd41.google.com X-Spam_score_int: -175 X-Spam_score: -17.6 X-Spam_bar: ----------------- X-Spam_report: (-17.6 / 5.0 requ) BAYES_00=-1.9, DKIMWL_WL_MED=-0.001, DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, DKIM_VALID_EF=-0.1, ENV_AND_HDR_SPF_MATCH=-0.5, RCVD_IN_DNSWL_NONE=-0.0001, SPF_HELO_NONE=0.001, SPF_PASS=-0.001, USER_IN_DEF_DKIM_WL=-7.5, USER_IN_DEF_SPF_WL=-7.5 autolearn=unavailable autolearn_force=no X-Spam_action: no action X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.23 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: qemu-devel-bounces+qemu-devel=archiver.kernel.org@nongnu.org Sender: "Qemu-devel" On Tue, Dec 1, 2020 at 3:16 AM Alexander Graf wrote: > > Hi Peter, > > On 01.12.20 09:21, Peter Collingbourne wrote: > > Sleep on WFx until the VTIMER is due but allow ourselves to be woken > > up on IPI. > > > > Signed-off-by: Peter Collingbourne > > > Thanks a bunch! > > > > --- > > Alexander Graf wrote: > >> I would love to take a patch from you here :). I'll still be stuck for a > >> while with the sysreg sync rework that Peter asked for before I can look > >> at WFI again. > > Okay, here's a patch :) It's a relatively straightforward adaptation > > of what we have in our fork, which can now boot Android to GUI while > > remaining at around 4% CPU when idle. > > > > I'm not set up to boot a full Linux distribution at the moment so I > > tested it on upstream QEMU by running a recent mainline Linux kernel > > with a rootfs containing an init program that just does sleep(5) > > and verified that the qemu process remains at low CPU usage during > > the sleep. This was on top of your v2 plus the last patch of your v1 > > since it doesn't look like you have a replacement for that logic yet. > > > > accel/hvf/hvf-cpus.c | 5 +-- > > include/sysemu/hvf_int.h | 3 +- > > target/arm/hvf/hvf.c | 94 +++++++++++----------------------------- > > 3 files changed, 28 insertions(+), 74 deletions(-) > > > > diff --git a/accel/hvf/hvf-cpus.c b/accel/hvf/hvf-cpus.c > > index 4360f64671..b2c8fb57f6 100644 > > --- a/accel/hvf/hvf-cpus.c > > +++ b/accel/hvf/hvf-cpus.c > > @@ -344,9 +344,8 @@ static int hvf_init_vcpu(CPUState *cpu) > > sigact.sa_handler = dummy_signal; > > sigaction(SIG_IPI, &sigact, NULL); > > > > - pthread_sigmask(SIG_BLOCK, NULL, &set); > > - sigdelset(&set, SIG_IPI); > > - pthread_sigmask(SIG_SETMASK, &set, NULL); > > + pthread_sigmask(SIG_BLOCK, NULL, &cpu->hvf->unblock_ipi_mask); > > + sigdelset(&cpu->hvf->unblock_ipi_mask, SIG_IPI); > > > What will this do to the x86 hvf implementation? We're now not > unblocking SIG_IPI again for that, right? Yes and that was the case before your patch series. > > > > #ifdef __aarch64__ > > r = hv_vcpu_create(&cpu->hvf->fd, (hv_vcpu_exit_t **)&cpu->hvf->exit, NULL); > > diff --git a/include/sysemu/hvf_int.h b/include/sysemu/hvf_int.h > > index c56baa3ae8..13adf6ea77 100644 > > --- a/include/sysemu/hvf_int.h > > +++ b/include/sysemu/hvf_int.h > > @@ -62,8 +62,7 @@ extern HVFState *hvf_state; > > struct hvf_vcpu_state { > > uint64_t fd; > > void *exit; > > - struct timespec ts; > > - bool sleeping; > > + sigset_t unblock_ipi_mask; > > }; > > > > void assert_hvf_ok(hv_return_t ret); > > diff --git a/target/arm/hvf/hvf.c b/target/arm/hvf/hvf.c > > index 8fe10966d2..60a361ff38 100644 > > --- a/target/arm/hvf/hvf.c > > +++ b/target/arm/hvf/hvf.c > > @@ -2,6 +2,7 @@ > > * QEMU Hypervisor.framework support for Apple Silicon > > > > * Copyright 2020 Alexander Graf > > + * Copyright 2020 Google LLC > > * > > * This work is licensed under the terms of the GNU GPL, version 2 or later. > > * See the COPYING file in the top-level directory. > > @@ -18,6 +19,7 @@ > > #include "sysemu/hw_accel.h" > > > > #include > > +#include > > > > #include "exec/address-spaces.h" > > #include "hw/irq.h" > > @@ -320,18 +322,8 @@ int hvf_arch_init_vcpu(CPUState *cpu) > > > > void hvf_kick_vcpu_thread(CPUState *cpu) > > { > > - if (cpu->hvf->sleeping) { > > - /* > > - * When sleeping, make sure we always send signals. Also, clear the > > - * timespec, so that an IPI that arrives between setting hvf->sleeping > > - * and the nanosleep syscall still aborts the sleep. > > - */ > > - cpu->thread_kicked = false; > > - cpu->hvf->ts = (struct timespec){ }; > > - cpus_kick_thread(cpu); > > - } else { > > - hv_vcpus_exit(&cpu->hvf->fd, 1); > > - } > > + cpus_kick_thread(cpu); > > + hv_vcpus_exit(&cpu->hvf->fd, 1); > > > This means your first WFI will almost always return immediately due to a > pending signal, because there probably was an IRQ pending before on the > same CPU, no? That's right. Any approach involving the "sleeping" field would need to be implemented carefully to avoid races that may result in missed wakeups so for simplicity I just decided to send both kinds of wakeups. In particular the approach in the updated patch you sent is racy and I'll elaborate more in the reply to that patch. > > } > > > > static int hvf_inject_interrupts(CPUState *cpu) > > @@ -385,18 +377,19 @@ int hvf_vcpu_exec(CPUState *cpu) > > uint64_t syndrome = hvf_exit->exception.syndrome; > > uint32_t ec = syn_get_ec(syndrome); > > > > + qemu_mutex_lock_iothread(); > > > Is there a particular reason you're moving the iothread lock out again > from the individual bits? I would really like to keep a notion of fast > path exits. We still need to lock at least once no matter the exit reason to check the interrupts so I don't think it's worth it to try and avoid locking like this. It also makes the implementation easier to reason about and therefore more likely to be correct. In our implementation we just stay locked the whole time unless we're in hv_vcpu_run() or pselect(). > > switch (exit_reason) { > > case HV_EXIT_REASON_EXCEPTION: > > /* This is the main one, handle below. */ > > break; > > case HV_EXIT_REASON_VTIMER_ACTIVATED: > > - qemu_mutex_lock_iothread(); > > current_cpu = cpu; > > qemu_set_irq(arm_cpu->gt_timer_outputs[GTIMER_VIRT], 1); > > qemu_mutex_unlock_iothread(); > > continue; > > case HV_EXIT_REASON_CANCELED: > > /* we got kicked, no exit to process */ > > + qemu_mutex_unlock_iothread(); > > continue; > > default: > > assert(0); > > @@ -413,7 +406,6 @@ int hvf_vcpu_exec(CPUState *cpu) > > uint32_t srt = (syndrome >> 16) & 0x1f; > > uint64_t val = 0; > > > > - qemu_mutex_lock_iothread(); > > current_cpu = cpu; > > > > DPRINTF("data abort: [pc=0x%llx va=0x%016llx pa=0x%016llx isv=%x " > > @@ -446,8 +438,6 @@ int hvf_vcpu_exec(CPUState *cpu) > > hvf_set_reg(cpu, srt, val); > > } > > > > - qemu_mutex_unlock_iothread(); > > - > > advance_pc = true; > > break; > > } > > @@ -493,68 +483,36 @@ int hvf_vcpu_exec(CPUState *cpu) > > case EC_WFX_TRAP: > > if (!(syndrome & WFX_IS_WFE) && !(cpu->interrupt_request & > > (CPU_INTERRUPT_HARD | CPU_INTERRUPT_FIQ))) { > > - uint64_t cval, ctl, val, diff, now; > > + uint64_t cval; > > > > - /* Set up a local timer for vtimer if necessary ... */ > > - r = hv_vcpu_get_sys_reg(cpu->hvf->fd, HV_SYS_REG_CNTV_CTL_EL0, &ctl); > > - assert_hvf_ok(r); > > r = hv_vcpu_get_sys_reg(cpu->hvf->fd, HV_SYS_REG_CNTV_CVAL_EL0, &cval); > > assert_hvf_ok(r); > > > > - asm volatile("mrs %0, cntvct_el0" : "=r"(val)); > > - diff = cval - val; > > - > > - now = qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) / > > - gt_cntfrq_period_ns(arm_cpu); > > - > > - /* Timer disabled or masked, just wait for long */ > > - if (!(ctl & 1) || (ctl & 2)) { > > - diff = (120 * NANOSECONDS_PER_SECOND) / > > - gt_cntfrq_period_ns(arm_cpu); > > + int64_t ticks_to_sleep = cval - mach_absolute_time(); > > + if (ticks_to_sleep < 0) { > > + break; > > > This will loop at 100% for Windows, which configures the vtimer as > cval=0 ctl=7, so with IRQ mask bit set. Okay, but the 120s is kind of arbitrary so we should just sleep until we get a signal. That can be done by passing null as the timespec argument to pselect(). > > > Alex > > > > } > > > > - if (diff < INT64_MAX) { > > - uint64_t ns = diff * gt_cntfrq_period_ns(arm_cpu); > > - struct timespec *ts = &cpu->hvf->ts; > > - > > - *ts = (struct timespec){ > > - .tv_sec = ns / NANOSECONDS_PER_SECOND, > > - .tv_nsec = ns % NANOSECONDS_PER_SECOND, > > - }; > > - > > - /* > > - * Waking up easily takes 1ms, don't go to sleep for smaller > > - * time periods than 2ms. > > - */ > > - if (!ts->tv_sec && (ts->tv_nsec < (SCALE_MS * 2))) { > > > I put this logic here on purpose. A pselect(1 ns) easily takes 1-2ms to > return. Without logic like this, super short WFIs will hurt performance > quite badly. I don't think that's accurate. According to this benchmark it's a few hundred nanoseconds at most. pcc@pac-mini /tmp> cat pselect.c #include #include int main() { sigset_t mask, orig_mask; pthread_sigmask(SIG_SETMASK, 0, &mask); sigaddset(&mask, SIGUSR1); pthread_sigmask(SIG_SETMASK, &mask, &orig_mask); for (int i = 0; i != 1000000; ++i) { struct timespec ts = { 0, 1 }; pselect(0, 0, 0, 0, &ts, &orig_mask); } } pcc@pac-mini /tmp> time ./pselect ________________________________________________________ Executed in 179.87 millis fish external usr time 77.68 millis 57.00 micros 77.62 millis sys time 101.37 millis 852.00 micros 100.52 millis Besides, all that you're really saving here is the single pselect call. There are no doubt more expensive syscalls involved in exiting and entering the VCPU that would dominate here. Peter > > > Alex > > > - advance_pc = true; > > - break; > > - } > > - > > - /* Set cpu->hvf->sleeping so that we get a SIG_IPI signal. */ > > - cpu->hvf->sleeping = true; > > - smp_mb(); > > - > > - /* Bail out if we received an IRQ meanwhile */ > > - if (cpu->thread_kicked || (cpu->interrupt_request & > > - (CPU_INTERRUPT_HARD | CPU_INTERRUPT_FIQ))) { > > - cpu->hvf->sleeping = false; > > - break; > > - } > > - > > - /* nanosleep returns on signal, so we wake up on kick. */ > > - nanosleep(ts, NULL); > > - > > - /* Out of sleep - either naturally or because of a kick */ > > - cpu->hvf->sleeping = false; > > - } > > + uint64_t seconds = ticks_to_sleep / arm_cpu->gt_cntfrq_hz; > > + uint64_t nanos = > > + (ticks_to_sleep - arm_cpu->gt_cntfrq_hz * seconds) * > > + 1000000000 / arm_cpu->gt_cntfrq_hz; > > + struct timespec ts = { seconds, nanos }; > > + > > + /* > > + * Use pselect to sleep so that other threads can IPI us while > > + * we're sleeping. > > + */ > > + qatomic_mb_set(&cpu->thread_kicked, false); > > + qemu_mutex_unlock_iothread(); > > + pselect(0, 0, 0, 0, &ts, &cpu->hvf->unblock_ipi_mask); > > + qemu_mutex_lock_iothread(); > > > > advance_pc = true; > > } > > break; > > case EC_AA64_HVC: > > cpu_synchronize_state(cpu); > > - qemu_mutex_lock_iothread(); > > current_cpu = cpu; > > if (arm_is_psci_call(arm_cpu, EXCP_HVC)) { > > arm_handle_psci_call(arm_cpu); > > @@ -562,11 +520,9 @@ int hvf_vcpu_exec(CPUState *cpu) > > DPRINTF("unknown HVC! %016llx", env->xregs[0]); > > env->xregs[0] = -1; > > } > > - qemu_mutex_unlock_iothread(); > > break; > > case EC_AA64_SMC: > > cpu_synchronize_state(cpu); > > - qemu_mutex_lock_iothread(); > > current_cpu = cpu; > > if (arm_is_psci_call(arm_cpu, EXCP_SMC)) { > > arm_handle_psci_call(arm_cpu); > > @@ -575,7 +531,6 @@ int hvf_vcpu_exec(CPUState *cpu) > > env->xregs[0] = -1; > > env->pc += 4; > > } > > - qemu_mutex_unlock_iothread(); > > break; > > default: > > cpu_synchronize_state(cpu); > > @@ -594,6 +549,7 @@ int hvf_vcpu_exec(CPUState *cpu) > > r = hv_vcpu_set_reg(cpu->hvf->fd, HV_REG_PC, pc); > > assert_hvf_ok(r); > > } > > + qemu_mutex_unlock_iothread(); > > } while (ret == 0); > > > > qemu_mutex_lock_iothread();