From: "André Przywara" <andre.przywara@arm.com> To: Christoffer Dall <christoffer.dall@arm.com> Cc: Marc Zyngier <marc.zyngier@arm.com>, kvmarm@lists.cs.columbia.edu, linux-arm-kernel@lists.infradead.org, kvm@vger.kernel.org Subject: Re: [PATCH 11/14] KVM: arm/arm64: timer: Rework data structures for multiple timers Date: Mon, 18 Feb 2019 15:10:16 +0000 [thread overview] Message-ID: <d44bd102-5ad6-6e08-5e47-8a8386a139c6@arm.com> (raw) In-Reply-To: <20190124140032.8588-12-christoffer.dall@arm.com> On Thu, 24 Jan 2019 15:00:29 +0100 Christoffer Dall <christoffer.dall@arm.com> wrote: Hi, I already looked at most of these patches earlier, without finding serious issues, but figured I would give those without any Reviewed-by: or Acked-by: tags a closer look. (This patch just carries a S-o-b: tag from Marc in the kvm-arm git repo.) > Prepare for having 4 timer data structures (2 for now). > > Change loaded to an enum so that we know not just whether *some* state > is loaded on the CPU, but also *which* state is loaded. > > Move loaded to the cpu data structure and not the individual timer > structure, in preparation for assigning the EL1 phys timer as well. > > Signed-off-by: Christoffer Dall <christoffer.dall@arm.com> > Acked-by: Marc Zyngier <marc.zyngier@arm.com> > --- > include/kvm/arm_arch_timer.h | 44 ++++++++++++++------------- > virt/kvm/arm/arch_timer.c | 58 +++++++++++++++++++----------------- > 2 files changed, 54 insertions(+), 48 deletions(-) > > diff --git a/include/kvm/arm_arch_timer.h b/include/kvm/arm_arch_timer.h > index d26b7fde9935..d40fe57a2d0d 100644 > --- a/include/kvm/arm_arch_timer.h > +++ b/include/kvm/arm_arch_timer.h > @@ -36,6 +36,8 @@ enum kvm_arch_timer_regs { > }; > > struct arch_timer_context { > + struct kvm_vcpu *vcpu; > + > /* Registers: control register, timer value */ > u32 cnt_ctl; > u64 cnt_cval; > @@ -43,32 +45,34 @@ struct arch_timer_context { > /* Timer IRQ */ > struct kvm_irq_level irq; > > - /* > - * We have multiple paths which can save/restore the timer state > - * onto the hardware, so we need some way of keeping track of > - * where the latest state is. > - * > - * loaded == true: State is loaded on the hardware registers. > - * loaded == false: State is stored in memory. > - */ > - bool loaded; > - > /* Virtual offset */ > - u64 cntvoff; > + u64 cntvoff; > + > + /* Emulated Timer (may be unused) */ > + struct hrtimer hrtimer; > +}; > + > +enum loaded_timer_state { > + TIMER_NOT_LOADED, > + TIMER_EL1_LOADED, So this gets reverted in PATCH 13/14, and I don't see it reappearing in the nv series later on. Is that just needed for assigning the phys timer in the next patch, and gets obsolete with the timer_map? Or is this a rebase artefact and we don't actually need this? The rest of the patch looks like valid transformations to me. Cheers, Andre. > }; > > struct arch_timer_cpu { > - struct arch_timer_context vtimer; > - struct arch_timer_context ptimer; > + struct arch_timer_context timers[NR_KVM_TIMERS]; > > /* Background timer used when the guest is not running */ > struct hrtimer bg_timer; > > - /* Physical timer emulation */ > - struct hrtimer phys_timer; > - > /* Is the timer enabled */ > bool enabled; > + > + /* > + * We have multiple paths which can save/restore the timer state > + * onto the hardware, and for nested virt the EL1 hardware timers can > + * contain state from either the VM's EL1 timers or EL2 timers, so we > + * need some way of keeping track of where the latest state is. > + */ > + enum loaded_timer_state loaded; > }; > > int kvm_timer_hyp_init(bool); > @@ -98,10 +102,10 @@ void kvm_timer_init_vhe(void); > > bool kvm_arch_timer_get_input_level(int vintid); > > -#define vcpu_vtimer(v) (&(v)->arch.timer_cpu.vtimer) > -#define vcpu_ptimer(v) (&(v)->arch.timer_cpu.ptimer) > -#define vcpu_get_timer(v,t) \ > - (t == TIMER_VTIMER ? vcpu_vtimer(v) : vcpu_ptimer(v)) > +#define vcpu_timer(v) (&(v)->arch.timer_cpu) > +#define vcpu_get_timer(v,t) (&vcpu_timer(v)->timers[(t)]) > +#define vcpu_vtimer(v) (&(v)->arch.timer_cpu.timers[TIMER_VTIMER]) > +#define vcpu_ptimer(v) (&(v)->arch.timer_cpu.timers[TIMER_PTIMER]) > > u64 kvm_arm_timer_read_sysreg(struct kvm_vcpu *vcpu, > enum kvm_arch_timers tmr, > diff --git a/virt/kvm/arm/arch_timer.c b/virt/kvm/arm/arch_timer.c > index 9502bb91776b..8b0eca5fbad1 100644 > --- a/virt/kvm/arm/arch_timer.c > +++ b/virt/kvm/arm/arch_timer.c > @@ -184,13 +184,11 @@ static enum hrtimer_restart kvm_bg_timer_expire(struct hrtimer *hrt) > static enum hrtimer_restart kvm_phys_timer_expire(struct hrtimer *hrt) > { > struct arch_timer_context *ptimer; > - struct arch_timer_cpu *timer; > struct kvm_vcpu *vcpu; > u64 ns; > > - timer = container_of(hrt, struct arch_timer_cpu, phys_timer); > - vcpu = container_of(timer, struct kvm_vcpu, arch.timer_cpu); > - ptimer = vcpu_ptimer(vcpu); > + ptimer = container_of(hrt, struct arch_timer_context, hrtimer); > + vcpu = ptimer->vcpu; > > /* > * Check that the timer has really expired from the guest's > @@ -209,9 +207,10 @@ static enum hrtimer_restart kvm_phys_timer_expire(struct hrtimer *hrt) > > static bool kvm_timer_should_fire(struct arch_timer_context *timer_ctx) > { > + struct arch_timer_cpu *timer = vcpu_timer(timer_ctx->vcpu); > u64 cval, now; > > - if (timer_ctx->loaded) { > + if (timer->loaded == TIMER_EL1_LOADED) { > u32 cnt_ctl; > > /* Only the virtual timer can be loaded so far */ > @@ -280,7 +279,6 @@ static void kvm_timer_update_irq(struct kvm_vcpu *vcpu, bool new_level, > /* Schedule the background timer for the emulated timer. */ > static void phys_timer_emulate(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > struct arch_timer_context *ptimer = vcpu_ptimer(vcpu); > > /* > @@ -289,11 +287,11 @@ static void phys_timer_emulate(struct kvm_vcpu *vcpu) > * then we also don't need a soft timer. > */ > if (kvm_timer_should_fire(ptimer) || !kvm_timer_irq_can_fire(ptimer)) { > - soft_timer_cancel(&timer->phys_timer); > + soft_timer_cancel(&ptimer->hrtimer); > return; > } > > - soft_timer_start(&timer->phys_timer, kvm_timer_compute_delta(ptimer)); > + soft_timer_start(&ptimer->hrtimer, kvm_timer_compute_delta(ptimer)); > } > > /* > @@ -303,7 +301,7 @@ static void phys_timer_emulate(struct kvm_vcpu *vcpu) > */ > static void kvm_timer_update_state(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > + struct arch_timer_cpu *timer = vcpu_timer(vcpu); > struct arch_timer_context *vtimer = vcpu_vtimer(vcpu); > struct arch_timer_context *ptimer = vcpu_ptimer(vcpu); > bool level; > @@ -329,13 +327,13 @@ static void kvm_timer_update_state(struct kvm_vcpu *vcpu) > > static void vtimer_save_state(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > + struct arch_timer_cpu *timer = vcpu_timer(vcpu); > struct arch_timer_context *vtimer = vcpu_vtimer(vcpu); > unsigned long flags; > > local_irq_save(flags); > > - if (!vtimer->loaded) > + if (timer->loaded == TIMER_NOT_LOADED) > goto out; > > if (timer->enabled) { > @@ -347,7 +345,7 @@ static void vtimer_save_state(struct kvm_vcpu *vcpu) > write_sysreg_el0(0, cntv_ctl); > isb(); > > - vtimer->loaded = false; > + timer->loaded = TIMER_NOT_LOADED; > out: > local_irq_restore(flags); > } > @@ -359,7 +357,7 @@ static void vtimer_save_state(struct kvm_vcpu *vcpu) > */ > static void kvm_timer_blocking(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > + struct arch_timer_cpu *timer = vcpu_timer(vcpu); > struct arch_timer_context *vtimer = vcpu_vtimer(vcpu); > struct arch_timer_context *ptimer = vcpu_ptimer(vcpu); > > @@ -379,20 +377,20 @@ static void kvm_timer_blocking(struct kvm_vcpu *vcpu) > > static void kvm_timer_unblocking(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > + struct arch_timer_cpu *timer = vcpu_timer(vcpu); > > soft_timer_cancel(&timer->bg_timer); > } > > static void vtimer_restore_state(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > + struct arch_timer_cpu *timer = vcpu_timer(vcpu); > struct arch_timer_context *vtimer = vcpu_vtimer(vcpu); > unsigned long flags; > > local_irq_save(flags); > > - if (vtimer->loaded) > + if (timer->loaded == TIMER_EL1_LOADED) > goto out; > > if (timer->enabled) { > @@ -401,7 +399,7 @@ static void vtimer_restore_state(struct kvm_vcpu *vcpu) > write_sysreg_el0(vtimer->cnt_ctl, cntv_ctl); > } > > - vtimer->loaded = true; > + timer->loaded = TIMER_EL1_LOADED; > out: > local_irq_restore(flags); > } > @@ -462,7 +460,7 @@ static void kvm_timer_vcpu_load_nogic(struct kvm_vcpu *vcpu) > > void kvm_timer_vcpu_load(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > + struct arch_timer_cpu *timer = vcpu_timer(vcpu); > struct arch_timer_context *vtimer = vcpu_vtimer(vcpu); > struct arch_timer_context *ptimer = vcpu_ptimer(vcpu); > > @@ -507,7 +505,8 @@ bool kvm_timer_should_notify_user(struct kvm_vcpu *vcpu) > > void kvm_timer_vcpu_put(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > + struct arch_timer_cpu *timer = vcpu_timer(vcpu); > + struct arch_timer_context *ptimer = vcpu_ptimer(vcpu); > > if (unlikely(!timer->enabled)) > return; > @@ -523,7 +522,7 @@ void kvm_timer_vcpu_put(struct kvm_vcpu *vcpu) > * In any case, we re-schedule the hrtimer for the physical timer when > * coming back to the VCPU thread in kvm_timer_vcpu_load(). > */ > - soft_timer_cancel(&timer->phys_timer); > + soft_timer_cancel(&ptimer->hrtimer); > > if (swait_active(kvm_arch_vcpu_wq(vcpu))) > kvm_timer_blocking(vcpu); > @@ -559,7 +558,7 @@ static void unmask_vtimer_irq_user(struct kvm_vcpu *vcpu) > > void kvm_timer_sync_hwstate(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > + struct arch_timer_cpu *timer = vcpu_timer(vcpu); > > if (unlikely(!timer->enabled)) > return; > @@ -570,7 +569,7 @@ void kvm_timer_sync_hwstate(struct kvm_vcpu *vcpu) > > int kvm_timer_vcpu_reset(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > + struct arch_timer_cpu *timer = vcpu_timer(vcpu); > struct arch_timer_context *vtimer = vcpu_vtimer(vcpu); > struct arch_timer_context *ptimer = vcpu_ptimer(vcpu); > > @@ -611,22 +610,25 @@ static void update_vtimer_cntvoff(struct kvm_vcpu *vcpu, u64 cntvoff) > > void kvm_timer_vcpu_init(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > + struct arch_timer_cpu *timer = vcpu_timer(vcpu); > struct arch_timer_context *vtimer = vcpu_vtimer(vcpu); > struct arch_timer_context *ptimer = vcpu_ptimer(vcpu); > > /* Synchronize cntvoff across all vtimers of a VM. */ > update_vtimer_cntvoff(vcpu, kvm_phys_timer_read()); > - vcpu_ptimer(vcpu)->cntvoff = 0; > + ptimer->cntvoff = 0; > > hrtimer_init(&timer->bg_timer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS); > timer->bg_timer.function = kvm_bg_timer_expire; > > - hrtimer_init(&timer->phys_timer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS); > - timer->phys_timer.function = kvm_phys_timer_expire; > + hrtimer_init(&ptimer->hrtimer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS); > + ptimer->hrtimer.function = kvm_phys_timer_expire; > > vtimer->irq.irq = default_vtimer_irq.irq; > ptimer->irq.irq = default_ptimer_irq.irq; > + > + vtimer->vcpu = vcpu; > + ptimer->vcpu = vcpu; > } > > static void kvm_timer_init_interrupt(void *info) > @@ -859,7 +861,7 @@ int kvm_timer_hyp_init(bool has_gic) > > void kvm_timer_vcpu_terminate(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > + struct arch_timer_cpu *timer = vcpu_timer(vcpu); > > soft_timer_cancel(&timer->bg_timer); > } > @@ -903,7 +905,7 @@ bool kvm_arch_timer_get_input_level(int vintid) > > int kvm_timer_enable(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > + struct arch_timer_cpu *timer = vcpu_timer(vcpu); > struct arch_timer_context *vtimer = vcpu_vtimer(vcpu); > int ret; >
WARNING: multiple messages have this Message-ID (diff)
From: "André Przywara" <andre.przywara@arm.com> To: Christoffer Dall <christoffer.dall@arm.com> Cc: Marc Zyngier <marc.zyngier@arm.com>, kvmarm@lists.cs.columbia.edu, linux-arm-kernel@lists.infradead.org, kvm@vger.kernel.org Subject: Re: [PATCH 11/14] KVM: arm/arm64: timer: Rework data structures for multiple timers Date: Mon, 18 Feb 2019 15:10:16 +0000 [thread overview] Message-ID: <d44bd102-5ad6-6e08-5e47-8a8386a139c6@arm.com> (raw) In-Reply-To: <20190124140032.8588-12-christoffer.dall@arm.com> On Thu, 24 Jan 2019 15:00:29 +0100 Christoffer Dall <christoffer.dall@arm.com> wrote: Hi, I already looked at most of these patches earlier, without finding serious issues, but figured I would give those without any Reviewed-by: or Acked-by: tags a closer look. (This patch just carries a S-o-b: tag from Marc in the kvm-arm git repo.) > Prepare for having 4 timer data structures (2 for now). > > Change loaded to an enum so that we know not just whether *some* state > is loaded on the CPU, but also *which* state is loaded. > > Move loaded to the cpu data structure and not the individual timer > structure, in preparation for assigning the EL1 phys timer as well. > > Signed-off-by: Christoffer Dall <christoffer.dall@arm.com> > Acked-by: Marc Zyngier <marc.zyngier@arm.com> > --- > include/kvm/arm_arch_timer.h | 44 ++++++++++++++------------- > virt/kvm/arm/arch_timer.c | 58 +++++++++++++++++++----------------- > 2 files changed, 54 insertions(+), 48 deletions(-) > > diff --git a/include/kvm/arm_arch_timer.h b/include/kvm/arm_arch_timer.h > index d26b7fde9935..d40fe57a2d0d 100644 > --- a/include/kvm/arm_arch_timer.h > +++ b/include/kvm/arm_arch_timer.h > @@ -36,6 +36,8 @@ enum kvm_arch_timer_regs { > }; > > struct arch_timer_context { > + struct kvm_vcpu *vcpu; > + > /* Registers: control register, timer value */ > u32 cnt_ctl; > u64 cnt_cval; > @@ -43,32 +45,34 @@ struct arch_timer_context { > /* Timer IRQ */ > struct kvm_irq_level irq; > > - /* > - * We have multiple paths which can save/restore the timer state > - * onto the hardware, so we need some way of keeping track of > - * where the latest state is. > - * > - * loaded == true: State is loaded on the hardware registers. > - * loaded == false: State is stored in memory. > - */ > - bool loaded; > - > /* Virtual offset */ > - u64 cntvoff; > + u64 cntvoff; > + > + /* Emulated Timer (may be unused) */ > + struct hrtimer hrtimer; > +}; > + > +enum loaded_timer_state { > + TIMER_NOT_LOADED, > + TIMER_EL1_LOADED, So this gets reverted in PATCH 13/14, and I don't see it reappearing in the nv series later on. Is that just needed for assigning the phys timer in the next patch, and gets obsolete with the timer_map? Or is this a rebase artefact and we don't actually need this? The rest of the patch looks like valid transformations to me. Cheers, Andre. > }; > > struct arch_timer_cpu { > - struct arch_timer_context vtimer; > - struct arch_timer_context ptimer; > + struct arch_timer_context timers[NR_KVM_TIMERS]; > > /* Background timer used when the guest is not running */ > struct hrtimer bg_timer; > > - /* Physical timer emulation */ > - struct hrtimer phys_timer; > - > /* Is the timer enabled */ > bool enabled; > + > + /* > + * We have multiple paths which can save/restore the timer state > + * onto the hardware, and for nested virt the EL1 hardware timers can > + * contain state from either the VM's EL1 timers or EL2 timers, so we > + * need some way of keeping track of where the latest state is. > + */ > + enum loaded_timer_state loaded; > }; > > int kvm_timer_hyp_init(bool); > @@ -98,10 +102,10 @@ void kvm_timer_init_vhe(void); > > bool kvm_arch_timer_get_input_level(int vintid); > > -#define vcpu_vtimer(v) (&(v)->arch.timer_cpu.vtimer) > -#define vcpu_ptimer(v) (&(v)->arch.timer_cpu.ptimer) > -#define vcpu_get_timer(v,t) \ > - (t == TIMER_VTIMER ? vcpu_vtimer(v) : vcpu_ptimer(v)) > +#define vcpu_timer(v) (&(v)->arch.timer_cpu) > +#define vcpu_get_timer(v,t) (&vcpu_timer(v)->timers[(t)]) > +#define vcpu_vtimer(v) (&(v)->arch.timer_cpu.timers[TIMER_VTIMER]) > +#define vcpu_ptimer(v) (&(v)->arch.timer_cpu.timers[TIMER_PTIMER]) > > u64 kvm_arm_timer_read_sysreg(struct kvm_vcpu *vcpu, > enum kvm_arch_timers tmr, > diff --git a/virt/kvm/arm/arch_timer.c b/virt/kvm/arm/arch_timer.c > index 9502bb91776b..8b0eca5fbad1 100644 > --- a/virt/kvm/arm/arch_timer.c > +++ b/virt/kvm/arm/arch_timer.c > @@ -184,13 +184,11 @@ static enum hrtimer_restart kvm_bg_timer_expire(struct hrtimer *hrt) > static enum hrtimer_restart kvm_phys_timer_expire(struct hrtimer *hrt) > { > struct arch_timer_context *ptimer; > - struct arch_timer_cpu *timer; > struct kvm_vcpu *vcpu; > u64 ns; > > - timer = container_of(hrt, struct arch_timer_cpu, phys_timer); > - vcpu = container_of(timer, struct kvm_vcpu, arch.timer_cpu); > - ptimer = vcpu_ptimer(vcpu); > + ptimer = container_of(hrt, struct arch_timer_context, hrtimer); > + vcpu = ptimer->vcpu; > > /* > * Check that the timer has really expired from the guest's > @@ -209,9 +207,10 @@ static enum hrtimer_restart kvm_phys_timer_expire(struct hrtimer *hrt) > > static bool kvm_timer_should_fire(struct arch_timer_context *timer_ctx) > { > + struct arch_timer_cpu *timer = vcpu_timer(timer_ctx->vcpu); > u64 cval, now; > > - if (timer_ctx->loaded) { > + if (timer->loaded == TIMER_EL1_LOADED) { > u32 cnt_ctl; > > /* Only the virtual timer can be loaded so far */ > @@ -280,7 +279,6 @@ static void kvm_timer_update_irq(struct kvm_vcpu *vcpu, bool new_level, > /* Schedule the background timer for the emulated timer. */ > static void phys_timer_emulate(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > struct arch_timer_context *ptimer = vcpu_ptimer(vcpu); > > /* > @@ -289,11 +287,11 @@ static void phys_timer_emulate(struct kvm_vcpu *vcpu) > * then we also don't need a soft timer. > */ > if (kvm_timer_should_fire(ptimer) || !kvm_timer_irq_can_fire(ptimer)) { > - soft_timer_cancel(&timer->phys_timer); > + soft_timer_cancel(&ptimer->hrtimer); > return; > } > > - soft_timer_start(&timer->phys_timer, kvm_timer_compute_delta(ptimer)); > + soft_timer_start(&ptimer->hrtimer, kvm_timer_compute_delta(ptimer)); > } > > /* > @@ -303,7 +301,7 @@ static void phys_timer_emulate(struct kvm_vcpu *vcpu) > */ > static void kvm_timer_update_state(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > + struct arch_timer_cpu *timer = vcpu_timer(vcpu); > struct arch_timer_context *vtimer = vcpu_vtimer(vcpu); > struct arch_timer_context *ptimer = vcpu_ptimer(vcpu); > bool level; > @@ -329,13 +327,13 @@ static void kvm_timer_update_state(struct kvm_vcpu *vcpu) > > static void vtimer_save_state(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > + struct arch_timer_cpu *timer = vcpu_timer(vcpu); > struct arch_timer_context *vtimer = vcpu_vtimer(vcpu); > unsigned long flags; > > local_irq_save(flags); > > - if (!vtimer->loaded) > + if (timer->loaded == TIMER_NOT_LOADED) > goto out; > > if (timer->enabled) { > @@ -347,7 +345,7 @@ static void vtimer_save_state(struct kvm_vcpu *vcpu) > write_sysreg_el0(0, cntv_ctl); > isb(); > > - vtimer->loaded = false; > + timer->loaded = TIMER_NOT_LOADED; > out: > local_irq_restore(flags); > } > @@ -359,7 +357,7 @@ static void vtimer_save_state(struct kvm_vcpu *vcpu) > */ > static void kvm_timer_blocking(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > + struct arch_timer_cpu *timer = vcpu_timer(vcpu); > struct arch_timer_context *vtimer = vcpu_vtimer(vcpu); > struct arch_timer_context *ptimer = vcpu_ptimer(vcpu); > > @@ -379,20 +377,20 @@ static void kvm_timer_blocking(struct kvm_vcpu *vcpu) > > static void kvm_timer_unblocking(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > + struct arch_timer_cpu *timer = vcpu_timer(vcpu); > > soft_timer_cancel(&timer->bg_timer); > } > > static void vtimer_restore_state(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > + struct arch_timer_cpu *timer = vcpu_timer(vcpu); > struct arch_timer_context *vtimer = vcpu_vtimer(vcpu); > unsigned long flags; > > local_irq_save(flags); > > - if (vtimer->loaded) > + if (timer->loaded == TIMER_EL1_LOADED) > goto out; > > if (timer->enabled) { > @@ -401,7 +399,7 @@ static void vtimer_restore_state(struct kvm_vcpu *vcpu) > write_sysreg_el0(vtimer->cnt_ctl, cntv_ctl); > } > > - vtimer->loaded = true; > + timer->loaded = TIMER_EL1_LOADED; > out: > local_irq_restore(flags); > } > @@ -462,7 +460,7 @@ static void kvm_timer_vcpu_load_nogic(struct kvm_vcpu *vcpu) > > void kvm_timer_vcpu_load(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > + struct arch_timer_cpu *timer = vcpu_timer(vcpu); > struct arch_timer_context *vtimer = vcpu_vtimer(vcpu); > struct arch_timer_context *ptimer = vcpu_ptimer(vcpu); > > @@ -507,7 +505,8 @@ bool kvm_timer_should_notify_user(struct kvm_vcpu *vcpu) > > void kvm_timer_vcpu_put(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > + struct arch_timer_cpu *timer = vcpu_timer(vcpu); > + struct arch_timer_context *ptimer = vcpu_ptimer(vcpu); > > if (unlikely(!timer->enabled)) > return; > @@ -523,7 +522,7 @@ void kvm_timer_vcpu_put(struct kvm_vcpu *vcpu) > * In any case, we re-schedule the hrtimer for the physical timer when > * coming back to the VCPU thread in kvm_timer_vcpu_load(). > */ > - soft_timer_cancel(&timer->phys_timer); > + soft_timer_cancel(&ptimer->hrtimer); > > if (swait_active(kvm_arch_vcpu_wq(vcpu))) > kvm_timer_blocking(vcpu); > @@ -559,7 +558,7 @@ static void unmask_vtimer_irq_user(struct kvm_vcpu *vcpu) > > void kvm_timer_sync_hwstate(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > + struct arch_timer_cpu *timer = vcpu_timer(vcpu); > > if (unlikely(!timer->enabled)) > return; > @@ -570,7 +569,7 @@ void kvm_timer_sync_hwstate(struct kvm_vcpu *vcpu) > > int kvm_timer_vcpu_reset(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > + struct arch_timer_cpu *timer = vcpu_timer(vcpu); > struct arch_timer_context *vtimer = vcpu_vtimer(vcpu); > struct arch_timer_context *ptimer = vcpu_ptimer(vcpu); > > @@ -611,22 +610,25 @@ static void update_vtimer_cntvoff(struct kvm_vcpu *vcpu, u64 cntvoff) > > void kvm_timer_vcpu_init(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > + struct arch_timer_cpu *timer = vcpu_timer(vcpu); > struct arch_timer_context *vtimer = vcpu_vtimer(vcpu); > struct arch_timer_context *ptimer = vcpu_ptimer(vcpu); > > /* Synchronize cntvoff across all vtimers of a VM. */ > update_vtimer_cntvoff(vcpu, kvm_phys_timer_read()); > - vcpu_ptimer(vcpu)->cntvoff = 0; > + ptimer->cntvoff = 0; > > hrtimer_init(&timer->bg_timer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS); > timer->bg_timer.function = kvm_bg_timer_expire; > > - hrtimer_init(&timer->phys_timer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS); > - timer->phys_timer.function = kvm_phys_timer_expire; > + hrtimer_init(&ptimer->hrtimer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS); > + ptimer->hrtimer.function = kvm_phys_timer_expire; > > vtimer->irq.irq = default_vtimer_irq.irq; > ptimer->irq.irq = default_ptimer_irq.irq; > + > + vtimer->vcpu = vcpu; > + ptimer->vcpu = vcpu; > } > > static void kvm_timer_init_interrupt(void *info) > @@ -859,7 +861,7 @@ int kvm_timer_hyp_init(bool has_gic) > > void kvm_timer_vcpu_terminate(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > + struct arch_timer_cpu *timer = vcpu_timer(vcpu); > > soft_timer_cancel(&timer->bg_timer); > } > @@ -903,7 +905,7 @@ bool kvm_arch_timer_get_input_level(int vintid) > > int kvm_timer_enable(struct kvm_vcpu *vcpu) > { > - struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu; > + struct arch_timer_cpu *timer = vcpu_timer(vcpu); > struct arch_timer_context *vtimer = vcpu_vtimer(vcpu); > int ret; > _______________________________________________ linux-arm-kernel mailing list linux-arm-kernel@lists.infradead.org http://lists.infradead.org/mailman/listinfo/linux-arm-kernel
next prev parent reply other threads:[~2019-02-18 15:10 UTC|newest] Thread overview: 65+ messages / expand[flat|nested] mbox.gz Atom feed top 2019-01-24 14:00 [PATCH 00/14] KVM: arm/arm64: Various rework in preparation of nested virt support Christoffer Dall 2019-01-24 14:00 ` Christoffer Dall 2019-01-24 14:00 ` [PATCH 01/14] arm/arm64: KVM: Introduce kvm_call_hyp_ret() Christoffer Dall 2019-01-24 14:00 ` Christoffer Dall 2019-01-24 14:00 ` [PATCH 02/14] arm64: KVM: Allow for direct call of HYP functions when using VHE Christoffer Dall 2019-01-24 14:00 ` Christoffer Dall 2019-01-24 14:00 ` [PATCH 03/14] arm64: KVM: Drop VHE-specific HYP call stub Christoffer Dall 2019-01-24 14:00 ` Christoffer Dall 2019-01-24 14:00 ` [PATCH 04/14] ARM: KVM: Teach some form of type-safety to kvm_call_hyp Christoffer Dall 2019-01-24 14:00 ` Christoffer Dall 2019-01-24 14:00 ` [PATCH 05/14] arm/arm64: KVM: Statically configure the host's view of MPIDR Christoffer Dall 2019-01-24 14:00 ` Christoffer Dall 2019-01-24 14:00 ` [PATCH 06/14] KVM: arm/arm64: Factor out VMID into struct kvm_vmid Christoffer Dall 2019-01-24 14:00 ` Christoffer Dall 2019-01-24 19:01 ` James Morse 2019-01-24 19:01 ` James Morse 2019-01-25 10:09 ` Marc Zyngier 2019-01-25 10:09 ` Marc Zyngier 2019-01-25 11:05 ` Julien Thierry 2019-01-25 11:05 ` Julien Thierry 2019-01-31 13:01 ` Marc Zyngier 2019-01-31 13:01 ` Marc Zyngier 2019-02-21 11:02 ` Julien Grall 2019-02-21 11:02 ` Julien Grall 2019-02-22 9:18 ` Marc Zyngier 2019-02-22 9:18 ` Marc Zyngier 2019-02-22 11:42 ` Julien Grall 2019-02-22 11:42 ` Julien Grall 2019-02-22 12:14 ` Marc Zyngier 2019-02-22 12:14 ` Marc Zyngier 2019-01-24 14:00 ` [PATCH 07/14] KVM: arm/arm64: Simplify bg_timer programming Christoffer Dall 2019-01-24 14:00 ` Christoffer Dall 2019-01-24 14:00 ` [PATCH 08/14] KVM: arm64: Fix ICH_ELRSR_EL2 sysreg naming Christoffer Dall 2019-01-24 14:00 ` Christoffer Dall 2019-01-24 14:00 ` [PATCH 09/14] KVM: arm64: Reuse sys_reg() macro when searching the trap table Christoffer Dall 2019-01-24 14:00 ` Christoffer Dall 2019-01-30 8:57 ` André Przywara 2019-01-30 8:57 ` André Przywara 2019-01-24 14:00 ` [PATCH 10/14] KVM: arm/arm64: consolidate arch timer trap handlers Christoffer Dall 2019-01-24 14:00 ` Christoffer Dall 2019-01-25 12:33 ` Julien Thierry 2019-01-25 12:33 ` Julien Thierry 2019-01-30 17:38 ` Marc Zyngier 2019-01-30 17:38 ` Marc Zyngier 2019-01-24 14:00 ` [PATCH 11/14] KVM: arm/arm64: timer: Rework data structures for multiple timers Christoffer Dall 2019-01-24 14:00 ` Christoffer Dall 2019-02-18 15:10 ` André Przywara [this message] 2019-02-18 15:10 ` André Przywara 2019-02-19 12:27 ` Christoffer Dall 2019-02-19 12:27 ` Christoffer Dall 2019-01-24 14:00 ` [PATCH 12/14] KVM: arm/arm64: arch_timer: Assign the phys timer on VHE systems Christoffer Dall 2019-01-24 14:00 ` Christoffer Dall 2019-02-18 15:10 ` André Przywara 2019-02-18 15:10 ` André Przywara 2019-02-19 12:43 ` Christoffer Dall 2019-02-19 12:43 ` Christoffer Dall 2019-02-20 17:58 ` Andre Przywara 2019-02-19 11:39 ` Alexandru Elisei 2019-02-19 11:39 ` Alexandru Elisei 2019-02-19 13:03 ` Christoffer Dall 2019-02-19 13:03 ` Christoffer Dall 2019-01-24 14:00 ` [PATCH 13/14] KVM: arm/arm64: Rework the timer code to use a timer_map Christoffer Dall 2019-01-24 14:00 ` Christoffer Dall 2019-01-24 14:00 ` [PATCH 14/14] KVM: arm/arm64: Move kvm_is_write_fault to header file Christoffer Dall 2019-01-24 14:00 ` Christoffer Dall
Reply instructions: You may reply publicly to this message via plain-text email using any one of the following methods: * Save the following mbox file, import it into your mail client, and reply-to-all from there: mbox Avoid top-posting and favor interleaved quoting: https://en.wikipedia.org/wiki/Posting_style#Interleaved_style * Reply using the --to, --cc, and --in-reply-to switches of git-send-email(1): git send-email \ --in-reply-to=d44bd102-5ad6-6e08-5e47-8a8386a139c6@arm.com \ --to=andre.przywara@arm.com \ --cc=christoffer.dall@arm.com \ --cc=kvm@vger.kernel.org \ --cc=kvmarm@lists.cs.columbia.edu \ --cc=linux-arm-kernel@lists.infradead.org \ --cc=marc.zyngier@arm.com \ /path/to/YOUR_REPLY https://kernel.org/pub/software/scm/git/docs/git-send-email.html * If your mail client supports setting the In-Reply-To header via mailto: links, try the mailto: linkBe sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes, see mirroring instructions on how to clone and mirror all data and code used by this external index.