* [PATCH v5 1/8] x86/cpu: Introduce vmce_save_vcpu_ctxt_one() func @ 2018-05-29 14:58 Alexandru Isaila 2018-05-29 14:58 ` [PATCH v5 2/8] x86/hvm: Introduce hvm_save_tsc_adjust_one() func Alexandru Isaila ` (6 more replies) 0 siblings, 7 replies; 12+ messages in thread From: Alexandru Isaila @ 2018-05-29 14:58 UTC (permalink / raw) To: xen-devel Cc: wei.liu2, andrew.cooper3, ian.jackson, paul.durrant, jbeulich, Alexandru Isaila This is used to save data from a single instance. Signed-off-by: Alexandru Isaila <aisaila@bitdefender.com> --- xen/arch/x86/cpu/mcheck/vmce.c | 16 ++++++++++------ 1 file changed, 10 insertions(+), 6 deletions(-) diff --git a/xen/arch/x86/cpu/mcheck/vmce.c b/xen/arch/x86/cpu/mcheck/vmce.c index e07cd2f..404f27e 100644 --- a/xen/arch/x86/cpu/mcheck/vmce.c +++ b/xen/arch/x86/cpu/mcheck/vmce.c @@ -349,6 +349,14 @@ int vmce_wrmsr(uint32_t msr, uint64_t val) return ret; } +static void vmce_save_vcpu_ctxt_one(struct vcpu *v, struct hvm_vmce_vcpu *ctxt) +{ + ctxt->caps = v->arch.vmce.mcg_cap; + ctxt->mci_ctl2_bank0 = v->arch.vmce.bank[0].mci_ctl2; + ctxt->mci_ctl2_bank1 = v->arch.vmce.bank[1].mci_ctl2; + ctxt->mcg_ext_ctl = v->arch.vmce.mcg_ext_ctl; +} + static int vmce_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h) { struct vcpu *v; @@ -356,13 +364,9 @@ static int vmce_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h) for_each_vcpu ( d, v ) { - struct hvm_vmce_vcpu ctxt = { - .caps = v->arch.vmce.mcg_cap, - .mci_ctl2_bank0 = v->arch.vmce.bank[0].mci_ctl2, - .mci_ctl2_bank1 = v->arch.vmce.bank[1].mci_ctl2, - .mcg_ext_ctl = v->arch.vmce.mcg_ext_ctl, - }; + struct hvm_vmce_vcpu ctxt; + vmce_save_vcpu_ctxt_one(v, &ctxt); err = hvm_save_entry(VMCE_VCPU, v->vcpu_id, h, &ctxt); if ( err ) break; -- 2.7.4 _______________________________________________ Xen-devel mailing list Xen-devel@lists.xenproject.org https://lists.xenproject.org/mailman/listinfo/xen-devel ^ permalink raw reply related [flat|nested] 12+ messages in thread
* [PATCH v5 2/8] x86/hvm: Introduce hvm_save_tsc_adjust_one() func 2018-05-29 14:58 [PATCH v5 1/8] x86/cpu: Introduce vmce_save_vcpu_ctxt_one() func Alexandru Isaila @ 2018-05-29 14:58 ` Alexandru Isaila 2018-05-29 14:58 ` [PATCH v5 3/8] x86/hvm: Introduce hvm_save_cpu_ctxt_one func Alexandru Isaila ` (5 subsequent siblings) 6 siblings, 0 replies; 12+ messages in thread From: Alexandru Isaila @ 2018-05-29 14:58 UTC (permalink / raw) To: xen-devel Cc: wei.liu2, andrew.cooper3, ian.jackson, paul.durrant, jbeulich, Alexandru Isaila This is used to save data from a single instance. Signed-off-by: Alexandru Isaila <aisaila@bitdefender.com> --- xen/arch/x86/hvm/hvm.c | 7 ++++++- 1 file changed, 6 insertions(+), 1 deletion(-) diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c index c23983c..76f7db9 100644 --- a/xen/arch/x86/hvm/hvm.c +++ b/xen/arch/x86/hvm/hvm.c @@ -740,6 +740,11 @@ void hvm_domain_destroy(struct domain *d) destroy_vpci_mmcfg(d); } +static void hvm_save_tsc_adjust_one(struct vcpu *v, struct hvm_tsc_adjust *ctxt) +{ + ctxt->tsc_adjust = v->arch.hvm_vcpu.msr_tsc_adjust; +} + static int hvm_save_tsc_adjust(struct domain *d, hvm_domain_context_t *h) { struct vcpu *v; @@ -748,7 +753,7 @@ static int hvm_save_tsc_adjust(struct domain *d, hvm_domain_context_t *h) for_each_vcpu ( d, v ) { - ctxt.tsc_adjust = v->arch.hvm_vcpu.msr_tsc_adjust; + hvm_save_tsc_adjust_one(v, &ctxt); err = hvm_save_entry(TSC_ADJUST, v->vcpu_id, h, &ctxt); if ( err ) break; -- 2.7.4 _______________________________________________ Xen-devel mailing list Xen-devel@lists.xenproject.org https://lists.xenproject.org/mailman/listinfo/xen-devel ^ permalink raw reply related [flat|nested] 12+ messages in thread
* [PATCH v5 3/8] x86/hvm: Introduce hvm_save_cpu_ctxt_one func 2018-05-29 14:58 [PATCH v5 1/8] x86/cpu: Introduce vmce_save_vcpu_ctxt_one() func Alexandru Isaila 2018-05-29 14:58 ` [PATCH v5 2/8] x86/hvm: Introduce hvm_save_tsc_adjust_one() func Alexandru Isaila @ 2018-05-29 14:58 ` Alexandru Isaila 2018-05-29 14:58 ` [PATCH v5 4/8] x86/hvm: Introduce hvm_save_cpu_xsave_states_one Alexandru Isaila ` (4 subsequent siblings) 6 siblings, 0 replies; 12+ messages in thread From: Alexandru Isaila @ 2018-05-29 14:58 UTC (permalink / raw) To: xen-devel Cc: wei.liu2, andrew.cooper3, ian.jackson, paul.durrant, jbeulich, Alexandru Isaila This is used to save data from a single instance. Signed-off-by: Alexandru Isaila <aisaila@bitdefender.com> --- xen/arch/x86/hvm/hvm.c | 195 +++++++++++++++++++++++++------------------------ 1 file changed, 100 insertions(+), 95 deletions(-) diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c index 76f7db9..b254378 100644 --- a/xen/arch/x86/hvm/hvm.c +++ b/xen/arch/x86/hvm/hvm.c @@ -785,11 +785,109 @@ static int hvm_load_tsc_adjust(struct domain *d, hvm_domain_context_t *h) HVM_REGISTER_SAVE_RESTORE(TSC_ADJUST, hvm_save_tsc_adjust, hvm_load_tsc_adjust, 1, HVMSR_PER_VCPU); +static void hvm_save_cpu_ctxt_one(struct vcpu *v, struct hvm_hw_cpu *ctxt) +{ + struct segment_register seg; + + /* Architecture-specific vmcs/vmcb bits */ + hvm_funcs.save_cpu_ctxt(v, ctxt); + + ctxt->tsc = hvm_get_guest_tsc_fixed(v, v->domain->arch.hvm_domain.sync_tsc); + + ctxt->msr_tsc_aux = hvm_msr_tsc_aux(v); + + hvm_get_segment_register(v, x86_seg_idtr, &seg); + ctxt->idtr_limit = seg.limit; + ctxt->idtr_base = seg.base; + + hvm_get_segment_register(v, x86_seg_gdtr, &seg); + ctxt->gdtr_limit = seg.limit; + ctxt->gdtr_base = seg.base; + + hvm_get_segment_register(v, x86_seg_cs, &seg); + ctxt->cs_sel = seg.sel; + ctxt->cs_limit = seg.limit; + ctxt->cs_base = seg.base; + ctxt->cs_arbytes = seg.attr; + + hvm_get_segment_register(v, x86_seg_ds, &seg); + ctxt->ds_sel = seg.sel; + ctxt->ds_limit = seg.limit; + ctxt->ds_base = seg.base; + ctxt->ds_arbytes = seg.attr; + + hvm_get_segment_register(v, x86_seg_es, &seg); + ctxt->es_sel = seg.sel; + ctxt->es_limit = seg.limit; + ctxt->es_base = seg.base; + ctxt->es_arbytes = seg.attr; + + hvm_get_segment_register(v, x86_seg_ss, &seg); + ctxt->ss_sel = seg.sel; + ctxt->ss_limit = seg.limit; + ctxt->ss_base = seg.base; + ctxt->ss_arbytes = seg.attr; + + hvm_get_segment_register(v, x86_seg_fs, &seg); + ctxt->fs_sel = seg.sel; + ctxt->fs_limit = seg.limit; + ctxt->fs_base = seg.base; + ctxt->fs_arbytes = seg.attr; + + hvm_get_segment_register(v, x86_seg_gs, &seg); + ctxt->gs_sel = seg.sel; + ctxt->gs_limit = seg.limit; + ctxt->gs_base = seg.base; + ctxt->gs_arbytes = seg.attr; + + hvm_get_segment_register(v, x86_seg_tr, &seg); + ctxt->tr_sel = seg.sel; + ctxt->tr_limit = seg.limit; + ctxt->tr_base = seg.base; + ctxt->tr_arbytes = seg.attr; + + hvm_get_segment_register(v, x86_seg_ldtr, &seg); + ctxt->ldtr_sel = seg.sel; + ctxt->ldtr_limit = seg.limit; + ctxt->ldtr_base = seg.base; + ctxt->ldtr_arbytes = seg.attr; + + if ( v->fpu_initialised ) + { + memcpy(ctxt->fpu_regs, v->arch.fpu_ctxt, sizeof(ctxt->fpu_regs)); + ctxt->flags = XEN_X86_FPU_INITIALISED; + } + + ctxt->rax = v->arch.user_regs.rax; + ctxt->rbx = v->arch.user_regs.rbx; + ctxt->rcx = v->arch.user_regs.rcx; + ctxt->rdx = v->arch.user_regs.rdx; + ctxt->rbp = v->arch.user_regs.rbp; + ctxt->rsi = v->arch.user_regs.rsi; + ctxt->rdi = v->arch.user_regs.rdi; + ctxt->rsp = v->arch.user_regs.rsp; + ctxt->rip = v->arch.user_regs.rip; + ctxt->rflags = v->arch.user_regs.rflags; + ctxt->r8 = v->arch.user_regs.r8; + ctxt->r9 = v->arch.user_regs.r9; + ctxt->r10 = v->arch.user_regs.r10; + ctxt->r11 = v->arch.user_regs.r11; + ctxt->r12 = v->arch.user_regs.r12; + ctxt->r13 = v->arch.user_regs.r13; + ctxt->r14 = v->arch.user_regs.r14; + ctxt->r15 = v->arch.user_regs.r15; + ctxt->dr0 = v->arch.debugreg[0]; + ctxt->dr1 = v->arch.debugreg[1]; + ctxt->dr2 = v->arch.debugreg[2]; + ctxt->dr3 = v->arch.debugreg[3]; + ctxt->dr6 = v->arch.debugreg[6]; + ctxt->dr7 = v->arch.debugreg[7]; +} + static int hvm_save_cpu_ctxt(struct domain *d, hvm_domain_context_t *h) { struct vcpu *v; struct hvm_hw_cpu ctxt; - struct segment_register seg; for_each_vcpu ( d, v ) { @@ -799,100 +897,7 @@ static int hvm_save_cpu_ctxt(struct domain *d, hvm_domain_context_t *h) continue; memset(&ctxt, 0, sizeof(ctxt)); - - /* Architecture-specific vmcs/vmcb bits */ - hvm_funcs.save_cpu_ctxt(v, &ctxt); - - ctxt.tsc = hvm_get_guest_tsc_fixed(v, d->arch.hvm_domain.sync_tsc); - - ctxt.msr_tsc_aux = hvm_msr_tsc_aux(v); - - hvm_get_segment_register(v, x86_seg_idtr, &seg); - ctxt.idtr_limit = seg.limit; - ctxt.idtr_base = seg.base; - - hvm_get_segment_register(v, x86_seg_gdtr, &seg); - ctxt.gdtr_limit = seg.limit; - ctxt.gdtr_base = seg.base; - - hvm_get_segment_register(v, x86_seg_cs, &seg); - ctxt.cs_sel = seg.sel; - ctxt.cs_limit = seg.limit; - ctxt.cs_base = seg.base; - ctxt.cs_arbytes = seg.attr; - - hvm_get_segment_register(v, x86_seg_ds, &seg); - ctxt.ds_sel = seg.sel; - ctxt.ds_limit = seg.limit; - ctxt.ds_base = seg.base; - ctxt.ds_arbytes = seg.attr; - - hvm_get_segment_register(v, x86_seg_es, &seg); - ctxt.es_sel = seg.sel; - ctxt.es_limit = seg.limit; - ctxt.es_base = seg.base; - ctxt.es_arbytes = seg.attr; - - hvm_get_segment_register(v, x86_seg_ss, &seg); - ctxt.ss_sel = seg.sel; - ctxt.ss_limit = seg.limit; - ctxt.ss_base = seg.base; - ctxt.ss_arbytes = seg.attr; - - hvm_get_segment_register(v, x86_seg_fs, &seg); - ctxt.fs_sel = seg.sel; - ctxt.fs_limit = seg.limit; - ctxt.fs_base = seg.base; - ctxt.fs_arbytes = seg.attr; - - hvm_get_segment_register(v, x86_seg_gs, &seg); - ctxt.gs_sel = seg.sel; - ctxt.gs_limit = seg.limit; - ctxt.gs_base = seg.base; - ctxt.gs_arbytes = seg.attr; - - hvm_get_segment_register(v, x86_seg_tr, &seg); - ctxt.tr_sel = seg.sel; - ctxt.tr_limit = seg.limit; - ctxt.tr_base = seg.base; - ctxt.tr_arbytes = seg.attr; - - hvm_get_segment_register(v, x86_seg_ldtr, &seg); - ctxt.ldtr_sel = seg.sel; - ctxt.ldtr_limit = seg.limit; - ctxt.ldtr_base = seg.base; - ctxt.ldtr_arbytes = seg.attr; - - if ( v->fpu_initialised ) - { - memcpy(ctxt.fpu_regs, v->arch.fpu_ctxt, sizeof(ctxt.fpu_regs)); - ctxt.flags = XEN_X86_FPU_INITIALISED; - } - - ctxt.rax = v->arch.user_regs.rax; - ctxt.rbx = v->arch.user_regs.rbx; - ctxt.rcx = v->arch.user_regs.rcx; - ctxt.rdx = v->arch.user_regs.rdx; - ctxt.rbp = v->arch.user_regs.rbp; - ctxt.rsi = v->arch.user_regs.rsi; - ctxt.rdi = v->arch.user_regs.rdi; - ctxt.rsp = v->arch.user_regs.rsp; - ctxt.rip = v->arch.user_regs.rip; - ctxt.rflags = v->arch.user_regs.rflags; - ctxt.r8 = v->arch.user_regs.r8; - ctxt.r9 = v->arch.user_regs.r9; - ctxt.r10 = v->arch.user_regs.r10; - ctxt.r11 = v->arch.user_regs.r11; - ctxt.r12 = v->arch.user_regs.r12; - ctxt.r13 = v->arch.user_regs.r13; - ctxt.r14 = v->arch.user_regs.r14; - ctxt.r15 = v->arch.user_regs.r15; - ctxt.dr0 = v->arch.debugreg[0]; - ctxt.dr1 = v->arch.debugreg[1]; - ctxt.dr2 = v->arch.debugreg[2]; - ctxt.dr3 = v->arch.debugreg[3]; - ctxt.dr6 = v->arch.debugreg[6]; - ctxt.dr7 = v->arch.debugreg[7]; + hvm_save_cpu_ctxt_one(v, &ctxt); if ( hvm_save_entry(CPU, v->vcpu_id, h, &ctxt) != 0 ) return 1; -- 2.7.4 _______________________________________________ Xen-devel mailing list Xen-devel@lists.xenproject.org https://lists.xenproject.org/mailman/listinfo/xen-devel ^ permalink raw reply related [flat|nested] 12+ messages in thread
* [PATCH v5 4/8] x86/hvm: Introduce hvm_save_cpu_xsave_states_one 2018-05-29 14:58 [PATCH v5 1/8] x86/cpu: Introduce vmce_save_vcpu_ctxt_one() func Alexandru Isaila 2018-05-29 14:58 ` [PATCH v5 2/8] x86/hvm: Introduce hvm_save_tsc_adjust_one() func Alexandru Isaila 2018-05-29 14:58 ` [PATCH v5 3/8] x86/hvm: Introduce hvm_save_cpu_ctxt_one func Alexandru Isaila @ 2018-05-29 14:58 ` Alexandru Isaila 2018-05-29 14:58 ` [PATCH v5 5/8] x86/hvm: Introduce hvm_save_cpu_msrs_one func Alexandru Isaila ` (3 subsequent siblings) 6 siblings, 0 replies; 12+ messages in thread From: Alexandru Isaila @ 2018-05-29 14:58 UTC (permalink / raw) To: xen-devel Cc: wei.liu2, andrew.cooper3, ian.jackson, paul.durrant, jbeulich, Alexandru Isaila This is used to save data from a single instance. Signed-off-by: Alexandru Isaila <aisaila@bitdefender.com> --- xen/arch/x86/hvm/hvm.c | 11 ++++++++--- 1 file changed, 8 insertions(+), 3 deletions(-) diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c index b254378..e8ecabf 100644 --- a/xen/arch/x86/hvm/hvm.c +++ b/xen/arch/x86/hvm/hvm.c @@ -1183,6 +1183,13 @@ HVM_REGISTER_SAVE_RESTORE(CPU, hvm_save_cpu_ctxt, hvm_load_cpu_ctxt, save_area) + \ xstate_ctxt_size(xcr0)) +static void hvm_save_cpu_xsave_states_one(struct vcpu *v, struct hvm_hw_cpu_xsave *ctxt) +{ + ctxt->xfeature_mask = xfeature_mask; + ctxt->xcr0 = v->arch.xcr0; + ctxt->xcr0_accum = v->arch.xcr0_accum; +} + static int hvm_save_cpu_xsave_states(struct domain *d, hvm_domain_context_t *h) { struct vcpu *v; @@ -1202,9 +1209,7 @@ static int hvm_save_cpu_xsave_states(struct domain *d, hvm_domain_context_t *h) ctxt = (struct hvm_hw_cpu_xsave *)&h->data[h->cur]; h->cur += size; - ctxt->xfeature_mask = xfeature_mask; - ctxt->xcr0 = v->arch.xcr0; - ctxt->xcr0_accum = v->arch.xcr0_accum; + hvm_save_cpu_xsave_states_one(v, ctxt); expand_xsave_states(v, &ctxt->save_area, size - offsetof(typeof(*ctxt), save_area)); } -- 2.7.4 _______________________________________________ Xen-devel mailing list Xen-devel@lists.xenproject.org https://lists.xenproject.org/mailman/listinfo/xen-devel ^ permalink raw reply related [flat|nested] 12+ messages in thread
* [PATCH v5 5/8] x86/hvm: Introduce hvm_save_cpu_msrs_one func 2018-05-29 14:58 [PATCH v5 1/8] x86/cpu: Introduce vmce_save_vcpu_ctxt_one() func Alexandru Isaila ` (2 preceding siblings ...) 2018-05-29 14:58 ` [PATCH v5 4/8] x86/hvm: Introduce hvm_save_cpu_xsave_states_one Alexandru Isaila @ 2018-05-29 14:58 ` Alexandru Isaila 2018-06-04 9:08 ` Paul Durrant 2018-05-29 14:58 ` [PATCH v5 6/8] x86/hvm: Introduce hvm_save_mtrr_msr_one func Alexandru Isaila ` (2 subsequent siblings) 6 siblings, 1 reply; 12+ messages in thread From: Alexandru Isaila @ 2018-05-29 14:58 UTC (permalink / raw) To: xen-devel Cc: wei.liu2, andrew.cooper3, ian.jackson, paul.durrant, jbeulich, Alexandru Isaila This is used to save data from a single instance. Signed-off-by: Alexandru Isaila <aisaila@bitdefender.com> --- xen/arch/x86/hvm/hvm.c | 59 ++++++++++++++++++++++++++++---------------------- 1 file changed, 33 insertions(+), 26 deletions(-) diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c index e8ecabf..4a22283 100644 --- a/xen/arch/x86/hvm/hvm.c +++ b/xen/arch/x86/hvm/hvm.c @@ -1354,6 +1354,38 @@ static const uint32_t msrs_to_send[] = { }; static unsigned int __read_mostly msr_count_max = ARRAY_SIZE(msrs_to_send); +static int hvm_save_cpu_msrs_one(struct vcpu *v, struct hvm_msr *ctxt) +{ + unsigned int i; + + for ( i = 0; i < ARRAY_SIZE(msrs_to_send); ++i ) + { + uint64_t val; + int rc = guest_rdmsr(v, msrs_to_send[i], &val); + + /* + * It is the programmers responsibility to ensure that + * msrs_to_send[] contain generally-read/write MSRs. + * X86EMUL_EXCEPTION here implies a missing feature, and that the + * guest doesn't have access to the MSR. + */ + if ( rc == X86EMUL_EXCEPTION ) + continue; + + if ( rc != X86EMUL_OKAY ) + { + ASSERT_UNREACHABLE(); + return -ENXIO; + } + + if ( !val ) + continue; /* Skip empty MSRs. */ + ctxt->msr[ctxt->count].index = msrs_to_send[i]; + ctxt->msr[ctxt->count++].val = val; + } + return 0; +} + static int hvm_save_cpu_msrs(struct domain *d, hvm_domain_context_t *h) { struct vcpu *v; @@ -1370,32 +1402,7 @@ static int hvm_save_cpu_msrs(struct domain *d, hvm_domain_context_t *h) ctxt = (struct hvm_msr *)&h->data[h->cur]; ctxt->count = 0; - for ( i = 0; i < ARRAY_SIZE(msrs_to_send); ++i ) - { - uint64_t val; - int rc = guest_rdmsr(v, msrs_to_send[i], &val); - - /* - * It is the programmers responsibility to ensure that - * msrs_to_send[] contain generally-read/write MSRs. - * X86EMUL_EXCEPTION here implies a missing feature, and that the - * guest doesn't have access to the MSR. - */ - if ( rc == X86EMUL_EXCEPTION ) - continue; - - if ( rc != X86EMUL_OKAY ) - { - ASSERT_UNREACHABLE(); - return -ENXIO; - } - - if ( !val ) - continue; /* Skip empty MSRs. */ - - ctxt->msr[ctxt->count].index = msrs_to_send[i]; - ctxt->msr[ctxt->count++].val = val; - } + hvm_save_cpu_msrs_one(v, ctxt); if ( hvm_funcs.save_msr ) hvm_funcs.save_msr(v, ctxt); -- 2.7.4 _______________________________________________ Xen-devel mailing list Xen-devel@lists.xenproject.org https://lists.xenproject.org/mailman/listinfo/xen-devel ^ permalink raw reply related [flat|nested] 12+ messages in thread
* Re: [PATCH v5 5/8] x86/hvm: Introduce hvm_save_cpu_msrs_one func 2018-05-29 14:58 ` [PATCH v5 5/8] x86/hvm: Introduce hvm_save_cpu_msrs_one func Alexandru Isaila @ 2018-06-04 9:08 ` Paul Durrant 0 siblings, 0 replies; 12+ messages in thread From: Paul Durrant @ 2018-06-04 9:08 UTC (permalink / raw) To: 'Alexandru Isaila', xen-devel Cc: Ian Jackson, Wei Liu, jbeulich, Andrew Cooper > -----Original Message----- > From: Alexandru Isaila [mailto:aisaila@bitdefender.com] > Sent: 29 May 2018 15:59 > To: xen-devel@lists.xen.org > Cc: Ian Jackson <Ian.Jackson@citrix.com>; Wei Liu <wei.liu2@citrix.com>; > jbeulich@suse.com; Andrew Cooper <Andrew.Cooper3@citrix.com>; Paul > Durrant <Paul.Durrant@citrix.com>; Alexandru Isaila > <aisaila@bitdefender.com> > Subject: [PATCH v5 5/8] x86/hvm: Introduce hvm_save_cpu_msrs_one func > > This is used to save data from a single instance. > > Signed-off-by: Alexandru Isaila <aisaila@bitdefender.com> > --- > xen/arch/x86/hvm/hvm.c | 59 ++++++++++++++++++++++++++++----------- > ----------- > 1 file changed, 33 insertions(+), 26 deletions(-) > > diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c > index e8ecabf..4a22283 100644 > --- a/xen/arch/x86/hvm/hvm.c > +++ b/xen/arch/x86/hvm/hvm.c > @@ -1354,6 +1354,38 @@ static const uint32_t msrs_to_send[] = { > }; > static unsigned int __read_mostly msr_count_max = > ARRAY_SIZE(msrs_to_send); > > +static int hvm_save_cpu_msrs_one(struct vcpu *v, struct hvm_msr *ctxt) > +{ > + unsigned int i; > + > + for ( i = 0; i < ARRAY_SIZE(msrs_to_send); ++i ) > + { > + uint64_t val; > + int rc = guest_rdmsr(v, msrs_to_send[i], &val); > + > + /* > + * It is the programmers responsibility to ensure that > + * msrs_to_send[] contain generally-read/write MSRs. > + * X86EMUL_EXCEPTION here implies a missing feature, and that the > + * guest doesn't have access to the MSR. > + */ > + if ( rc == X86EMUL_EXCEPTION ) > + continue; > + > + if ( rc != X86EMUL_OKAY ) > + { > + ASSERT_UNREACHABLE(); > + return -ENXIO; > + } > + > + if ( !val ) > + continue; /* Skip empty MSRs. */ > + ctxt->msr[ctxt->count].index = msrs_to_send[i]; > + ctxt->msr[ctxt->count++].val = val; > + } > + return 0; > +} > + > static int hvm_save_cpu_msrs(struct domain *d, hvm_domain_context_t > *h) > { > struct vcpu *v; > @@ -1370,32 +1402,7 @@ static int hvm_save_cpu_msrs(struct domain *d, > hvm_domain_context_t *h) > ctxt = (struct hvm_msr *)&h->data[h->cur]; > ctxt->count = 0; > > - for ( i = 0; i < ARRAY_SIZE(msrs_to_send); ++i ) > - { > - uint64_t val; > - int rc = guest_rdmsr(v, msrs_to_send[i], &val); > - > - /* > - * It is the programmers responsibility to ensure that > - * msrs_to_send[] contain generally-read/write MSRs. > - * X86EMUL_EXCEPTION here implies a missing feature, and that the > - * guest doesn't have access to the MSR. > - */ > - if ( rc == X86EMUL_EXCEPTION ) > - continue; > - > - if ( rc != X86EMUL_OKAY ) > - { > - ASSERT_UNREACHABLE(); > - return -ENXIO; > - } > - > - if ( !val ) > - continue; /* Skip empty MSRs. */ > - > - ctxt->msr[ctxt->count].index = msrs_to_send[i]; > - ctxt->msr[ctxt->count++].val = val; > - } > + hvm_save_cpu_msrs_one(v, ctxt); No check for the return value here means you're not going to propagate a potential -ENXIO. Paul > > if ( hvm_funcs.save_msr ) > hvm_funcs.save_msr(v, ctxt); > -- > 2.7.4 _______________________________________________ Xen-devel mailing list Xen-devel@lists.xenproject.org https://lists.xenproject.org/mailman/listinfo/xen-devel ^ permalink raw reply [flat|nested] 12+ messages in thread
* [PATCH v5 6/8] x86/hvm: Introduce hvm_save_mtrr_msr_one func 2018-05-29 14:58 [PATCH v5 1/8] x86/cpu: Introduce vmce_save_vcpu_ctxt_one() func Alexandru Isaila ` (3 preceding siblings ...) 2018-05-29 14:58 ` [PATCH v5 5/8] x86/hvm: Introduce hvm_save_cpu_msrs_one func Alexandru Isaila @ 2018-05-29 14:58 ` Alexandru Isaila 2018-05-29 14:58 ` [PATCH v5 7/8] x86/hvm: Introduce viridian_save_vcpu_ctxt_one() func Alexandru Isaila 2018-05-29 14:58 ` [PATCH v5 8/8] x86/domctl: Don't pause the whole domain if only getting vcpu state Alexandru Isaila 6 siblings, 0 replies; 12+ messages in thread From: Alexandru Isaila @ 2018-05-29 14:58 UTC (permalink / raw) To: xen-devel Cc: wei.liu2, andrew.cooper3, ian.jackson, paul.durrant, jbeulich, Alexandru Isaila This is used to save data from a single instance. Signed-off-by: Alexandru Isaila <aisaila@bitdefender.com> --- xen/arch/x86/hvm/mtrr.c | 52 +++++++++++++++++++++++++++---------------------- 1 file changed, 29 insertions(+), 23 deletions(-) diff --git a/xen/arch/x86/hvm/mtrr.c b/xen/arch/x86/hvm/mtrr.c index b721c63..d311031 100644 --- a/xen/arch/x86/hvm/mtrr.c +++ b/xen/arch/x86/hvm/mtrr.c @@ -666,36 +666,42 @@ int hvm_set_mem_pinned_cacheattr(struct domain *d, uint64_t gfn_start, return 0; } -static int hvm_save_mtrr_msr(struct domain *d, hvm_domain_context_t *h) +static void hvm_save_mtrr_msr_one(struct vcpu *v, struct hvm_hw_mtrr *hw_mtrr) { + struct mtrr_state *mtrr_state = &v->arch.hvm_vcpu.mtrr; int i; - struct vcpu *v; - struct hvm_hw_mtrr hw_mtrr; - struct mtrr_state *mtrr_state; - /* save mtrr&pat */ - for_each_vcpu(d, v) + + hvm_get_guest_pat(v, &hw_mtrr->msr_pat_cr); + + hw_mtrr->msr_mtrr_def_type = mtrr_state->def_type + | (mtrr_state->enabled << 10); + hw_mtrr->msr_mtrr_cap = mtrr_state->mtrr_cap; + + for ( i = 0; i < MTRR_VCNT; i++ ) { - mtrr_state = &v->arch.hvm_vcpu.mtrr; + /* save physbase */ + hw_mtrr->msr_mtrr_var[i*2] = + ((uint64_t*)mtrr_state->var_ranges)[i*2]; + /* save physmask */ + hw_mtrr->msr_mtrr_var[i*2+1] = + ((uint64_t*)mtrr_state->var_ranges)[i*2+1]; + } - hvm_get_guest_pat(v, &hw_mtrr.msr_pat_cr); + for ( i = 0; i < NUM_FIXED_MSR; i++ ) + hw_mtrr->msr_mtrr_fixed[i] = + ((uint64_t*)mtrr_state->fixed_ranges)[i]; - hw_mtrr.msr_mtrr_def_type = mtrr_state->def_type - | (mtrr_state->enabled << 10); - hw_mtrr.msr_mtrr_cap = mtrr_state->mtrr_cap; +} - for ( i = 0; i < MTRR_VCNT; i++ ) - { - /* save physbase */ - hw_mtrr.msr_mtrr_var[i*2] = - ((uint64_t*)mtrr_state->var_ranges)[i*2]; - /* save physmask */ - hw_mtrr.msr_mtrr_var[i*2+1] = - ((uint64_t*)mtrr_state->var_ranges)[i*2+1]; - } +static int hvm_save_mtrr_msr(struct domain *d, hvm_domain_context_t *h) +{ + struct vcpu *v; + struct hvm_hw_mtrr hw_mtrr; + /* save mtrr&pat */ - for ( i = 0; i < NUM_FIXED_MSR; i++ ) - hw_mtrr.msr_mtrr_fixed[i] = - ((uint64_t*)mtrr_state->fixed_ranges)[i]; + for_each_vcpu(d, v) + { + hvm_save_mtrr_msr_one(v, &hw_mtrr); if ( hvm_save_entry(MTRR, v->vcpu_id, h, &hw_mtrr) != 0 ) return 1; -- 2.7.4 _______________________________________________ Xen-devel mailing list Xen-devel@lists.xenproject.org https://lists.xenproject.org/mailman/listinfo/xen-devel ^ permalink raw reply related [flat|nested] 12+ messages in thread
* [PATCH v5 7/8] x86/hvm: Introduce viridian_save_vcpu_ctxt_one() func 2018-05-29 14:58 [PATCH v5 1/8] x86/cpu: Introduce vmce_save_vcpu_ctxt_one() func Alexandru Isaila ` (4 preceding siblings ...) 2018-05-29 14:58 ` [PATCH v5 6/8] x86/hvm: Introduce hvm_save_mtrr_msr_one func Alexandru Isaila @ 2018-05-29 14:58 ` Alexandru Isaila 2018-06-04 9:11 ` Paul Durrant 2018-05-29 14:58 ` [PATCH v5 8/8] x86/domctl: Don't pause the whole domain if only getting vcpu state Alexandru Isaila 6 siblings, 1 reply; 12+ messages in thread From: Alexandru Isaila @ 2018-05-29 14:58 UTC (permalink / raw) To: xen-devel Cc: wei.liu2, andrew.cooper3, ian.jackson, paul.durrant, jbeulich, Alexandru Isaila This is used to save data from a single instance. Signed-off-by: Alexandru Isaila <aisaila@bitdefender.com> --- xen/arch/x86/hvm/viridian.c | 14 ++++++++++---- 1 file changed, 10 insertions(+), 4 deletions(-) diff --git a/xen/arch/x86/hvm/viridian.c b/xen/arch/x86/hvm/viridian.c index 694eae6..d22c8ac 100644 --- a/xen/arch/x86/hvm/viridian.c +++ b/xen/arch/x86/hvm/viridian.c @@ -1026,6 +1026,13 @@ static int viridian_load_domain_ctxt(struct domain *d, hvm_domain_context_t *h) HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_DOMAIN, viridian_save_domain_ctxt, viridian_load_domain_ctxt, 1, HVMSR_PER_DOM); +static void viridian_save_vcpu_ctxt_one(struct vcpu *v, struct hvm_viridian_vcpu_context *ctxt) +{ + ctxt->vp_assist_msr = v->arch.hvm_vcpu.viridian.vp_assist.msr.raw; + ctxt->vp_assist_pending = v->arch.hvm_vcpu.viridian.vp_assist.pending; + *ctxt->_pad = 0; +} + static int viridian_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h) { struct vcpu *v; @@ -1034,10 +1041,9 @@ static int viridian_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h) return 0; for_each_vcpu( d, v ) { - struct hvm_viridian_vcpu_context ctxt = { - .vp_assist_msr = v->arch.hvm_vcpu.viridian.vp_assist.msr.raw, - .vp_assist_pending = v->arch.hvm_vcpu.viridian.vp_assist.pending, - }; + struct hvm_viridian_vcpu_context ctxt; + + viridian_save_vcpu_ctxt_one(v, &ctxt); if ( hvm_save_entry(VIRIDIAN_VCPU, v->vcpu_id, h, &ctxt) != 0 ) return 1; -- 2.7.4 _______________________________________________ Xen-devel mailing list Xen-devel@lists.xenproject.org https://lists.xenproject.org/mailman/listinfo/xen-devel ^ permalink raw reply related [flat|nested] 12+ messages in thread
* Re: [PATCH v5 7/8] x86/hvm: Introduce viridian_save_vcpu_ctxt_one() func 2018-05-29 14:58 ` [PATCH v5 7/8] x86/hvm: Introduce viridian_save_vcpu_ctxt_one() func Alexandru Isaila @ 2018-06-04 9:11 ` Paul Durrant 0 siblings, 0 replies; 12+ messages in thread From: Paul Durrant @ 2018-06-04 9:11 UTC (permalink / raw) To: 'Alexandru Isaila', xen-devel Cc: Ian Jackson, Wei Liu, jbeulich, Andrew Cooper > -----Original Message----- > From: Alexandru Isaila [mailto:aisaila@bitdefender.com] > Sent: 29 May 2018 15:59 > To: xen-devel@lists.xen.org > Cc: Ian Jackson <Ian.Jackson@citrix.com>; Wei Liu <wei.liu2@citrix.com>; > jbeulich@suse.com; Andrew Cooper <Andrew.Cooper3@citrix.com>; Paul > Durrant <Paul.Durrant@citrix.com>; Alexandru Isaila > <aisaila@bitdefender.com> > Subject: [PATCH v5 7/8] x86/hvm: Introduce viridian_save_vcpu_ctxt_one() > func > > This is used to save data from a single instance. > > Signed-off-by: Alexandru Isaila <aisaila@bitdefender.com> > --- > xen/arch/x86/hvm/viridian.c | 14 ++++++++++---- > 1 file changed, 10 insertions(+), 4 deletions(-) > > diff --git a/xen/arch/x86/hvm/viridian.c b/xen/arch/x86/hvm/viridian.c > index 694eae6..d22c8ac 100644 > --- a/xen/arch/x86/hvm/viridian.c > +++ b/xen/arch/x86/hvm/viridian.c > @@ -1026,6 +1026,13 @@ static int viridian_load_domain_ctxt(struct domain > *d, hvm_domain_context_t *h) > HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_DOMAIN, > viridian_save_domain_ctxt, > viridian_load_domain_ctxt, 1, HVMSR_PER_DOM); > > +static void viridian_save_vcpu_ctxt_one(struct vcpu *v, struct > hvm_viridian_vcpu_context *ctxt) > +{ > + ctxt->vp_assist_msr = v->arch.hvm_vcpu.viridian.vp_assist.msr.raw; > + ctxt->vp_assist_pending = v->arch.hvm_vcpu.viridian.vp_assist.pending; > + *ctxt->_pad = 0; I'd prefer a memset() to zero the context prior to setting fields. It's more robust if fields (and more padding) are added in future. Paul > +} > + > static int viridian_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t > *h) > { > struct vcpu *v; > @@ -1034,10 +1041,9 @@ static int viridian_save_vcpu_ctxt(struct domain > *d, hvm_domain_context_t *h) > return 0; > > for_each_vcpu( d, v ) { > - struct hvm_viridian_vcpu_context ctxt = { > - .vp_assist_msr = v->arch.hvm_vcpu.viridian.vp_assist.msr.raw, > - .vp_assist_pending = v->arch.hvm_vcpu.viridian.vp_assist.pending, > - }; > + struct hvm_viridian_vcpu_context ctxt; > + > + viridian_save_vcpu_ctxt_one(v, &ctxt); > > if ( hvm_save_entry(VIRIDIAN_VCPU, v->vcpu_id, h, &ctxt) != 0 ) > return 1; > -- > 2.7.4 _______________________________________________ Xen-devel mailing list Xen-devel@lists.xenproject.org https://lists.xenproject.org/mailman/listinfo/xen-devel ^ permalink raw reply [flat|nested] 12+ messages in thread
* [PATCH v5 8/8] x86/domctl: Don't pause the whole domain if only getting vcpu state 2018-05-29 14:58 [PATCH v5 1/8] x86/cpu: Introduce vmce_save_vcpu_ctxt_one() func Alexandru Isaila ` (5 preceding siblings ...) 2018-05-29 14:58 ` [PATCH v5 7/8] x86/hvm: Introduce viridian_save_vcpu_ctxt_one() func Alexandru Isaila @ 2018-05-29 14:58 ` Alexandru Isaila 2018-06-04 9:21 ` Paul Durrant 2018-06-06 12:35 ` Jan Beulich 6 siblings, 2 replies; 12+ messages in thread From: Alexandru Isaila @ 2018-05-29 14:58 UTC (permalink / raw) To: xen-devel Cc: wei.liu2, andrew.cooper3, ian.jackson, paul.durrant, jbeulich, Alexandru Isaila This patch is focused on merging the *save() funcs to the *save_one() funcs to remove redundancy. Also the for loop is moved to the caller so now we can save info for a single vcpu instance. Signed-off-by: Alexandru Isaila <aisaila@bitdefender.com> --- Changes since V4: - Save funcs are deleted - vcpu si added to the param list - save_one funcs are now static - Add blank line after "/* save mtrr&pat */" comment --- xen/arch/x86/cpu/mcheck/vmce.c | 28 ++-- xen/arch/x86/hvm/hpet.c | 3 +- xen/arch/x86/hvm/hvm.c | 287 +++++++++++++++++------------------------ xen/arch/x86/hvm/i8254.c | 3 +- xen/arch/x86/hvm/irq.c | 9 +- xen/arch/x86/hvm/mtrr.c | 36 ++---- xen/arch/x86/hvm/pmtimer.c | 3 +- xen/arch/x86/hvm/rtc.c | 3 +- xen/arch/x86/hvm/save.c | 125 +++++++++++++----- xen/arch/x86/hvm/vioapic.c | 3 +- xen/arch/x86/hvm/viridian.c | 30 ++--- xen/arch/x86/hvm/vlapic.c | 28 ++-- xen/arch/x86/hvm/vpic.c | 3 +- xen/include/asm-x86/hvm/save.h | 2 +- 14 files changed, 271 insertions(+), 292 deletions(-) diff --git a/xen/arch/x86/cpu/mcheck/vmce.c b/xen/arch/x86/cpu/mcheck/vmce.c index 404f27e..7bde9ff 100644 --- a/xen/arch/x86/cpu/mcheck/vmce.c +++ b/xen/arch/x86/cpu/mcheck/vmce.c @@ -349,29 +349,17 @@ int vmce_wrmsr(uint32_t msr, uint64_t val) return ret; } -static void vmce_save_vcpu_ctxt_one(struct vcpu *v, struct hvm_vmce_vcpu *ctxt) +static int vmce_save_vcpu_ctxt_one(struct vcpu *v, hvm_domain_context_t *h) { - ctxt->caps = v->arch.vmce.mcg_cap; - ctxt->mci_ctl2_bank0 = v->arch.vmce.bank[0].mci_ctl2; - ctxt->mci_ctl2_bank1 = v->arch.vmce.bank[1].mci_ctl2; - ctxt->mcg_ext_ctl = v->arch.vmce.mcg_ext_ctl; -} - -static int vmce_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h) -{ - struct vcpu *v; int err = 0; + struct hvm_vmce_vcpu ctxt; - for_each_vcpu ( d, v ) - { - struct hvm_vmce_vcpu ctxt; - - vmce_save_vcpu_ctxt_one(v, &ctxt); - err = hvm_save_entry(VMCE_VCPU, v->vcpu_id, h, &ctxt); - if ( err ) - break; - } + ctxt.caps = v->arch.vmce.mcg_cap; + ctxt.mci_ctl2_bank0 = v->arch.vmce.bank[0].mci_ctl2; + ctxt.mci_ctl2_bank1 = v->arch.vmce.bank[1].mci_ctl2; + ctxt.mcg_ext_ctl = v->arch.vmce.mcg_ext_ctl; + err = hvm_save_entry(VMCE_VCPU, v->vcpu_id, h, &ctxt); return err; } @@ -394,7 +382,7 @@ static int vmce_load_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h) return err ?: vmce_restore_vcpu(v, &ctxt); } -HVM_REGISTER_SAVE_RESTORE(VMCE_VCPU, vmce_save_vcpu_ctxt, +HVM_REGISTER_SAVE_RESTORE(VMCE_VCPU, vmce_save_vcpu_ctxt_one, vmce_load_vcpu_ctxt, 1, HVMSR_PER_VCPU); /* diff --git a/xen/arch/x86/hvm/hpet.c b/xen/arch/x86/hvm/hpet.c index 2837709..3ed6547 100644 --- a/xen/arch/x86/hvm/hpet.c +++ b/xen/arch/x86/hvm/hpet.c @@ -516,8 +516,9 @@ static const struct hvm_mmio_ops hpet_mmio_ops = { }; -static int hpet_save(struct domain *d, hvm_domain_context_t *h) +static int hpet_save(struct vcpu *vcpu, hvm_domain_context_t *h) { + struct domain *d = vcpu->domain; HPETState *hp = domain_vhpet(d); struct vcpu *v = pt_global_vcpu_target(d); int rc; diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c index 4a22283..ea27055 100644 --- a/xen/arch/x86/hvm/hvm.c +++ b/xen/arch/x86/hvm/hvm.c @@ -740,24 +740,13 @@ void hvm_domain_destroy(struct domain *d) destroy_vpci_mmcfg(d); } -static void hvm_save_tsc_adjust_one(struct vcpu *v, struct hvm_tsc_adjust *ctxt) +static int hvm_save_tsc_adjust_one(struct vcpu *v, hvm_domain_context_t *h) { - ctxt->tsc_adjust = v->arch.hvm_vcpu.msr_tsc_adjust; -} - -static int hvm_save_tsc_adjust(struct domain *d, hvm_domain_context_t *h) -{ - struct vcpu *v; - struct hvm_tsc_adjust ctxt; int err = 0; + struct hvm_tsc_adjust ctxt; - for_each_vcpu ( d, v ) - { - hvm_save_tsc_adjust_one(v, &ctxt); - err = hvm_save_entry(TSC_ADJUST, v->vcpu_id, h, &ctxt); - if ( err ) - break; - } + ctxt.tsc_adjust = v->arch.hvm_vcpu.msr_tsc_adjust; + err = hvm_save_entry(TSC_ADJUST, v->vcpu_id, h, &ctxt); return err; } @@ -782,126 +771,116 @@ static int hvm_load_tsc_adjust(struct domain *d, hvm_domain_context_t *h) return 0; } -HVM_REGISTER_SAVE_RESTORE(TSC_ADJUST, hvm_save_tsc_adjust, +HVM_REGISTER_SAVE_RESTORE(TSC_ADJUST, hvm_save_tsc_adjust_one, hvm_load_tsc_adjust, 1, HVMSR_PER_VCPU); -static void hvm_save_cpu_ctxt_one(struct vcpu *v, struct hvm_hw_cpu *ctxt) +static int hvm_save_cpu_ctxt_one(struct vcpu *v, hvm_domain_context_t *h) { struct segment_register seg; + struct hvm_hw_cpu ctxt = {}; + + /* We don't need to save state for a vcpu that is down; the restore + * code will leave it down if there is nothing saved. */ + if ( v->pause_flags & VPF_down ) + return 2; /* Architecture-specific vmcs/vmcb bits */ - hvm_funcs.save_cpu_ctxt(v, ctxt); + hvm_funcs.save_cpu_ctxt(v, &ctxt); - ctxt->tsc = hvm_get_guest_tsc_fixed(v, v->domain->arch.hvm_domain.sync_tsc); + ctxt.tsc = hvm_get_guest_tsc_fixed(v, v->domain->arch.hvm_domain.sync_tsc); - ctxt->msr_tsc_aux = hvm_msr_tsc_aux(v); + ctxt.msr_tsc_aux = hvm_msr_tsc_aux(v); hvm_get_segment_register(v, x86_seg_idtr, &seg); - ctxt->idtr_limit = seg.limit; - ctxt->idtr_base = seg.base; + ctxt.idtr_limit = seg.limit; + ctxt.idtr_base = seg.base; hvm_get_segment_register(v, x86_seg_gdtr, &seg); - ctxt->gdtr_limit = seg.limit; - ctxt->gdtr_base = seg.base; + ctxt.gdtr_limit = seg.limit; + ctxt.gdtr_base = seg.base; hvm_get_segment_register(v, x86_seg_cs, &seg); - ctxt->cs_sel = seg.sel; - ctxt->cs_limit = seg.limit; - ctxt->cs_base = seg.base; - ctxt->cs_arbytes = seg.attr; + ctxt.cs_sel = seg.sel; + ctxt.cs_limit = seg.limit; + ctxt.cs_base = seg.base; + ctxt.cs_arbytes = seg.attr; hvm_get_segment_register(v, x86_seg_ds, &seg); - ctxt->ds_sel = seg.sel; - ctxt->ds_limit = seg.limit; - ctxt->ds_base = seg.base; - ctxt->ds_arbytes = seg.attr; + ctxt.ds_sel = seg.sel; + ctxt.ds_limit = seg.limit; + ctxt.ds_base = seg.base; + ctxt.ds_arbytes = seg.attr; hvm_get_segment_register(v, x86_seg_es, &seg); - ctxt->es_sel = seg.sel; - ctxt->es_limit = seg.limit; - ctxt->es_base = seg.base; - ctxt->es_arbytes = seg.attr; + ctxt.es_sel = seg.sel; + ctxt.es_limit = seg.limit; + ctxt.es_base = seg.base; + ctxt.es_arbytes = seg.attr; hvm_get_segment_register(v, x86_seg_ss, &seg); - ctxt->ss_sel = seg.sel; - ctxt->ss_limit = seg.limit; - ctxt->ss_base = seg.base; - ctxt->ss_arbytes = seg.attr; + ctxt.ss_sel = seg.sel; + ctxt.ss_limit = seg.limit; + ctxt.ss_base = seg.base; + ctxt.ss_arbytes = seg.attr; hvm_get_segment_register(v, x86_seg_fs, &seg); - ctxt->fs_sel = seg.sel; - ctxt->fs_limit = seg.limit; - ctxt->fs_base = seg.base; - ctxt->fs_arbytes = seg.attr; + ctxt.fs_sel = seg.sel; + ctxt.fs_limit = seg.limit; + ctxt.fs_base = seg.base; + ctxt.fs_arbytes = seg.attr; hvm_get_segment_register(v, x86_seg_gs, &seg); - ctxt->gs_sel = seg.sel; - ctxt->gs_limit = seg.limit; - ctxt->gs_base = seg.base; - ctxt->gs_arbytes = seg.attr; + ctxt.gs_sel = seg.sel; + ctxt.gs_limit = seg.limit; + ctxt.gs_base = seg.base; + ctxt.gs_arbytes = seg.attr; hvm_get_segment_register(v, x86_seg_tr, &seg); - ctxt->tr_sel = seg.sel; - ctxt->tr_limit = seg.limit; - ctxt->tr_base = seg.base; - ctxt->tr_arbytes = seg.attr; + ctxt.tr_sel = seg.sel; + ctxt.tr_limit = seg.limit; + ctxt.tr_base = seg.base; + ctxt.tr_arbytes = seg.attr; hvm_get_segment_register(v, x86_seg_ldtr, &seg); - ctxt->ldtr_sel = seg.sel; - ctxt->ldtr_limit = seg.limit; - ctxt->ldtr_base = seg.base; - ctxt->ldtr_arbytes = seg.attr; + ctxt.ldtr_sel = seg.sel; + ctxt.ldtr_limit = seg.limit; + ctxt.ldtr_base = seg.base; + ctxt.ldtr_arbytes = seg.attr; if ( v->fpu_initialised ) { - memcpy(ctxt->fpu_regs, v->arch.fpu_ctxt, sizeof(ctxt->fpu_regs)); - ctxt->flags = XEN_X86_FPU_INITIALISED; - } - - ctxt->rax = v->arch.user_regs.rax; - ctxt->rbx = v->arch.user_regs.rbx; - ctxt->rcx = v->arch.user_regs.rcx; - ctxt->rdx = v->arch.user_regs.rdx; - ctxt->rbp = v->arch.user_regs.rbp; - ctxt->rsi = v->arch.user_regs.rsi; - ctxt->rdi = v->arch.user_regs.rdi; - ctxt->rsp = v->arch.user_regs.rsp; - ctxt->rip = v->arch.user_regs.rip; - ctxt->rflags = v->arch.user_regs.rflags; - ctxt->r8 = v->arch.user_regs.r8; - ctxt->r9 = v->arch.user_regs.r9; - ctxt->r10 = v->arch.user_regs.r10; - ctxt->r11 = v->arch.user_regs.r11; - ctxt->r12 = v->arch.user_regs.r12; - ctxt->r13 = v->arch.user_regs.r13; - ctxt->r14 = v->arch.user_regs.r14; - ctxt->r15 = v->arch.user_regs.r15; - ctxt->dr0 = v->arch.debugreg[0]; - ctxt->dr1 = v->arch.debugreg[1]; - ctxt->dr2 = v->arch.debugreg[2]; - ctxt->dr3 = v->arch.debugreg[3]; - ctxt->dr6 = v->arch.debugreg[6]; - ctxt->dr7 = v->arch.debugreg[7]; -} - -static int hvm_save_cpu_ctxt(struct domain *d, hvm_domain_context_t *h) -{ - struct vcpu *v; - struct hvm_hw_cpu ctxt; - - for_each_vcpu ( d, v ) - { - /* We don't need to save state for a vcpu that is down; the restore - * code will leave it down if there is nothing saved. */ - if ( v->pause_flags & VPF_down ) - continue; - - memset(&ctxt, 0, sizeof(ctxt)); - hvm_save_cpu_ctxt_one(v, &ctxt); + memcpy(ctxt.fpu_regs, v->arch.fpu_ctxt, sizeof(ctxt.fpu_regs)); + ctxt.flags = XEN_X86_FPU_INITIALISED; + } + + ctxt.rax = v->arch.user_regs.rax; + ctxt.rbx = v->arch.user_regs.rbx; + ctxt.rcx = v->arch.user_regs.rcx; + ctxt.rdx = v->arch.user_regs.rdx; + ctxt.rbp = v->arch.user_regs.rbp; + ctxt.rsi = v->arch.user_regs.rsi; + ctxt.rdi = v->arch.user_regs.rdi; + ctxt.rsp = v->arch.user_regs.rsp; + ctxt.rip = v->arch.user_regs.rip; + ctxt.rflags = v->arch.user_regs.rflags; + ctxt.r8 = v->arch.user_regs.r8; + ctxt.r9 = v->arch.user_regs.r9; + ctxt.r10 = v->arch.user_regs.r10; + ctxt.r11 = v->arch.user_regs.r11; + ctxt.r12 = v->arch.user_regs.r12; + ctxt.r13 = v->arch.user_regs.r13; + ctxt.r14 = v->arch.user_regs.r14; + ctxt.r15 = v->arch.user_regs.r15; + ctxt.dr0 = v->arch.debugreg[0]; + ctxt.dr1 = v->arch.debugreg[1]; + ctxt.dr2 = v->arch.debugreg[2]; + ctxt.dr3 = v->arch.debugreg[3]; + ctxt.dr6 = v->arch.debugreg[6]; + ctxt.dr7 = v->arch.debugreg[7]; + + if ( hvm_save_entry(CPU, v->vcpu_id, h, &ctxt) != 0 ) + return 1; - if ( hvm_save_entry(CPU, v->vcpu_id, h, &ctxt) != 0 ) - return 1; - } return 0; } @@ -1176,43 +1155,29 @@ static int hvm_load_cpu_ctxt(struct domain *d, hvm_domain_context_t *h) return 0; } -HVM_REGISTER_SAVE_RESTORE(CPU, hvm_save_cpu_ctxt, hvm_load_cpu_ctxt, +HVM_REGISTER_SAVE_RESTORE(CPU, hvm_save_cpu_ctxt_one, hvm_load_cpu_ctxt, 1, HVMSR_PER_VCPU); #define HVM_CPU_XSAVE_SIZE(xcr0) (offsetof(struct hvm_hw_cpu_xsave, \ save_area) + \ xstate_ctxt_size(xcr0)) -static void hvm_save_cpu_xsave_states_one(struct vcpu *v, struct hvm_hw_cpu_xsave *ctxt) +static int hvm_save_cpu_xsave_states_one(struct vcpu *v, hvm_domain_context_t *h) { + unsigned int size = HVM_CPU_XSAVE_SIZE(v->arch.xcr0_accum); + struct hvm_hw_cpu_xsave *ctxt; + + if ( !xsave_enabled(v) ) + return 2; + if ( _hvm_init_entry(h, CPU_XSAVE_CODE, v->vcpu_id, size) ) + return 1; + ctxt = (struct hvm_hw_cpu_xsave *)&h->data[h->cur]; + h->cur += size; ctxt->xfeature_mask = xfeature_mask; ctxt->xcr0 = v->arch.xcr0; ctxt->xcr0_accum = v->arch.xcr0_accum; -} - -static int hvm_save_cpu_xsave_states(struct domain *d, hvm_domain_context_t *h) -{ - struct vcpu *v; - struct hvm_hw_cpu_xsave *ctxt; - - if ( !cpu_has_xsave ) - return 0; /* do nothing */ - - for_each_vcpu ( d, v ) - { - unsigned int size = HVM_CPU_XSAVE_SIZE(v->arch.xcr0_accum); - - if ( !xsave_enabled(v) ) - continue; - if ( _hvm_init_entry(h, CPU_XSAVE_CODE, v->vcpu_id, size) ) - return 1; - ctxt = (struct hvm_hw_cpu_xsave *)&h->data[h->cur]; - h->cur += size; - - hvm_save_cpu_xsave_states_one(v, ctxt); - expand_xsave_states(v, &ctxt->save_area, - size - offsetof(typeof(*ctxt), save_area)); - } + expand_xsave_states(v, &ctxt->save_area, + size - offsetof(typeof(*ctxt), save_area)); return 0; } @@ -1354,10 +1319,18 @@ static const uint32_t msrs_to_send[] = { }; static unsigned int __read_mostly msr_count_max = ARRAY_SIZE(msrs_to_send); -static int hvm_save_cpu_msrs_one(struct vcpu *v, struct hvm_msr *ctxt) +static int hvm_save_cpu_msrs_one(struct vcpu *v, hvm_domain_context_t *h) { + + struct hvm_save_descriptor *desc = _p(&h->data[h->cur]); + struct hvm_msr *ctxt; unsigned int i; + if ( _hvm_init_entry(h, CPU_MSR_CODE, v->vcpu_id, + HVM_CPU_MSR_SIZE(msr_count_max)) ) + return 1; + ctxt = (struct hvm_msr *)&h->data[h->cur]; + ctxt->count = 0; for ( i = 0; i < ARRAY_SIZE(msrs_to_send); ++i ) { uint64_t val; @@ -1383,45 +1356,23 @@ static int hvm_save_cpu_msrs_one(struct vcpu *v, struct hvm_msr *ctxt) ctxt->msr[ctxt->count].index = msrs_to_send[i]; ctxt->msr[ctxt->count++].val = val; } - return 0; -} - -static int hvm_save_cpu_msrs(struct domain *d, hvm_domain_context_t *h) -{ - struct vcpu *v; - - for_each_vcpu ( d, v ) - { - struct hvm_save_descriptor *desc = _p(&h->data[h->cur]); - struct hvm_msr *ctxt; - unsigned int i; - - if ( _hvm_init_entry(h, CPU_MSR_CODE, v->vcpu_id, - HVM_CPU_MSR_SIZE(msr_count_max)) ) - return 1; - ctxt = (struct hvm_msr *)&h->data[h->cur]; - ctxt->count = 0; + if ( hvm_funcs.save_msr ) + hvm_funcs.save_msr(v, ctxt); - hvm_save_cpu_msrs_one(v, ctxt); + ASSERT(ctxt->count <= msr_count_max); - if ( hvm_funcs.save_msr ) - hvm_funcs.save_msr(v, ctxt); - - ASSERT(ctxt->count <= msr_count_max); - - for ( i = 0; i < ctxt->count; ++i ) - ctxt->msr[i]._rsvd = 0; + for ( i = 0; i < ctxt->count; ++i ) + ctxt->msr[i]._rsvd = 0; - if ( ctxt->count ) - { - /* Rewrite length to indicate how much space we actually used. */ - desc->length = HVM_CPU_MSR_SIZE(ctxt->count); - h->cur += HVM_CPU_MSR_SIZE(ctxt->count); - } - else - /* or rewind and remove the descriptor from the stream. */ - h->cur -= sizeof(struct hvm_save_descriptor); + if ( ctxt->count ) + { + /* Rewrite length to indicate how much space we actually used. */ + desc->length = HVM_CPU_MSR_SIZE(ctxt->count); + h->cur += HVM_CPU_MSR_SIZE(ctxt->count); } + else + /* or rewind and remove the descriptor from the stream. */ + h->cur -= sizeof(struct hvm_save_descriptor); return 0; } @@ -1517,7 +1468,7 @@ static int __init hvm_register_CPU_save_and_restore(void) { hvm_register_savevm(CPU_XSAVE_CODE, "CPU_XSAVE", - hvm_save_cpu_xsave_states, + hvm_save_cpu_xsave_states_one, hvm_load_cpu_xsave_states, HVM_CPU_XSAVE_SIZE(xfeature_mask) + sizeof(struct hvm_save_descriptor), @@ -1529,7 +1480,7 @@ static int __init hvm_register_CPU_save_and_restore(void) if ( msr_count_max ) hvm_register_savevm(CPU_MSR_CODE, "CPU_MSR", - hvm_save_cpu_msrs, + hvm_save_cpu_msrs_one, hvm_load_cpu_msrs, HVM_CPU_MSR_SIZE(msr_count_max) + sizeof(struct hvm_save_descriptor), diff --git a/xen/arch/x86/hvm/i8254.c b/xen/arch/x86/hvm/i8254.c index 992f08d..e0d2255 100644 --- a/xen/arch/x86/hvm/i8254.c +++ b/xen/arch/x86/hvm/i8254.c @@ -390,8 +390,9 @@ void pit_stop_channel0_irq(PITState *pit) spin_unlock(&pit->lock); } -static int pit_save(struct domain *d, hvm_domain_context_t *h) +static int pit_save(struct vcpu *v, hvm_domain_context_t *h) { + struct domain *d = v->domain; PITState *pit = domain_vpit(d); int rc; diff --git a/xen/arch/x86/hvm/irq.c b/xen/arch/x86/hvm/irq.c index c85d004..72acb73 100644 --- a/xen/arch/x86/hvm/irq.c +++ b/xen/arch/x86/hvm/irq.c @@ -630,8 +630,9 @@ static int __init dump_irq_info_key_init(void) } __initcall(dump_irq_info_key_init); -static int irq_save_pci(struct domain *d, hvm_domain_context_t *h) +static int irq_save_pci(struct vcpu *v, hvm_domain_context_t *h) { + struct domain *d = v->domain; struct hvm_irq *hvm_irq = hvm_domain_irq(d); unsigned int asserted, pdev, pintx; int rc; @@ -662,16 +663,18 @@ static int irq_save_pci(struct domain *d, hvm_domain_context_t *h) return rc; } -static int irq_save_isa(struct domain *d, hvm_domain_context_t *h) +static int irq_save_isa(struct vcpu *v, hvm_domain_context_t *h) { + struct domain *d = v->domain; struct hvm_irq *hvm_irq = hvm_domain_irq(d); /* Save ISA IRQ lines */ return ( hvm_save_entry(ISA_IRQ, 0, h, &hvm_irq->isa_irq) ); } -static int irq_save_link(struct domain *d, hvm_domain_context_t *h) +static int irq_save_link(struct vcpu *v, hvm_domain_context_t *h) { + struct domain *d = v->domain; struct hvm_irq *hvm_irq = hvm_domain_irq(d); /* Save PCI-ISA link state */ diff --git a/xen/arch/x86/hvm/mtrr.c b/xen/arch/x86/hvm/mtrr.c index d311031..b0e0a60 100644 --- a/xen/arch/x86/hvm/mtrr.c +++ b/xen/arch/x86/hvm/mtrr.c @@ -666,46 +666,36 @@ int hvm_set_mem_pinned_cacheattr(struct domain *d, uint64_t gfn_start, return 0; } -static void hvm_save_mtrr_msr_one(struct vcpu *v, struct hvm_hw_mtrr *hw_mtrr) +static int hvm_save_mtrr_msr_one(struct vcpu *v, hvm_domain_context_t *h) { - struct mtrr_state *mtrr_state = &v->arch.hvm_vcpu.mtrr; int i; + struct mtrr_state *mtrr_state = &v->arch.hvm_vcpu.mtrr; + struct hvm_hw_mtrr hw_mtrr; + /* save mtrr&pat */ - hvm_get_guest_pat(v, &hw_mtrr->msr_pat_cr); + hvm_get_guest_pat(v, &hw_mtrr.msr_pat_cr); - hw_mtrr->msr_mtrr_def_type = mtrr_state->def_type + hw_mtrr.msr_mtrr_def_type = mtrr_state->def_type | (mtrr_state->enabled << 10); - hw_mtrr->msr_mtrr_cap = mtrr_state->mtrr_cap; + hw_mtrr.msr_mtrr_cap = mtrr_state->mtrr_cap; for ( i = 0; i < MTRR_VCNT; i++ ) { /* save physbase */ - hw_mtrr->msr_mtrr_var[i*2] = + hw_mtrr.msr_mtrr_var[i*2] = ((uint64_t*)mtrr_state->var_ranges)[i*2]; /* save physmask */ - hw_mtrr->msr_mtrr_var[i*2+1] = + hw_mtrr.msr_mtrr_var[i*2+1] = ((uint64_t*)mtrr_state->var_ranges)[i*2+1]; } for ( i = 0; i < NUM_FIXED_MSR; i++ ) - hw_mtrr->msr_mtrr_fixed[i] = + hw_mtrr.msr_mtrr_fixed[i] = ((uint64_t*)mtrr_state->fixed_ranges)[i]; -} - -static int hvm_save_mtrr_msr(struct domain *d, hvm_domain_context_t *h) -{ - struct vcpu *v; - struct hvm_hw_mtrr hw_mtrr; - /* save mtrr&pat */ - - for_each_vcpu(d, v) - { - hvm_save_mtrr_msr_one(v, &hw_mtrr); + if ( hvm_save_entry(MTRR, v->vcpu_id, h, &hw_mtrr) != 0 ) + return 1; - if ( hvm_save_entry(MTRR, v->vcpu_id, h, &hw_mtrr) != 0 ) - return 1; - } return 0; } @@ -751,7 +741,7 @@ static int hvm_load_mtrr_msr(struct domain *d, hvm_domain_context_t *h) return 0; } -HVM_REGISTER_SAVE_RESTORE(MTRR, hvm_save_mtrr_msr, hvm_load_mtrr_msr, +HVM_REGISTER_SAVE_RESTORE(MTRR, hvm_save_mtrr_msr_one, hvm_load_mtrr_msr, 1, HVMSR_PER_VCPU); void memory_type_changed(struct domain *d) diff --git a/xen/arch/x86/hvm/pmtimer.c b/xen/arch/x86/hvm/pmtimer.c index 435647f..d8dcbc2 100644 --- a/xen/arch/x86/hvm/pmtimer.c +++ b/xen/arch/x86/hvm/pmtimer.c @@ -249,8 +249,9 @@ static int handle_pmt_io( return X86EMUL_OKAY; } -static int acpi_save(struct domain *d, hvm_domain_context_t *h) +static int acpi_save(struct vcpu *v, hvm_domain_context_t *h) { + struct domain *d = v->domain; struct hvm_hw_acpi *acpi = &d->arch.hvm_domain.acpi; PMTState *s = &d->arch.hvm_domain.pl_time->vpmt; uint32_t x, msb = acpi->tmr_val & TMR_VAL_MSB; diff --git a/xen/arch/x86/hvm/rtc.c b/xen/arch/x86/hvm/rtc.c index cb75b99..58b70fc 100644 --- a/xen/arch/x86/hvm/rtc.c +++ b/xen/arch/x86/hvm/rtc.c @@ -737,8 +737,9 @@ void rtc_migrate_timers(struct vcpu *v) } /* Save RTC hardware state */ -static int rtc_save(struct domain *d, hvm_domain_context_t *h) +static int rtc_save(struct vcpu *v, hvm_domain_context_t *h) { + struct domain *d = v->domain; RTCState *s = domain_vrtc(d); int rc; diff --git a/xen/arch/x86/hvm/save.c b/xen/arch/x86/hvm/save.c index 8984a23..fe24f08 100644 --- a/xen/arch/x86/hvm/save.c +++ b/xen/arch/x86/hvm/save.c @@ -135,9 +135,12 @@ size_t hvm_save_size(struct domain *d) int hvm_save_one(struct domain *d, unsigned int typecode, unsigned int instance, XEN_GUEST_HANDLE_64(uint8) handle, uint64_t *bufsz) { - int rv; + int rv = 0; hvm_domain_context_t ctxt = { }; const struct hvm_save_descriptor *desc; + bool is_single_instance = false; + uint32_t off = 0; + struct vcpu *v; if ( d->is_dying || typecode > HVM_SAVE_CODE_MAX || @@ -145,43 +148,89 @@ int hvm_save_one(struct domain *d, unsigned int typecode, unsigned int instance, !hvm_sr_handlers[typecode].save ) return -EINVAL; + if( hvm_sr_handlers[typecode].kind == HVMSR_PER_VCPU && + instance < d->max_vcpus ) + is_single_instance = true; + ctxt.size = hvm_sr_handlers[typecode].size; - if ( hvm_sr_handlers[typecode].kind == HVMSR_PER_VCPU ) + if ( hvm_sr_handlers[typecode].kind == HVMSR_PER_VCPU && + instance == d->max_vcpus ) ctxt.size *= d->max_vcpus; ctxt.data = xmalloc_bytes(ctxt.size); if ( !ctxt.data ) return -ENOMEM; - if ( (rv = hvm_sr_handlers[typecode].save(d, &ctxt)) != 0 ) - printk(XENLOG_G_ERR "HVM%d save: failed to save type %"PRIu16" (%d)\n", - d->domain_id, typecode, rv); - else if ( rv = -ENOENT, ctxt.cur >= sizeof(*desc) ) - { - uint32_t off; - for ( off = 0; off <= (ctxt.cur - sizeof(*desc)); off += desc->length ) + if( is_single_instance ) + vcpu_pause(d->vcpu[instance]); + else + domain_pause(d); + + if( is_single_instance ) + { + if ( (rv = hvm_sr_handlers[typecode].save(d->vcpu[instance], + &ctxt)) != 0 ) { - desc = (void *)(ctxt.data + off); - /* Move past header */ - off += sizeof(*desc); - if ( ctxt.cur < desc->length || - off > ctxt.cur - desc->length ) - break; - if ( instance == desc->instance ) - { - rv = 0; - if ( guest_handle_is_null(handle) ) - *bufsz = desc->length; - else if ( *bufsz < desc->length ) - rv = -ENOBUFS; - else if ( copy_to_guest(handle, ctxt.data + off, desc->length) ) - rv = -EFAULT; - else - *bufsz = desc->length; - break; - } + printk(XENLOG_G_ERR "HVM%d save: failed to save type %"PRIu16" (%d)\n", + d->domain_id, typecode, rv); + vcpu_unpause(d->vcpu[instance]); + } + else if ( rv = -ENOENT, ctxt.cur >= sizeof(*desc) ) + { + desc = (void *)(ctxt.data); + /* Move past header */ + off = sizeof(*desc); + if ( ctxt.cur < desc->length || + off > ctxt.cur - desc->length ) + rv = -EFAULT; + rv = 0; + if ( guest_handle_is_null(handle) ) + *bufsz = desc->length; + else if ( *bufsz < desc->length ) + rv = -ENOBUFS; + else if ( copy_to_guest(handle, ctxt.data + off, desc->length) ) + rv = -EFAULT; + else + *bufsz = desc->length; + vcpu_unpause(d->vcpu[instance]); } } + else + { + for_each_vcpu ( d, v ) + { + if ( (rv = hvm_sr_handlers[typecode].save(d->vcpu[instance], + &ctxt)) != 0 ) + { + printk(XENLOG_G_ERR "HVM%d save: failed to save type %"PRIu16" (%d)\n", + d->domain_id, typecode, rv); + } + else if ( rv = -ENOENT, ctxt.cur >= sizeof(*desc) ) + { + desc = (void *)(ctxt.data + off); + /* Move past header */ + off += sizeof(*desc); + if ( ctxt.cur < desc->length || + off > ctxt.cur - desc->length ) + break; + if ( instance == desc->instance ) + { + rv = 0; + if ( guest_handle_is_null(handle) ) + *bufsz = desc->length; + else if ( *bufsz < desc->length ) + rv = -ENOBUFS; + else if ( copy_to_guest(handle, ctxt.data + off, desc->length) ) + rv = -EFAULT; + else + *bufsz = desc->length; + break; + } + off += desc->length; + } + } + domain_unpause(d); + } xfree(ctxt.data); return rv; @@ -193,7 +242,8 @@ int hvm_save(struct domain *d, hvm_domain_context_t *h) struct hvm_save_header hdr; struct hvm_save_end end; hvm_save_handler handler; - unsigned int i; + unsigned int i, rc; + struct vcpu *v = NULL; if ( d->is_dying ) return -EINVAL; @@ -225,12 +275,19 @@ int hvm_save(struct domain *d, hvm_domain_context_t *h) { printk(XENLOG_G_INFO "HVM%d save: %s\n", d->domain_id, hvm_sr_handlers[i].name); - if ( handler(d, h) != 0 ) + for_each_vcpu ( d, v ) { - printk(XENLOG_G_ERR - "HVM%d save: failed to save type %"PRIu16"\n", - d->domain_id, i); - return -EFAULT; + rc = handler(v, h); + if( rc == 2 ) + continue; + + if( rc != 0 ) + { + printk(XENLOG_G_ERR + "HVM%d save: failed to save type %"PRIu16"\n", + d->domain_id, i); + return -EFAULT; + } } } } diff --git a/xen/arch/x86/hvm/vioapic.c b/xen/arch/x86/hvm/vioapic.c index 97b419f..86d02cf 100644 --- a/xen/arch/x86/hvm/vioapic.c +++ b/xen/arch/x86/hvm/vioapic.c @@ -569,8 +569,9 @@ int vioapic_get_trigger_mode(const struct domain *d, unsigned int gsi) return vioapic->redirtbl[pin].fields.trig_mode; } -static int ioapic_save(struct domain *d, hvm_domain_context_t *h) +static int ioapic_save(struct vcpu *v, hvm_domain_context_t *h) { + struct domain *d = v->domain; struct hvm_vioapic *s; if ( !has_vioapic(d) ) diff --git a/xen/arch/x86/hvm/viridian.c b/xen/arch/x86/hvm/viridian.c index d22c8ac..8a7c592 100644 --- a/xen/arch/x86/hvm/viridian.c +++ b/xen/arch/x86/hvm/viridian.c @@ -990,8 +990,9 @@ out: return HVM_HCALL_completed; } -static int viridian_save_domain_ctxt(struct domain *d, hvm_domain_context_t *h) +static int viridian_save_domain_ctxt(struct vcpu *v, hvm_domain_context_t *h) { + struct domain *d = v->domain; struct hvm_viridian_domain_context ctxt = { .time_ref_count = d->arch.hvm_domain.viridian.time_ref_count.val, .hypercall_gpa = d->arch.hvm_domain.viridian.hypercall_gpa.raw, @@ -1026,28 +1027,19 @@ static int viridian_load_domain_ctxt(struct domain *d, hvm_domain_context_t *h) HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_DOMAIN, viridian_save_domain_ctxt, viridian_load_domain_ctxt, 1, HVMSR_PER_DOM); -static void viridian_save_vcpu_ctxt_one(struct vcpu *v, struct hvm_viridian_vcpu_context *ctxt) +static int viridian_save_vcpu_ctxt_one(struct vcpu *v, hvm_domain_context_t *h) { - ctxt->vp_assist_msr = v->arch.hvm_vcpu.viridian.vp_assist.msr.raw; - ctxt->vp_assist_pending = v->arch.hvm_vcpu.viridian.vp_assist.pending; - *ctxt->_pad = 0; -} - -static int viridian_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h) -{ - struct vcpu *v; + struct domain *d = v->domain; + struct hvm_viridian_vcpu_context ctxt = { + .vp_assist_msr = v->arch.hvm_vcpu.viridian.vp_assist.msr.raw, + .vp_assist_pending = v->arch.hvm_vcpu.viridian.vp_assist.pending, + }; if ( !is_viridian_domain(d) ) return 0; - for_each_vcpu( d, v ) { - struct hvm_viridian_vcpu_context ctxt; - - viridian_save_vcpu_ctxt_one(v, &ctxt); - - if ( hvm_save_entry(VIRIDIAN_VCPU, v->vcpu_id, h, &ctxt) != 0 ) - return 1; - } + if ( hvm_save_entry(VIRIDIAN_VCPU, v->vcpu_id, h, &ctxt) != 0 ) + return 1; return 0; } @@ -1082,7 +1074,7 @@ static int viridian_load_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h) return 0; } -HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_VCPU, viridian_save_vcpu_ctxt, +HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_VCPU, viridian_save_vcpu_ctxt_one, viridian_load_vcpu_ctxt, 1, HVMSR_PER_VCPU); static int __init parse_viridian_version(const char *arg) diff --git a/xen/arch/x86/hvm/vlapic.c b/xen/arch/x86/hvm/vlapic.c index 1b9f00a..ab35400 100644 --- a/xen/arch/x86/hvm/vlapic.c +++ b/xen/arch/x86/hvm/vlapic.c @@ -1435,43 +1435,35 @@ static void lapic_rearm(struct vlapic *s) s->timer_last_update = s->pt.last_plt_gtime; } -static int lapic_save_hidden(struct domain *d, hvm_domain_context_t *h) +static int lapic_save_hidden(struct vcpu *v, hvm_domain_context_t *h) { - struct vcpu *v; + struct domain *d = v->domain; struct vlapic *s; int rc = 0; if ( !has_vlapic(d) ) return 0; - for_each_vcpu ( d, v ) - { - s = vcpu_vlapic(v); - if ( (rc = hvm_save_entry(LAPIC, v->vcpu_id, h, &s->hw)) != 0 ) - break; - } + s = vcpu_vlapic(v); + rc = hvm_save_entry(LAPIC, v->vcpu_id, h, &s->hw); return rc; } -static int lapic_save_regs(struct domain *d, hvm_domain_context_t *h) +static int lapic_save_regs(struct vcpu *v, hvm_domain_context_t *h) { - struct vcpu *v; + struct domain *d = v->domain; struct vlapic *s; int rc = 0; if ( !has_vlapic(d) ) return 0; - for_each_vcpu ( d, v ) - { - if ( hvm_funcs.sync_pir_to_irr ) - hvm_funcs.sync_pir_to_irr(v); + if ( hvm_funcs.sync_pir_to_irr ) + hvm_funcs.sync_pir_to_irr(v); - s = vcpu_vlapic(v); - if ( (rc = hvm_save_entry(LAPIC_REGS, v->vcpu_id, h, s->regs)) != 0 ) - break; - } + s = vcpu_vlapic(v); + rc = hvm_save_entry(LAPIC_REGS, v->vcpu_id, h, s->regs); return rc; } diff --git a/xen/arch/x86/hvm/vpic.c b/xen/arch/x86/hvm/vpic.c index e160bbd..bad5066 100644 --- a/xen/arch/x86/hvm/vpic.c +++ b/xen/arch/x86/hvm/vpic.c @@ -371,8 +371,9 @@ static int vpic_intercept_elcr_io( return X86EMUL_OKAY; } -static int vpic_save(struct domain *d, hvm_domain_context_t *h) +static int vpic_save(struct vcpu *v, hvm_domain_context_t *h) { + struct domain *d = v->domain; struct hvm_hw_vpic *s; int i; diff --git a/xen/include/asm-x86/hvm/save.h b/xen/include/asm-x86/hvm/save.h index f889e8f..fe642ab 100644 --- a/xen/include/asm-x86/hvm/save.h +++ b/xen/include/asm-x86/hvm/save.h @@ -95,7 +95,7 @@ static inline uint16_t hvm_load_instance(struct hvm_domain_context *h) * The save handler may save multiple instances of a type into the buffer; * the load handler will be called once for each instance found when * restoring. Both return non-zero on error. */ -typedef int (*hvm_save_handler) (struct domain *d, +typedef int (*hvm_save_handler) (struct vcpu *v, hvm_domain_context_t *h); typedef int (*hvm_load_handler) (struct domain *d, hvm_domain_context_t *h); -- 2.7.4 _______________________________________________ Xen-devel mailing list Xen-devel@lists.xenproject.org https://lists.xenproject.org/mailman/listinfo/xen-devel ^ permalink raw reply related [flat|nested] 12+ messages in thread
* Re: [PATCH v5 8/8] x86/domctl: Don't pause the whole domain if only getting vcpu state 2018-05-29 14:58 ` [PATCH v5 8/8] x86/domctl: Don't pause the whole domain if only getting vcpu state Alexandru Isaila @ 2018-06-04 9:21 ` Paul Durrant 2018-06-06 12:35 ` Jan Beulich 1 sibling, 0 replies; 12+ messages in thread From: Paul Durrant @ 2018-06-04 9:21 UTC (permalink / raw) To: 'Alexandru Isaila', xen-devel Cc: Ian Jackson, Wei Liu, jbeulich, Andrew Cooper > -----Original Message----- > From: Alexandru Isaila [mailto:aisaila@bitdefender.com] > Sent: 29 May 2018 15:59 > To: xen-devel@lists.xen.org > Cc: Ian Jackson <Ian.Jackson@citrix.com>; Wei Liu <wei.liu2@citrix.com>; > jbeulich@suse.com; Andrew Cooper <Andrew.Cooper3@citrix.com>; Paul > Durrant <Paul.Durrant@citrix.com>; Alexandru Isaila > <aisaila@bitdefender.com> > Subject: [PATCH v5 8/8] x86/domctl: Don't pause the whole domain if only > getting vcpu state > > This patch is focused on merging the *save() funcs to the *save_one() > funcs to remove redundancy. Also the for loop is moved to the caller so > now we can save info for a single vcpu instance. > > Signed-off-by: Alexandru Isaila <aisaila@bitdefender.com> > > --- > Changes since V4: > - Save funcs are deleted > - vcpu si added to the param list > - save_one funcs are now static > - Add blank line after "/* save mtrr&pat */" comment > --- > xen/arch/x86/cpu/mcheck/vmce.c | 28 ++-- > xen/arch/x86/hvm/hpet.c | 3 +- > xen/arch/x86/hvm/hvm.c | 287 +++++++++++++++++---------------------- > -- > xen/arch/x86/hvm/i8254.c | 3 +- > xen/arch/x86/hvm/irq.c | 9 +- > xen/arch/x86/hvm/mtrr.c | 36 ++---- > xen/arch/x86/hvm/pmtimer.c | 3 +- > xen/arch/x86/hvm/rtc.c | 3 +- > xen/arch/x86/hvm/save.c | 125 +++++++++++++----- > xen/arch/x86/hvm/vioapic.c | 3 +- > xen/arch/x86/hvm/viridian.c | 30 ++--- > xen/arch/x86/hvm/vlapic.c | 28 ++-- > xen/arch/x86/hvm/vpic.c | 3 +- > xen/include/asm-x86/hvm/save.h | 2 +- > 14 files changed, 271 insertions(+), 292 deletions(-) > > diff --git a/xen/arch/x86/cpu/mcheck/vmce.c > b/xen/arch/x86/cpu/mcheck/vmce.c > index 404f27e..7bde9ff 100644 > --- a/xen/arch/x86/cpu/mcheck/vmce.c > +++ b/xen/arch/x86/cpu/mcheck/vmce.c > @@ -349,29 +349,17 @@ int vmce_wrmsr(uint32_t msr, uint64_t val) > return ret; > } > > -static void vmce_save_vcpu_ctxt_one(struct vcpu *v, struct > hvm_vmce_vcpu *ctxt) > +static int vmce_save_vcpu_ctxt_one(struct vcpu *v, > hvm_domain_context_t *h) > { > - ctxt->caps = v->arch.vmce.mcg_cap; > - ctxt->mci_ctl2_bank0 = v->arch.vmce.bank[0].mci_ctl2; > - ctxt->mci_ctl2_bank1 = v->arch.vmce.bank[1].mci_ctl2; > - ctxt->mcg_ext_ctl = v->arch.vmce.mcg_ext_ctl; > -} > - > -static int vmce_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t > *h) > -{ > - struct vcpu *v; > int err = 0; > + struct hvm_vmce_vcpu ctxt; > > - for_each_vcpu ( d, v ) > - { > - struct hvm_vmce_vcpu ctxt; > - > - vmce_save_vcpu_ctxt_one(v, &ctxt); > - err = hvm_save_entry(VMCE_VCPU, v->vcpu_id, h, &ctxt); > - if ( err ) > - break; > - } > + ctxt.caps = v->arch.vmce.mcg_cap; > + ctxt.mci_ctl2_bank0 = v->arch.vmce.bank[0].mci_ctl2; > + ctxt.mci_ctl2_bank1 = v->arch.vmce.bank[1].mci_ctl2; > + ctxt.mcg_ext_ctl = v->arch.vmce.mcg_ext_ctl; > > + err = hvm_save_entry(VMCE_VCPU, v->vcpu_id, h, &ctxt); Can't you just return hvm_save_entry(...) directly here and drop err? > return err; > } > > @@ -394,7 +382,7 @@ static int vmce_load_vcpu_ctxt(struct domain *d, > hvm_domain_context_t *h) > return err ?: vmce_restore_vcpu(v, &ctxt); > } > > -HVM_REGISTER_SAVE_RESTORE(VMCE_VCPU, vmce_save_vcpu_ctxt, > +HVM_REGISTER_SAVE_RESTORE(VMCE_VCPU, > vmce_save_vcpu_ctxt_one, > vmce_load_vcpu_ctxt, 1, HVMSR_PER_VCPU); > > /* > diff --git a/xen/arch/x86/hvm/hpet.c b/xen/arch/x86/hvm/hpet.c > index 2837709..3ed6547 100644 > --- a/xen/arch/x86/hvm/hpet.c > +++ b/xen/arch/x86/hvm/hpet.c > @@ -516,8 +516,9 @@ static const struct hvm_mmio_ops hpet_mmio_ops = > { > }; > > > -static int hpet_save(struct domain *d, hvm_domain_context_t *h) > +static int hpet_save(struct vcpu *vcpu, hvm_domain_context_t *h) > { > + struct domain *d = vcpu->domain; > HPETState *hp = domain_vhpet(d); > struct vcpu *v = pt_global_vcpu_target(d); > int rc; > diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c > index 4a22283..ea27055 100644 > --- a/xen/arch/x86/hvm/hvm.c > +++ b/xen/arch/x86/hvm/hvm.c > @@ -740,24 +740,13 @@ void hvm_domain_destroy(struct domain *d) > destroy_vpci_mmcfg(d); > } > > -static void hvm_save_tsc_adjust_one(struct vcpu *v, struct hvm_tsc_adjust > *ctxt) > +static int hvm_save_tsc_adjust_one(struct vcpu *v, > hvm_domain_context_t *h) > { > - ctxt->tsc_adjust = v->arch.hvm_vcpu.msr_tsc_adjust; > -} > - > -static int hvm_save_tsc_adjust(struct domain *d, hvm_domain_context_t > *h) > -{ > - struct vcpu *v; > - struct hvm_tsc_adjust ctxt; > int err = 0; > + struct hvm_tsc_adjust ctxt; > > - for_each_vcpu ( d, v ) > - { > - hvm_save_tsc_adjust_one(v, &ctxt); > - err = hvm_save_entry(TSC_ADJUST, v->vcpu_id, h, &ctxt); > - if ( err ) > - break; > - } > + ctxt.tsc_adjust = v->arch.hvm_vcpu.msr_tsc_adjust; > + err = hvm_save_entry(TSC_ADJUST, v->vcpu_id, h, &ctxt); > Looks like you can lose err here too. > return err; > } > @@ -782,126 +771,116 @@ static int hvm_load_tsc_adjust(struct domain *d, > hvm_domain_context_t *h) > return 0; > } > > -HVM_REGISTER_SAVE_RESTORE(TSC_ADJUST, hvm_save_tsc_adjust, > +HVM_REGISTER_SAVE_RESTORE(TSC_ADJUST, hvm_save_tsc_adjust_one, > hvm_load_tsc_adjust, 1, HVMSR_PER_VCPU); > > -static void hvm_save_cpu_ctxt_one(struct vcpu *v, struct hvm_hw_cpu > *ctxt) > +static int hvm_save_cpu_ctxt_one(struct vcpu *v, hvm_domain_context_t > *h) > { > struct segment_register seg; > + struct hvm_hw_cpu ctxt = {}; > + > + /* We don't need to save state for a vcpu that is down; the restore > + * code will leave it down if there is nothing saved. */ > + if ( v->pause_flags & VPF_down ) > + return 2; > > /* Architecture-specific vmcs/vmcb bits */ > - hvm_funcs.save_cpu_ctxt(v, ctxt); > + hvm_funcs.save_cpu_ctxt(v, &ctxt); > > - ctxt->tsc = hvm_get_guest_tsc_fixed(v, v->domain- > >arch.hvm_domain.sync_tsc); > + ctxt.tsc = hvm_get_guest_tsc_fixed(v, v->domain- > >arch.hvm_domain.sync_tsc); > > - ctxt->msr_tsc_aux = hvm_msr_tsc_aux(v); > + ctxt.msr_tsc_aux = hvm_msr_tsc_aux(v); > > hvm_get_segment_register(v, x86_seg_idtr, &seg); > - ctxt->idtr_limit = seg.limit; > - ctxt->idtr_base = seg.base; > + ctxt.idtr_limit = seg.limit; > + ctxt.idtr_base = seg.base; > > hvm_get_segment_register(v, x86_seg_gdtr, &seg); > - ctxt->gdtr_limit = seg.limit; > - ctxt->gdtr_base = seg.base; > + ctxt.gdtr_limit = seg.limit; > + ctxt.gdtr_base = seg.base; > > hvm_get_segment_register(v, x86_seg_cs, &seg); > - ctxt->cs_sel = seg.sel; > - ctxt->cs_limit = seg.limit; > - ctxt->cs_base = seg.base; > - ctxt->cs_arbytes = seg.attr; > + ctxt.cs_sel = seg.sel; > + ctxt.cs_limit = seg.limit; > + ctxt.cs_base = seg.base; > + ctxt.cs_arbytes = seg.attr; > > hvm_get_segment_register(v, x86_seg_ds, &seg); > - ctxt->ds_sel = seg.sel; > - ctxt->ds_limit = seg.limit; > - ctxt->ds_base = seg.base; > - ctxt->ds_arbytes = seg.attr; > + ctxt.ds_sel = seg.sel; > + ctxt.ds_limit = seg.limit; > + ctxt.ds_base = seg.base; > + ctxt.ds_arbytes = seg.attr; > > hvm_get_segment_register(v, x86_seg_es, &seg); > - ctxt->es_sel = seg.sel; > - ctxt->es_limit = seg.limit; > - ctxt->es_base = seg.base; > - ctxt->es_arbytes = seg.attr; > + ctxt.es_sel = seg.sel; > + ctxt.es_limit = seg.limit; > + ctxt.es_base = seg.base; > + ctxt.es_arbytes = seg.attr; > > hvm_get_segment_register(v, x86_seg_ss, &seg); > - ctxt->ss_sel = seg.sel; > - ctxt->ss_limit = seg.limit; > - ctxt->ss_base = seg.base; > - ctxt->ss_arbytes = seg.attr; > + ctxt.ss_sel = seg.sel; > + ctxt.ss_limit = seg.limit; > + ctxt.ss_base = seg.base; > + ctxt.ss_arbytes = seg.attr; > > hvm_get_segment_register(v, x86_seg_fs, &seg); > - ctxt->fs_sel = seg.sel; > - ctxt->fs_limit = seg.limit; > - ctxt->fs_base = seg.base; > - ctxt->fs_arbytes = seg.attr; > + ctxt.fs_sel = seg.sel; > + ctxt.fs_limit = seg.limit; > + ctxt.fs_base = seg.base; > + ctxt.fs_arbytes = seg.attr; > > hvm_get_segment_register(v, x86_seg_gs, &seg); > - ctxt->gs_sel = seg.sel; > - ctxt->gs_limit = seg.limit; > - ctxt->gs_base = seg.base; > - ctxt->gs_arbytes = seg.attr; > + ctxt.gs_sel = seg.sel; > + ctxt.gs_limit = seg.limit; > + ctxt.gs_base = seg.base; > + ctxt.gs_arbytes = seg.attr; > > hvm_get_segment_register(v, x86_seg_tr, &seg); > - ctxt->tr_sel = seg.sel; > - ctxt->tr_limit = seg.limit; > - ctxt->tr_base = seg.base; > - ctxt->tr_arbytes = seg.attr; > + ctxt.tr_sel = seg.sel; > + ctxt.tr_limit = seg.limit; > + ctxt.tr_base = seg.base; > + ctxt.tr_arbytes = seg.attr; > > hvm_get_segment_register(v, x86_seg_ldtr, &seg); > - ctxt->ldtr_sel = seg.sel; > - ctxt->ldtr_limit = seg.limit; > - ctxt->ldtr_base = seg.base; > - ctxt->ldtr_arbytes = seg.attr; > + ctxt.ldtr_sel = seg.sel; > + ctxt.ldtr_limit = seg.limit; > + ctxt.ldtr_base = seg.base; > + ctxt.ldtr_arbytes = seg.attr; > > if ( v->fpu_initialised ) > { > - memcpy(ctxt->fpu_regs, v->arch.fpu_ctxt, sizeof(ctxt->fpu_regs)); > - ctxt->flags = XEN_X86_FPU_INITIALISED; > - } > - > - ctxt->rax = v->arch.user_regs.rax; > - ctxt->rbx = v->arch.user_regs.rbx; > - ctxt->rcx = v->arch.user_regs.rcx; > - ctxt->rdx = v->arch.user_regs.rdx; > - ctxt->rbp = v->arch.user_regs.rbp; > - ctxt->rsi = v->arch.user_regs.rsi; > - ctxt->rdi = v->arch.user_regs.rdi; > - ctxt->rsp = v->arch.user_regs.rsp; > - ctxt->rip = v->arch.user_regs.rip; > - ctxt->rflags = v->arch.user_regs.rflags; > - ctxt->r8 = v->arch.user_regs.r8; > - ctxt->r9 = v->arch.user_regs.r9; > - ctxt->r10 = v->arch.user_regs.r10; > - ctxt->r11 = v->arch.user_regs.r11; > - ctxt->r12 = v->arch.user_regs.r12; > - ctxt->r13 = v->arch.user_regs.r13; > - ctxt->r14 = v->arch.user_regs.r14; > - ctxt->r15 = v->arch.user_regs.r15; > - ctxt->dr0 = v->arch.debugreg[0]; > - ctxt->dr1 = v->arch.debugreg[1]; > - ctxt->dr2 = v->arch.debugreg[2]; > - ctxt->dr3 = v->arch.debugreg[3]; > - ctxt->dr6 = v->arch.debugreg[6]; > - ctxt->dr7 = v->arch.debugreg[7]; > -} > - > -static int hvm_save_cpu_ctxt(struct domain *d, hvm_domain_context_t > *h) > -{ > - struct vcpu *v; > - struct hvm_hw_cpu ctxt; > - > - for_each_vcpu ( d, v ) > - { > - /* We don't need to save state for a vcpu that is down; the restore > - * code will leave it down if there is nothing saved. */ > - if ( v->pause_flags & VPF_down ) > - continue; > - > - memset(&ctxt, 0, sizeof(ctxt)); > - hvm_save_cpu_ctxt_one(v, &ctxt); > + memcpy(ctxt.fpu_regs, v->arch.fpu_ctxt, sizeof(ctxt.fpu_regs)); > + ctxt.flags = XEN_X86_FPU_INITIALISED; > + } > + > + ctxt.rax = v->arch.user_regs.rax; > + ctxt.rbx = v->arch.user_regs.rbx; > + ctxt.rcx = v->arch.user_regs.rcx; > + ctxt.rdx = v->arch.user_regs.rdx; > + ctxt.rbp = v->arch.user_regs.rbp; > + ctxt.rsi = v->arch.user_regs.rsi; > + ctxt.rdi = v->arch.user_regs.rdi; > + ctxt.rsp = v->arch.user_regs.rsp; > + ctxt.rip = v->arch.user_regs.rip; > + ctxt.rflags = v->arch.user_regs.rflags; > + ctxt.r8 = v->arch.user_regs.r8; > + ctxt.r9 = v->arch.user_regs.r9; > + ctxt.r10 = v->arch.user_regs.r10; > + ctxt.r11 = v->arch.user_regs.r11; > + ctxt.r12 = v->arch.user_regs.r12; > + ctxt.r13 = v->arch.user_regs.r13; > + ctxt.r14 = v->arch.user_regs.r14; > + ctxt.r15 = v->arch.user_regs.r15; > + ctxt.dr0 = v->arch.debugreg[0]; > + ctxt.dr1 = v->arch.debugreg[1]; > + ctxt.dr2 = v->arch.debugreg[2]; > + ctxt.dr3 = v->arch.debugreg[3]; > + ctxt.dr6 = v->arch.debugreg[6]; > + ctxt.dr7 = v->arch.debugreg[7]; > + > + if ( hvm_save_entry(CPU, v->vcpu_id, h, &ctxt) != 0 ) > + return 1; > > - if ( hvm_save_entry(CPU, v->vcpu_id, h, &ctxt) != 0 ) > - return 1; > - } > return 0; > } > > @@ -1176,43 +1155,29 @@ static int hvm_load_cpu_ctxt(struct domain *d, > hvm_domain_context_t *h) > return 0; > } > > -HVM_REGISTER_SAVE_RESTORE(CPU, hvm_save_cpu_ctxt, > hvm_load_cpu_ctxt, > +HVM_REGISTER_SAVE_RESTORE(CPU, hvm_save_cpu_ctxt_one, > hvm_load_cpu_ctxt, > 1, HVMSR_PER_VCPU); > > #define HVM_CPU_XSAVE_SIZE(xcr0) (offsetof(struct hvm_hw_cpu_xsave, > \ > save_area) + \ > xstate_ctxt_size(xcr0)) > > -static void hvm_save_cpu_xsave_states_one(struct vcpu *v, struct > hvm_hw_cpu_xsave *ctxt) > +static int hvm_save_cpu_xsave_states_one(struct vcpu *v, > hvm_domain_context_t *h) > { > + unsigned int size = HVM_CPU_XSAVE_SIZE(v->arch.xcr0_accum); > + struct hvm_hw_cpu_xsave *ctxt; > + > + if ( !xsave_enabled(v) ) > + return 2; > + if ( _hvm_init_entry(h, CPU_XSAVE_CODE, v->vcpu_id, size) ) > + return 1; > + ctxt = (struct hvm_hw_cpu_xsave *)&h->data[h->cur]; > + h->cur += size; > ctxt->xfeature_mask = xfeature_mask; > ctxt->xcr0 = v->arch.xcr0; > ctxt->xcr0_accum = v->arch.xcr0_accum; > -} > - > -static int hvm_save_cpu_xsave_states(struct domain *d, > hvm_domain_context_t *h) > -{ > - struct vcpu *v; > - struct hvm_hw_cpu_xsave *ctxt; > - > - if ( !cpu_has_xsave ) > - return 0; /* do nothing */ > - > - for_each_vcpu ( d, v ) > - { > - unsigned int size = HVM_CPU_XSAVE_SIZE(v->arch.xcr0_accum); > - > - if ( !xsave_enabled(v) ) > - continue; > - if ( _hvm_init_entry(h, CPU_XSAVE_CODE, v->vcpu_id, size) ) > - return 1; > - ctxt = (struct hvm_hw_cpu_xsave *)&h->data[h->cur]; > - h->cur += size; > - > - hvm_save_cpu_xsave_states_one(v, ctxt); > - expand_xsave_states(v, &ctxt->save_area, > - size - offsetof(typeof(*ctxt), save_area)); > - } > + expand_xsave_states(v, &ctxt->save_area, > + size - offsetof(typeof(*ctxt), save_area)); > > return 0; > } > @@ -1354,10 +1319,18 @@ static const uint32_t msrs_to_send[] = { > }; > static unsigned int __read_mostly msr_count_max = > ARRAY_SIZE(msrs_to_send); > > -static int hvm_save_cpu_msrs_one(struct vcpu *v, struct hvm_msr *ctxt) > +static int hvm_save_cpu_msrs_one(struct vcpu *v, hvm_domain_context_t > *h) > { > + > + struct hvm_save_descriptor *desc = _p(&h->data[h->cur]); > + struct hvm_msr *ctxt; > unsigned int i; > > + if ( _hvm_init_entry(h, CPU_MSR_CODE, v->vcpu_id, > + HVM_CPU_MSR_SIZE(msr_count_max)) ) > + return 1; > + ctxt = (struct hvm_msr *)&h->data[h->cur]; > + ctxt->count = 0; > for ( i = 0; i < ARRAY_SIZE(msrs_to_send); ++i ) > { > uint64_t val; > @@ -1383,45 +1356,23 @@ static int hvm_save_cpu_msrs_one(struct vcpu > *v, struct hvm_msr *ctxt) > ctxt->msr[ctxt->count].index = msrs_to_send[i]; > ctxt->msr[ctxt->count++].val = val; > } > - return 0; > -} > - > -static int hvm_save_cpu_msrs(struct domain *d, hvm_domain_context_t > *h) > -{ > - struct vcpu *v; > - > - for_each_vcpu ( d, v ) > - { > - struct hvm_save_descriptor *desc = _p(&h->data[h->cur]); > - struct hvm_msr *ctxt; > - unsigned int i; > - > - if ( _hvm_init_entry(h, CPU_MSR_CODE, v->vcpu_id, > - HVM_CPU_MSR_SIZE(msr_count_max)) ) > - return 1; > - ctxt = (struct hvm_msr *)&h->data[h->cur]; > - ctxt->count = 0; > + if ( hvm_funcs.save_msr ) > + hvm_funcs.save_msr(v, ctxt); > > - hvm_save_cpu_msrs_one(v, ctxt); > + ASSERT(ctxt->count <= msr_count_max); > > - if ( hvm_funcs.save_msr ) > - hvm_funcs.save_msr(v, ctxt); > - > - ASSERT(ctxt->count <= msr_count_max); > - > - for ( i = 0; i < ctxt->count; ++i ) > - ctxt->msr[i]._rsvd = 0; > + for ( i = 0; i < ctxt->count; ++i ) > + ctxt->msr[i]._rsvd = 0; > > - if ( ctxt->count ) > - { > - /* Rewrite length to indicate how much space we actually used. */ > - desc->length = HVM_CPU_MSR_SIZE(ctxt->count); > - h->cur += HVM_CPU_MSR_SIZE(ctxt->count); > - } > - else > - /* or rewind and remove the descriptor from the stream. */ > - h->cur -= sizeof(struct hvm_save_descriptor); > + if ( ctxt->count ) > + { > + /* Rewrite length to indicate how much space we actually used. */ > + desc->length = HVM_CPU_MSR_SIZE(ctxt->count); > + h->cur += HVM_CPU_MSR_SIZE(ctxt->count); > } > + else > + /* or rewind and remove the descriptor from the stream. */ > + h->cur -= sizeof(struct hvm_save_descriptor); > > return 0; > } > @@ -1517,7 +1468,7 @@ static int __init > hvm_register_CPU_save_and_restore(void) > { > hvm_register_savevm(CPU_XSAVE_CODE, > "CPU_XSAVE", > - hvm_save_cpu_xsave_states, > + hvm_save_cpu_xsave_states_one, > hvm_load_cpu_xsave_states, > HVM_CPU_XSAVE_SIZE(xfeature_mask) + > sizeof(struct hvm_save_descriptor), > @@ -1529,7 +1480,7 @@ static int __init > hvm_register_CPU_save_and_restore(void) > if ( msr_count_max ) > hvm_register_savevm(CPU_MSR_CODE, > "CPU_MSR", > - hvm_save_cpu_msrs, > + hvm_save_cpu_msrs_one, > hvm_load_cpu_msrs, > HVM_CPU_MSR_SIZE(msr_count_max) + > sizeof(struct hvm_save_descriptor), > diff --git a/xen/arch/x86/hvm/i8254.c b/xen/arch/x86/hvm/i8254.c > index 992f08d..e0d2255 100644 > --- a/xen/arch/x86/hvm/i8254.c > +++ b/xen/arch/x86/hvm/i8254.c > @@ -390,8 +390,9 @@ void pit_stop_channel0_irq(PITState *pit) > spin_unlock(&pit->lock); > } > > -static int pit_save(struct domain *d, hvm_domain_context_t *h) > +static int pit_save(struct vcpu *v, hvm_domain_context_t *h) > { > + struct domain *d = v->domain; > PITState *pit = domain_vpit(d); > int rc; > > diff --git a/xen/arch/x86/hvm/irq.c b/xen/arch/x86/hvm/irq.c > index c85d004..72acb73 100644 > --- a/xen/arch/x86/hvm/irq.c > +++ b/xen/arch/x86/hvm/irq.c > @@ -630,8 +630,9 @@ static int __init dump_irq_info_key_init(void) > } > __initcall(dump_irq_info_key_init); > > -static int irq_save_pci(struct domain *d, hvm_domain_context_t *h) > +static int irq_save_pci(struct vcpu *v, hvm_domain_context_t *h) > { > + struct domain *d = v->domain; > struct hvm_irq *hvm_irq = hvm_domain_irq(d); > unsigned int asserted, pdev, pintx; > int rc; > @@ -662,16 +663,18 @@ static int irq_save_pci(struct domain *d, > hvm_domain_context_t *h) > return rc; > } > > -static int irq_save_isa(struct domain *d, hvm_domain_context_t *h) > +static int irq_save_isa(struct vcpu *v, hvm_domain_context_t *h) > { > + struct domain *d = v->domain; > struct hvm_irq *hvm_irq = hvm_domain_irq(d); > > /* Save ISA IRQ lines */ > return ( hvm_save_entry(ISA_IRQ, 0, h, &hvm_irq->isa_irq) ); > } > > -static int irq_save_link(struct domain *d, hvm_domain_context_t *h) > +static int irq_save_link(struct vcpu *v, hvm_domain_context_t *h) > { > + struct domain *d = v->domain; > struct hvm_irq *hvm_irq = hvm_domain_irq(d); > > /* Save PCI-ISA link state */ > diff --git a/xen/arch/x86/hvm/mtrr.c b/xen/arch/x86/hvm/mtrr.c > index d311031..b0e0a60 100644 > --- a/xen/arch/x86/hvm/mtrr.c > +++ b/xen/arch/x86/hvm/mtrr.c > @@ -666,46 +666,36 @@ int hvm_set_mem_pinned_cacheattr(struct > domain *d, uint64_t gfn_start, > return 0; > } > > -static void hvm_save_mtrr_msr_one(struct vcpu *v, struct hvm_hw_mtrr > *hw_mtrr) > +static int hvm_save_mtrr_msr_one(struct vcpu *v, hvm_domain_context_t > *h) > { > - struct mtrr_state *mtrr_state = &v->arch.hvm_vcpu.mtrr; > int i; > + struct mtrr_state *mtrr_state = &v->arch.hvm_vcpu.mtrr; > + struct hvm_hw_mtrr hw_mtrr; > + /* save mtrr&pat */ > > - hvm_get_guest_pat(v, &hw_mtrr->msr_pat_cr); > + hvm_get_guest_pat(v, &hw_mtrr.msr_pat_cr); > > - hw_mtrr->msr_mtrr_def_type = mtrr_state->def_type > + hw_mtrr.msr_mtrr_def_type = mtrr_state->def_type > | (mtrr_state->enabled << 10); > - hw_mtrr->msr_mtrr_cap = mtrr_state->mtrr_cap; > + hw_mtrr.msr_mtrr_cap = mtrr_state->mtrr_cap; > > for ( i = 0; i < MTRR_VCNT; i++ ) > { > /* save physbase */ > - hw_mtrr->msr_mtrr_var[i*2] = > + hw_mtrr.msr_mtrr_var[i*2] = > ((uint64_t*)mtrr_state->var_ranges)[i*2]; > /* save physmask */ > - hw_mtrr->msr_mtrr_var[i*2+1] = > + hw_mtrr.msr_mtrr_var[i*2+1] = > ((uint64_t*)mtrr_state->var_ranges)[i*2+1]; > } > > for ( i = 0; i < NUM_FIXED_MSR; i++ ) > - hw_mtrr->msr_mtrr_fixed[i] = > + hw_mtrr.msr_mtrr_fixed[i] = > ((uint64_t*)mtrr_state->fixed_ranges)[i]; > > -} > - > -static int hvm_save_mtrr_msr(struct domain *d, hvm_domain_context_t > *h) > -{ > - struct vcpu *v; > - struct hvm_hw_mtrr hw_mtrr; > - /* save mtrr&pat */ > - > - for_each_vcpu(d, v) > - { > - hvm_save_mtrr_msr_one(v, &hw_mtrr); > + if ( hvm_save_entry(MTRR, v->vcpu_id, h, &hw_mtrr) != 0 ) > + return 1; > > - if ( hvm_save_entry(MTRR, v->vcpu_id, h, &hw_mtrr) != 0 ) > - return 1; > - } > return 0; > } > > @@ -751,7 +741,7 @@ static int hvm_load_mtrr_msr(struct domain *d, > hvm_domain_context_t *h) > return 0; > } > > -HVM_REGISTER_SAVE_RESTORE(MTRR, hvm_save_mtrr_msr, > hvm_load_mtrr_msr, > +HVM_REGISTER_SAVE_RESTORE(MTRR, hvm_save_mtrr_msr_one, > hvm_load_mtrr_msr, > 1, HVMSR_PER_VCPU); > > void memory_type_changed(struct domain *d) > diff --git a/xen/arch/x86/hvm/pmtimer.c b/xen/arch/x86/hvm/pmtimer.c > index 435647f..d8dcbc2 100644 > --- a/xen/arch/x86/hvm/pmtimer.c > +++ b/xen/arch/x86/hvm/pmtimer.c > @@ -249,8 +249,9 @@ static int handle_pmt_io( > return X86EMUL_OKAY; > } > > -static int acpi_save(struct domain *d, hvm_domain_context_t *h) > +static int acpi_save(struct vcpu *v, hvm_domain_context_t *h) > { > + struct domain *d = v->domain; > struct hvm_hw_acpi *acpi = &d->arch.hvm_domain.acpi; > PMTState *s = &d->arch.hvm_domain.pl_time->vpmt; > uint32_t x, msb = acpi->tmr_val & TMR_VAL_MSB; > diff --git a/xen/arch/x86/hvm/rtc.c b/xen/arch/x86/hvm/rtc.c > index cb75b99..58b70fc 100644 > --- a/xen/arch/x86/hvm/rtc.c > +++ b/xen/arch/x86/hvm/rtc.c > @@ -737,8 +737,9 @@ void rtc_migrate_timers(struct vcpu *v) > } > > /* Save RTC hardware state */ > -static int rtc_save(struct domain *d, hvm_domain_context_t *h) > +static int rtc_save(struct vcpu *v, hvm_domain_context_t *h) > { > + struct domain *d = v->domain; > RTCState *s = domain_vrtc(d); > int rc; > > diff --git a/xen/arch/x86/hvm/save.c b/xen/arch/x86/hvm/save.c > index 8984a23..fe24f08 100644 > --- a/xen/arch/x86/hvm/save.c > +++ b/xen/arch/x86/hvm/save.c > @@ -135,9 +135,12 @@ size_t hvm_save_size(struct domain *d) > int hvm_save_one(struct domain *d, unsigned int typecode, unsigned int > instance, > XEN_GUEST_HANDLE_64(uint8) handle, uint64_t *bufsz) > { > - int rv; > + int rv = 0; > hvm_domain_context_t ctxt = { }; > const struct hvm_save_descriptor *desc; > + bool is_single_instance = false; > + uint32_t off = 0; > + struct vcpu *v; > > if ( d->is_dying || > typecode > HVM_SAVE_CODE_MAX || > @@ -145,43 +148,89 @@ int hvm_save_one(struct domain *d, unsigned int > typecode, unsigned int instance, > !hvm_sr_handlers[typecode].save ) > return -EINVAL; > > + if( hvm_sr_handlers[typecode].kind == HVMSR_PER_VCPU && > + instance < d->max_vcpus ) > + is_single_instance = true; > + > ctxt.size = hvm_sr_handlers[typecode].size; > - if ( hvm_sr_handlers[typecode].kind == HVMSR_PER_VCPU ) > + if ( hvm_sr_handlers[typecode].kind == HVMSR_PER_VCPU && > + instance == d->max_vcpus ) > ctxt.size *= d->max_vcpus; > ctxt.data = xmalloc_bytes(ctxt.size); > if ( !ctxt.data ) > return -ENOMEM; > > - if ( (rv = hvm_sr_handlers[typecode].save(d, &ctxt)) != 0 ) > - printk(XENLOG_G_ERR "HVM%d save: failed to save type %"PRIu16" > (%d)\n", > - d->domain_id, typecode, rv); > - else if ( rv = -ENOENT, ctxt.cur >= sizeof(*desc) ) > - { > - uint32_t off; > > - for ( off = 0; off <= (ctxt.cur - sizeof(*desc)); off += desc->length ) > + if( is_single_instance ) > + vcpu_pause(d->vcpu[instance]); > + else > + domain_pause(d); > + > + if( is_single_instance ) > + { > + if ( (rv = hvm_sr_handlers[typecode].save(d->vcpu[instance], > + &ctxt)) != 0 ) > { > - desc = (void *)(ctxt.data + off); > - /* Move past header */ > - off += sizeof(*desc); > - if ( ctxt.cur < desc->length || > - off > ctxt.cur - desc->length ) > - break; > - if ( instance == desc->instance ) > - { > - rv = 0; > - if ( guest_handle_is_null(handle) ) > - *bufsz = desc->length; > - else if ( *bufsz < desc->length ) > - rv = -ENOBUFS; > - else if ( copy_to_guest(handle, ctxt.data + off, desc->length) ) > - rv = -EFAULT; > - else > - *bufsz = desc->length; > - break; > - } > + printk(XENLOG_G_ERR "HVM%d save: failed to save type %"PRIu16" > (%d)\n", > + d->domain_id, typecode, rv); > + vcpu_unpause(d->vcpu[instance]); > + } > + else if ( rv = -ENOENT, ctxt.cur >= sizeof(*desc) ) Not a fan of this style. Any reason you can't init rv inside the else-if clause below? > + { > + desc = (void *)(ctxt.data); > + /* Move past header */ > + off = sizeof(*desc); > + if ( ctxt.cur < desc->length || > + off > ctxt.cur - desc->length ) > + rv = -EFAULT; > + rv = 0; > + if ( guest_handle_is_null(handle) ) > + *bufsz = desc->length; > + else if ( *bufsz < desc->length ) > + rv = -ENOBUFS; > + else if ( copy_to_guest(handle, ctxt.data + off, desc->length) ) > + rv = -EFAULT; > + else > + *bufsz = desc->length; > + vcpu_unpause(d->vcpu[instance]); > } > } > + else > + { > + for_each_vcpu ( d, v ) > + { > + if ( (rv = hvm_sr_handlers[typecode].save(d->vcpu[instance], > + &ctxt)) != 0 ) > + { > + printk(XENLOG_G_ERR "HVM%d save: failed to save type > %"PRIu16" (%d)\n", > + d->domain_id, typecode, rv); > + } > + else if ( rv = -ENOENT, ctxt.cur >= sizeof(*desc) ) Same here. > + { > + desc = (void *)(ctxt.data + off); > + /* Move past header */ > + off += sizeof(*desc); > + if ( ctxt.cur < desc->length || > + off > ctxt.cur - desc->length ) > + break; > + if ( instance == desc->instance ) > + { > + rv = 0; > + if ( guest_handle_is_null(handle) ) > + *bufsz = desc->length; > + else if ( *bufsz < desc->length ) > + rv = -ENOBUFS; > + else if ( copy_to_guest(handle, ctxt.data + off, desc->length) ) > + rv = -EFAULT; > + else > + *bufsz = desc->length; > + break; > + } > + off += desc->length; > + } > + } > + domain_unpause(d); > + } > > xfree(ctxt.data); > return rv; > @@ -193,7 +242,8 @@ int hvm_save(struct domain *d, > hvm_domain_context_t *h) > struct hvm_save_header hdr; > struct hvm_save_end end; > hvm_save_handler handler; > - unsigned int i; > + unsigned int i, rc; > + struct vcpu *v = NULL; > > if ( d->is_dying ) > return -EINVAL; > @@ -225,12 +275,19 @@ int hvm_save(struct domain *d, > hvm_domain_context_t *h) > { > printk(XENLOG_G_INFO "HVM%d save: %s\n", > d->domain_id, hvm_sr_handlers[i].name); > - if ( handler(d, h) != 0 ) > + for_each_vcpu ( d, v ) > { > - printk(XENLOG_G_ERR > - "HVM%d save: failed to save type %"PRIu16"\n", > - d->domain_id, i); > - return -EFAULT; > + rc = handler(v, h); > + if( rc == 2 ) What does the magic '2' mean? I think we need at least a comment here, if not a #define or enum return. Paul > + continue; > + > + if( rc != 0 ) > + { > + printk(XENLOG_G_ERR > + "HVM%d save: failed to save type %"PRIu16"\n", > + d->domain_id, i); > + return -EFAULT; > + } > } > } > } > diff --git a/xen/arch/x86/hvm/vioapic.c b/xen/arch/x86/hvm/vioapic.c > index 97b419f..86d02cf 100644 > --- a/xen/arch/x86/hvm/vioapic.c > +++ b/xen/arch/x86/hvm/vioapic.c > @@ -569,8 +569,9 @@ int vioapic_get_trigger_mode(const struct domain *d, > unsigned int gsi) > return vioapic->redirtbl[pin].fields.trig_mode; > } > > -static int ioapic_save(struct domain *d, hvm_domain_context_t *h) > +static int ioapic_save(struct vcpu *v, hvm_domain_context_t *h) > { > + struct domain *d = v->domain; > struct hvm_vioapic *s; > > if ( !has_vioapic(d) ) > diff --git a/xen/arch/x86/hvm/viridian.c b/xen/arch/x86/hvm/viridian.c > index d22c8ac..8a7c592 100644 > --- a/xen/arch/x86/hvm/viridian.c > +++ b/xen/arch/x86/hvm/viridian.c > @@ -990,8 +990,9 @@ out: > return HVM_HCALL_completed; > } > > -static int viridian_save_domain_ctxt(struct domain *d, > hvm_domain_context_t *h) > +static int viridian_save_domain_ctxt(struct vcpu *v, > hvm_domain_context_t *h) > { > + struct domain *d = v->domain; > struct hvm_viridian_domain_context ctxt = { > .time_ref_count = d->arch.hvm_domain.viridian.time_ref_count.val, > .hypercall_gpa = d->arch.hvm_domain.viridian.hypercall_gpa.raw, > @@ -1026,28 +1027,19 @@ static int viridian_load_domain_ctxt(struct > domain *d, hvm_domain_context_t *h) > HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_DOMAIN, > viridian_save_domain_ctxt, > viridian_load_domain_ctxt, 1, HVMSR_PER_DOM); > > -static void viridian_save_vcpu_ctxt_one(struct vcpu *v, struct > hvm_viridian_vcpu_context *ctxt) > +static int viridian_save_vcpu_ctxt_one(struct vcpu *v, > hvm_domain_context_t *h) > { > - ctxt->vp_assist_msr = v->arch.hvm_vcpu.viridian.vp_assist.msr.raw; > - ctxt->vp_assist_pending = v->arch.hvm_vcpu.viridian.vp_assist.pending; > - *ctxt->_pad = 0; > -} > - > -static int viridian_save_vcpu_ctxt(struct domain *d, > hvm_domain_context_t *h) > -{ > - struct vcpu *v; > + struct domain *d = v->domain; > + struct hvm_viridian_vcpu_context ctxt = { > + .vp_assist_msr = v->arch.hvm_vcpu.viridian.vp_assist.msr.raw, > + .vp_assist_pending = v->arch.hvm_vcpu.viridian.vp_assist.pending, > + }; > > if ( !is_viridian_domain(d) ) > return 0; > > - for_each_vcpu( d, v ) { > - struct hvm_viridian_vcpu_context ctxt; > - > - viridian_save_vcpu_ctxt_one(v, &ctxt); > - > - if ( hvm_save_entry(VIRIDIAN_VCPU, v->vcpu_id, h, &ctxt) != 0 ) > - return 1; > - } > + if ( hvm_save_entry(VIRIDIAN_VCPU, v->vcpu_id, h, &ctxt) != 0 ) > + return 1; > > return 0; > } > @@ -1082,7 +1074,7 @@ static int viridian_load_vcpu_ctxt(struct domain *d, > hvm_domain_context_t *h) > return 0; > } > > -HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_VCPU, viridian_save_vcpu_ctxt, > +HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_VCPU, > viridian_save_vcpu_ctxt_one, > viridian_load_vcpu_ctxt, 1, HVMSR_PER_VCPU); > > static int __init parse_viridian_version(const char *arg) > diff --git a/xen/arch/x86/hvm/vlapic.c b/xen/arch/x86/hvm/vlapic.c > index 1b9f00a..ab35400 100644 > --- a/xen/arch/x86/hvm/vlapic.c > +++ b/xen/arch/x86/hvm/vlapic.c > @@ -1435,43 +1435,35 @@ static void lapic_rearm(struct vlapic *s) > s->timer_last_update = s->pt.last_plt_gtime; > } > > -static int lapic_save_hidden(struct domain *d, hvm_domain_context_t *h) > +static int lapic_save_hidden(struct vcpu *v, hvm_domain_context_t *h) > { > - struct vcpu *v; > + struct domain *d = v->domain; > struct vlapic *s; > int rc = 0; > > if ( !has_vlapic(d) ) > return 0; > > - for_each_vcpu ( d, v ) > - { > - s = vcpu_vlapic(v); > - if ( (rc = hvm_save_entry(LAPIC, v->vcpu_id, h, &s->hw)) != 0 ) > - break; > - } > + s = vcpu_vlapic(v); > + rc = hvm_save_entry(LAPIC, v->vcpu_id, h, &s->hw); > > return rc; > } > > -static int lapic_save_regs(struct domain *d, hvm_domain_context_t *h) > +static int lapic_save_regs(struct vcpu *v, hvm_domain_context_t *h) > { > - struct vcpu *v; > + struct domain *d = v->domain; > struct vlapic *s; > int rc = 0; > > if ( !has_vlapic(d) ) > return 0; > > - for_each_vcpu ( d, v ) > - { > - if ( hvm_funcs.sync_pir_to_irr ) > - hvm_funcs.sync_pir_to_irr(v); > + if ( hvm_funcs.sync_pir_to_irr ) > + hvm_funcs.sync_pir_to_irr(v); > > - s = vcpu_vlapic(v); > - if ( (rc = hvm_save_entry(LAPIC_REGS, v->vcpu_id, h, s->regs)) != 0 ) > - break; > - } > + s = vcpu_vlapic(v); > + rc = hvm_save_entry(LAPIC_REGS, v->vcpu_id, h, s->regs); > > return rc; > } > diff --git a/xen/arch/x86/hvm/vpic.c b/xen/arch/x86/hvm/vpic.c > index e160bbd..bad5066 100644 > --- a/xen/arch/x86/hvm/vpic.c > +++ b/xen/arch/x86/hvm/vpic.c > @@ -371,8 +371,9 @@ static int vpic_intercept_elcr_io( > return X86EMUL_OKAY; > } > > -static int vpic_save(struct domain *d, hvm_domain_context_t *h) > +static int vpic_save(struct vcpu *v, hvm_domain_context_t *h) > { > + struct domain *d = v->domain; > struct hvm_hw_vpic *s; > int i; > > diff --git a/xen/include/asm-x86/hvm/save.h b/xen/include/asm- > x86/hvm/save.h > index f889e8f..fe642ab 100644 > --- a/xen/include/asm-x86/hvm/save.h > +++ b/xen/include/asm-x86/hvm/save.h > @@ -95,7 +95,7 @@ static inline uint16_t hvm_load_instance(struct > hvm_domain_context *h) > * The save handler may save multiple instances of a type into the buffer; > * the load handler will be called once for each instance found when > * restoring. Both return non-zero on error. */ > -typedef int (*hvm_save_handler) (struct domain *d, > +typedef int (*hvm_save_handler) (struct vcpu *v, > hvm_domain_context_t *h); > typedef int (*hvm_load_handler) (struct domain *d, > hvm_domain_context_t *h); > -- > 2.7.4 _______________________________________________ Xen-devel mailing list Xen-devel@lists.xenproject.org https://lists.xenproject.org/mailman/listinfo/xen-devel ^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: [PATCH v5 8/8] x86/domctl: Don't pause the whole domain if only getting vcpu state 2018-05-29 14:58 ` [PATCH v5 8/8] x86/domctl: Don't pause the whole domain if only getting vcpu state Alexandru Isaila 2018-06-04 9:21 ` Paul Durrant @ 2018-06-06 12:35 ` Jan Beulich 1 sibling, 0 replies; 12+ messages in thread From: Jan Beulich @ 2018-06-06 12:35 UTC (permalink / raw) To: aisaila; +Cc: Andrew Cooper, Paul Durrant, Wei Liu, Ian Jackson, xen-devel >>> On 29.05.18 at 16:58, <aisaila@bitdefender.com> wrote: > This patch is focused on merging the *save() funcs to the *save_one() > funcs to remove redundancy. Also the for loop is moved to the caller so > now we can save info for a single vcpu instance. > > Signed-off-by: Alexandru Isaila <aisaila@bitdefender.com> > > --- > Changes since V4: > - Save funcs are deleted > - vcpu si added to the param list > - save_one funcs are now static > - Add blank line after "/* save mtrr&pat */" comment > --- > xen/arch/x86/cpu/mcheck/vmce.c | 28 ++-- > xen/arch/x86/hvm/hpet.c | 3 +- > xen/arch/x86/hvm/hvm.c | 287 +++++++++++++++++------------------------ > xen/arch/x86/hvm/i8254.c | 3 +- > xen/arch/x86/hvm/irq.c | 9 +- > xen/arch/x86/hvm/mtrr.c | 36 ++---- > xen/arch/x86/hvm/pmtimer.c | 3 +- > xen/arch/x86/hvm/rtc.c | 3 +- > xen/arch/x86/hvm/save.c | 125 +++++++++++++----- > xen/arch/x86/hvm/vioapic.c | 3 +- > xen/arch/x86/hvm/viridian.c | 30 ++--- > xen/arch/x86/hvm/vlapic.c | 28 ++-- > xen/arch/x86/hvm/vpic.c | 3 +- > xen/include/asm-x86/hvm/save.h | 2 +- > 14 files changed, 271 insertions(+), 292 deletions(-) Since I see you've already sent v6: This patch is still way too large for its subject. It should, as said before, in particular have no need to touch most of the files above. IOW - there's more splitting work to be done afaict. Jan _______________________________________________ Xen-devel mailing list Xen-devel@lists.xenproject.org https://lists.xenproject.org/mailman/listinfo/xen-devel ^ permalink raw reply [flat|nested] 12+ messages in thread
end of thread, other threads:[~2018-06-06 12:35 UTC | newest] Thread overview: 12+ messages (download: mbox.gz / follow: Atom feed) -- links below jump to the message on this page -- 2018-05-29 14:58 [PATCH v5 1/8] x86/cpu: Introduce vmce_save_vcpu_ctxt_one() func Alexandru Isaila 2018-05-29 14:58 ` [PATCH v5 2/8] x86/hvm: Introduce hvm_save_tsc_adjust_one() func Alexandru Isaila 2018-05-29 14:58 ` [PATCH v5 3/8] x86/hvm: Introduce hvm_save_cpu_ctxt_one func Alexandru Isaila 2018-05-29 14:58 ` [PATCH v5 4/8] x86/hvm: Introduce hvm_save_cpu_xsave_states_one Alexandru Isaila 2018-05-29 14:58 ` [PATCH v5 5/8] x86/hvm: Introduce hvm_save_cpu_msrs_one func Alexandru Isaila 2018-06-04 9:08 ` Paul Durrant 2018-05-29 14:58 ` [PATCH v5 6/8] x86/hvm: Introduce hvm_save_mtrr_msr_one func Alexandru Isaila 2018-05-29 14:58 ` [PATCH v5 7/8] x86/hvm: Introduce viridian_save_vcpu_ctxt_one() func Alexandru Isaila 2018-06-04 9:11 ` Paul Durrant 2018-05-29 14:58 ` [PATCH v5 8/8] x86/domctl: Don't pause the whole domain if only getting vcpu state Alexandru Isaila 2018-06-04 9:21 ` Paul Durrant 2018-06-06 12:35 ` Jan Beulich
This is an external index of several public inboxes, see mirroring instructions on how to clone and mirror all data and code used by this external index.