* [PATCH v5 1/8] x86/cpu: Introduce vmce_save_vcpu_ctxt_one() func
@ 2018-05-29 14:58 Alexandru Isaila
2018-05-29 14:58 ` [PATCH v5 2/8] x86/hvm: Introduce hvm_save_tsc_adjust_one() func Alexandru Isaila
` (6 more replies)
0 siblings, 7 replies; 12+ messages in thread
From: Alexandru Isaila @ 2018-05-29 14:58 UTC (permalink / raw)
To: xen-devel
Cc: wei.liu2, andrew.cooper3, ian.jackson, paul.durrant, jbeulich,
Alexandru Isaila
This is used to save data from a single instance.
Signed-off-by: Alexandru Isaila <aisaila@bitdefender.com>
---
xen/arch/x86/cpu/mcheck/vmce.c | 16 ++++++++++------
1 file changed, 10 insertions(+), 6 deletions(-)
diff --git a/xen/arch/x86/cpu/mcheck/vmce.c b/xen/arch/x86/cpu/mcheck/vmce.c
index e07cd2f..404f27e 100644
--- a/xen/arch/x86/cpu/mcheck/vmce.c
+++ b/xen/arch/x86/cpu/mcheck/vmce.c
@@ -349,6 +349,14 @@ int vmce_wrmsr(uint32_t msr, uint64_t val)
return ret;
}
+static void vmce_save_vcpu_ctxt_one(struct vcpu *v, struct hvm_vmce_vcpu *ctxt)
+{
+ ctxt->caps = v->arch.vmce.mcg_cap;
+ ctxt->mci_ctl2_bank0 = v->arch.vmce.bank[0].mci_ctl2;
+ ctxt->mci_ctl2_bank1 = v->arch.vmce.bank[1].mci_ctl2;
+ ctxt->mcg_ext_ctl = v->arch.vmce.mcg_ext_ctl;
+}
+
static int vmce_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h)
{
struct vcpu *v;
@@ -356,13 +364,9 @@ static int vmce_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h)
for_each_vcpu ( d, v )
{
- struct hvm_vmce_vcpu ctxt = {
- .caps = v->arch.vmce.mcg_cap,
- .mci_ctl2_bank0 = v->arch.vmce.bank[0].mci_ctl2,
- .mci_ctl2_bank1 = v->arch.vmce.bank[1].mci_ctl2,
- .mcg_ext_ctl = v->arch.vmce.mcg_ext_ctl,
- };
+ struct hvm_vmce_vcpu ctxt;
+ vmce_save_vcpu_ctxt_one(v, &ctxt);
err = hvm_save_entry(VMCE_VCPU, v->vcpu_id, h, &ctxt);
if ( err )
break;
--
2.7.4
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel
^ permalink raw reply related [flat|nested] 12+ messages in thread
* [PATCH v5 2/8] x86/hvm: Introduce hvm_save_tsc_adjust_one() func
2018-05-29 14:58 [PATCH v5 1/8] x86/cpu: Introduce vmce_save_vcpu_ctxt_one() func Alexandru Isaila
@ 2018-05-29 14:58 ` Alexandru Isaila
2018-05-29 14:58 ` [PATCH v5 3/8] x86/hvm: Introduce hvm_save_cpu_ctxt_one func Alexandru Isaila
` (5 subsequent siblings)
6 siblings, 0 replies; 12+ messages in thread
From: Alexandru Isaila @ 2018-05-29 14:58 UTC (permalink / raw)
To: xen-devel
Cc: wei.liu2, andrew.cooper3, ian.jackson, paul.durrant, jbeulich,
Alexandru Isaila
This is used to save data from a single instance.
Signed-off-by: Alexandru Isaila <aisaila@bitdefender.com>
---
xen/arch/x86/hvm/hvm.c | 7 ++++++-
1 file changed, 6 insertions(+), 1 deletion(-)
diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c
index c23983c..76f7db9 100644
--- a/xen/arch/x86/hvm/hvm.c
+++ b/xen/arch/x86/hvm/hvm.c
@@ -740,6 +740,11 @@ void hvm_domain_destroy(struct domain *d)
destroy_vpci_mmcfg(d);
}
+static void hvm_save_tsc_adjust_one(struct vcpu *v, struct hvm_tsc_adjust *ctxt)
+{
+ ctxt->tsc_adjust = v->arch.hvm_vcpu.msr_tsc_adjust;
+}
+
static int hvm_save_tsc_adjust(struct domain *d, hvm_domain_context_t *h)
{
struct vcpu *v;
@@ -748,7 +753,7 @@ static int hvm_save_tsc_adjust(struct domain *d, hvm_domain_context_t *h)
for_each_vcpu ( d, v )
{
- ctxt.tsc_adjust = v->arch.hvm_vcpu.msr_tsc_adjust;
+ hvm_save_tsc_adjust_one(v, &ctxt);
err = hvm_save_entry(TSC_ADJUST, v->vcpu_id, h, &ctxt);
if ( err )
break;
--
2.7.4
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel
^ permalink raw reply related [flat|nested] 12+ messages in thread
* [PATCH v5 3/8] x86/hvm: Introduce hvm_save_cpu_ctxt_one func
2018-05-29 14:58 [PATCH v5 1/8] x86/cpu: Introduce vmce_save_vcpu_ctxt_one() func Alexandru Isaila
2018-05-29 14:58 ` [PATCH v5 2/8] x86/hvm: Introduce hvm_save_tsc_adjust_one() func Alexandru Isaila
@ 2018-05-29 14:58 ` Alexandru Isaila
2018-05-29 14:58 ` [PATCH v5 4/8] x86/hvm: Introduce hvm_save_cpu_xsave_states_one Alexandru Isaila
` (4 subsequent siblings)
6 siblings, 0 replies; 12+ messages in thread
From: Alexandru Isaila @ 2018-05-29 14:58 UTC (permalink / raw)
To: xen-devel
Cc: wei.liu2, andrew.cooper3, ian.jackson, paul.durrant, jbeulich,
Alexandru Isaila
This is used to save data from a single instance.
Signed-off-by: Alexandru Isaila <aisaila@bitdefender.com>
---
xen/arch/x86/hvm/hvm.c | 195 +++++++++++++++++++++++++------------------------
1 file changed, 100 insertions(+), 95 deletions(-)
diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c
index 76f7db9..b254378 100644
--- a/xen/arch/x86/hvm/hvm.c
+++ b/xen/arch/x86/hvm/hvm.c
@@ -785,11 +785,109 @@ static int hvm_load_tsc_adjust(struct domain *d, hvm_domain_context_t *h)
HVM_REGISTER_SAVE_RESTORE(TSC_ADJUST, hvm_save_tsc_adjust,
hvm_load_tsc_adjust, 1, HVMSR_PER_VCPU);
+static void hvm_save_cpu_ctxt_one(struct vcpu *v, struct hvm_hw_cpu *ctxt)
+{
+ struct segment_register seg;
+
+ /* Architecture-specific vmcs/vmcb bits */
+ hvm_funcs.save_cpu_ctxt(v, ctxt);
+
+ ctxt->tsc = hvm_get_guest_tsc_fixed(v, v->domain->arch.hvm_domain.sync_tsc);
+
+ ctxt->msr_tsc_aux = hvm_msr_tsc_aux(v);
+
+ hvm_get_segment_register(v, x86_seg_idtr, &seg);
+ ctxt->idtr_limit = seg.limit;
+ ctxt->idtr_base = seg.base;
+
+ hvm_get_segment_register(v, x86_seg_gdtr, &seg);
+ ctxt->gdtr_limit = seg.limit;
+ ctxt->gdtr_base = seg.base;
+
+ hvm_get_segment_register(v, x86_seg_cs, &seg);
+ ctxt->cs_sel = seg.sel;
+ ctxt->cs_limit = seg.limit;
+ ctxt->cs_base = seg.base;
+ ctxt->cs_arbytes = seg.attr;
+
+ hvm_get_segment_register(v, x86_seg_ds, &seg);
+ ctxt->ds_sel = seg.sel;
+ ctxt->ds_limit = seg.limit;
+ ctxt->ds_base = seg.base;
+ ctxt->ds_arbytes = seg.attr;
+
+ hvm_get_segment_register(v, x86_seg_es, &seg);
+ ctxt->es_sel = seg.sel;
+ ctxt->es_limit = seg.limit;
+ ctxt->es_base = seg.base;
+ ctxt->es_arbytes = seg.attr;
+
+ hvm_get_segment_register(v, x86_seg_ss, &seg);
+ ctxt->ss_sel = seg.sel;
+ ctxt->ss_limit = seg.limit;
+ ctxt->ss_base = seg.base;
+ ctxt->ss_arbytes = seg.attr;
+
+ hvm_get_segment_register(v, x86_seg_fs, &seg);
+ ctxt->fs_sel = seg.sel;
+ ctxt->fs_limit = seg.limit;
+ ctxt->fs_base = seg.base;
+ ctxt->fs_arbytes = seg.attr;
+
+ hvm_get_segment_register(v, x86_seg_gs, &seg);
+ ctxt->gs_sel = seg.sel;
+ ctxt->gs_limit = seg.limit;
+ ctxt->gs_base = seg.base;
+ ctxt->gs_arbytes = seg.attr;
+
+ hvm_get_segment_register(v, x86_seg_tr, &seg);
+ ctxt->tr_sel = seg.sel;
+ ctxt->tr_limit = seg.limit;
+ ctxt->tr_base = seg.base;
+ ctxt->tr_arbytes = seg.attr;
+
+ hvm_get_segment_register(v, x86_seg_ldtr, &seg);
+ ctxt->ldtr_sel = seg.sel;
+ ctxt->ldtr_limit = seg.limit;
+ ctxt->ldtr_base = seg.base;
+ ctxt->ldtr_arbytes = seg.attr;
+
+ if ( v->fpu_initialised )
+ {
+ memcpy(ctxt->fpu_regs, v->arch.fpu_ctxt, sizeof(ctxt->fpu_regs));
+ ctxt->flags = XEN_X86_FPU_INITIALISED;
+ }
+
+ ctxt->rax = v->arch.user_regs.rax;
+ ctxt->rbx = v->arch.user_regs.rbx;
+ ctxt->rcx = v->arch.user_regs.rcx;
+ ctxt->rdx = v->arch.user_regs.rdx;
+ ctxt->rbp = v->arch.user_regs.rbp;
+ ctxt->rsi = v->arch.user_regs.rsi;
+ ctxt->rdi = v->arch.user_regs.rdi;
+ ctxt->rsp = v->arch.user_regs.rsp;
+ ctxt->rip = v->arch.user_regs.rip;
+ ctxt->rflags = v->arch.user_regs.rflags;
+ ctxt->r8 = v->arch.user_regs.r8;
+ ctxt->r9 = v->arch.user_regs.r9;
+ ctxt->r10 = v->arch.user_regs.r10;
+ ctxt->r11 = v->arch.user_regs.r11;
+ ctxt->r12 = v->arch.user_regs.r12;
+ ctxt->r13 = v->arch.user_regs.r13;
+ ctxt->r14 = v->arch.user_regs.r14;
+ ctxt->r15 = v->arch.user_regs.r15;
+ ctxt->dr0 = v->arch.debugreg[0];
+ ctxt->dr1 = v->arch.debugreg[1];
+ ctxt->dr2 = v->arch.debugreg[2];
+ ctxt->dr3 = v->arch.debugreg[3];
+ ctxt->dr6 = v->arch.debugreg[6];
+ ctxt->dr7 = v->arch.debugreg[7];
+}
+
static int hvm_save_cpu_ctxt(struct domain *d, hvm_domain_context_t *h)
{
struct vcpu *v;
struct hvm_hw_cpu ctxt;
- struct segment_register seg;
for_each_vcpu ( d, v )
{
@@ -799,100 +897,7 @@ static int hvm_save_cpu_ctxt(struct domain *d, hvm_domain_context_t *h)
continue;
memset(&ctxt, 0, sizeof(ctxt));
-
- /* Architecture-specific vmcs/vmcb bits */
- hvm_funcs.save_cpu_ctxt(v, &ctxt);
-
- ctxt.tsc = hvm_get_guest_tsc_fixed(v, d->arch.hvm_domain.sync_tsc);
-
- ctxt.msr_tsc_aux = hvm_msr_tsc_aux(v);
-
- hvm_get_segment_register(v, x86_seg_idtr, &seg);
- ctxt.idtr_limit = seg.limit;
- ctxt.idtr_base = seg.base;
-
- hvm_get_segment_register(v, x86_seg_gdtr, &seg);
- ctxt.gdtr_limit = seg.limit;
- ctxt.gdtr_base = seg.base;
-
- hvm_get_segment_register(v, x86_seg_cs, &seg);
- ctxt.cs_sel = seg.sel;
- ctxt.cs_limit = seg.limit;
- ctxt.cs_base = seg.base;
- ctxt.cs_arbytes = seg.attr;
-
- hvm_get_segment_register(v, x86_seg_ds, &seg);
- ctxt.ds_sel = seg.sel;
- ctxt.ds_limit = seg.limit;
- ctxt.ds_base = seg.base;
- ctxt.ds_arbytes = seg.attr;
-
- hvm_get_segment_register(v, x86_seg_es, &seg);
- ctxt.es_sel = seg.sel;
- ctxt.es_limit = seg.limit;
- ctxt.es_base = seg.base;
- ctxt.es_arbytes = seg.attr;
-
- hvm_get_segment_register(v, x86_seg_ss, &seg);
- ctxt.ss_sel = seg.sel;
- ctxt.ss_limit = seg.limit;
- ctxt.ss_base = seg.base;
- ctxt.ss_arbytes = seg.attr;
-
- hvm_get_segment_register(v, x86_seg_fs, &seg);
- ctxt.fs_sel = seg.sel;
- ctxt.fs_limit = seg.limit;
- ctxt.fs_base = seg.base;
- ctxt.fs_arbytes = seg.attr;
-
- hvm_get_segment_register(v, x86_seg_gs, &seg);
- ctxt.gs_sel = seg.sel;
- ctxt.gs_limit = seg.limit;
- ctxt.gs_base = seg.base;
- ctxt.gs_arbytes = seg.attr;
-
- hvm_get_segment_register(v, x86_seg_tr, &seg);
- ctxt.tr_sel = seg.sel;
- ctxt.tr_limit = seg.limit;
- ctxt.tr_base = seg.base;
- ctxt.tr_arbytes = seg.attr;
-
- hvm_get_segment_register(v, x86_seg_ldtr, &seg);
- ctxt.ldtr_sel = seg.sel;
- ctxt.ldtr_limit = seg.limit;
- ctxt.ldtr_base = seg.base;
- ctxt.ldtr_arbytes = seg.attr;
-
- if ( v->fpu_initialised )
- {
- memcpy(ctxt.fpu_regs, v->arch.fpu_ctxt, sizeof(ctxt.fpu_regs));
- ctxt.flags = XEN_X86_FPU_INITIALISED;
- }
-
- ctxt.rax = v->arch.user_regs.rax;
- ctxt.rbx = v->arch.user_regs.rbx;
- ctxt.rcx = v->arch.user_regs.rcx;
- ctxt.rdx = v->arch.user_regs.rdx;
- ctxt.rbp = v->arch.user_regs.rbp;
- ctxt.rsi = v->arch.user_regs.rsi;
- ctxt.rdi = v->arch.user_regs.rdi;
- ctxt.rsp = v->arch.user_regs.rsp;
- ctxt.rip = v->arch.user_regs.rip;
- ctxt.rflags = v->arch.user_regs.rflags;
- ctxt.r8 = v->arch.user_regs.r8;
- ctxt.r9 = v->arch.user_regs.r9;
- ctxt.r10 = v->arch.user_regs.r10;
- ctxt.r11 = v->arch.user_regs.r11;
- ctxt.r12 = v->arch.user_regs.r12;
- ctxt.r13 = v->arch.user_regs.r13;
- ctxt.r14 = v->arch.user_regs.r14;
- ctxt.r15 = v->arch.user_regs.r15;
- ctxt.dr0 = v->arch.debugreg[0];
- ctxt.dr1 = v->arch.debugreg[1];
- ctxt.dr2 = v->arch.debugreg[2];
- ctxt.dr3 = v->arch.debugreg[3];
- ctxt.dr6 = v->arch.debugreg[6];
- ctxt.dr7 = v->arch.debugreg[7];
+ hvm_save_cpu_ctxt_one(v, &ctxt);
if ( hvm_save_entry(CPU, v->vcpu_id, h, &ctxt) != 0 )
return 1;
--
2.7.4
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel
^ permalink raw reply related [flat|nested] 12+ messages in thread
* [PATCH v5 4/8] x86/hvm: Introduce hvm_save_cpu_xsave_states_one
2018-05-29 14:58 [PATCH v5 1/8] x86/cpu: Introduce vmce_save_vcpu_ctxt_one() func Alexandru Isaila
2018-05-29 14:58 ` [PATCH v5 2/8] x86/hvm: Introduce hvm_save_tsc_adjust_one() func Alexandru Isaila
2018-05-29 14:58 ` [PATCH v5 3/8] x86/hvm: Introduce hvm_save_cpu_ctxt_one func Alexandru Isaila
@ 2018-05-29 14:58 ` Alexandru Isaila
2018-05-29 14:58 ` [PATCH v5 5/8] x86/hvm: Introduce hvm_save_cpu_msrs_one func Alexandru Isaila
` (3 subsequent siblings)
6 siblings, 0 replies; 12+ messages in thread
From: Alexandru Isaila @ 2018-05-29 14:58 UTC (permalink / raw)
To: xen-devel
Cc: wei.liu2, andrew.cooper3, ian.jackson, paul.durrant, jbeulich,
Alexandru Isaila
This is used to save data from a single instance.
Signed-off-by: Alexandru Isaila <aisaila@bitdefender.com>
---
xen/arch/x86/hvm/hvm.c | 11 ++++++++---
1 file changed, 8 insertions(+), 3 deletions(-)
diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c
index b254378..e8ecabf 100644
--- a/xen/arch/x86/hvm/hvm.c
+++ b/xen/arch/x86/hvm/hvm.c
@@ -1183,6 +1183,13 @@ HVM_REGISTER_SAVE_RESTORE(CPU, hvm_save_cpu_ctxt, hvm_load_cpu_ctxt,
save_area) + \
xstate_ctxt_size(xcr0))
+static void hvm_save_cpu_xsave_states_one(struct vcpu *v, struct hvm_hw_cpu_xsave *ctxt)
+{
+ ctxt->xfeature_mask = xfeature_mask;
+ ctxt->xcr0 = v->arch.xcr0;
+ ctxt->xcr0_accum = v->arch.xcr0_accum;
+}
+
static int hvm_save_cpu_xsave_states(struct domain *d, hvm_domain_context_t *h)
{
struct vcpu *v;
@@ -1202,9 +1209,7 @@ static int hvm_save_cpu_xsave_states(struct domain *d, hvm_domain_context_t *h)
ctxt = (struct hvm_hw_cpu_xsave *)&h->data[h->cur];
h->cur += size;
- ctxt->xfeature_mask = xfeature_mask;
- ctxt->xcr0 = v->arch.xcr0;
- ctxt->xcr0_accum = v->arch.xcr0_accum;
+ hvm_save_cpu_xsave_states_one(v, ctxt);
expand_xsave_states(v, &ctxt->save_area,
size - offsetof(typeof(*ctxt), save_area));
}
--
2.7.4
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel
^ permalink raw reply related [flat|nested] 12+ messages in thread
* [PATCH v5 5/8] x86/hvm: Introduce hvm_save_cpu_msrs_one func
2018-05-29 14:58 [PATCH v5 1/8] x86/cpu: Introduce vmce_save_vcpu_ctxt_one() func Alexandru Isaila
` (2 preceding siblings ...)
2018-05-29 14:58 ` [PATCH v5 4/8] x86/hvm: Introduce hvm_save_cpu_xsave_states_one Alexandru Isaila
@ 2018-05-29 14:58 ` Alexandru Isaila
2018-06-04 9:08 ` Paul Durrant
2018-05-29 14:58 ` [PATCH v5 6/8] x86/hvm: Introduce hvm_save_mtrr_msr_one func Alexandru Isaila
` (2 subsequent siblings)
6 siblings, 1 reply; 12+ messages in thread
From: Alexandru Isaila @ 2018-05-29 14:58 UTC (permalink / raw)
To: xen-devel
Cc: wei.liu2, andrew.cooper3, ian.jackson, paul.durrant, jbeulich,
Alexandru Isaila
This is used to save data from a single instance.
Signed-off-by: Alexandru Isaila <aisaila@bitdefender.com>
---
xen/arch/x86/hvm/hvm.c | 59 ++++++++++++++++++++++++++++----------------------
1 file changed, 33 insertions(+), 26 deletions(-)
diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c
index e8ecabf..4a22283 100644
--- a/xen/arch/x86/hvm/hvm.c
+++ b/xen/arch/x86/hvm/hvm.c
@@ -1354,6 +1354,38 @@ static const uint32_t msrs_to_send[] = {
};
static unsigned int __read_mostly msr_count_max = ARRAY_SIZE(msrs_to_send);
+static int hvm_save_cpu_msrs_one(struct vcpu *v, struct hvm_msr *ctxt)
+{
+ unsigned int i;
+
+ for ( i = 0; i < ARRAY_SIZE(msrs_to_send); ++i )
+ {
+ uint64_t val;
+ int rc = guest_rdmsr(v, msrs_to_send[i], &val);
+
+ /*
+ * It is the programmers responsibility to ensure that
+ * msrs_to_send[] contain generally-read/write MSRs.
+ * X86EMUL_EXCEPTION here implies a missing feature, and that the
+ * guest doesn't have access to the MSR.
+ */
+ if ( rc == X86EMUL_EXCEPTION )
+ continue;
+
+ if ( rc != X86EMUL_OKAY )
+ {
+ ASSERT_UNREACHABLE();
+ return -ENXIO;
+ }
+
+ if ( !val )
+ continue; /* Skip empty MSRs. */
+ ctxt->msr[ctxt->count].index = msrs_to_send[i];
+ ctxt->msr[ctxt->count++].val = val;
+ }
+ return 0;
+}
+
static int hvm_save_cpu_msrs(struct domain *d, hvm_domain_context_t *h)
{
struct vcpu *v;
@@ -1370,32 +1402,7 @@ static int hvm_save_cpu_msrs(struct domain *d, hvm_domain_context_t *h)
ctxt = (struct hvm_msr *)&h->data[h->cur];
ctxt->count = 0;
- for ( i = 0; i < ARRAY_SIZE(msrs_to_send); ++i )
- {
- uint64_t val;
- int rc = guest_rdmsr(v, msrs_to_send[i], &val);
-
- /*
- * It is the programmers responsibility to ensure that
- * msrs_to_send[] contain generally-read/write MSRs.
- * X86EMUL_EXCEPTION here implies a missing feature, and that the
- * guest doesn't have access to the MSR.
- */
- if ( rc == X86EMUL_EXCEPTION )
- continue;
-
- if ( rc != X86EMUL_OKAY )
- {
- ASSERT_UNREACHABLE();
- return -ENXIO;
- }
-
- if ( !val )
- continue; /* Skip empty MSRs. */
-
- ctxt->msr[ctxt->count].index = msrs_to_send[i];
- ctxt->msr[ctxt->count++].val = val;
- }
+ hvm_save_cpu_msrs_one(v, ctxt);
if ( hvm_funcs.save_msr )
hvm_funcs.save_msr(v, ctxt);
--
2.7.4
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel
^ permalink raw reply related [flat|nested] 12+ messages in thread
* [PATCH v5 6/8] x86/hvm: Introduce hvm_save_mtrr_msr_one func
2018-05-29 14:58 [PATCH v5 1/8] x86/cpu: Introduce vmce_save_vcpu_ctxt_one() func Alexandru Isaila
` (3 preceding siblings ...)
2018-05-29 14:58 ` [PATCH v5 5/8] x86/hvm: Introduce hvm_save_cpu_msrs_one func Alexandru Isaila
@ 2018-05-29 14:58 ` Alexandru Isaila
2018-05-29 14:58 ` [PATCH v5 7/8] x86/hvm: Introduce viridian_save_vcpu_ctxt_one() func Alexandru Isaila
2018-05-29 14:58 ` [PATCH v5 8/8] x86/domctl: Don't pause the whole domain if only getting vcpu state Alexandru Isaila
6 siblings, 0 replies; 12+ messages in thread
From: Alexandru Isaila @ 2018-05-29 14:58 UTC (permalink / raw)
To: xen-devel
Cc: wei.liu2, andrew.cooper3, ian.jackson, paul.durrant, jbeulich,
Alexandru Isaila
This is used to save data from a single instance.
Signed-off-by: Alexandru Isaila <aisaila@bitdefender.com>
---
xen/arch/x86/hvm/mtrr.c | 52 +++++++++++++++++++++++++++----------------------
1 file changed, 29 insertions(+), 23 deletions(-)
diff --git a/xen/arch/x86/hvm/mtrr.c b/xen/arch/x86/hvm/mtrr.c
index b721c63..d311031 100644
--- a/xen/arch/x86/hvm/mtrr.c
+++ b/xen/arch/x86/hvm/mtrr.c
@@ -666,36 +666,42 @@ int hvm_set_mem_pinned_cacheattr(struct domain *d, uint64_t gfn_start,
return 0;
}
-static int hvm_save_mtrr_msr(struct domain *d, hvm_domain_context_t *h)
+static void hvm_save_mtrr_msr_one(struct vcpu *v, struct hvm_hw_mtrr *hw_mtrr)
{
+ struct mtrr_state *mtrr_state = &v->arch.hvm_vcpu.mtrr;
int i;
- struct vcpu *v;
- struct hvm_hw_mtrr hw_mtrr;
- struct mtrr_state *mtrr_state;
- /* save mtrr&pat */
- for_each_vcpu(d, v)
+
+ hvm_get_guest_pat(v, &hw_mtrr->msr_pat_cr);
+
+ hw_mtrr->msr_mtrr_def_type = mtrr_state->def_type
+ | (mtrr_state->enabled << 10);
+ hw_mtrr->msr_mtrr_cap = mtrr_state->mtrr_cap;
+
+ for ( i = 0; i < MTRR_VCNT; i++ )
{
- mtrr_state = &v->arch.hvm_vcpu.mtrr;
+ /* save physbase */
+ hw_mtrr->msr_mtrr_var[i*2] =
+ ((uint64_t*)mtrr_state->var_ranges)[i*2];
+ /* save physmask */
+ hw_mtrr->msr_mtrr_var[i*2+1] =
+ ((uint64_t*)mtrr_state->var_ranges)[i*2+1];
+ }
- hvm_get_guest_pat(v, &hw_mtrr.msr_pat_cr);
+ for ( i = 0; i < NUM_FIXED_MSR; i++ )
+ hw_mtrr->msr_mtrr_fixed[i] =
+ ((uint64_t*)mtrr_state->fixed_ranges)[i];
- hw_mtrr.msr_mtrr_def_type = mtrr_state->def_type
- | (mtrr_state->enabled << 10);
- hw_mtrr.msr_mtrr_cap = mtrr_state->mtrr_cap;
+}
- for ( i = 0; i < MTRR_VCNT; i++ )
- {
- /* save physbase */
- hw_mtrr.msr_mtrr_var[i*2] =
- ((uint64_t*)mtrr_state->var_ranges)[i*2];
- /* save physmask */
- hw_mtrr.msr_mtrr_var[i*2+1] =
- ((uint64_t*)mtrr_state->var_ranges)[i*2+1];
- }
+static int hvm_save_mtrr_msr(struct domain *d, hvm_domain_context_t *h)
+{
+ struct vcpu *v;
+ struct hvm_hw_mtrr hw_mtrr;
+ /* save mtrr&pat */
- for ( i = 0; i < NUM_FIXED_MSR; i++ )
- hw_mtrr.msr_mtrr_fixed[i] =
- ((uint64_t*)mtrr_state->fixed_ranges)[i];
+ for_each_vcpu(d, v)
+ {
+ hvm_save_mtrr_msr_one(v, &hw_mtrr);
if ( hvm_save_entry(MTRR, v->vcpu_id, h, &hw_mtrr) != 0 )
return 1;
--
2.7.4
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel
^ permalink raw reply related [flat|nested] 12+ messages in thread
* [PATCH v5 7/8] x86/hvm: Introduce viridian_save_vcpu_ctxt_one() func
2018-05-29 14:58 [PATCH v5 1/8] x86/cpu: Introduce vmce_save_vcpu_ctxt_one() func Alexandru Isaila
` (4 preceding siblings ...)
2018-05-29 14:58 ` [PATCH v5 6/8] x86/hvm: Introduce hvm_save_mtrr_msr_one func Alexandru Isaila
@ 2018-05-29 14:58 ` Alexandru Isaila
2018-06-04 9:11 ` Paul Durrant
2018-05-29 14:58 ` [PATCH v5 8/8] x86/domctl: Don't pause the whole domain if only getting vcpu state Alexandru Isaila
6 siblings, 1 reply; 12+ messages in thread
From: Alexandru Isaila @ 2018-05-29 14:58 UTC (permalink / raw)
To: xen-devel
Cc: wei.liu2, andrew.cooper3, ian.jackson, paul.durrant, jbeulich,
Alexandru Isaila
This is used to save data from a single instance.
Signed-off-by: Alexandru Isaila <aisaila@bitdefender.com>
---
xen/arch/x86/hvm/viridian.c | 14 ++++++++++----
1 file changed, 10 insertions(+), 4 deletions(-)
diff --git a/xen/arch/x86/hvm/viridian.c b/xen/arch/x86/hvm/viridian.c
index 694eae6..d22c8ac 100644
--- a/xen/arch/x86/hvm/viridian.c
+++ b/xen/arch/x86/hvm/viridian.c
@@ -1026,6 +1026,13 @@ static int viridian_load_domain_ctxt(struct domain *d, hvm_domain_context_t *h)
HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_DOMAIN, viridian_save_domain_ctxt,
viridian_load_domain_ctxt, 1, HVMSR_PER_DOM);
+static void viridian_save_vcpu_ctxt_one(struct vcpu *v, struct hvm_viridian_vcpu_context *ctxt)
+{
+ ctxt->vp_assist_msr = v->arch.hvm_vcpu.viridian.vp_assist.msr.raw;
+ ctxt->vp_assist_pending = v->arch.hvm_vcpu.viridian.vp_assist.pending;
+ *ctxt->_pad = 0;
+}
+
static int viridian_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h)
{
struct vcpu *v;
@@ -1034,10 +1041,9 @@ static int viridian_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h)
return 0;
for_each_vcpu( d, v ) {
- struct hvm_viridian_vcpu_context ctxt = {
- .vp_assist_msr = v->arch.hvm_vcpu.viridian.vp_assist.msr.raw,
- .vp_assist_pending = v->arch.hvm_vcpu.viridian.vp_assist.pending,
- };
+ struct hvm_viridian_vcpu_context ctxt;
+
+ viridian_save_vcpu_ctxt_one(v, &ctxt);
if ( hvm_save_entry(VIRIDIAN_VCPU, v->vcpu_id, h, &ctxt) != 0 )
return 1;
--
2.7.4
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel
^ permalink raw reply related [flat|nested] 12+ messages in thread
* [PATCH v5 8/8] x86/domctl: Don't pause the whole domain if only getting vcpu state
2018-05-29 14:58 [PATCH v5 1/8] x86/cpu: Introduce vmce_save_vcpu_ctxt_one() func Alexandru Isaila
` (5 preceding siblings ...)
2018-05-29 14:58 ` [PATCH v5 7/8] x86/hvm: Introduce viridian_save_vcpu_ctxt_one() func Alexandru Isaila
@ 2018-05-29 14:58 ` Alexandru Isaila
2018-06-04 9:21 ` Paul Durrant
2018-06-06 12:35 ` Jan Beulich
6 siblings, 2 replies; 12+ messages in thread
From: Alexandru Isaila @ 2018-05-29 14:58 UTC (permalink / raw)
To: xen-devel
Cc: wei.liu2, andrew.cooper3, ian.jackson, paul.durrant, jbeulich,
Alexandru Isaila
This patch is focused on merging the *save() funcs to the *save_one()
funcs to remove redundancy. Also the for loop is moved to the caller so
now we can save info for a single vcpu instance.
Signed-off-by: Alexandru Isaila <aisaila@bitdefender.com>
---
Changes since V4:
- Save funcs are deleted
- vcpu si added to the param list
- save_one funcs are now static
- Add blank line after "/* save mtrr&pat */" comment
---
xen/arch/x86/cpu/mcheck/vmce.c | 28 ++--
xen/arch/x86/hvm/hpet.c | 3 +-
xen/arch/x86/hvm/hvm.c | 287 +++++++++++++++++------------------------
xen/arch/x86/hvm/i8254.c | 3 +-
xen/arch/x86/hvm/irq.c | 9 +-
xen/arch/x86/hvm/mtrr.c | 36 ++----
xen/arch/x86/hvm/pmtimer.c | 3 +-
xen/arch/x86/hvm/rtc.c | 3 +-
xen/arch/x86/hvm/save.c | 125 +++++++++++++-----
xen/arch/x86/hvm/vioapic.c | 3 +-
xen/arch/x86/hvm/viridian.c | 30 ++---
xen/arch/x86/hvm/vlapic.c | 28 ++--
xen/arch/x86/hvm/vpic.c | 3 +-
xen/include/asm-x86/hvm/save.h | 2 +-
14 files changed, 271 insertions(+), 292 deletions(-)
diff --git a/xen/arch/x86/cpu/mcheck/vmce.c b/xen/arch/x86/cpu/mcheck/vmce.c
index 404f27e..7bde9ff 100644
--- a/xen/arch/x86/cpu/mcheck/vmce.c
+++ b/xen/arch/x86/cpu/mcheck/vmce.c
@@ -349,29 +349,17 @@ int vmce_wrmsr(uint32_t msr, uint64_t val)
return ret;
}
-static void vmce_save_vcpu_ctxt_one(struct vcpu *v, struct hvm_vmce_vcpu *ctxt)
+static int vmce_save_vcpu_ctxt_one(struct vcpu *v, hvm_domain_context_t *h)
{
- ctxt->caps = v->arch.vmce.mcg_cap;
- ctxt->mci_ctl2_bank0 = v->arch.vmce.bank[0].mci_ctl2;
- ctxt->mci_ctl2_bank1 = v->arch.vmce.bank[1].mci_ctl2;
- ctxt->mcg_ext_ctl = v->arch.vmce.mcg_ext_ctl;
-}
-
-static int vmce_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h)
-{
- struct vcpu *v;
int err = 0;
+ struct hvm_vmce_vcpu ctxt;
- for_each_vcpu ( d, v )
- {
- struct hvm_vmce_vcpu ctxt;
-
- vmce_save_vcpu_ctxt_one(v, &ctxt);
- err = hvm_save_entry(VMCE_VCPU, v->vcpu_id, h, &ctxt);
- if ( err )
- break;
- }
+ ctxt.caps = v->arch.vmce.mcg_cap;
+ ctxt.mci_ctl2_bank0 = v->arch.vmce.bank[0].mci_ctl2;
+ ctxt.mci_ctl2_bank1 = v->arch.vmce.bank[1].mci_ctl2;
+ ctxt.mcg_ext_ctl = v->arch.vmce.mcg_ext_ctl;
+ err = hvm_save_entry(VMCE_VCPU, v->vcpu_id, h, &ctxt);
return err;
}
@@ -394,7 +382,7 @@ static int vmce_load_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h)
return err ?: vmce_restore_vcpu(v, &ctxt);
}
-HVM_REGISTER_SAVE_RESTORE(VMCE_VCPU, vmce_save_vcpu_ctxt,
+HVM_REGISTER_SAVE_RESTORE(VMCE_VCPU, vmce_save_vcpu_ctxt_one,
vmce_load_vcpu_ctxt, 1, HVMSR_PER_VCPU);
/*
diff --git a/xen/arch/x86/hvm/hpet.c b/xen/arch/x86/hvm/hpet.c
index 2837709..3ed6547 100644
--- a/xen/arch/x86/hvm/hpet.c
+++ b/xen/arch/x86/hvm/hpet.c
@@ -516,8 +516,9 @@ static const struct hvm_mmio_ops hpet_mmio_ops = {
};
-static int hpet_save(struct domain *d, hvm_domain_context_t *h)
+static int hpet_save(struct vcpu *vcpu, hvm_domain_context_t *h)
{
+ struct domain *d = vcpu->domain;
HPETState *hp = domain_vhpet(d);
struct vcpu *v = pt_global_vcpu_target(d);
int rc;
diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c
index 4a22283..ea27055 100644
--- a/xen/arch/x86/hvm/hvm.c
+++ b/xen/arch/x86/hvm/hvm.c
@@ -740,24 +740,13 @@ void hvm_domain_destroy(struct domain *d)
destroy_vpci_mmcfg(d);
}
-static void hvm_save_tsc_adjust_one(struct vcpu *v, struct hvm_tsc_adjust *ctxt)
+static int hvm_save_tsc_adjust_one(struct vcpu *v, hvm_domain_context_t *h)
{
- ctxt->tsc_adjust = v->arch.hvm_vcpu.msr_tsc_adjust;
-}
-
-static int hvm_save_tsc_adjust(struct domain *d, hvm_domain_context_t *h)
-{
- struct vcpu *v;
- struct hvm_tsc_adjust ctxt;
int err = 0;
+ struct hvm_tsc_adjust ctxt;
- for_each_vcpu ( d, v )
- {
- hvm_save_tsc_adjust_one(v, &ctxt);
- err = hvm_save_entry(TSC_ADJUST, v->vcpu_id, h, &ctxt);
- if ( err )
- break;
- }
+ ctxt.tsc_adjust = v->arch.hvm_vcpu.msr_tsc_adjust;
+ err = hvm_save_entry(TSC_ADJUST, v->vcpu_id, h, &ctxt);
return err;
}
@@ -782,126 +771,116 @@ static int hvm_load_tsc_adjust(struct domain *d, hvm_domain_context_t *h)
return 0;
}
-HVM_REGISTER_SAVE_RESTORE(TSC_ADJUST, hvm_save_tsc_adjust,
+HVM_REGISTER_SAVE_RESTORE(TSC_ADJUST, hvm_save_tsc_adjust_one,
hvm_load_tsc_adjust, 1, HVMSR_PER_VCPU);
-static void hvm_save_cpu_ctxt_one(struct vcpu *v, struct hvm_hw_cpu *ctxt)
+static int hvm_save_cpu_ctxt_one(struct vcpu *v, hvm_domain_context_t *h)
{
struct segment_register seg;
+ struct hvm_hw_cpu ctxt = {};
+
+ /* We don't need to save state for a vcpu that is down; the restore
+ * code will leave it down if there is nothing saved. */
+ if ( v->pause_flags & VPF_down )
+ return 2;
/* Architecture-specific vmcs/vmcb bits */
- hvm_funcs.save_cpu_ctxt(v, ctxt);
+ hvm_funcs.save_cpu_ctxt(v, &ctxt);
- ctxt->tsc = hvm_get_guest_tsc_fixed(v, v->domain->arch.hvm_domain.sync_tsc);
+ ctxt.tsc = hvm_get_guest_tsc_fixed(v, v->domain->arch.hvm_domain.sync_tsc);
- ctxt->msr_tsc_aux = hvm_msr_tsc_aux(v);
+ ctxt.msr_tsc_aux = hvm_msr_tsc_aux(v);
hvm_get_segment_register(v, x86_seg_idtr, &seg);
- ctxt->idtr_limit = seg.limit;
- ctxt->idtr_base = seg.base;
+ ctxt.idtr_limit = seg.limit;
+ ctxt.idtr_base = seg.base;
hvm_get_segment_register(v, x86_seg_gdtr, &seg);
- ctxt->gdtr_limit = seg.limit;
- ctxt->gdtr_base = seg.base;
+ ctxt.gdtr_limit = seg.limit;
+ ctxt.gdtr_base = seg.base;
hvm_get_segment_register(v, x86_seg_cs, &seg);
- ctxt->cs_sel = seg.sel;
- ctxt->cs_limit = seg.limit;
- ctxt->cs_base = seg.base;
- ctxt->cs_arbytes = seg.attr;
+ ctxt.cs_sel = seg.sel;
+ ctxt.cs_limit = seg.limit;
+ ctxt.cs_base = seg.base;
+ ctxt.cs_arbytes = seg.attr;
hvm_get_segment_register(v, x86_seg_ds, &seg);
- ctxt->ds_sel = seg.sel;
- ctxt->ds_limit = seg.limit;
- ctxt->ds_base = seg.base;
- ctxt->ds_arbytes = seg.attr;
+ ctxt.ds_sel = seg.sel;
+ ctxt.ds_limit = seg.limit;
+ ctxt.ds_base = seg.base;
+ ctxt.ds_arbytes = seg.attr;
hvm_get_segment_register(v, x86_seg_es, &seg);
- ctxt->es_sel = seg.sel;
- ctxt->es_limit = seg.limit;
- ctxt->es_base = seg.base;
- ctxt->es_arbytes = seg.attr;
+ ctxt.es_sel = seg.sel;
+ ctxt.es_limit = seg.limit;
+ ctxt.es_base = seg.base;
+ ctxt.es_arbytes = seg.attr;
hvm_get_segment_register(v, x86_seg_ss, &seg);
- ctxt->ss_sel = seg.sel;
- ctxt->ss_limit = seg.limit;
- ctxt->ss_base = seg.base;
- ctxt->ss_arbytes = seg.attr;
+ ctxt.ss_sel = seg.sel;
+ ctxt.ss_limit = seg.limit;
+ ctxt.ss_base = seg.base;
+ ctxt.ss_arbytes = seg.attr;
hvm_get_segment_register(v, x86_seg_fs, &seg);
- ctxt->fs_sel = seg.sel;
- ctxt->fs_limit = seg.limit;
- ctxt->fs_base = seg.base;
- ctxt->fs_arbytes = seg.attr;
+ ctxt.fs_sel = seg.sel;
+ ctxt.fs_limit = seg.limit;
+ ctxt.fs_base = seg.base;
+ ctxt.fs_arbytes = seg.attr;
hvm_get_segment_register(v, x86_seg_gs, &seg);
- ctxt->gs_sel = seg.sel;
- ctxt->gs_limit = seg.limit;
- ctxt->gs_base = seg.base;
- ctxt->gs_arbytes = seg.attr;
+ ctxt.gs_sel = seg.sel;
+ ctxt.gs_limit = seg.limit;
+ ctxt.gs_base = seg.base;
+ ctxt.gs_arbytes = seg.attr;
hvm_get_segment_register(v, x86_seg_tr, &seg);
- ctxt->tr_sel = seg.sel;
- ctxt->tr_limit = seg.limit;
- ctxt->tr_base = seg.base;
- ctxt->tr_arbytes = seg.attr;
+ ctxt.tr_sel = seg.sel;
+ ctxt.tr_limit = seg.limit;
+ ctxt.tr_base = seg.base;
+ ctxt.tr_arbytes = seg.attr;
hvm_get_segment_register(v, x86_seg_ldtr, &seg);
- ctxt->ldtr_sel = seg.sel;
- ctxt->ldtr_limit = seg.limit;
- ctxt->ldtr_base = seg.base;
- ctxt->ldtr_arbytes = seg.attr;
+ ctxt.ldtr_sel = seg.sel;
+ ctxt.ldtr_limit = seg.limit;
+ ctxt.ldtr_base = seg.base;
+ ctxt.ldtr_arbytes = seg.attr;
if ( v->fpu_initialised )
{
- memcpy(ctxt->fpu_regs, v->arch.fpu_ctxt, sizeof(ctxt->fpu_regs));
- ctxt->flags = XEN_X86_FPU_INITIALISED;
- }
-
- ctxt->rax = v->arch.user_regs.rax;
- ctxt->rbx = v->arch.user_regs.rbx;
- ctxt->rcx = v->arch.user_regs.rcx;
- ctxt->rdx = v->arch.user_regs.rdx;
- ctxt->rbp = v->arch.user_regs.rbp;
- ctxt->rsi = v->arch.user_regs.rsi;
- ctxt->rdi = v->arch.user_regs.rdi;
- ctxt->rsp = v->arch.user_regs.rsp;
- ctxt->rip = v->arch.user_regs.rip;
- ctxt->rflags = v->arch.user_regs.rflags;
- ctxt->r8 = v->arch.user_regs.r8;
- ctxt->r9 = v->arch.user_regs.r9;
- ctxt->r10 = v->arch.user_regs.r10;
- ctxt->r11 = v->arch.user_regs.r11;
- ctxt->r12 = v->arch.user_regs.r12;
- ctxt->r13 = v->arch.user_regs.r13;
- ctxt->r14 = v->arch.user_regs.r14;
- ctxt->r15 = v->arch.user_regs.r15;
- ctxt->dr0 = v->arch.debugreg[0];
- ctxt->dr1 = v->arch.debugreg[1];
- ctxt->dr2 = v->arch.debugreg[2];
- ctxt->dr3 = v->arch.debugreg[3];
- ctxt->dr6 = v->arch.debugreg[6];
- ctxt->dr7 = v->arch.debugreg[7];
-}
-
-static int hvm_save_cpu_ctxt(struct domain *d, hvm_domain_context_t *h)
-{
- struct vcpu *v;
- struct hvm_hw_cpu ctxt;
-
- for_each_vcpu ( d, v )
- {
- /* We don't need to save state for a vcpu that is down; the restore
- * code will leave it down if there is nothing saved. */
- if ( v->pause_flags & VPF_down )
- continue;
-
- memset(&ctxt, 0, sizeof(ctxt));
- hvm_save_cpu_ctxt_one(v, &ctxt);
+ memcpy(ctxt.fpu_regs, v->arch.fpu_ctxt, sizeof(ctxt.fpu_regs));
+ ctxt.flags = XEN_X86_FPU_INITIALISED;
+ }
+
+ ctxt.rax = v->arch.user_regs.rax;
+ ctxt.rbx = v->arch.user_regs.rbx;
+ ctxt.rcx = v->arch.user_regs.rcx;
+ ctxt.rdx = v->arch.user_regs.rdx;
+ ctxt.rbp = v->arch.user_regs.rbp;
+ ctxt.rsi = v->arch.user_regs.rsi;
+ ctxt.rdi = v->arch.user_regs.rdi;
+ ctxt.rsp = v->arch.user_regs.rsp;
+ ctxt.rip = v->arch.user_regs.rip;
+ ctxt.rflags = v->arch.user_regs.rflags;
+ ctxt.r8 = v->arch.user_regs.r8;
+ ctxt.r9 = v->arch.user_regs.r9;
+ ctxt.r10 = v->arch.user_regs.r10;
+ ctxt.r11 = v->arch.user_regs.r11;
+ ctxt.r12 = v->arch.user_regs.r12;
+ ctxt.r13 = v->arch.user_regs.r13;
+ ctxt.r14 = v->arch.user_regs.r14;
+ ctxt.r15 = v->arch.user_regs.r15;
+ ctxt.dr0 = v->arch.debugreg[0];
+ ctxt.dr1 = v->arch.debugreg[1];
+ ctxt.dr2 = v->arch.debugreg[2];
+ ctxt.dr3 = v->arch.debugreg[3];
+ ctxt.dr6 = v->arch.debugreg[6];
+ ctxt.dr7 = v->arch.debugreg[7];
+
+ if ( hvm_save_entry(CPU, v->vcpu_id, h, &ctxt) != 0 )
+ return 1;
- if ( hvm_save_entry(CPU, v->vcpu_id, h, &ctxt) != 0 )
- return 1;
- }
return 0;
}
@@ -1176,43 +1155,29 @@ static int hvm_load_cpu_ctxt(struct domain *d, hvm_domain_context_t *h)
return 0;
}
-HVM_REGISTER_SAVE_RESTORE(CPU, hvm_save_cpu_ctxt, hvm_load_cpu_ctxt,
+HVM_REGISTER_SAVE_RESTORE(CPU, hvm_save_cpu_ctxt_one, hvm_load_cpu_ctxt,
1, HVMSR_PER_VCPU);
#define HVM_CPU_XSAVE_SIZE(xcr0) (offsetof(struct hvm_hw_cpu_xsave, \
save_area) + \
xstate_ctxt_size(xcr0))
-static void hvm_save_cpu_xsave_states_one(struct vcpu *v, struct hvm_hw_cpu_xsave *ctxt)
+static int hvm_save_cpu_xsave_states_one(struct vcpu *v, hvm_domain_context_t *h)
{
+ unsigned int size = HVM_CPU_XSAVE_SIZE(v->arch.xcr0_accum);
+ struct hvm_hw_cpu_xsave *ctxt;
+
+ if ( !xsave_enabled(v) )
+ return 2;
+ if ( _hvm_init_entry(h, CPU_XSAVE_CODE, v->vcpu_id, size) )
+ return 1;
+ ctxt = (struct hvm_hw_cpu_xsave *)&h->data[h->cur];
+ h->cur += size;
ctxt->xfeature_mask = xfeature_mask;
ctxt->xcr0 = v->arch.xcr0;
ctxt->xcr0_accum = v->arch.xcr0_accum;
-}
-
-static int hvm_save_cpu_xsave_states(struct domain *d, hvm_domain_context_t *h)
-{
- struct vcpu *v;
- struct hvm_hw_cpu_xsave *ctxt;
-
- if ( !cpu_has_xsave )
- return 0; /* do nothing */
-
- for_each_vcpu ( d, v )
- {
- unsigned int size = HVM_CPU_XSAVE_SIZE(v->arch.xcr0_accum);
-
- if ( !xsave_enabled(v) )
- continue;
- if ( _hvm_init_entry(h, CPU_XSAVE_CODE, v->vcpu_id, size) )
- return 1;
- ctxt = (struct hvm_hw_cpu_xsave *)&h->data[h->cur];
- h->cur += size;
-
- hvm_save_cpu_xsave_states_one(v, ctxt);
- expand_xsave_states(v, &ctxt->save_area,
- size - offsetof(typeof(*ctxt), save_area));
- }
+ expand_xsave_states(v, &ctxt->save_area,
+ size - offsetof(typeof(*ctxt), save_area));
return 0;
}
@@ -1354,10 +1319,18 @@ static const uint32_t msrs_to_send[] = {
};
static unsigned int __read_mostly msr_count_max = ARRAY_SIZE(msrs_to_send);
-static int hvm_save_cpu_msrs_one(struct vcpu *v, struct hvm_msr *ctxt)
+static int hvm_save_cpu_msrs_one(struct vcpu *v, hvm_domain_context_t *h)
{
+
+ struct hvm_save_descriptor *desc = _p(&h->data[h->cur]);
+ struct hvm_msr *ctxt;
unsigned int i;
+ if ( _hvm_init_entry(h, CPU_MSR_CODE, v->vcpu_id,
+ HVM_CPU_MSR_SIZE(msr_count_max)) )
+ return 1;
+ ctxt = (struct hvm_msr *)&h->data[h->cur];
+ ctxt->count = 0;
for ( i = 0; i < ARRAY_SIZE(msrs_to_send); ++i )
{
uint64_t val;
@@ -1383,45 +1356,23 @@ static int hvm_save_cpu_msrs_one(struct vcpu *v, struct hvm_msr *ctxt)
ctxt->msr[ctxt->count].index = msrs_to_send[i];
ctxt->msr[ctxt->count++].val = val;
}
- return 0;
-}
-
-static int hvm_save_cpu_msrs(struct domain *d, hvm_domain_context_t *h)
-{
- struct vcpu *v;
-
- for_each_vcpu ( d, v )
- {
- struct hvm_save_descriptor *desc = _p(&h->data[h->cur]);
- struct hvm_msr *ctxt;
- unsigned int i;
-
- if ( _hvm_init_entry(h, CPU_MSR_CODE, v->vcpu_id,
- HVM_CPU_MSR_SIZE(msr_count_max)) )
- return 1;
- ctxt = (struct hvm_msr *)&h->data[h->cur];
- ctxt->count = 0;
+ if ( hvm_funcs.save_msr )
+ hvm_funcs.save_msr(v, ctxt);
- hvm_save_cpu_msrs_one(v, ctxt);
+ ASSERT(ctxt->count <= msr_count_max);
- if ( hvm_funcs.save_msr )
- hvm_funcs.save_msr(v, ctxt);
-
- ASSERT(ctxt->count <= msr_count_max);
-
- for ( i = 0; i < ctxt->count; ++i )
- ctxt->msr[i]._rsvd = 0;
+ for ( i = 0; i < ctxt->count; ++i )
+ ctxt->msr[i]._rsvd = 0;
- if ( ctxt->count )
- {
- /* Rewrite length to indicate how much space we actually used. */
- desc->length = HVM_CPU_MSR_SIZE(ctxt->count);
- h->cur += HVM_CPU_MSR_SIZE(ctxt->count);
- }
- else
- /* or rewind and remove the descriptor from the stream. */
- h->cur -= sizeof(struct hvm_save_descriptor);
+ if ( ctxt->count )
+ {
+ /* Rewrite length to indicate how much space we actually used. */
+ desc->length = HVM_CPU_MSR_SIZE(ctxt->count);
+ h->cur += HVM_CPU_MSR_SIZE(ctxt->count);
}
+ else
+ /* or rewind and remove the descriptor from the stream. */
+ h->cur -= sizeof(struct hvm_save_descriptor);
return 0;
}
@@ -1517,7 +1468,7 @@ static int __init hvm_register_CPU_save_and_restore(void)
{
hvm_register_savevm(CPU_XSAVE_CODE,
"CPU_XSAVE",
- hvm_save_cpu_xsave_states,
+ hvm_save_cpu_xsave_states_one,
hvm_load_cpu_xsave_states,
HVM_CPU_XSAVE_SIZE(xfeature_mask) +
sizeof(struct hvm_save_descriptor),
@@ -1529,7 +1480,7 @@ static int __init hvm_register_CPU_save_and_restore(void)
if ( msr_count_max )
hvm_register_savevm(CPU_MSR_CODE,
"CPU_MSR",
- hvm_save_cpu_msrs,
+ hvm_save_cpu_msrs_one,
hvm_load_cpu_msrs,
HVM_CPU_MSR_SIZE(msr_count_max) +
sizeof(struct hvm_save_descriptor),
diff --git a/xen/arch/x86/hvm/i8254.c b/xen/arch/x86/hvm/i8254.c
index 992f08d..e0d2255 100644
--- a/xen/arch/x86/hvm/i8254.c
+++ b/xen/arch/x86/hvm/i8254.c
@@ -390,8 +390,9 @@ void pit_stop_channel0_irq(PITState *pit)
spin_unlock(&pit->lock);
}
-static int pit_save(struct domain *d, hvm_domain_context_t *h)
+static int pit_save(struct vcpu *v, hvm_domain_context_t *h)
{
+ struct domain *d = v->domain;
PITState *pit = domain_vpit(d);
int rc;
diff --git a/xen/arch/x86/hvm/irq.c b/xen/arch/x86/hvm/irq.c
index c85d004..72acb73 100644
--- a/xen/arch/x86/hvm/irq.c
+++ b/xen/arch/x86/hvm/irq.c
@@ -630,8 +630,9 @@ static int __init dump_irq_info_key_init(void)
}
__initcall(dump_irq_info_key_init);
-static int irq_save_pci(struct domain *d, hvm_domain_context_t *h)
+static int irq_save_pci(struct vcpu *v, hvm_domain_context_t *h)
{
+ struct domain *d = v->domain;
struct hvm_irq *hvm_irq = hvm_domain_irq(d);
unsigned int asserted, pdev, pintx;
int rc;
@@ -662,16 +663,18 @@ static int irq_save_pci(struct domain *d, hvm_domain_context_t *h)
return rc;
}
-static int irq_save_isa(struct domain *d, hvm_domain_context_t *h)
+static int irq_save_isa(struct vcpu *v, hvm_domain_context_t *h)
{
+ struct domain *d = v->domain;
struct hvm_irq *hvm_irq = hvm_domain_irq(d);
/* Save ISA IRQ lines */
return ( hvm_save_entry(ISA_IRQ, 0, h, &hvm_irq->isa_irq) );
}
-static int irq_save_link(struct domain *d, hvm_domain_context_t *h)
+static int irq_save_link(struct vcpu *v, hvm_domain_context_t *h)
{
+ struct domain *d = v->domain;
struct hvm_irq *hvm_irq = hvm_domain_irq(d);
/* Save PCI-ISA link state */
diff --git a/xen/arch/x86/hvm/mtrr.c b/xen/arch/x86/hvm/mtrr.c
index d311031..b0e0a60 100644
--- a/xen/arch/x86/hvm/mtrr.c
+++ b/xen/arch/x86/hvm/mtrr.c
@@ -666,46 +666,36 @@ int hvm_set_mem_pinned_cacheattr(struct domain *d, uint64_t gfn_start,
return 0;
}
-static void hvm_save_mtrr_msr_one(struct vcpu *v, struct hvm_hw_mtrr *hw_mtrr)
+static int hvm_save_mtrr_msr_one(struct vcpu *v, hvm_domain_context_t *h)
{
- struct mtrr_state *mtrr_state = &v->arch.hvm_vcpu.mtrr;
int i;
+ struct mtrr_state *mtrr_state = &v->arch.hvm_vcpu.mtrr;
+ struct hvm_hw_mtrr hw_mtrr;
+ /* save mtrr&pat */
- hvm_get_guest_pat(v, &hw_mtrr->msr_pat_cr);
+ hvm_get_guest_pat(v, &hw_mtrr.msr_pat_cr);
- hw_mtrr->msr_mtrr_def_type = mtrr_state->def_type
+ hw_mtrr.msr_mtrr_def_type = mtrr_state->def_type
| (mtrr_state->enabled << 10);
- hw_mtrr->msr_mtrr_cap = mtrr_state->mtrr_cap;
+ hw_mtrr.msr_mtrr_cap = mtrr_state->mtrr_cap;
for ( i = 0; i < MTRR_VCNT; i++ )
{
/* save physbase */
- hw_mtrr->msr_mtrr_var[i*2] =
+ hw_mtrr.msr_mtrr_var[i*2] =
((uint64_t*)mtrr_state->var_ranges)[i*2];
/* save physmask */
- hw_mtrr->msr_mtrr_var[i*2+1] =
+ hw_mtrr.msr_mtrr_var[i*2+1] =
((uint64_t*)mtrr_state->var_ranges)[i*2+1];
}
for ( i = 0; i < NUM_FIXED_MSR; i++ )
- hw_mtrr->msr_mtrr_fixed[i] =
+ hw_mtrr.msr_mtrr_fixed[i] =
((uint64_t*)mtrr_state->fixed_ranges)[i];
-}
-
-static int hvm_save_mtrr_msr(struct domain *d, hvm_domain_context_t *h)
-{
- struct vcpu *v;
- struct hvm_hw_mtrr hw_mtrr;
- /* save mtrr&pat */
-
- for_each_vcpu(d, v)
- {
- hvm_save_mtrr_msr_one(v, &hw_mtrr);
+ if ( hvm_save_entry(MTRR, v->vcpu_id, h, &hw_mtrr) != 0 )
+ return 1;
- if ( hvm_save_entry(MTRR, v->vcpu_id, h, &hw_mtrr) != 0 )
- return 1;
- }
return 0;
}
@@ -751,7 +741,7 @@ static int hvm_load_mtrr_msr(struct domain *d, hvm_domain_context_t *h)
return 0;
}
-HVM_REGISTER_SAVE_RESTORE(MTRR, hvm_save_mtrr_msr, hvm_load_mtrr_msr,
+HVM_REGISTER_SAVE_RESTORE(MTRR, hvm_save_mtrr_msr_one, hvm_load_mtrr_msr,
1, HVMSR_PER_VCPU);
void memory_type_changed(struct domain *d)
diff --git a/xen/arch/x86/hvm/pmtimer.c b/xen/arch/x86/hvm/pmtimer.c
index 435647f..d8dcbc2 100644
--- a/xen/arch/x86/hvm/pmtimer.c
+++ b/xen/arch/x86/hvm/pmtimer.c
@@ -249,8 +249,9 @@ static int handle_pmt_io(
return X86EMUL_OKAY;
}
-static int acpi_save(struct domain *d, hvm_domain_context_t *h)
+static int acpi_save(struct vcpu *v, hvm_domain_context_t *h)
{
+ struct domain *d = v->domain;
struct hvm_hw_acpi *acpi = &d->arch.hvm_domain.acpi;
PMTState *s = &d->arch.hvm_domain.pl_time->vpmt;
uint32_t x, msb = acpi->tmr_val & TMR_VAL_MSB;
diff --git a/xen/arch/x86/hvm/rtc.c b/xen/arch/x86/hvm/rtc.c
index cb75b99..58b70fc 100644
--- a/xen/arch/x86/hvm/rtc.c
+++ b/xen/arch/x86/hvm/rtc.c
@@ -737,8 +737,9 @@ void rtc_migrate_timers(struct vcpu *v)
}
/* Save RTC hardware state */
-static int rtc_save(struct domain *d, hvm_domain_context_t *h)
+static int rtc_save(struct vcpu *v, hvm_domain_context_t *h)
{
+ struct domain *d = v->domain;
RTCState *s = domain_vrtc(d);
int rc;
diff --git a/xen/arch/x86/hvm/save.c b/xen/arch/x86/hvm/save.c
index 8984a23..fe24f08 100644
--- a/xen/arch/x86/hvm/save.c
+++ b/xen/arch/x86/hvm/save.c
@@ -135,9 +135,12 @@ size_t hvm_save_size(struct domain *d)
int hvm_save_one(struct domain *d, unsigned int typecode, unsigned int instance,
XEN_GUEST_HANDLE_64(uint8) handle, uint64_t *bufsz)
{
- int rv;
+ int rv = 0;
hvm_domain_context_t ctxt = { };
const struct hvm_save_descriptor *desc;
+ bool is_single_instance = false;
+ uint32_t off = 0;
+ struct vcpu *v;
if ( d->is_dying ||
typecode > HVM_SAVE_CODE_MAX ||
@@ -145,43 +148,89 @@ int hvm_save_one(struct domain *d, unsigned int typecode, unsigned int instance,
!hvm_sr_handlers[typecode].save )
return -EINVAL;
+ if( hvm_sr_handlers[typecode].kind == HVMSR_PER_VCPU &&
+ instance < d->max_vcpus )
+ is_single_instance = true;
+
ctxt.size = hvm_sr_handlers[typecode].size;
- if ( hvm_sr_handlers[typecode].kind == HVMSR_PER_VCPU )
+ if ( hvm_sr_handlers[typecode].kind == HVMSR_PER_VCPU &&
+ instance == d->max_vcpus )
ctxt.size *= d->max_vcpus;
ctxt.data = xmalloc_bytes(ctxt.size);
if ( !ctxt.data )
return -ENOMEM;
- if ( (rv = hvm_sr_handlers[typecode].save(d, &ctxt)) != 0 )
- printk(XENLOG_G_ERR "HVM%d save: failed to save type %"PRIu16" (%d)\n",
- d->domain_id, typecode, rv);
- else if ( rv = -ENOENT, ctxt.cur >= sizeof(*desc) )
- {
- uint32_t off;
- for ( off = 0; off <= (ctxt.cur - sizeof(*desc)); off += desc->length )
+ if( is_single_instance )
+ vcpu_pause(d->vcpu[instance]);
+ else
+ domain_pause(d);
+
+ if( is_single_instance )
+ {
+ if ( (rv = hvm_sr_handlers[typecode].save(d->vcpu[instance],
+ &ctxt)) != 0 )
{
- desc = (void *)(ctxt.data + off);
- /* Move past header */
- off += sizeof(*desc);
- if ( ctxt.cur < desc->length ||
- off > ctxt.cur - desc->length )
- break;
- if ( instance == desc->instance )
- {
- rv = 0;
- if ( guest_handle_is_null(handle) )
- *bufsz = desc->length;
- else if ( *bufsz < desc->length )
- rv = -ENOBUFS;
- else if ( copy_to_guest(handle, ctxt.data + off, desc->length) )
- rv = -EFAULT;
- else
- *bufsz = desc->length;
- break;
- }
+ printk(XENLOG_G_ERR "HVM%d save: failed to save type %"PRIu16" (%d)\n",
+ d->domain_id, typecode, rv);
+ vcpu_unpause(d->vcpu[instance]);
+ }
+ else if ( rv = -ENOENT, ctxt.cur >= sizeof(*desc) )
+ {
+ desc = (void *)(ctxt.data);
+ /* Move past header */
+ off = sizeof(*desc);
+ if ( ctxt.cur < desc->length ||
+ off > ctxt.cur - desc->length )
+ rv = -EFAULT;
+ rv = 0;
+ if ( guest_handle_is_null(handle) )
+ *bufsz = desc->length;
+ else if ( *bufsz < desc->length )
+ rv = -ENOBUFS;
+ else if ( copy_to_guest(handle, ctxt.data + off, desc->length) )
+ rv = -EFAULT;
+ else
+ *bufsz = desc->length;
+ vcpu_unpause(d->vcpu[instance]);
}
}
+ else
+ {
+ for_each_vcpu ( d, v )
+ {
+ if ( (rv = hvm_sr_handlers[typecode].save(d->vcpu[instance],
+ &ctxt)) != 0 )
+ {
+ printk(XENLOG_G_ERR "HVM%d save: failed to save type %"PRIu16" (%d)\n",
+ d->domain_id, typecode, rv);
+ }
+ else if ( rv = -ENOENT, ctxt.cur >= sizeof(*desc) )
+ {
+ desc = (void *)(ctxt.data + off);
+ /* Move past header */
+ off += sizeof(*desc);
+ if ( ctxt.cur < desc->length ||
+ off > ctxt.cur - desc->length )
+ break;
+ if ( instance == desc->instance )
+ {
+ rv = 0;
+ if ( guest_handle_is_null(handle) )
+ *bufsz = desc->length;
+ else if ( *bufsz < desc->length )
+ rv = -ENOBUFS;
+ else if ( copy_to_guest(handle, ctxt.data + off, desc->length) )
+ rv = -EFAULT;
+ else
+ *bufsz = desc->length;
+ break;
+ }
+ off += desc->length;
+ }
+ }
+ domain_unpause(d);
+ }
xfree(ctxt.data);
return rv;
@@ -193,7 +242,8 @@ int hvm_save(struct domain *d, hvm_domain_context_t *h)
struct hvm_save_header hdr;
struct hvm_save_end end;
hvm_save_handler handler;
- unsigned int i;
+ unsigned int i, rc;
+ struct vcpu *v = NULL;
if ( d->is_dying )
return -EINVAL;
@@ -225,12 +275,19 @@ int hvm_save(struct domain *d, hvm_domain_context_t *h)
{
printk(XENLOG_G_INFO "HVM%d save: %s\n",
d->domain_id, hvm_sr_handlers[i].name);
- if ( handler(d, h) != 0 )
+ for_each_vcpu ( d, v )
{
- printk(XENLOG_G_ERR
- "HVM%d save: failed to save type %"PRIu16"\n",
- d->domain_id, i);
- return -EFAULT;
+ rc = handler(v, h);
+ if( rc == 2 )
+ continue;
+
+ if( rc != 0 )
+ {
+ printk(XENLOG_G_ERR
+ "HVM%d save: failed to save type %"PRIu16"\n",
+ d->domain_id, i);
+ return -EFAULT;
+ }
}
}
}
diff --git a/xen/arch/x86/hvm/vioapic.c b/xen/arch/x86/hvm/vioapic.c
index 97b419f..86d02cf 100644
--- a/xen/arch/x86/hvm/vioapic.c
+++ b/xen/arch/x86/hvm/vioapic.c
@@ -569,8 +569,9 @@ int vioapic_get_trigger_mode(const struct domain *d, unsigned int gsi)
return vioapic->redirtbl[pin].fields.trig_mode;
}
-static int ioapic_save(struct domain *d, hvm_domain_context_t *h)
+static int ioapic_save(struct vcpu *v, hvm_domain_context_t *h)
{
+ struct domain *d = v->domain;
struct hvm_vioapic *s;
if ( !has_vioapic(d) )
diff --git a/xen/arch/x86/hvm/viridian.c b/xen/arch/x86/hvm/viridian.c
index d22c8ac..8a7c592 100644
--- a/xen/arch/x86/hvm/viridian.c
+++ b/xen/arch/x86/hvm/viridian.c
@@ -990,8 +990,9 @@ out:
return HVM_HCALL_completed;
}
-static int viridian_save_domain_ctxt(struct domain *d, hvm_domain_context_t *h)
+static int viridian_save_domain_ctxt(struct vcpu *v, hvm_domain_context_t *h)
{
+ struct domain *d = v->domain;
struct hvm_viridian_domain_context ctxt = {
.time_ref_count = d->arch.hvm_domain.viridian.time_ref_count.val,
.hypercall_gpa = d->arch.hvm_domain.viridian.hypercall_gpa.raw,
@@ -1026,28 +1027,19 @@ static int viridian_load_domain_ctxt(struct domain *d, hvm_domain_context_t *h)
HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_DOMAIN, viridian_save_domain_ctxt,
viridian_load_domain_ctxt, 1, HVMSR_PER_DOM);
-static void viridian_save_vcpu_ctxt_one(struct vcpu *v, struct hvm_viridian_vcpu_context *ctxt)
+static int viridian_save_vcpu_ctxt_one(struct vcpu *v, hvm_domain_context_t *h)
{
- ctxt->vp_assist_msr = v->arch.hvm_vcpu.viridian.vp_assist.msr.raw;
- ctxt->vp_assist_pending = v->arch.hvm_vcpu.viridian.vp_assist.pending;
- *ctxt->_pad = 0;
-}
-
-static int viridian_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h)
-{
- struct vcpu *v;
+ struct domain *d = v->domain;
+ struct hvm_viridian_vcpu_context ctxt = {
+ .vp_assist_msr = v->arch.hvm_vcpu.viridian.vp_assist.msr.raw,
+ .vp_assist_pending = v->arch.hvm_vcpu.viridian.vp_assist.pending,
+ };
if ( !is_viridian_domain(d) )
return 0;
- for_each_vcpu( d, v ) {
- struct hvm_viridian_vcpu_context ctxt;
-
- viridian_save_vcpu_ctxt_one(v, &ctxt);
-
- if ( hvm_save_entry(VIRIDIAN_VCPU, v->vcpu_id, h, &ctxt) != 0 )
- return 1;
- }
+ if ( hvm_save_entry(VIRIDIAN_VCPU, v->vcpu_id, h, &ctxt) != 0 )
+ return 1;
return 0;
}
@@ -1082,7 +1074,7 @@ static int viridian_load_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h)
return 0;
}
-HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_VCPU, viridian_save_vcpu_ctxt,
+HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_VCPU, viridian_save_vcpu_ctxt_one,
viridian_load_vcpu_ctxt, 1, HVMSR_PER_VCPU);
static int __init parse_viridian_version(const char *arg)
diff --git a/xen/arch/x86/hvm/vlapic.c b/xen/arch/x86/hvm/vlapic.c
index 1b9f00a..ab35400 100644
--- a/xen/arch/x86/hvm/vlapic.c
+++ b/xen/arch/x86/hvm/vlapic.c
@@ -1435,43 +1435,35 @@ static void lapic_rearm(struct vlapic *s)
s->timer_last_update = s->pt.last_plt_gtime;
}
-static int lapic_save_hidden(struct domain *d, hvm_domain_context_t *h)
+static int lapic_save_hidden(struct vcpu *v, hvm_domain_context_t *h)
{
- struct vcpu *v;
+ struct domain *d = v->domain;
struct vlapic *s;
int rc = 0;
if ( !has_vlapic(d) )
return 0;
- for_each_vcpu ( d, v )
- {
- s = vcpu_vlapic(v);
- if ( (rc = hvm_save_entry(LAPIC, v->vcpu_id, h, &s->hw)) != 0 )
- break;
- }
+ s = vcpu_vlapic(v);
+ rc = hvm_save_entry(LAPIC, v->vcpu_id, h, &s->hw);
return rc;
}
-static int lapic_save_regs(struct domain *d, hvm_domain_context_t *h)
+static int lapic_save_regs(struct vcpu *v, hvm_domain_context_t *h)
{
- struct vcpu *v;
+ struct domain *d = v->domain;
struct vlapic *s;
int rc = 0;
if ( !has_vlapic(d) )
return 0;
- for_each_vcpu ( d, v )
- {
- if ( hvm_funcs.sync_pir_to_irr )
- hvm_funcs.sync_pir_to_irr(v);
+ if ( hvm_funcs.sync_pir_to_irr )
+ hvm_funcs.sync_pir_to_irr(v);
- s = vcpu_vlapic(v);
- if ( (rc = hvm_save_entry(LAPIC_REGS, v->vcpu_id, h, s->regs)) != 0 )
- break;
- }
+ s = vcpu_vlapic(v);
+ rc = hvm_save_entry(LAPIC_REGS, v->vcpu_id, h, s->regs);
return rc;
}
diff --git a/xen/arch/x86/hvm/vpic.c b/xen/arch/x86/hvm/vpic.c
index e160bbd..bad5066 100644
--- a/xen/arch/x86/hvm/vpic.c
+++ b/xen/arch/x86/hvm/vpic.c
@@ -371,8 +371,9 @@ static int vpic_intercept_elcr_io(
return X86EMUL_OKAY;
}
-static int vpic_save(struct domain *d, hvm_domain_context_t *h)
+static int vpic_save(struct vcpu *v, hvm_domain_context_t *h)
{
+ struct domain *d = v->domain;
struct hvm_hw_vpic *s;
int i;
diff --git a/xen/include/asm-x86/hvm/save.h b/xen/include/asm-x86/hvm/save.h
index f889e8f..fe642ab 100644
--- a/xen/include/asm-x86/hvm/save.h
+++ b/xen/include/asm-x86/hvm/save.h
@@ -95,7 +95,7 @@ static inline uint16_t hvm_load_instance(struct hvm_domain_context *h)
* The save handler may save multiple instances of a type into the buffer;
* the load handler will be called once for each instance found when
* restoring. Both return non-zero on error. */
-typedef int (*hvm_save_handler) (struct domain *d,
+typedef int (*hvm_save_handler) (struct vcpu *v,
hvm_domain_context_t *h);
typedef int (*hvm_load_handler) (struct domain *d,
hvm_domain_context_t *h);
--
2.7.4
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel
^ permalink raw reply related [flat|nested] 12+ messages in thread
* Re: [PATCH v5 5/8] x86/hvm: Introduce hvm_save_cpu_msrs_one func
2018-05-29 14:58 ` [PATCH v5 5/8] x86/hvm: Introduce hvm_save_cpu_msrs_one func Alexandru Isaila
@ 2018-06-04 9:08 ` Paul Durrant
0 siblings, 0 replies; 12+ messages in thread
From: Paul Durrant @ 2018-06-04 9:08 UTC (permalink / raw)
To: 'Alexandru Isaila', xen-devel
Cc: Ian Jackson, Wei Liu, jbeulich, Andrew Cooper
> -----Original Message-----
> From: Alexandru Isaila [mailto:aisaila@bitdefender.com]
> Sent: 29 May 2018 15:59
> To: xen-devel@lists.xen.org
> Cc: Ian Jackson <Ian.Jackson@citrix.com>; Wei Liu <wei.liu2@citrix.com>;
> jbeulich@suse.com; Andrew Cooper <Andrew.Cooper3@citrix.com>; Paul
> Durrant <Paul.Durrant@citrix.com>; Alexandru Isaila
> <aisaila@bitdefender.com>
> Subject: [PATCH v5 5/8] x86/hvm: Introduce hvm_save_cpu_msrs_one func
>
> This is used to save data from a single instance.
>
> Signed-off-by: Alexandru Isaila <aisaila@bitdefender.com>
> ---
> xen/arch/x86/hvm/hvm.c | 59 ++++++++++++++++++++++++++++-----------
> -----------
> 1 file changed, 33 insertions(+), 26 deletions(-)
>
> diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c
> index e8ecabf..4a22283 100644
> --- a/xen/arch/x86/hvm/hvm.c
> +++ b/xen/arch/x86/hvm/hvm.c
> @@ -1354,6 +1354,38 @@ static const uint32_t msrs_to_send[] = {
> };
> static unsigned int __read_mostly msr_count_max =
> ARRAY_SIZE(msrs_to_send);
>
> +static int hvm_save_cpu_msrs_one(struct vcpu *v, struct hvm_msr *ctxt)
> +{
> + unsigned int i;
> +
> + for ( i = 0; i < ARRAY_SIZE(msrs_to_send); ++i )
> + {
> + uint64_t val;
> + int rc = guest_rdmsr(v, msrs_to_send[i], &val);
> +
> + /*
> + * It is the programmers responsibility to ensure that
> + * msrs_to_send[] contain generally-read/write MSRs.
> + * X86EMUL_EXCEPTION here implies a missing feature, and that the
> + * guest doesn't have access to the MSR.
> + */
> + if ( rc == X86EMUL_EXCEPTION )
> + continue;
> +
> + if ( rc != X86EMUL_OKAY )
> + {
> + ASSERT_UNREACHABLE();
> + return -ENXIO;
> + }
> +
> + if ( !val )
> + continue; /* Skip empty MSRs. */
> + ctxt->msr[ctxt->count].index = msrs_to_send[i];
> + ctxt->msr[ctxt->count++].val = val;
> + }
> + return 0;
> +}
> +
> static int hvm_save_cpu_msrs(struct domain *d, hvm_domain_context_t
> *h)
> {
> struct vcpu *v;
> @@ -1370,32 +1402,7 @@ static int hvm_save_cpu_msrs(struct domain *d,
> hvm_domain_context_t *h)
> ctxt = (struct hvm_msr *)&h->data[h->cur];
> ctxt->count = 0;
>
> - for ( i = 0; i < ARRAY_SIZE(msrs_to_send); ++i )
> - {
> - uint64_t val;
> - int rc = guest_rdmsr(v, msrs_to_send[i], &val);
> -
> - /*
> - * It is the programmers responsibility to ensure that
> - * msrs_to_send[] contain generally-read/write MSRs.
> - * X86EMUL_EXCEPTION here implies a missing feature, and that the
> - * guest doesn't have access to the MSR.
> - */
> - if ( rc == X86EMUL_EXCEPTION )
> - continue;
> -
> - if ( rc != X86EMUL_OKAY )
> - {
> - ASSERT_UNREACHABLE();
> - return -ENXIO;
> - }
> -
> - if ( !val )
> - continue; /* Skip empty MSRs. */
> -
> - ctxt->msr[ctxt->count].index = msrs_to_send[i];
> - ctxt->msr[ctxt->count++].val = val;
> - }
> + hvm_save_cpu_msrs_one(v, ctxt);
No check for the return value here means you're not going to propagate a potential -ENXIO.
Paul
>
> if ( hvm_funcs.save_msr )
> hvm_funcs.save_msr(v, ctxt);
> --
> 2.7.4
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: [PATCH v5 7/8] x86/hvm: Introduce viridian_save_vcpu_ctxt_one() func
2018-05-29 14:58 ` [PATCH v5 7/8] x86/hvm: Introduce viridian_save_vcpu_ctxt_one() func Alexandru Isaila
@ 2018-06-04 9:11 ` Paul Durrant
0 siblings, 0 replies; 12+ messages in thread
From: Paul Durrant @ 2018-06-04 9:11 UTC (permalink / raw)
To: 'Alexandru Isaila', xen-devel
Cc: Ian Jackson, Wei Liu, jbeulich, Andrew Cooper
> -----Original Message-----
> From: Alexandru Isaila [mailto:aisaila@bitdefender.com]
> Sent: 29 May 2018 15:59
> To: xen-devel@lists.xen.org
> Cc: Ian Jackson <Ian.Jackson@citrix.com>; Wei Liu <wei.liu2@citrix.com>;
> jbeulich@suse.com; Andrew Cooper <Andrew.Cooper3@citrix.com>; Paul
> Durrant <Paul.Durrant@citrix.com>; Alexandru Isaila
> <aisaila@bitdefender.com>
> Subject: [PATCH v5 7/8] x86/hvm: Introduce viridian_save_vcpu_ctxt_one()
> func
>
> This is used to save data from a single instance.
>
> Signed-off-by: Alexandru Isaila <aisaila@bitdefender.com>
> ---
> xen/arch/x86/hvm/viridian.c | 14 ++++++++++----
> 1 file changed, 10 insertions(+), 4 deletions(-)
>
> diff --git a/xen/arch/x86/hvm/viridian.c b/xen/arch/x86/hvm/viridian.c
> index 694eae6..d22c8ac 100644
> --- a/xen/arch/x86/hvm/viridian.c
> +++ b/xen/arch/x86/hvm/viridian.c
> @@ -1026,6 +1026,13 @@ static int viridian_load_domain_ctxt(struct domain
> *d, hvm_domain_context_t *h)
> HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_DOMAIN,
> viridian_save_domain_ctxt,
> viridian_load_domain_ctxt, 1, HVMSR_PER_DOM);
>
> +static void viridian_save_vcpu_ctxt_one(struct vcpu *v, struct
> hvm_viridian_vcpu_context *ctxt)
> +{
> + ctxt->vp_assist_msr = v->arch.hvm_vcpu.viridian.vp_assist.msr.raw;
> + ctxt->vp_assist_pending = v->arch.hvm_vcpu.viridian.vp_assist.pending;
> + *ctxt->_pad = 0;
I'd prefer a memset() to zero the context prior to setting fields. It's more robust if fields (and more padding) are added in future.
Paul
> +}
> +
> static int viridian_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t
> *h)
> {
> struct vcpu *v;
> @@ -1034,10 +1041,9 @@ static int viridian_save_vcpu_ctxt(struct domain
> *d, hvm_domain_context_t *h)
> return 0;
>
> for_each_vcpu( d, v ) {
> - struct hvm_viridian_vcpu_context ctxt = {
> - .vp_assist_msr = v->arch.hvm_vcpu.viridian.vp_assist.msr.raw,
> - .vp_assist_pending = v->arch.hvm_vcpu.viridian.vp_assist.pending,
> - };
> + struct hvm_viridian_vcpu_context ctxt;
> +
> + viridian_save_vcpu_ctxt_one(v, &ctxt);
>
> if ( hvm_save_entry(VIRIDIAN_VCPU, v->vcpu_id, h, &ctxt) != 0 )
> return 1;
> --
> 2.7.4
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: [PATCH v5 8/8] x86/domctl: Don't pause the whole domain if only getting vcpu state
2018-05-29 14:58 ` [PATCH v5 8/8] x86/domctl: Don't pause the whole domain if only getting vcpu state Alexandru Isaila
@ 2018-06-04 9:21 ` Paul Durrant
2018-06-06 12:35 ` Jan Beulich
1 sibling, 0 replies; 12+ messages in thread
From: Paul Durrant @ 2018-06-04 9:21 UTC (permalink / raw)
To: 'Alexandru Isaila', xen-devel
Cc: Ian Jackson, Wei Liu, jbeulich, Andrew Cooper
> -----Original Message-----
> From: Alexandru Isaila [mailto:aisaila@bitdefender.com]
> Sent: 29 May 2018 15:59
> To: xen-devel@lists.xen.org
> Cc: Ian Jackson <Ian.Jackson@citrix.com>; Wei Liu <wei.liu2@citrix.com>;
> jbeulich@suse.com; Andrew Cooper <Andrew.Cooper3@citrix.com>; Paul
> Durrant <Paul.Durrant@citrix.com>; Alexandru Isaila
> <aisaila@bitdefender.com>
> Subject: [PATCH v5 8/8] x86/domctl: Don't pause the whole domain if only
> getting vcpu state
>
> This patch is focused on merging the *save() funcs to the *save_one()
> funcs to remove redundancy. Also the for loop is moved to the caller so
> now we can save info for a single vcpu instance.
>
> Signed-off-by: Alexandru Isaila <aisaila@bitdefender.com>
>
> ---
> Changes since V4:
> - Save funcs are deleted
> - vcpu si added to the param list
> - save_one funcs are now static
> - Add blank line after "/* save mtrr&pat */" comment
> ---
> xen/arch/x86/cpu/mcheck/vmce.c | 28 ++--
> xen/arch/x86/hvm/hpet.c | 3 +-
> xen/arch/x86/hvm/hvm.c | 287 +++++++++++++++++----------------------
> --
> xen/arch/x86/hvm/i8254.c | 3 +-
> xen/arch/x86/hvm/irq.c | 9 +-
> xen/arch/x86/hvm/mtrr.c | 36 ++----
> xen/arch/x86/hvm/pmtimer.c | 3 +-
> xen/arch/x86/hvm/rtc.c | 3 +-
> xen/arch/x86/hvm/save.c | 125 +++++++++++++-----
> xen/arch/x86/hvm/vioapic.c | 3 +-
> xen/arch/x86/hvm/viridian.c | 30 ++---
> xen/arch/x86/hvm/vlapic.c | 28 ++--
> xen/arch/x86/hvm/vpic.c | 3 +-
> xen/include/asm-x86/hvm/save.h | 2 +-
> 14 files changed, 271 insertions(+), 292 deletions(-)
>
> diff --git a/xen/arch/x86/cpu/mcheck/vmce.c
> b/xen/arch/x86/cpu/mcheck/vmce.c
> index 404f27e..7bde9ff 100644
> --- a/xen/arch/x86/cpu/mcheck/vmce.c
> +++ b/xen/arch/x86/cpu/mcheck/vmce.c
> @@ -349,29 +349,17 @@ int vmce_wrmsr(uint32_t msr, uint64_t val)
> return ret;
> }
>
> -static void vmce_save_vcpu_ctxt_one(struct vcpu *v, struct
> hvm_vmce_vcpu *ctxt)
> +static int vmce_save_vcpu_ctxt_one(struct vcpu *v,
> hvm_domain_context_t *h)
> {
> - ctxt->caps = v->arch.vmce.mcg_cap;
> - ctxt->mci_ctl2_bank0 = v->arch.vmce.bank[0].mci_ctl2;
> - ctxt->mci_ctl2_bank1 = v->arch.vmce.bank[1].mci_ctl2;
> - ctxt->mcg_ext_ctl = v->arch.vmce.mcg_ext_ctl;
> -}
> -
> -static int vmce_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t
> *h)
> -{
> - struct vcpu *v;
> int err = 0;
> + struct hvm_vmce_vcpu ctxt;
>
> - for_each_vcpu ( d, v )
> - {
> - struct hvm_vmce_vcpu ctxt;
> -
> - vmce_save_vcpu_ctxt_one(v, &ctxt);
> - err = hvm_save_entry(VMCE_VCPU, v->vcpu_id, h, &ctxt);
> - if ( err )
> - break;
> - }
> + ctxt.caps = v->arch.vmce.mcg_cap;
> + ctxt.mci_ctl2_bank0 = v->arch.vmce.bank[0].mci_ctl2;
> + ctxt.mci_ctl2_bank1 = v->arch.vmce.bank[1].mci_ctl2;
> + ctxt.mcg_ext_ctl = v->arch.vmce.mcg_ext_ctl;
>
> + err = hvm_save_entry(VMCE_VCPU, v->vcpu_id, h, &ctxt);
Can't you just return hvm_save_entry(...) directly here and drop err?
> return err;
> }
>
> @@ -394,7 +382,7 @@ static int vmce_load_vcpu_ctxt(struct domain *d,
> hvm_domain_context_t *h)
> return err ?: vmce_restore_vcpu(v, &ctxt);
> }
>
> -HVM_REGISTER_SAVE_RESTORE(VMCE_VCPU, vmce_save_vcpu_ctxt,
> +HVM_REGISTER_SAVE_RESTORE(VMCE_VCPU,
> vmce_save_vcpu_ctxt_one,
> vmce_load_vcpu_ctxt, 1, HVMSR_PER_VCPU);
>
> /*
> diff --git a/xen/arch/x86/hvm/hpet.c b/xen/arch/x86/hvm/hpet.c
> index 2837709..3ed6547 100644
> --- a/xen/arch/x86/hvm/hpet.c
> +++ b/xen/arch/x86/hvm/hpet.c
> @@ -516,8 +516,9 @@ static const struct hvm_mmio_ops hpet_mmio_ops =
> {
> };
>
>
> -static int hpet_save(struct domain *d, hvm_domain_context_t *h)
> +static int hpet_save(struct vcpu *vcpu, hvm_domain_context_t *h)
> {
> + struct domain *d = vcpu->domain;
> HPETState *hp = domain_vhpet(d);
> struct vcpu *v = pt_global_vcpu_target(d);
> int rc;
> diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c
> index 4a22283..ea27055 100644
> --- a/xen/arch/x86/hvm/hvm.c
> +++ b/xen/arch/x86/hvm/hvm.c
> @@ -740,24 +740,13 @@ void hvm_domain_destroy(struct domain *d)
> destroy_vpci_mmcfg(d);
> }
>
> -static void hvm_save_tsc_adjust_one(struct vcpu *v, struct hvm_tsc_adjust
> *ctxt)
> +static int hvm_save_tsc_adjust_one(struct vcpu *v,
> hvm_domain_context_t *h)
> {
> - ctxt->tsc_adjust = v->arch.hvm_vcpu.msr_tsc_adjust;
> -}
> -
> -static int hvm_save_tsc_adjust(struct domain *d, hvm_domain_context_t
> *h)
> -{
> - struct vcpu *v;
> - struct hvm_tsc_adjust ctxt;
> int err = 0;
> + struct hvm_tsc_adjust ctxt;
>
> - for_each_vcpu ( d, v )
> - {
> - hvm_save_tsc_adjust_one(v, &ctxt);
> - err = hvm_save_entry(TSC_ADJUST, v->vcpu_id, h, &ctxt);
> - if ( err )
> - break;
> - }
> + ctxt.tsc_adjust = v->arch.hvm_vcpu.msr_tsc_adjust;
> + err = hvm_save_entry(TSC_ADJUST, v->vcpu_id, h, &ctxt);
>
Looks like you can lose err here too.
> return err;
> }
> @@ -782,126 +771,116 @@ static int hvm_load_tsc_adjust(struct domain *d,
> hvm_domain_context_t *h)
> return 0;
> }
>
> -HVM_REGISTER_SAVE_RESTORE(TSC_ADJUST, hvm_save_tsc_adjust,
> +HVM_REGISTER_SAVE_RESTORE(TSC_ADJUST, hvm_save_tsc_adjust_one,
> hvm_load_tsc_adjust, 1, HVMSR_PER_VCPU);
>
> -static void hvm_save_cpu_ctxt_one(struct vcpu *v, struct hvm_hw_cpu
> *ctxt)
> +static int hvm_save_cpu_ctxt_one(struct vcpu *v, hvm_domain_context_t
> *h)
> {
> struct segment_register seg;
> + struct hvm_hw_cpu ctxt = {};
> +
> + /* We don't need to save state for a vcpu that is down; the restore
> + * code will leave it down if there is nothing saved. */
> + if ( v->pause_flags & VPF_down )
> + return 2;
>
> /* Architecture-specific vmcs/vmcb bits */
> - hvm_funcs.save_cpu_ctxt(v, ctxt);
> + hvm_funcs.save_cpu_ctxt(v, &ctxt);
>
> - ctxt->tsc = hvm_get_guest_tsc_fixed(v, v->domain-
> >arch.hvm_domain.sync_tsc);
> + ctxt.tsc = hvm_get_guest_tsc_fixed(v, v->domain-
> >arch.hvm_domain.sync_tsc);
>
> - ctxt->msr_tsc_aux = hvm_msr_tsc_aux(v);
> + ctxt.msr_tsc_aux = hvm_msr_tsc_aux(v);
>
> hvm_get_segment_register(v, x86_seg_idtr, &seg);
> - ctxt->idtr_limit = seg.limit;
> - ctxt->idtr_base = seg.base;
> + ctxt.idtr_limit = seg.limit;
> + ctxt.idtr_base = seg.base;
>
> hvm_get_segment_register(v, x86_seg_gdtr, &seg);
> - ctxt->gdtr_limit = seg.limit;
> - ctxt->gdtr_base = seg.base;
> + ctxt.gdtr_limit = seg.limit;
> + ctxt.gdtr_base = seg.base;
>
> hvm_get_segment_register(v, x86_seg_cs, &seg);
> - ctxt->cs_sel = seg.sel;
> - ctxt->cs_limit = seg.limit;
> - ctxt->cs_base = seg.base;
> - ctxt->cs_arbytes = seg.attr;
> + ctxt.cs_sel = seg.sel;
> + ctxt.cs_limit = seg.limit;
> + ctxt.cs_base = seg.base;
> + ctxt.cs_arbytes = seg.attr;
>
> hvm_get_segment_register(v, x86_seg_ds, &seg);
> - ctxt->ds_sel = seg.sel;
> - ctxt->ds_limit = seg.limit;
> - ctxt->ds_base = seg.base;
> - ctxt->ds_arbytes = seg.attr;
> + ctxt.ds_sel = seg.sel;
> + ctxt.ds_limit = seg.limit;
> + ctxt.ds_base = seg.base;
> + ctxt.ds_arbytes = seg.attr;
>
> hvm_get_segment_register(v, x86_seg_es, &seg);
> - ctxt->es_sel = seg.sel;
> - ctxt->es_limit = seg.limit;
> - ctxt->es_base = seg.base;
> - ctxt->es_arbytes = seg.attr;
> + ctxt.es_sel = seg.sel;
> + ctxt.es_limit = seg.limit;
> + ctxt.es_base = seg.base;
> + ctxt.es_arbytes = seg.attr;
>
> hvm_get_segment_register(v, x86_seg_ss, &seg);
> - ctxt->ss_sel = seg.sel;
> - ctxt->ss_limit = seg.limit;
> - ctxt->ss_base = seg.base;
> - ctxt->ss_arbytes = seg.attr;
> + ctxt.ss_sel = seg.sel;
> + ctxt.ss_limit = seg.limit;
> + ctxt.ss_base = seg.base;
> + ctxt.ss_arbytes = seg.attr;
>
> hvm_get_segment_register(v, x86_seg_fs, &seg);
> - ctxt->fs_sel = seg.sel;
> - ctxt->fs_limit = seg.limit;
> - ctxt->fs_base = seg.base;
> - ctxt->fs_arbytes = seg.attr;
> + ctxt.fs_sel = seg.sel;
> + ctxt.fs_limit = seg.limit;
> + ctxt.fs_base = seg.base;
> + ctxt.fs_arbytes = seg.attr;
>
> hvm_get_segment_register(v, x86_seg_gs, &seg);
> - ctxt->gs_sel = seg.sel;
> - ctxt->gs_limit = seg.limit;
> - ctxt->gs_base = seg.base;
> - ctxt->gs_arbytes = seg.attr;
> + ctxt.gs_sel = seg.sel;
> + ctxt.gs_limit = seg.limit;
> + ctxt.gs_base = seg.base;
> + ctxt.gs_arbytes = seg.attr;
>
> hvm_get_segment_register(v, x86_seg_tr, &seg);
> - ctxt->tr_sel = seg.sel;
> - ctxt->tr_limit = seg.limit;
> - ctxt->tr_base = seg.base;
> - ctxt->tr_arbytes = seg.attr;
> + ctxt.tr_sel = seg.sel;
> + ctxt.tr_limit = seg.limit;
> + ctxt.tr_base = seg.base;
> + ctxt.tr_arbytes = seg.attr;
>
> hvm_get_segment_register(v, x86_seg_ldtr, &seg);
> - ctxt->ldtr_sel = seg.sel;
> - ctxt->ldtr_limit = seg.limit;
> - ctxt->ldtr_base = seg.base;
> - ctxt->ldtr_arbytes = seg.attr;
> + ctxt.ldtr_sel = seg.sel;
> + ctxt.ldtr_limit = seg.limit;
> + ctxt.ldtr_base = seg.base;
> + ctxt.ldtr_arbytes = seg.attr;
>
> if ( v->fpu_initialised )
> {
> - memcpy(ctxt->fpu_regs, v->arch.fpu_ctxt, sizeof(ctxt->fpu_regs));
> - ctxt->flags = XEN_X86_FPU_INITIALISED;
> - }
> -
> - ctxt->rax = v->arch.user_regs.rax;
> - ctxt->rbx = v->arch.user_regs.rbx;
> - ctxt->rcx = v->arch.user_regs.rcx;
> - ctxt->rdx = v->arch.user_regs.rdx;
> - ctxt->rbp = v->arch.user_regs.rbp;
> - ctxt->rsi = v->arch.user_regs.rsi;
> - ctxt->rdi = v->arch.user_regs.rdi;
> - ctxt->rsp = v->arch.user_regs.rsp;
> - ctxt->rip = v->arch.user_regs.rip;
> - ctxt->rflags = v->arch.user_regs.rflags;
> - ctxt->r8 = v->arch.user_regs.r8;
> - ctxt->r9 = v->arch.user_regs.r9;
> - ctxt->r10 = v->arch.user_regs.r10;
> - ctxt->r11 = v->arch.user_regs.r11;
> - ctxt->r12 = v->arch.user_regs.r12;
> - ctxt->r13 = v->arch.user_regs.r13;
> - ctxt->r14 = v->arch.user_regs.r14;
> - ctxt->r15 = v->arch.user_regs.r15;
> - ctxt->dr0 = v->arch.debugreg[0];
> - ctxt->dr1 = v->arch.debugreg[1];
> - ctxt->dr2 = v->arch.debugreg[2];
> - ctxt->dr3 = v->arch.debugreg[3];
> - ctxt->dr6 = v->arch.debugreg[6];
> - ctxt->dr7 = v->arch.debugreg[7];
> -}
> -
> -static int hvm_save_cpu_ctxt(struct domain *d, hvm_domain_context_t
> *h)
> -{
> - struct vcpu *v;
> - struct hvm_hw_cpu ctxt;
> -
> - for_each_vcpu ( d, v )
> - {
> - /* We don't need to save state for a vcpu that is down; the restore
> - * code will leave it down if there is nothing saved. */
> - if ( v->pause_flags & VPF_down )
> - continue;
> -
> - memset(&ctxt, 0, sizeof(ctxt));
> - hvm_save_cpu_ctxt_one(v, &ctxt);
> + memcpy(ctxt.fpu_regs, v->arch.fpu_ctxt, sizeof(ctxt.fpu_regs));
> + ctxt.flags = XEN_X86_FPU_INITIALISED;
> + }
> +
> + ctxt.rax = v->arch.user_regs.rax;
> + ctxt.rbx = v->arch.user_regs.rbx;
> + ctxt.rcx = v->arch.user_regs.rcx;
> + ctxt.rdx = v->arch.user_regs.rdx;
> + ctxt.rbp = v->arch.user_regs.rbp;
> + ctxt.rsi = v->arch.user_regs.rsi;
> + ctxt.rdi = v->arch.user_regs.rdi;
> + ctxt.rsp = v->arch.user_regs.rsp;
> + ctxt.rip = v->arch.user_regs.rip;
> + ctxt.rflags = v->arch.user_regs.rflags;
> + ctxt.r8 = v->arch.user_regs.r8;
> + ctxt.r9 = v->arch.user_regs.r9;
> + ctxt.r10 = v->arch.user_regs.r10;
> + ctxt.r11 = v->arch.user_regs.r11;
> + ctxt.r12 = v->arch.user_regs.r12;
> + ctxt.r13 = v->arch.user_regs.r13;
> + ctxt.r14 = v->arch.user_regs.r14;
> + ctxt.r15 = v->arch.user_regs.r15;
> + ctxt.dr0 = v->arch.debugreg[0];
> + ctxt.dr1 = v->arch.debugreg[1];
> + ctxt.dr2 = v->arch.debugreg[2];
> + ctxt.dr3 = v->arch.debugreg[3];
> + ctxt.dr6 = v->arch.debugreg[6];
> + ctxt.dr7 = v->arch.debugreg[7];
> +
> + if ( hvm_save_entry(CPU, v->vcpu_id, h, &ctxt) != 0 )
> + return 1;
>
> - if ( hvm_save_entry(CPU, v->vcpu_id, h, &ctxt) != 0 )
> - return 1;
> - }
> return 0;
> }
>
> @@ -1176,43 +1155,29 @@ static int hvm_load_cpu_ctxt(struct domain *d,
> hvm_domain_context_t *h)
> return 0;
> }
>
> -HVM_REGISTER_SAVE_RESTORE(CPU, hvm_save_cpu_ctxt,
> hvm_load_cpu_ctxt,
> +HVM_REGISTER_SAVE_RESTORE(CPU, hvm_save_cpu_ctxt_one,
> hvm_load_cpu_ctxt,
> 1, HVMSR_PER_VCPU);
>
> #define HVM_CPU_XSAVE_SIZE(xcr0) (offsetof(struct hvm_hw_cpu_xsave,
> \
> save_area) + \
> xstate_ctxt_size(xcr0))
>
> -static void hvm_save_cpu_xsave_states_one(struct vcpu *v, struct
> hvm_hw_cpu_xsave *ctxt)
> +static int hvm_save_cpu_xsave_states_one(struct vcpu *v,
> hvm_domain_context_t *h)
> {
> + unsigned int size = HVM_CPU_XSAVE_SIZE(v->arch.xcr0_accum);
> + struct hvm_hw_cpu_xsave *ctxt;
> +
> + if ( !xsave_enabled(v) )
> + return 2;
> + if ( _hvm_init_entry(h, CPU_XSAVE_CODE, v->vcpu_id, size) )
> + return 1;
> + ctxt = (struct hvm_hw_cpu_xsave *)&h->data[h->cur];
> + h->cur += size;
> ctxt->xfeature_mask = xfeature_mask;
> ctxt->xcr0 = v->arch.xcr0;
> ctxt->xcr0_accum = v->arch.xcr0_accum;
> -}
> -
> -static int hvm_save_cpu_xsave_states(struct domain *d,
> hvm_domain_context_t *h)
> -{
> - struct vcpu *v;
> - struct hvm_hw_cpu_xsave *ctxt;
> -
> - if ( !cpu_has_xsave )
> - return 0; /* do nothing */
> -
> - for_each_vcpu ( d, v )
> - {
> - unsigned int size = HVM_CPU_XSAVE_SIZE(v->arch.xcr0_accum);
> -
> - if ( !xsave_enabled(v) )
> - continue;
> - if ( _hvm_init_entry(h, CPU_XSAVE_CODE, v->vcpu_id, size) )
> - return 1;
> - ctxt = (struct hvm_hw_cpu_xsave *)&h->data[h->cur];
> - h->cur += size;
> -
> - hvm_save_cpu_xsave_states_one(v, ctxt);
> - expand_xsave_states(v, &ctxt->save_area,
> - size - offsetof(typeof(*ctxt), save_area));
> - }
> + expand_xsave_states(v, &ctxt->save_area,
> + size - offsetof(typeof(*ctxt), save_area));
>
> return 0;
> }
> @@ -1354,10 +1319,18 @@ static const uint32_t msrs_to_send[] = {
> };
> static unsigned int __read_mostly msr_count_max =
> ARRAY_SIZE(msrs_to_send);
>
> -static int hvm_save_cpu_msrs_one(struct vcpu *v, struct hvm_msr *ctxt)
> +static int hvm_save_cpu_msrs_one(struct vcpu *v, hvm_domain_context_t
> *h)
> {
> +
> + struct hvm_save_descriptor *desc = _p(&h->data[h->cur]);
> + struct hvm_msr *ctxt;
> unsigned int i;
>
> + if ( _hvm_init_entry(h, CPU_MSR_CODE, v->vcpu_id,
> + HVM_CPU_MSR_SIZE(msr_count_max)) )
> + return 1;
> + ctxt = (struct hvm_msr *)&h->data[h->cur];
> + ctxt->count = 0;
> for ( i = 0; i < ARRAY_SIZE(msrs_to_send); ++i )
> {
> uint64_t val;
> @@ -1383,45 +1356,23 @@ static int hvm_save_cpu_msrs_one(struct vcpu
> *v, struct hvm_msr *ctxt)
> ctxt->msr[ctxt->count].index = msrs_to_send[i];
> ctxt->msr[ctxt->count++].val = val;
> }
> - return 0;
> -}
> -
> -static int hvm_save_cpu_msrs(struct domain *d, hvm_domain_context_t
> *h)
> -{
> - struct vcpu *v;
> -
> - for_each_vcpu ( d, v )
> - {
> - struct hvm_save_descriptor *desc = _p(&h->data[h->cur]);
> - struct hvm_msr *ctxt;
> - unsigned int i;
> -
> - if ( _hvm_init_entry(h, CPU_MSR_CODE, v->vcpu_id,
> - HVM_CPU_MSR_SIZE(msr_count_max)) )
> - return 1;
> - ctxt = (struct hvm_msr *)&h->data[h->cur];
> - ctxt->count = 0;
> + if ( hvm_funcs.save_msr )
> + hvm_funcs.save_msr(v, ctxt);
>
> - hvm_save_cpu_msrs_one(v, ctxt);
> + ASSERT(ctxt->count <= msr_count_max);
>
> - if ( hvm_funcs.save_msr )
> - hvm_funcs.save_msr(v, ctxt);
> -
> - ASSERT(ctxt->count <= msr_count_max);
> -
> - for ( i = 0; i < ctxt->count; ++i )
> - ctxt->msr[i]._rsvd = 0;
> + for ( i = 0; i < ctxt->count; ++i )
> + ctxt->msr[i]._rsvd = 0;
>
> - if ( ctxt->count )
> - {
> - /* Rewrite length to indicate how much space we actually used. */
> - desc->length = HVM_CPU_MSR_SIZE(ctxt->count);
> - h->cur += HVM_CPU_MSR_SIZE(ctxt->count);
> - }
> - else
> - /* or rewind and remove the descriptor from the stream. */
> - h->cur -= sizeof(struct hvm_save_descriptor);
> + if ( ctxt->count )
> + {
> + /* Rewrite length to indicate how much space we actually used. */
> + desc->length = HVM_CPU_MSR_SIZE(ctxt->count);
> + h->cur += HVM_CPU_MSR_SIZE(ctxt->count);
> }
> + else
> + /* or rewind and remove the descriptor from the stream. */
> + h->cur -= sizeof(struct hvm_save_descriptor);
>
> return 0;
> }
> @@ -1517,7 +1468,7 @@ static int __init
> hvm_register_CPU_save_and_restore(void)
> {
> hvm_register_savevm(CPU_XSAVE_CODE,
> "CPU_XSAVE",
> - hvm_save_cpu_xsave_states,
> + hvm_save_cpu_xsave_states_one,
> hvm_load_cpu_xsave_states,
> HVM_CPU_XSAVE_SIZE(xfeature_mask) +
> sizeof(struct hvm_save_descriptor),
> @@ -1529,7 +1480,7 @@ static int __init
> hvm_register_CPU_save_and_restore(void)
> if ( msr_count_max )
> hvm_register_savevm(CPU_MSR_CODE,
> "CPU_MSR",
> - hvm_save_cpu_msrs,
> + hvm_save_cpu_msrs_one,
> hvm_load_cpu_msrs,
> HVM_CPU_MSR_SIZE(msr_count_max) +
> sizeof(struct hvm_save_descriptor),
> diff --git a/xen/arch/x86/hvm/i8254.c b/xen/arch/x86/hvm/i8254.c
> index 992f08d..e0d2255 100644
> --- a/xen/arch/x86/hvm/i8254.c
> +++ b/xen/arch/x86/hvm/i8254.c
> @@ -390,8 +390,9 @@ void pit_stop_channel0_irq(PITState *pit)
> spin_unlock(&pit->lock);
> }
>
> -static int pit_save(struct domain *d, hvm_domain_context_t *h)
> +static int pit_save(struct vcpu *v, hvm_domain_context_t *h)
> {
> + struct domain *d = v->domain;
> PITState *pit = domain_vpit(d);
> int rc;
>
> diff --git a/xen/arch/x86/hvm/irq.c b/xen/arch/x86/hvm/irq.c
> index c85d004..72acb73 100644
> --- a/xen/arch/x86/hvm/irq.c
> +++ b/xen/arch/x86/hvm/irq.c
> @@ -630,8 +630,9 @@ static int __init dump_irq_info_key_init(void)
> }
> __initcall(dump_irq_info_key_init);
>
> -static int irq_save_pci(struct domain *d, hvm_domain_context_t *h)
> +static int irq_save_pci(struct vcpu *v, hvm_domain_context_t *h)
> {
> + struct domain *d = v->domain;
> struct hvm_irq *hvm_irq = hvm_domain_irq(d);
> unsigned int asserted, pdev, pintx;
> int rc;
> @@ -662,16 +663,18 @@ static int irq_save_pci(struct domain *d,
> hvm_domain_context_t *h)
> return rc;
> }
>
> -static int irq_save_isa(struct domain *d, hvm_domain_context_t *h)
> +static int irq_save_isa(struct vcpu *v, hvm_domain_context_t *h)
> {
> + struct domain *d = v->domain;
> struct hvm_irq *hvm_irq = hvm_domain_irq(d);
>
> /* Save ISA IRQ lines */
> return ( hvm_save_entry(ISA_IRQ, 0, h, &hvm_irq->isa_irq) );
> }
>
> -static int irq_save_link(struct domain *d, hvm_domain_context_t *h)
> +static int irq_save_link(struct vcpu *v, hvm_domain_context_t *h)
> {
> + struct domain *d = v->domain;
> struct hvm_irq *hvm_irq = hvm_domain_irq(d);
>
> /* Save PCI-ISA link state */
> diff --git a/xen/arch/x86/hvm/mtrr.c b/xen/arch/x86/hvm/mtrr.c
> index d311031..b0e0a60 100644
> --- a/xen/arch/x86/hvm/mtrr.c
> +++ b/xen/arch/x86/hvm/mtrr.c
> @@ -666,46 +666,36 @@ int hvm_set_mem_pinned_cacheattr(struct
> domain *d, uint64_t gfn_start,
> return 0;
> }
>
> -static void hvm_save_mtrr_msr_one(struct vcpu *v, struct hvm_hw_mtrr
> *hw_mtrr)
> +static int hvm_save_mtrr_msr_one(struct vcpu *v, hvm_domain_context_t
> *h)
> {
> - struct mtrr_state *mtrr_state = &v->arch.hvm_vcpu.mtrr;
> int i;
> + struct mtrr_state *mtrr_state = &v->arch.hvm_vcpu.mtrr;
> + struct hvm_hw_mtrr hw_mtrr;
> + /* save mtrr&pat */
>
> - hvm_get_guest_pat(v, &hw_mtrr->msr_pat_cr);
> + hvm_get_guest_pat(v, &hw_mtrr.msr_pat_cr);
>
> - hw_mtrr->msr_mtrr_def_type = mtrr_state->def_type
> + hw_mtrr.msr_mtrr_def_type = mtrr_state->def_type
> | (mtrr_state->enabled << 10);
> - hw_mtrr->msr_mtrr_cap = mtrr_state->mtrr_cap;
> + hw_mtrr.msr_mtrr_cap = mtrr_state->mtrr_cap;
>
> for ( i = 0; i < MTRR_VCNT; i++ )
> {
> /* save physbase */
> - hw_mtrr->msr_mtrr_var[i*2] =
> + hw_mtrr.msr_mtrr_var[i*2] =
> ((uint64_t*)mtrr_state->var_ranges)[i*2];
> /* save physmask */
> - hw_mtrr->msr_mtrr_var[i*2+1] =
> + hw_mtrr.msr_mtrr_var[i*2+1] =
> ((uint64_t*)mtrr_state->var_ranges)[i*2+1];
> }
>
> for ( i = 0; i < NUM_FIXED_MSR; i++ )
> - hw_mtrr->msr_mtrr_fixed[i] =
> + hw_mtrr.msr_mtrr_fixed[i] =
> ((uint64_t*)mtrr_state->fixed_ranges)[i];
>
> -}
> -
> -static int hvm_save_mtrr_msr(struct domain *d, hvm_domain_context_t
> *h)
> -{
> - struct vcpu *v;
> - struct hvm_hw_mtrr hw_mtrr;
> - /* save mtrr&pat */
> -
> - for_each_vcpu(d, v)
> - {
> - hvm_save_mtrr_msr_one(v, &hw_mtrr);
> + if ( hvm_save_entry(MTRR, v->vcpu_id, h, &hw_mtrr) != 0 )
> + return 1;
>
> - if ( hvm_save_entry(MTRR, v->vcpu_id, h, &hw_mtrr) != 0 )
> - return 1;
> - }
> return 0;
> }
>
> @@ -751,7 +741,7 @@ static int hvm_load_mtrr_msr(struct domain *d,
> hvm_domain_context_t *h)
> return 0;
> }
>
> -HVM_REGISTER_SAVE_RESTORE(MTRR, hvm_save_mtrr_msr,
> hvm_load_mtrr_msr,
> +HVM_REGISTER_SAVE_RESTORE(MTRR, hvm_save_mtrr_msr_one,
> hvm_load_mtrr_msr,
> 1, HVMSR_PER_VCPU);
>
> void memory_type_changed(struct domain *d)
> diff --git a/xen/arch/x86/hvm/pmtimer.c b/xen/arch/x86/hvm/pmtimer.c
> index 435647f..d8dcbc2 100644
> --- a/xen/arch/x86/hvm/pmtimer.c
> +++ b/xen/arch/x86/hvm/pmtimer.c
> @@ -249,8 +249,9 @@ static int handle_pmt_io(
> return X86EMUL_OKAY;
> }
>
> -static int acpi_save(struct domain *d, hvm_domain_context_t *h)
> +static int acpi_save(struct vcpu *v, hvm_domain_context_t *h)
> {
> + struct domain *d = v->domain;
> struct hvm_hw_acpi *acpi = &d->arch.hvm_domain.acpi;
> PMTState *s = &d->arch.hvm_domain.pl_time->vpmt;
> uint32_t x, msb = acpi->tmr_val & TMR_VAL_MSB;
> diff --git a/xen/arch/x86/hvm/rtc.c b/xen/arch/x86/hvm/rtc.c
> index cb75b99..58b70fc 100644
> --- a/xen/arch/x86/hvm/rtc.c
> +++ b/xen/arch/x86/hvm/rtc.c
> @@ -737,8 +737,9 @@ void rtc_migrate_timers(struct vcpu *v)
> }
>
> /* Save RTC hardware state */
> -static int rtc_save(struct domain *d, hvm_domain_context_t *h)
> +static int rtc_save(struct vcpu *v, hvm_domain_context_t *h)
> {
> + struct domain *d = v->domain;
> RTCState *s = domain_vrtc(d);
> int rc;
>
> diff --git a/xen/arch/x86/hvm/save.c b/xen/arch/x86/hvm/save.c
> index 8984a23..fe24f08 100644
> --- a/xen/arch/x86/hvm/save.c
> +++ b/xen/arch/x86/hvm/save.c
> @@ -135,9 +135,12 @@ size_t hvm_save_size(struct domain *d)
> int hvm_save_one(struct domain *d, unsigned int typecode, unsigned int
> instance,
> XEN_GUEST_HANDLE_64(uint8) handle, uint64_t *bufsz)
> {
> - int rv;
> + int rv = 0;
> hvm_domain_context_t ctxt = { };
> const struct hvm_save_descriptor *desc;
> + bool is_single_instance = false;
> + uint32_t off = 0;
> + struct vcpu *v;
>
> if ( d->is_dying ||
> typecode > HVM_SAVE_CODE_MAX ||
> @@ -145,43 +148,89 @@ int hvm_save_one(struct domain *d, unsigned int
> typecode, unsigned int instance,
> !hvm_sr_handlers[typecode].save )
> return -EINVAL;
>
> + if( hvm_sr_handlers[typecode].kind == HVMSR_PER_VCPU &&
> + instance < d->max_vcpus )
> + is_single_instance = true;
> +
> ctxt.size = hvm_sr_handlers[typecode].size;
> - if ( hvm_sr_handlers[typecode].kind == HVMSR_PER_VCPU )
> + if ( hvm_sr_handlers[typecode].kind == HVMSR_PER_VCPU &&
> + instance == d->max_vcpus )
> ctxt.size *= d->max_vcpus;
> ctxt.data = xmalloc_bytes(ctxt.size);
> if ( !ctxt.data )
> return -ENOMEM;
>
> - if ( (rv = hvm_sr_handlers[typecode].save(d, &ctxt)) != 0 )
> - printk(XENLOG_G_ERR "HVM%d save: failed to save type %"PRIu16"
> (%d)\n",
> - d->domain_id, typecode, rv);
> - else if ( rv = -ENOENT, ctxt.cur >= sizeof(*desc) )
> - {
> - uint32_t off;
>
> - for ( off = 0; off <= (ctxt.cur - sizeof(*desc)); off += desc->length )
> + if( is_single_instance )
> + vcpu_pause(d->vcpu[instance]);
> + else
> + domain_pause(d);
> +
> + if( is_single_instance )
> + {
> + if ( (rv = hvm_sr_handlers[typecode].save(d->vcpu[instance],
> + &ctxt)) != 0 )
> {
> - desc = (void *)(ctxt.data + off);
> - /* Move past header */
> - off += sizeof(*desc);
> - if ( ctxt.cur < desc->length ||
> - off > ctxt.cur - desc->length )
> - break;
> - if ( instance == desc->instance )
> - {
> - rv = 0;
> - if ( guest_handle_is_null(handle) )
> - *bufsz = desc->length;
> - else if ( *bufsz < desc->length )
> - rv = -ENOBUFS;
> - else if ( copy_to_guest(handle, ctxt.data + off, desc->length) )
> - rv = -EFAULT;
> - else
> - *bufsz = desc->length;
> - break;
> - }
> + printk(XENLOG_G_ERR "HVM%d save: failed to save type %"PRIu16"
> (%d)\n",
> + d->domain_id, typecode, rv);
> + vcpu_unpause(d->vcpu[instance]);
> + }
> + else if ( rv = -ENOENT, ctxt.cur >= sizeof(*desc) )
Not a fan of this style. Any reason you can't init rv inside the else-if clause below?
> + {
> + desc = (void *)(ctxt.data);
> + /* Move past header */
> + off = sizeof(*desc);
> + if ( ctxt.cur < desc->length ||
> + off > ctxt.cur - desc->length )
> + rv = -EFAULT;
> + rv = 0;
> + if ( guest_handle_is_null(handle) )
> + *bufsz = desc->length;
> + else if ( *bufsz < desc->length )
> + rv = -ENOBUFS;
> + else if ( copy_to_guest(handle, ctxt.data + off, desc->length) )
> + rv = -EFAULT;
> + else
> + *bufsz = desc->length;
> + vcpu_unpause(d->vcpu[instance]);
> }
> }
> + else
> + {
> + for_each_vcpu ( d, v )
> + {
> + if ( (rv = hvm_sr_handlers[typecode].save(d->vcpu[instance],
> + &ctxt)) != 0 )
> + {
> + printk(XENLOG_G_ERR "HVM%d save: failed to save type
> %"PRIu16" (%d)\n",
> + d->domain_id, typecode, rv);
> + }
> + else if ( rv = -ENOENT, ctxt.cur >= sizeof(*desc) )
Same here.
> + {
> + desc = (void *)(ctxt.data + off);
> + /* Move past header */
> + off += sizeof(*desc);
> + if ( ctxt.cur < desc->length ||
> + off > ctxt.cur - desc->length )
> + break;
> + if ( instance == desc->instance )
> + {
> + rv = 0;
> + if ( guest_handle_is_null(handle) )
> + *bufsz = desc->length;
> + else if ( *bufsz < desc->length )
> + rv = -ENOBUFS;
> + else if ( copy_to_guest(handle, ctxt.data + off, desc->length) )
> + rv = -EFAULT;
> + else
> + *bufsz = desc->length;
> + break;
> + }
> + off += desc->length;
> + }
> + }
> + domain_unpause(d);
> + }
>
> xfree(ctxt.data);
> return rv;
> @@ -193,7 +242,8 @@ int hvm_save(struct domain *d,
> hvm_domain_context_t *h)
> struct hvm_save_header hdr;
> struct hvm_save_end end;
> hvm_save_handler handler;
> - unsigned int i;
> + unsigned int i, rc;
> + struct vcpu *v = NULL;
>
> if ( d->is_dying )
> return -EINVAL;
> @@ -225,12 +275,19 @@ int hvm_save(struct domain *d,
> hvm_domain_context_t *h)
> {
> printk(XENLOG_G_INFO "HVM%d save: %s\n",
> d->domain_id, hvm_sr_handlers[i].name);
> - if ( handler(d, h) != 0 )
> + for_each_vcpu ( d, v )
> {
> - printk(XENLOG_G_ERR
> - "HVM%d save: failed to save type %"PRIu16"\n",
> - d->domain_id, i);
> - return -EFAULT;
> + rc = handler(v, h);
> + if( rc == 2 )
What does the magic '2' mean? I think we need at least a comment here, if not a #define or enum return.
Paul
> + continue;
> +
> + if( rc != 0 )
> + {
> + printk(XENLOG_G_ERR
> + "HVM%d save: failed to save type %"PRIu16"\n",
> + d->domain_id, i);
> + return -EFAULT;
> + }
> }
> }
> }
> diff --git a/xen/arch/x86/hvm/vioapic.c b/xen/arch/x86/hvm/vioapic.c
> index 97b419f..86d02cf 100644
> --- a/xen/arch/x86/hvm/vioapic.c
> +++ b/xen/arch/x86/hvm/vioapic.c
> @@ -569,8 +569,9 @@ int vioapic_get_trigger_mode(const struct domain *d,
> unsigned int gsi)
> return vioapic->redirtbl[pin].fields.trig_mode;
> }
>
> -static int ioapic_save(struct domain *d, hvm_domain_context_t *h)
> +static int ioapic_save(struct vcpu *v, hvm_domain_context_t *h)
> {
> + struct domain *d = v->domain;
> struct hvm_vioapic *s;
>
> if ( !has_vioapic(d) )
> diff --git a/xen/arch/x86/hvm/viridian.c b/xen/arch/x86/hvm/viridian.c
> index d22c8ac..8a7c592 100644
> --- a/xen/arch/x86/hvm/viridian.c
> +++ b/xen/arch/x86/hvm/viridian.c
> @@ -990,8 +990,9 @@ out:
> return HVM_HCALL_completed;
> }
>
> -static int viridian_save_domain_ctxt(struct domain *d,
> hvm_domain_context_t *h)
> +static int viridian_save_domain_ctxt(struct vcpu *v,
> hvm_domain_context_t *h)
> {
> + struct domain *d = v->domain;
> struct hvm_viridian_domain_context ctxt = {
> .time_ref_count = d->arch.hvm_domain.viridian.time_ref_count.val,
> .hypercall_gpa = d->arch.hvm_domain.viridian.hypercall_gpa.raw,
> @@ -1026,28 +1027,19 @@ static int viridian_load_domain_ctxt(struct
> domain *d, hvm_domain_context_t *h)
> HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_DOMAIN,
> viridian_save_domain_ctxt,
> viridian_load_domain_ctxt, 1, HVMSR_PER_DOM);
>
> -static void viridian_save_vcpu_ctxt_one(struct vcpu *v, struct
> hvm_viridian_vcpu_context *ctxt)
> +static int viridian_save_vcpu_ctxt_one(struct vcpu *v,
> hvm_domain_context_t *h)
> {
> - ctxt->vp_assist_msr = v->arch.hvm_vcpu.viridian.vp_assist.msr.raw;
> - ctxt->vp_assist_pending = v->arch.hvm_vcpu.viridian.vp_assist.pending;
> - *ctxt->_pad = 0;
> -}
> -
> -static int viridian_save_vcpu_ctxt(struct domain *d,
> hvm_domain_context_t *h)
> -{
> - struct vcpu *v;
> + struct domain *d = v->domain;
> + struct hvm_viridian_vcpu_context ctxt = {
> + .vp_assist_msr = v->arch.hvm_vcpu.viridian.vp_assist.msr.raw,
> + .vp_assist_pending = v->arch.hvm_vcpu.viridian.vp_assist.pending,
> + };
>
> if ( !is_viridian_domain(d) )
> return 0;
>
> - for_each_vcpu( d, v ) {
> - struct hvm_viridian_vcpu_context ctxt;
> -
> - viridian_save_vcpu_ctxt_one(v, &ctxt);
> -
> - if ( hvm_save_entry(VIRIDIAN_VCPU, v->vcpu_id, h, &ctxt) != 0 )
> - return 1;
> - }
> + if ( hvm_save_entry(VIRIDIAN_VCPU, v->vcpu_id, h, &ctxt) != 0 )
> + return 1;
>
> return 0;
> }
> @@ -1082,7 +1074,7 @@ static int viridian_load_vcpu_ctxt(struct domain *d,
> hvm_domain_context_t *h)
> return 0;
> }
>
> -HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_VCPU, viridian_save_vcpu_ctxt,
> +HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_VCPU,
> viridian_save_vcpu_ctxt_one,
> viridian_load_vcpu_ctxt, 1, HVMSR_PER_VCPU);
>
> static int __init parse_viridian_version(const char *arg)
> diff --git a/xen/arch/x86/hvm/vlapic.c b/xen/arch/x86/hvm/vlapic.c
> index 1b9f00a..ab35400 100644
> --- a/xen/arch/x86/hvm/vlapic.c
> +++ b/xen/arch/x86/hvm/vlapic.c
> @@ -1435,43 +1435,35 @@ static void lapic_rearm(struct vlapic *s)
> s->timer_last_update = s->pt.last_plt_gtime;
> }
>
> -static int lapic_save_hidden(struct domain *d, hvm_domain_context_t *h)
> +static int lapic_save_hidden(struct vcpu *v, hvm_domain_context_t *h)
> {
> - struct vcpu *v;
> + struct domain *d = v->domain;
> struct vlapic *s;
> int rc = 0;
>
> if ( !has_vlapic(d) )
> return 0;
>
> - for_each_vcpu ( d, v )
> - {
> - s = vcpu_vlapic(v);
> - if ( (rc = hvm_save_entry(LAPIC, v->vcpu_id, h, &s->hw)) != 0 )
> - break;
> - }
> + s = vcpu_vlapic(v);
> + rc = hvm_save_entry(LAPIC, v->vcpu_id, h, &s->hw);
>
> return rc;
> }
>
> -static int lapic_save_regs(struct domain *d, hvm_domain_context_t *h)
> +static int lapic_save_regs(struct vcpu *v, hvm_domain_context_t *h)
> {
> - struct vcpu *v;
> + struct domain *d = v->domain;
> struct vlapic *s;
> int rc = 0;
>
> if ( !has_vlapic(d) )
> return 0;
>
> - for_each_vcpu ( d, v )
> - {
> - if ( hvm_funcs.sync_pir_to_irr )
> - hvm_funcs.sync_pir_to_irr(v);
> + if ( hvm_funcs.sync_pir_to_irr )
> + hvm_funcs.sync_pir_to_irr(v);
>
> - s = vcpu_vlapic(v);
> - if ( (rc = hvm_save_entry(LAPIC_REGS, v->vcpu_id, h, s->regs)) != 0 )
> - break;
> - }
> + s = vcpu_vlapic(v);
> + rc = hvm_save_entry(LAPIC_REGS, v->vcpu_id, h, s->regs);
>
> return rc;
> }
> diff --git a/xen/arch/x86/hvm/vpic.c b/xen/arch/x86/hvm/vpic.c
> index e160bbd..bad5066 100644
> --- a/xen/arch/x86/hvm/vpic.c
> +++ b/xen/arch/x86/hvm/vpic.c
> @@ -371,8 +371,9 @@ static int vpic_intercept_elcr_io(
> return X86EMUL_OKAY;
> }
>
> -static int vpic_save(struct domain *d, hvm_domain_context_t *h)
> +static int vpic_save(struct vcpu *v, hvm_domain_context_t *h)
> {
> + struct domain *d = v->domain;
> struct hvm_hw_vpic *s;
> int i;
>
> diff --git a/xen/include/asm-x86/hvm/save.h b/xen/include/asm-
> x86/hvm/save.h
> index f889e8f..fe642ab 100644
> --- a/xen/include/asm-x86/hvm/save.h
> +++ b/xen/include/asm-x86/hvm/save.h
> @@ -95,7 +95,7 @@ static inline uint16_t hvm_load_instance(struct
> hvm_domain_context *h)
> * The save handler may save multiple instances of a type into the buffer;
> * the load handler will be called once for each instance found when
> * restoring. Both return non-zero on error. */
> -typedef int (*hvm_save_handler) (struct domain *d,
> +typedef int (*hvm_save_handler) (struct vcpu *v,
> hvm_domain_context_t *h);
> typedef int (*hvm_load_handler) (struct domain *d,
> hvm_domain_context_t *h);
> --
> 2.7.4
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: [PATCH v5 8/8] x86/domctl: Don't pause the whole domain if only getting vcpu state
2018-05-29 14:58 ` [PATCH v5 8/8] x86/domctl: Don't pause the whole domain if only getting vcpu state Alexandru Isaila
2018-06-04 9:21 ` Paul Durrant
@ 2018-06-06 12:35 ` Jan Beulich
1 sibling, 0 replies; 12+ messages in thread
From: Jan Beulich @ 2018-06-06 12:35 UTC (permalink / raw)
To: aisaila; +Cc: Andrew Cooper, Paul Durrant, Wei Liu, Ian Jackson, xen-devel
>>> On 29.05.18 at 16:58, <aisaila@bitdefender.com> wrote:
> This patch is focused on merging the *save() funcs to the *save_one()
> funcs to remove redundancy. Also the for loop is moved to the caller so
> now we can save info for a single vcpu instance.
>
> Signed-off-by: Alexandru Isaila <aisaila@bitdefender.com>
>
> ---
> Changes since V4:
> - Save funcs are deleted
> - vcpu si added to the param list
> - save_one funcs are now static
> - Add blank line after "/* save mtrr&pat */" comment
> ---
> xen/arch/x86/cpu/mcheck/vmce.c | 28 ++--
> xen/arch/x86/hvm/hpet.c | 3 +-
> xen/arch/x86/hvm/hvm.c | 287 +++++++++++++++++------------------------
> xen/arch/x86/hvm/i8254.c | 3 +-
> xen/arch/x86/hvm/irq.c | 9 +-
> xen/arch/x86/hvm/mtrr.c | 36 ++----
> xen/arch/x86/hvm/pmtimer.c | 3 +-
> xen/arch/x86/hvm/rtc.c | 3 +-
> xen/arch/x86/hvm/save.c | 125 +++++++++++++-----
> xen/arch/x86/hvm/vioapic.c | 3 +-
> xen/arch/x86/hvm/viridian.c | 30 ++---
> xen/arch/x86/hvm/vlapic.c | 28 ++--
> xen/arch/x86/hvm/vpic.c | 3 +-
> xen/include/asm-x86/hvm/save.h | 2 +-
> 14 files changed, 271 insertions(+), 292 deletions(-)
Since I see you've already sent v6: This patch is still way too large for its subject.
It should, as said before, in particular have no need to touch most of the files
above. IOW - there's more splitting work to be done afaict.
Jan
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel
^ permalink raw reply [flat|nested] 12+ messages in thread
end of thread, other threads:[~2018-06-06 12:35 UTC | newest]
Thread overview: 12+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2018-05-29 14:58 [PATCH v5 1/8] x86/cpu: Introduce vmce_save_vcpu_ctxt_one() func Alexandru Isaila
2018-05-29 14:58 ` [PATCH v5 2/8] x86/hvm: Introduce hvm_save_tsc_adjust_one() func Alexandru Isaila
2018-05-29 14:58 ` [PATCH v5 3/8] x86/hvm: Introduce hvm_save_cpu_ctxt_one func Alexandru Isaila
2018-05-29 14:58 ` [PATCH v5 4/8] x86/hvm: Introduce hvm_save_cpu_xsave_states_one Alexandru Isaila
2018-05-29 14:58 ` [PATCH v5 5/8] x86/hvm: Introduce hvm_save_cpu_msrs_one func Alexandru Isaila
2018-06-04 9:08 ` Paul Durrant
2018-05-29 14:58 ` [PATCH v5 6/8] x86/hvm: Introduce hvm_save_mtrr_msr_one func Alexandru Isaila
2018-05-29 14:58 ` [PATCH v5 7/8] x86/hvm: Introduce viridian_save_vcpu_ctxt_one() func Alexandru Isaila
2018-06-04 9:11 ` Paul Durrant
2018-05-29 14:58 ` [PATCH v5 8/8] x86/domctl: Don't pause the whole domain if only getting vcpu state Alexandru Isaila
2018-06-04 9:21 ` Paul Durrant
2018-06-06 12:35 ` Jan Beulich
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.