All of lore.kernel.org
 help / color / mirror / Atom feed
From: isaku.yamahata@intel.com
To: kvm@vger.kernel.org, linux-kernel@vger.kernel.org
Cc: isaku.yamahata@intel.com, isaku.yamahata@gmail.com,
	Paolo Bonzini <pbonzini@redhat.com>,
	erdemaktas@google.com, Sean Christopherson <seanjc@google.com>,
	Sagi Shahar <sagis@google.com>
Subject: [RFC PATCH v6 073/104] KVM: TDX: handle vcpu migration over logical processor
Date: Thu,  5 May 2022 11:15:07 -0700	[thread overview]
Message-ID: <96148d16e5c3b394bec4d7f51130b2d35e33352e.1651774250.git.isaku.yamahata@intel.com> (raw)
In-Reply-To: <cover.1651774250.git.isaku.yamahata@intel.com>

From: Isaku Yamahata <isaku.yamahata@intel.com>

For vcpu migration, in the case of VMX, VCMS is flushed on the source pcpu,
and load it on the target pcpu.  There are corresponding TDX SEAMCALL APIs,
call them on vcpu migration.  The logic is mostly same as VMX except the
TDX SEAMCALLs are used.

When shutting down the machine, (VMX or TDX) vcpus needs to be shutdown on
each pcpu.  Do the similar for TDX with TDX SEAMCALL APIs.

Signed-off-by: Isaku Yamahata <isaku.yamahata@intel.com>
---
 arch/x86/kvm/vmx/main.c    |  43 +++++++++++--
 arch/x86/kvm/vmx/tdx.c     | 121 +++++++++++++++++++++++++++++++++++++
 arch/x86/kvm/vmx/tdx.h     |   2 +
 arch/x86/kvm/vmx/x86_ops.h |   6 ++
 4 files changed, 168 insertions(+), 4 deletions(-)

diff --git a/arch/x86/kvm/vmx/main.c b/arch/x86/kvm/vmx/main.c
index f101f358d90c..ad09988c4faa 100644
--- a/arch/x86/kvm/vmx/main.c
+++ b/arch/x86/kvm/vmx/main.c
@@ -17,6 +17,25 @@ static bool vt_is_vm_type_supported(unsigned long type)
 		(enable_tdx && tdx_is_vm_type_supported(type));
 }
 
+static int vt_hardware_enable(void)
+{
+	int ret;
+
+	ret = vmx_hardware_enable();
+	if (ret)
+		return ret;
+
+	tdx_hardware_enable();
+	return 0;
+}
+
+static void vt_hardware_disable(void)
+{
+	/* Note, TDX *and* VMX need to be disabled if TDX is enabled. */
+	tdx_hardware_disable();
+	vmx_hardware_disable();
+}
+
 static __init int vt_hardware_setup(void)
 {
 	int ret;
@@ -151,6 +170,14 @@ static fastpath_t vt_vcpu_run(struct kvm_vcpu *vcpu)
 	return vmx_vcpu_run(vcpu);
 }
 
+static void vt_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
+{
+	if (is_td_vcpu(vcpu))
+		return tdx_vcpu_load(vcpu, cpu);
+
+	return vmx_vcpu_load(vcpu, cpu);
+}
+
 static void vt_flush_tlb_all(struct kvm_vcpu *vcpu)
 {
 	if (is_td_vcpu(vcpu))
@@ -192,6 +219,14 @@ static void vt_load_mmu_pgd(struct kvm_vcpu *vcpu, hpa_t root_hpa,
 	vmx_load_mmu_pgd(vcpu, root_hpa, pgd_level);
 }
 
+static void vt_sched_in(struct kvm_vcpu *vcpu, int cpu)
+{
+	if (is_td_vcpu(vcpu))
+		return;
+
+	vmx_sched_in(vcpu, cpu);
+}
+
 static int vt_mem_enc_ioctl(struct kvm *kvm, void __user *argp)
 {
 	if (!is_td(kvm))
@@ -214,8 +249,8 @@ struct kvm_x86_ops vt_x86_ops __initdata = {
 	.hardware_unsetup = vt_hardware_unsetup,
 	.check_processor_compatibility = vmx_check_processor_compatibility,
 
-	.hardware_enable = vmx_hardware_enable,
-	.hardware_disable = vmx_hardware_disable,
+	.hardware_enable = vt_hardware_enable,
+	.hardware_disable = vt_hardware_disable,
 	.has_emulated_msr = vmx_has_emulated_msr,
 
 	.is_vm_type_supported = vt_is_vm_type_supported,
@@ -231,7 +266,7 @@ struct kvm_x86_ops vt_x86_ops __initdata = {
 	.vcpu_reset = vt_vcpu_reset,
 
 	.prepare_switch_to_guest = vt_prepare_switch_to_guest,
-	.vcpu_load = vmx_vcpu_load,
+	.vcpu_load = vt_vcpu_load,
 	.vcpu_put = vt_vcpu_put,
 
 	.update_exception_bitmap = vmx_update_exception_bitmap,
@@ -317,7 +352,7 @@ struct kvm_x86_ops vt_x86_ops __initdata = {
 
 	.request_immediate_exit = vmx_request_immediate_exit,
 
-	.sched_in = vmx_sched_in,
+	.sched_in = vt_sched_in,
 
 	.cpu_dirty_log_size = PML_ENTITY_NUM,
 	.update_cpu_dirty_logging = vmx_update_cpu_dirty_logging,
diff --git a/arch/x86/kvm/vmx/tdx.c b/arch/x86/kvm/vmx/tdx.c
index f92941586b42..6233c65b2a48 100644
--- a/arch/x86/kvm/vmx/tdx.c
+++ b/arch/x86/kvm/vmx/tdx.c
@@ -61,6 +61,14 @@ static struct tdx_capabilities tdx_caps;
 static DEFINE_MUTEX(tdx_lock);
 static struct mutex *tdx_mng_key_config_lock;
 
+/*
+ * A per-CPU list of TD vCPUs associated with a given CPU.  Used when a CPU
+ * is brought down to invoke TDH_VP_FLUSH on the approapriate TD vCPUS.
+ * Protected by interrupt mask.  This list is manipulated in process context
+ * of vcpu and IPI callback.  See tdx_flush_vp_on_cpu().
+ */
+static DEFINE_PER_CPU(struct list_head, associated_tdvcpus);
+
 static __always_inline hpa_t set_hkid_to_hpa(hpa_t pa, u16 hkid)
 {
 	pa &= ~hkid_mask;
@@ -95,6 +103,36 @@ static inline bool is_td_finalized(struct kvm_tdx *kvm_tdx)
 	return kvm_tdx->finalized;
 }
 
+static inline void tdx_disassociate_vp(struct kvm_vcpu *vcpu)
+{
+	list_del(&to_tdx(vcpu)->cpu_list);
+
+	/*
+	 * Ensure tdx->cpu_list is updated is before setting vcpu->cpu to -1,
+	 * otherwise, a different CPU can see vcpu->cpu = -1 and add the vCPU
+	 * to its list before its deleted from this CPUs list.
+	 */
+	smp_wmb();
+
+	vcpu->cpu = -1;
+}
+
+void tdx_hardware_enable(void)
+{
+	INIT_LIST_HEAD(&per_cpu(associated_tdvcpus, raw_smp_processor_id()));
+}
+
+void tdx_hardware_disable(void)
+{
+	int cpu = raw_smp_processor_id();
+	struct list_head *tdvcpus = &per_cpu(associated_tdvcpus, cpu);
+	struct vcpu_tdx *tdx, *tmp;
+
+	/* Safe variant needed as tdx_disassociate_vp() deletes the entry. */
+	list_for_each_entry_safe(tdx, tmp, tdvcpus, cpu_list)
+		tdx_disassociate_vp(&tdx->vcpu);
+}
+
 static void tdx_clear_page(unsigned long page)
 {
 	const void *zero_page = (const void *) __va(page_to_phys(ZERO_PAGE(0)));
@@ -171,6 +209,41 @@ static void tdx_reclaim_td_page(struct tdx_td_page *page)
 	free_page(page->va);
 }
 
+static void tdx_flush_vp(void *arg)
+{
+	struct kvm_vcpu *vcpu = arg;
+	u64 err;
+
+	lockdep_assert_irqs_disabled();
+
+	/* Task migration can race with CPU offlining. */
+	if (vcpu->cpu != raw_smp_processor_id())
+		return;
+
+	/*
+	 * No need to do TDH_VP_FLUSH if the vCPU hasn't been initialized.  The
+	 * list tracking still needs to be updated so that it's correct if/when
+	 * the vCPU does get initialized.
+	 */
+	if (is_td_vcpu_created(to_tdx(vcpu))) {
+		err = tdh_vp_flush(to_tdx(vcpu)->tdvpr.pa);
+		if (unlikely(err && err != TDX_VCPU_NOT_ASSOCIATED)) {
+			if (WARN_ON_ONCE(err))
+				pr_tdx_error(TDH_VP_FLUSH, err, NULL);
+		}
+	}
+
+	tdx_disassociate_vp(vcpu);
+}
+
+static void tdx_flush_vp_on_cpu(struct kvm_vcpu *vcpu)
+{
+	if (unlikely(vcpu->cpu == -1))
+		return;
+
+	smp_call_function_single(vcpu->cpu, tdx_flush_vp, vcpu, 1);
+}
+
 static int tdx_do_tdh_phymem_cache_wb(void *param)
 {
 	u64 err = 0;
@@ -195,9 +268,11 @@ void tdx_mmu_release_hkid(struct kvm *kvm)
 	struct kvm_tdx *kvm_tdx = to_kvm_tdx(kvm);
 	cpumask_var_t packages;
 	bool cpumask_allocated;
+	struct kvm_vcpu *vcpu;
 	u64 err;
 	int ret;
 	int i;
+	unsigned long j;
 
 	if (!is_hkid_assigned(kvm_tdx))
 		return;
@@ -205,6 +280,19 @@ void tdx_mmu_release_hkid(struct kvm *kvm)
 	if (!is_td_created(kvm_tdx))
 		goto free_hkid;
 
+	kvm_for_each_vcpu(j, vcpu, kvm)
+		tdx_flush_vp_on_cpu(vcpu);
+
+	mutex_lock(&tdx_lock);
+	err = tdh_mng_vpflushdone(kvm_tdx->tdr.pa);
+	mutex_unlock(&tdx_lock);
+	if (WARN_ON_ONCE(err)) {
+		pr_tdx_error(TDH_MNG_VPFLUSHDONE, err, NULL);
+		pr_err("tdh_mng_vpflushdone failed. HKID %d is leaked.\n",
+			kvm_tdx->hkid);
+		return;
+	}
+
 	cpumask_allocated = zalloc_cpumask_var(&packages, GFP_KERNEL);
 	cpus_read_lock();
 	for_each_online_cpu(i) {
@@ -479,6 +567,26 @@ int tdx_vcpu_create(struct kvm_vcpu *vcpu)
 	return ret;
 }
 
+void tdx_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
+{
+	struct vcpu_tdx *tdx = to_tdx(vcpu);
+
+	if (vcpu->cpu == cpu)
+		return;
+
+	tdx_flush_vp_on_cpu(vcpu);
+
+	local_irq_disable();
+	/*
+	 * Pairs with the smp_wmb() in tdx_disassociate_vp() to ensure
+	 * vcpu->cpu is read before tdx->cpu_list.
+	 */
+	smp_rmb();
+
+	list_add(&tdx->cpu_list, &per_cpu(associated_tdvcpus, cpu));
+	local_irq_enable();
+}
+
 void tdx_prepare_switch_to_guest(struct kvm_vcpu *vcpu)
 {
 	struct vcpu_tdx *tdx = to_tdx(vcpu);
@@ -525,6 +633,19 @@ void tdx_vcpu_free(struct kvm_vcpu *vcpu)
 		tdx_reclaim_td_page(&tdx->tdvpx[i]);
 	kfree(tdx->tdvpx);
 	tdx_reclaim_td_page(&tdx->tdvpr);
+
+	/*
+	 * kvm_free_vcpus()
+	 *   -> kvm_unload_vcpu_mmu()
+	 *
+	 * does vcpu_load() for every vcpu after they already disassociated
+	 * from the per cpu list when tdx_vm_teardown(). So we need to
+	 * disassociate them again, otherwise the freed vcpu data will be
+	 * accessed when do list_{del,add}() on associated_tdvcpus list
+	 * later.
+	 */
+	tdx_flush_vp_on_cpu(vcpu);
+	WARN_ON(vcpu->cpu != -1);
 }
 
 void tdx_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event)
diff --git a/arch/x86/kvm/vmx/tdx.h b/arch/x86/kvm/vmx/tdx.h
index 414c15235ed0..32e05efa70f9 100644
--- a/arch/x86/kvm/vmx/tdx.h
+++ b/arch/x86/kvm/vmx/tdx.h
@@ -85,6 +85,8 @@ struct vcpu_tdx {
 	struct tdx_td_page tdvpr;
 	struct tdx_td_page *tdvpx;
 
+	struct list_head cpu_list;
+
 	union tdx_exit_reason exit_reason;
 
 	bool initialized;
diff --git a/arch/x86/kvm/vmx/x86_ops.h b/arch/x86/kvm/vmx/x86_ops.h
index a2deb42794c0..6594bcb717cd 100644
--- a/arch/x86/kvm/vmx/x86_ops.h
+++ b/arch/x86/kvm/vmx/x86_ops.h
@@ -132,6 +132,8 @@ void vmx_setup_mce(struct kvm_vcpu *vcpu);
 int __init tdx_hardware_setup(struct kvm_x86_ops *x86_ops);
 bool tdx_is_vm_type_supported(unsigned long type);
 void tdx_hardware_unsetup(void);
+void tdx_hardware_enable(void);
+void tdx_hardware_disable(void);
 int tdx_dev_ioctl(void __user *argp);
 
 int tdx_vm_init(struct kvm *kvm);
@@ -144,6 +146,7 @@ void tdx_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event);
 fastpath_t tdx_vcpu_run(struct kvm_vcpu *vcpu);
 void tdx_prepare_switch_to_guest(struct kvm_vcpu *vcpu);
 void tdx_vcpu_put(struct kvm_vcpu *vcpu);
+void tdx_vcpu_load(struct kvm_vcpu *vcpu, int cpu);
 
 int tdx_vm_ioctl(struct kvm *kvm, void __user *argp);
 int tdx_vcpu_ioctl(struct kvm_vcpu *vcpu, void __user *argp);
@@ -154,6 +157,8 @@ void tdx_load_mmu_pgd(struct kvm_vcpu *vcpu, hpa_t root_hpa, int root_level);
 static inline int tdx_hardware_setup(struct kvm_x86_ops *x86_ops) { return 0; }
 static inline bool tdx_is_vm_type_supported(unsigned long type) { return false; }
 static inline void tdx_hardware_unsetup(void) {}
+static inline void tdx_hardware_enable(void) {}
+static inline void tdx_hardware_disable(void) {}
 static inline int tdx_dev_ioctl(void __user *argp) { return -EOPNOTSUPP; };
 
 static inline int tdx_vm_init(struct kvm *kvm) { return -EOPNOTSUPP; }
@@ -167,6 +172,7 @@ static inline void tdx_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event) {}
 static inline fastpath_t tdx_vcpu_run(struct kvm_vcpu *vcpu) { return EXIT_FASTPATH_NONE; }
 static inline void tdx_prepare_switch_to_guest(struct kvm_vcpu *vcpu) {}
 static inline void tdx_vcpu_put(struct kvm_vcpu *vcpu) {}
+static inline void tdx_vcpu_load(struct kvm_vcpu *vcpu, int cpu) {}
 
 static inline int tdx_vm_ioctl(struct kvm *kvm, void __user *argp) { return -EOPNOTSUPP; }
 static inline int tdx_vcpu_ioctl(struct kvm_vcpu *vcpu, void __user *argp) { return -EOPNOTSUPP; }
-- 
2.25.1


  parent reply	other threads:[~2022-05-05 18:23 UTC|newest]

Thread overview: 146+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2022-05-05 18:13 [RFC PATCH v6 000/104] KVM TDX basic feature support isaku.yamahata
2022-05-05 18:13 ` [RFC PATCH v6 001/104] KVM: x86: Move check_processor_compatibility from init ops to runtime ops isaku.yamahata
2022-05-05 18:13 ` [RFC PATCH v6 002/104] Partially revert "KVM: Pass kvm_init()'s opaque param to additional arch funcs" isaku.yamahata
2022-05-05 18:13 ` [RFC PATCH v6 003/104] KVM: Refactor CPU compatibility check on module initialiization isaku.yamahata
2022-05-23 22:27   ` Sagi Shahar
2022-05-26 19:03     ` Isaku Yamahata
2022-05-05 18:13 ` [RFC PATCH v6 004/104] KVM: VMX: Move out vmx_x86_ops to 'main.c' to wrap VMX and TDX isaku.yamahata
2022-05-05 18:13 ` [RFC PATCH v6 005/104] x86/virt/vmx/tdx: export platform_has_tdx isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 006/104] KVM: TDX: Detect CPU feature on kernel module initialization isaku.yamahata
2022-05-23 23:47   ` Sagi Shahar
2022-05-26 19:28     ` Isaku Yamahata
2022-06-01 22:11       ` Kai Huang
2022-05-05 18:14 ` [RFC PATCH v6 007/104] KVM: Enable hardware before doing arch VM initialization isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 008/104] KVM: x86: Refactor KVM VMX module init/exit functions isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 009/104] KVM: TDX: Add placeholders for TDX VM/vcpu structure isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 010/104] x86/virt/tdx: Add a helper function to return system wide info about TDX module isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 011/104] KVM: TDX: Initialize TDX module when loading kvm_intel.ko isaku.yamahata
2022-05-06 13:57   ` Xiaoyao Li
2022-05-06 21:29     ` Isaku Yamahata
2022-05-05 18:14 ` [RFC PATCH v6 012/104] KVM: x86: Introduce vm_type to differentiate default VMs from confidential VMs isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 013/104] KVM: TDX: Make TDX VM type supported isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 014/104] [MARKER] The start of TDX KVM patch series: TDX architectural definitions isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 015/104] KVM: TDX: Define " isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 016/104] KVM: TDX: Add TDX "architectural" error codes isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 017/104] KVM: TDX: Add C wrapper functions for SEAMCALLs to the TDX module isaku.yamahata
2022-05-06  8:56   ` Xiaoyao Li
2022-05-06 21:35     ` Isaku Yamahata
2022-05-05 18:14 ` [RFC PATCH v6 018/104] KVM: TDX: Add helper functions to print TDX SEAMCALL error isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 019/104] [MARKER] The start of TDX KVM patch series: TD VM creation/destruction isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 020/104] KVM: TDX: Stub in tdx.h with structs, accessors, and VMCS helpers isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 021/104] x86/cpu: Add helper functions to allocate/free TDX private host key id isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 022/104] KVM: TDX: create/destroy VM structure isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 023/104] KVM: TDX: x86: Add ioctl to get TDX systemwide parameters isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 024/104] KVM: TDX: Add place holder for TDX VM specific mem_enc_op ioctl isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 025/104] KVM: TDX: initialize VM with TDX specific parameters isaku.yamahata
2022-05-06  5:27   ` Xiaoyao Li
2022-05-06 13:54   ` Xiaoyao Li
2022-05-09 15:18     ` Isaku Yamahata
2022-05-13 12:34       ` Kai Huang
2022-05-05 18:14 ` [RFC PATCH v6 026/104] KVM: TDX: Make KVM_CAP_SET_IDENTITY_MAP_ADDR unsupported for TDX isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 027/104] KVM: TDX: Make pmu_intel.c ignore guest TD case isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 028/104] [MARKER] The start of TDX KVM patch series: TD vcpu creation/destruction isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 029/104] KVM: TDX: allocate/free TDX vcpu structure isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 030/104] " isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 031/104] KVM: TDX: Do TDX specific vcpu initialization isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 032/104] [MARKER] The start of TDX KVM patch series: KVM MMU GPA shared bits isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 033/104] KVM: x86/mmu: introduce config for PRIVATE KVM MMU isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 034/104] KVM: x86/mmu: Add address conversion functions for TDX shared bits isaku.yamahata
2022-05-10  0:16   ` Kai Huang
2022-05-12  9:50     ` Isaku Yamahata
2022-05-05 18:14 ` [RFC PATCH v6 035/104] [MARKER] The start of TDX KVM patch series: KVM TDP refactoring for TDX isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 036/104] KVM: x86/mmu: Explicitly check for MMIO spte in fast page fault isaku.yamahata
2022-08-01 22:27   ` David Matlack
2022-08-01 23:27     ` Sean Christopherson
2022-08-02  1:46       ` Huang, Kai
2022-08-02 16:34         ` David Matlack
2022-08-03  0:28           ` Kai Huang
2022-05-05 18:14 ` [RFC PATCH v6 037/104] KVM: x86/mmu: Allow non-zero value for non-present SPTE isaku.yamahata
2022-08-04 22:54   ` David Matlack
2022-08-04 23:18     ` David Matlack
2022-08-05  0:03       ` Huang, Kai
2022-08-05 16:46         ` David Matlack
2022-08-05 17:14           ` Sean Christopherson
2022-08-04 23:23     ` Sean Christopherson
2022-08-04 23:43       ` David Matlack
2022-05-05 18:14 ` [RFC PATCH v6 038/104] KVM: x86/mmu: Track shadow MMIO value/mask on a per-VM basis isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 039/104] KVM: x86/mmu: Disallow fast page fault on private GPA isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 040/104] KVM: x86/mmu: Allow per-VM override of the TDP max page level isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 041/104] KVM: x86/mmu: Zap only leaf SPTEs for deleted/moved memslot for private mmu isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 042/104] KVM: VMX: Introduce test mode related to EPT violation VE isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 043/104] [MARKER] The start of TDX KVM patch series: KVM TDP MMU hooks isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 044/104] KVM: x86/mmu: Focibly use TDP MMU for TDX isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 045/104] KVM: x86/mmu: Add a private pointer to struct kvm_mmu_page isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 046/104] KVM: x86/tdp_mmu: refactor kvm_tdp_mmu_map() isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 047/104] KVM: x86/tdp_mmu: Support TDX private mapping for TDP MMU isaku.yamahata
2022-05-27 17:38   ` Paolo Bonzini
2022-06-01 20:49     ` Isaku Yamahata
2022-05-05 18:14 ` [RFC PATCH v6 048/104] [MARKER] The start of TDX KVM patch series: TDX EPT violation isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 049/104] KVM: x86/mmu: Disallow dirty logging for x86 TDX isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 050/104] KVM: x86/tdp_mmu: Ignore unsupported mmu operation on private GFNs isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 051/104] KVM: VMX: Split out guts of EPT violation to common/exposed function isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 052/104] KVM: VMX: Move setting of EPT MMU masks to common VT-x code isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 053/104] KVM: TDX: Add load_mmu_pgd method for TDX isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 054/104] KVM: TDX: don't request KVM_REQ_APIC_PAGE_RELOAD isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 055/104] KVM: TDX: TDP MMU TDX support isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 056/104] [MARKER] The start of TDX KVM patch series: KVM TDP MMU MapGPA isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 057/104] KVM: x86/mmu: steal software usable git to record if GFN is for shared or not isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 058/104] KVM: x86/tdp_mmu: implement MapGPA hypercall for TDX isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 059/104] KVM: x86/mmu: Introduce kvm_mmu_map_tdp_page() for use by TDX isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 060/104] [MARKER] The start of TDX KVM patch series: TD finalization isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 061/104] KVM: TDX: Create initial guest memory isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 062/104] KVM: TDX: Finalize VM initialization isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 063/104] [MARKER] The start of TDX KVM patch series: TD vcpu enter/exit isaku.yamahata
2022-05-05 18:14 ` [RFC PATCH v6 064/104] KVM: TDX: Add helper assembly function to TDX vcpu isaku.yamahata
2022-05-31 15:58   ` Paolo Bonzini
2022-06-01 20:50     ` Isaku Yamahata
2022-05-05 18:14 ` [RFC PATCH v6 065/104] KVM: TDX: Implement TDX vcpu enter/exit path isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 066/104] KVM: TDX: vcpu_run: save/restore host state(host kernel gs) isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 067/104] KVM: TDX: restore host xsave state when exit from the guest TD isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 068/104] KVM: x86: Allow to update cached values in kvm_user_return_msrs w/o wrmsr isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 069/104] KVM: TDX: restore user ret MSRs isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 070/104] [MARKER] The start of TDX KVM patch series: TD vcpu exits/interrupts/hypercalls isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 071/104] KVM: TDX: complete interrupts after tdexit isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 072/104] KVM: TDX: restore debug store when TD exit isaku.yamahata
2022-05-05 18:15 ` isaku.yamahata [this message]
2022-05-05 18:15 ` [RFC PATCH v6 074/104] KVM: x86: Add a switch_db_regs flag to handle TDX's auto-switched behavior isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 075/104] KVM: TDX: Add support for find pending IRQ in a protected local APIC isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 076/104] KVM: x86: Assume timer IRQ was injected if APIC state is proteced isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 077/104] KVM: TDX: remove use of struct vcpu_vmx from posted_interrupt.c isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 078/104] KVM: TDX: Implement interrupt injection isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 079/104] KVM: TDX: Implements vcpu request_immediate_exit isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 080/104] KVM: TDX: Implement methods to inject NMI isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 081/104] KVM: VMX: Modify NMI and INTR handlers to take intr_info as function argument isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 082/104] KVM: VMX: Move NMI/exception handler to common helper isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 083/104] KVM: x86: Split core of hypercall emulation to helper function isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 084/104] KVM: TDX: Add a place holder to handle TDX VM exit isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 085/104] KVM: TDX: handle EXIT_REASON_OTHER_SMI isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 086/104] KVM: TDX: handle ept violation/misconfig exit isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 087/104] KVM: TDX: handle EXCEPTION_NMI and EXTERNAL_INTERRUPT isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 088/104] KVM: TDX: Add a place holder for handler of TDX hypercalls (TDG.VP.VMCALL) isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 089/104] KVM: TDX: handle KVM hypercall with TDG.VP.VMCALL isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 090/104] KVM: TDX: Handle TDX PV CPUID hypercall isaku.yamahata
2022-06-14 18:15   ` Sagi Shahar
2022-06-29 10:13     ` Isaku Yamahata
2022-07-18 22:37       ` Sagi Shahar
2022-07-19 19:23         ` Sean Christopherson
2022-05-05 18:15 ` [RFC PATCH v6 091/104] KVM: TDX: Handle TDX PV HLT hypercall isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 092/104] KVM: TDX: Handle TDX PV port io hypercall isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 093/104] KVM: TDX: Handle TDX PV MMIO hypercall isaku.yamahata
2022-06-14 18:08   ` Sagi Shahar
2022-06-29 10:17     ` Isaku Yamahata
2022-05-05 18:15 ` [RFC PATCH v6 094/104] KVM: TDX: Implement callbacks for MSR operations for TDX isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 095/104] KVM: TDX: Handle TDX PV rdmsr/wrmsr hypercall isaku.yamahata
2022-06-10 21:04   ` Sagi Shahar
2022-06-29 10:24     ` Isaku Yamahata
2022-05-05 18:15 ` [RFC PATCH v6 096/104] KVM: TDX: Handle TDX PV report fatal error hypercall isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 097/104] KVM: TDX: Handle TDX PV map_gpa hypercall isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 098/104] KVM: TDX: Handle TDG.VP.VMCALL<GetTdVmCallInfo> hypercall isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 099/104] KVM: TDX: Silently discard SMI request isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 100/104] KVM: TDX: Silently ignore INIT/SIPI isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 101/104] KVM: TDX: Add methods to ignore accesses to CPU state isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 102/104] Documentation/virtual/kvm: Document on Trust Domain Extensions(TDX) isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 103/104] KVM: x86: design documentation on TDX support of x86 KVM TDP MMU isaku.yamahata
2022-05-05 18:15 ` [RFC PATCH v6 104/104] [MARKER] the end of (the first phase of) TDX KVM patch series isaku.yamahata
2022-05-31 14:46 ` [RFC PATCH v6 000/104] KVM TDX basic feature support Paolo Bonzini
2022-06-01 20:53   ` Isaku Yamahata

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=96148d16e5c3b394bec4d7f51130b2d35e33352e.1651774250.git.isaku.yamahata@intel.com \
    --to=isaku.yamahata@intel.com \
    --cc=erdemaktas@google.com \
    --cc=isaku.yamahata@gmail.com \
    --cc=kvm@vger.kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=pbonzini@redhat.com \
    --cc=sagis@google.com \
    --cc=seanjc@google.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.