From: "Adalbert Lazăr" <alazar@bitdefender.com>
To: kvm@vger.kernel.org
Cc: linux-mm@kvack.org, virtualization@lists.linux-foundation.org,
"Paolo Bonzini" <pbonzini@redhat.com>,
"Radim Krčmář" <rkrcmar@redhat.com>,
"Konrad Rzeszutek Wilk" <konrad.wilk@oracle.com>,
"Tamas K Lengyel" <tamas@tklengyel.com>,
"Mathieu Tarral" <mathieu.tarral@protonmail.com>,
"Samuel Laurén" <samuel.lauren@iki.fi>,
"Patrick Colp" <patrick.colp@oracle.com>,
"Jan Kiszka" <jan.kiszka@siemens.com>,
"Stefan Hajnoczi" <stefanha@redhat.com>,
"Weijiang Yang" <weijiang.yang@intel.com>,
Zhang@kvack.org, "Yu C" <yu.c.zhang@intel.com>,
"Mihai Donțu" <mdontu@bitdefender.com>,
"Adalbert Lazăr" <alazar@bitdefender.com>,
"Nicușor Cîțu" <ncitu@bitdefender.com>,
"Mircea Cîrjaliu" <mcirjaliu@bitdefender.com>,
"Marian Rotariu" <marian.c.rotariu@gmail.com>
Subject: [RFC PATCH v6 77/92] kvm: introspection: add trace functions
Date: Fri, 9 Aug 2019 19:00:32 +0300 [thread overview]
Message-ID: <20190809160047.8319-78-alazar@bitdefender.com> (raw)
In-Reply-To: <20190809160047.8319-1-alazar@bitdefender.com>
Co-developed-by: Nicușor Cîțu <ncitu@bitdefender.com>
Signed-off-by: Nicușor Cîțu <ncitu@bitdefender.com>
Co-developed-by: Mircea Cîrjaliu <mcirjaliu@bitdefender.com>
Signed-off-by: Mircea Cîrjaliu <mcirjaliu@bitdefender.com>
Co-developed-by: Marian Rotariu <marian.c.rotariu@gmail.com>
Signed-off-by: Marian Rotariu <marian.c.rotariu@gmail.com>
Co-developed-by: Adalbert Lazăr <alazar@bitdefender.com>
Signed-off-by: Adalbert Lazăr <alazar@bitdefender.com>
---
arch/x86/kvm/kvmi.c | 63 ++++
include/trace/events/kvmi.h | 680 ++++++++++++++++++++++++++++++++++++
virt/kvm/kvmi.c | 20 ++
virt/kvm/kvmi_mem.c | 5 +
virt/kvm/kvmi_msg.c | 16 +
5 files changed, 784 insertions(+)
create mode 100644 include/trace/events/kvmi.h
diff --git a/arch/x86/kvm/kvmi.c b/arch/x86/kvm/kvmi.c
index 5312f179af9c..171e76449271 100644
--- a/arch/x86/kvm/kvmi.c
+++ b/arch/x86/kvm/kvmi.c
@@ -9,6 +9,8 @@
#include <asm/vmx.h>
#include "../../../virt/kvm/kvmi_int.h"
+#include <trace/events/kvmi.h>
+
static unsigned long *msr_mask(struct kvm_vcpu *vcpu, unsigned int *msr)
{
switch (*msr) {
@@ -102,6 +104,9 @@ static bool __kvmi_msr_event(struct kvm_vcpu *vcpu, struct msr_data *msr)
if (old_msr.data == msr->data)
return true;
+ trace_kvmi_event_msr_send(vcpu->vcpu_id, msr->index, old_msr.data,
+ msr->data);
+
action = kvmi_send_msr(vcpu, msr->index, old_msr.data, msr->data,
&ret_value);
switch (action) {
@@ -113,6 +118,8 @@ static bool __kvmi_msr_event(struct kvm_vcpu *vcpu, struct msr_data *msr)
kvmi_handle_common_event_actions(vcpu, action, "MSR");
}
+ trace_kvmi_event_msr_recv(vcpu->vcpu_id, action, ret_value);
+
return ret;
}
@@ -387,6 +394,8 @@ static bool __kvmi_cr_event(struct kvm_vcpu *vcpu, unsigned int cr,
if (!test_bit(cr, IVCPU(vcpu)->cr_mask))
return true;
+ trace_kvmi_event_cr_send(vcpu->vcpu_id, cr, old_value, *new_value);
+
action = kvmi_send_cr(vcpu, cr, old_value, *new_value, &ret_value);
switch (action) {
case KVMI_EVENT_ACTION_CONTINUE:
@@ -397,6 +406,8 @@ static bool __kvmi_cr_event(struct kvm_vcpu *vcpu, unsigned int cr,
kvmi_handle_common_event_actions(vcpu, action, "CR");
}
+ trace_kvmi_event_cr_recv(vcpu->vcpu_id, action, ret_value);
+
return ret;
}
@@ -437,6 +448,8 @@ static void __kvmi_xsetbv_event(struct kvm_vcpu *vcpu)
{
u32 action;
+ trace_kvmi_event_xsetbv_send(vcpu->vcpu_id);
+
action = kvmi_send_xsetbv(vcpu);
switch (action) {
case KVMI_EVENT_ACTION_CONTINUE:
@@ -444,6 +457,8 @@ static void __kvmi_xsetbv_event(struct kvm_vcpu *vcpu)
default:
kvmi_handle_common_event_actions(vcpu, action, "XSETBV");
}
+
+ trace_kvmi_event_xsetbv_recv(vcpu->vcpu_id, action);
}
void kvmi_xsetbv_event(struct kvm_vcpu *vcpu)
@@ -460,12 +475,26 @@ void kvmi_xsetbv_event(struct kvm_vcpu *vcpu)
kvmi_put(vcpu->kvm);
}
+static u64 get_next_rip(struct kvm_vcpu *vcpu)
+{
+ struct kvmi_vcpu *ivcpu = IVCPU(vcpu);
+
+ if (ivcpu->have_delayed_regs)
+ return ivcpu->delayed_regs.rip;
+ else
+ return kvm_rip_read(vcpu);
+}
+
void kvmi_arch_breakpoint_event(struct kvm_vcpu *vcpu, u64 gva, u8 insn_len)
{
u32 action;
u64 gpa;
+ u64 old_rip;
gpa = kvm_mmu_gva_to_gpa_system(vcpu, gva, 0, NULL);
+ old_rip = kvm_rip_read(vcpu);
+
+ trace_kvmi_event_bp_send(vcpu->vcpu_id, gpa, old_rip);
action = kvmi_msg_send_bp(vcpu, gpa, insn_len);
switch (action) {
@@ -478,6 +507,8 @@ void kvmi_arch_breakpoint_event(struct kvm_vcpu *vcpu, u64 gva, u8 insn_len)
default:
kvmi_handle_common_event_actions(vcpu, action, "BP");
}
+
+ trace_kvmi_event_bp_recv(vcpu->vcpu_id, action, get_next_rip(vcpu));
}
#define KVM_HC_XEN_HVM_OP_GUEST_REQUEST_VM_EVENT 24
@@ -504,6 +535,8 @@ void kvmi_arch_hypercall_event(struct kvm_vcpu *vcpu)
{
u32 action;
+ trace_kvmi_event_hc_send(vcpu->vcpu_id);
+
action = kvmi_msg_send_hypercall(vcpu);
switch (action) {
case KVMI_EVENT_ACTION_CONTINUE:
@@ -511,6 +544,8 @@ void kvmi_arch_hypercall_event(struct kvm_vcpu *vcpu)
default:
kvmi_handle_common_event_actions(vcpu, action, "HYPERCALL");
}
+
+ trace_kvmi_event_hc_recv(vcpu->vcpu_id, action);
}
bool kvmi_arch_pf_event(struct kvm_vcpu *vcpu, gpa_t gpa, gva_t gva,
@@ -532,6 +567,9 @@ bool kvmi_arch_pf_event(struct kvm_vcpu *vcpu, gpa_t gpa, gva_t gva,
if (ivcpu->effective_rep_complete)
return true;
+ trace_kvmi_event_pf_send(vcpu->vcpu_id, gpa, gva, access,
+ kvm_rip_read(vcpu));
+
action = kvmi_msg_send_pf(vcpu, gpa, gva, access, &ivcpu->ss_requested,
&ivcpu->rep_complete, &ctx_addr,
ivcpu->ctx_data, &ctx_size);
@@ -553,6 +591,9 @@ bool kvmi_arch_pf_event(struct kvm_vcpu *vcpu, gpa_t gpa, gva_t gva,
kvmi_handle_common_event_actions(vcpu, action, "PF");
}
+ trace_kvmi_event_pf_recv(vcpu->vcpu_id, action, get_next_rip(vcpu),
+ ctx_size, ivcpu->ss_requested, ret);
+
return ret;
}
@@ -628,6 +669,11 @@ void kvmi_arch_trap_event(struct kvm_vcpu *vcpu)
err = 0;
}
+ trace_kvmi_event_trap_send(vcpu->vcpu_id, vector,
+ IVCPU(vcpu)->exception.nr,
+ err, IVCPU(vcpu)->exception.error_code,
+ vcpu->arch.cr2);
+
action = kvmi_send_trap(vcpu, vector, type, err, vcpu->arch.cr2);
switch (action) {
case KVMI_EVENT_ACTION_CONTINUE:
@@ -635,6 +681,8 @@ void kvmi_arch_trap_event(struct kvm_vcpu *vcpu)
default:
kvmi_handle_common_event_actions(vcpu, action, "TRAP");
}
+
+ trace_kvmi_event_trap_recv(vcpu->vcpu_id, action);
}
static bool __kvmi_descriptor_event(struct kvm_vcpu *vcpu, u8 descriptor,
@@ -643,6 +691,8 @@ static bool __kvmi_descriptor_event(struct kvm_vcpu *vcpu, u8 descriptor,
u32 action;
bool ret = false;
+ trace_kvmi_event_desc_send(vcpu->vcpu_id, descriptor, write);
+
action = kvmi_msg_send_descriptor(vcpu, descriptor, write);
switch (action) {
case KVMI_EVENT_ACTION_CONTINUE:
@@ -654,6 +704,8 @@ static bool __kvmi_descriptor_event(struct kvm_vcpu *vcpu, u8 descriptor,
kvmi_handle_common_event_actions(vcpu, action, "DESC");
}
+ trace_kvmi_event_desc_recv(vcpu->vcpu_id, action);
+
return ret;
}
@@ -718,6 +770,15 @@ int kvmi_arch_cmd_inject_exception(struct kvm_vcpu *vcpu, u8 vector,
bool error_code_valid,
u32 error_code, u64 address)
{
+ struct x86_exception e = {
+ .error_code_valid = error_code_valid,
+ .error_code = error_code,
+ .address = address,
+ .vector = vector,
+ };
+
+ trace_kvmi_cmd_inject_exception(vcpu, &e);
+
if (!(is_vector_valid(vector) && is_gva_valid(vcpu, address)))
return -KVM_EINVAL;
@@ -876,6 +937,8 @@ void kvmi_arch_update_page_tracking(struct kvm *kvm,
return;
}
+ trace_kvmi_set_gfn_access(m->gfn, m->access, m->write_bitmap, slot->id);
+
for (i = 0; i < ARRAY_SIZE(track_modes); i++) {
unsigned int allow_bit = track_modes[i].allow_bit;
enum kvm_page_track_mode mode = track_modes[i].track_mode;
diff --git a/include/trace/events/kvmi.h b/include/trace/events/kvmi.h
new file mode 100644
index 000000000000..442189437fe7
--- /dev/null
+++ b/include/trace/events/kvmi.h
@@ -0,0 +1,680 @@
+/* SPDX-License-Identifier: GPL-2.0 */
+#undef TRACE_SYSTEM
+#define TRACE_SYSTEM kvmi
+
+#if !defined(_TRACE_KVMI_H) || defined(TRACE_HEADER_MULTI_READ)
+#define _TRACE_KVMI_H
+
+#include <linux/tracepoint.h>
+
+#ifndef __TRACE_KVMI_STRUCTURES
+#define __TRACE_KVMI_STRUCTURES
+
+#undef EN
+#define EN(x) { x, #x }
+
+static const struct trace_print_flags kvmi_msg_id_symbol[] = {
+ EN(KVMI_GET_VERSION),
+ EN(KVMI_CHECK_COMMAND),
+ EN(KVMI_CHECK_EVENT),
+ EN(KVMI_GET_GUEST_INFO),
+ EN(KVMI_GET_VCPU_INFO),
+ EN(KVMI_GET_REGISTERS),
+ EN(KVMI_SET_REGISTERS),
+ EN(KVMI_GET_PAGE_ACCESS),
+ EN(KVMI_SET_PAGE_ACCESS),
+ EN(KVMI_GET_PAGE_WRITE_BITMAP),
+ EN(KVMI_SET_PAGE_WRITE_BITMAP),
+ EN(KVMI_INJECT_EXCEPTION),
+ EN(KVMI_READ_PHYSICAL),
+ EN(KVMI_WRITE_PHYSICAL),
+ EN(KVMI_GET_MAP_TOKEN),
+ EN(KVMI_CONTROL_EVENTS),
+ EN(KVMI_CONTROL_CR),
+ EN(KVMI_CONTROL_MSR),
+ EN(KVMI_EVENT),
+ EN(KVMI_EVENT_REPLY),
+ EN(KVMI_GET_CPUID),
+ EN(KVMI_GET_XSAVE),
+ EN(KVMI_PAUSE_VCPU),
+ EN(KVMI_CONTROL_VM_EVENTS),
+ EN(KVMI_GET_MTRR_TYPE),
+ EN(KVMI_CONTROL_SPP),
+ EN(KVMI_CONTROL_CMD_RESPONSE),
+ {-1, NULL}
+};
+
+static const struct trace_print_flags kvmi_descriptor_symbol[] = {
+ EN(KVMI_DESC_IDTR),
+ EN(KVMI_DESC_GDTR),
+ EN(KVMI_DESC_LDTR),
+ EN(KVMI_DESC_TR),
+ {-1, NULL}
+};
+
+static const struct trace_print_flags kvmi_event_symbol[] = {
+ EN(KVMI_EVENT_UNHOOK),
+ EN(KVMI_EVENT_CR),
+ EN(KVMI_EVENT_MSR),
+ EN(KVMI_EVENT_XSETBV),
+ EN(KVMI_EVENT_BREAKPOINT),
+ EN(KVMI_EVENT_HYPERCALL),
+ EN(KVMI_EVENT_PF),
+ EN(KVMI_EVENT_TRAP),
+ EN(KVMI_EVENT_DESCRIPTOR),
+ EN(KVMI_EVENT_CREATE_VCPU),
+ EN(KVMI_EVENT_PAUSE_VCPU),
+ EN(KVMI_EVENT_SINGLESTEP),
+ { -1, NULL }
+};
+
+static const struct trace_print_flags kvmi_action_symbol[] = {
+ {KVMI_EVENT_ACTION_CONTINUE, "continue"},
+ {KVMI_EVENT_ACTION_RETRY, "retry"},
+ {KVMI_EVENT_ACTION_CRASH, "crash"},
+ {-1, NULL}
+};
+
+#endif /* __TRACE_KVMI_STRUCTURES */
+
+TRACE_EVENT(
+ kvmi_vm_command,
+ TP_PROTO(__u16 id, __u32 seq),
+ TP_ARGS(id, seq),
+ TP_STRUCT__entry(
+ __field(__u16, id)
+ __field(__u32, seq)
+ ),
+ TP_fast_assign(
+ __entry->id = id;
+ __entry->seq = seq;
+ ),
+ TP_printk("%s seq %d",
+ trace_print_symbols_seq(p, __entry->id, kvmi_msg_id_symbol),
+ __entry->seq)
+);
+
+TRACE_EVENT(
+ kvmi_vm_reply,
+ TP_PROTO(__u16 id, __u32 seq, __s32 err),
+ TP_ARGS(id, seq, err),
+ TP_STRUCT__entry(
+ __field(__u16, id)
+ __field(__u32, seq)
+ __field(__s32, err)
+ ),
+ TP_fast_assign(
+ __entry->id = id;
+ __entry->seq = seq;
+ __entry->err = err;
+ ),
+ TP_printk("%s seq %d err %d",
+ trace_print_symbols_seq(p, __entry->id, kvmi_msg_id_symbol),
+ __entry->seq,
+ __entry->err)
+);
+
+TRACE_EVENT(
+ kvmi_vcpu_command,
+ TP_PROTO(__u16 vcpu, __u16 id, __u32 seq),
+ TP_ARGS(vcpu, id, seq),
+ TP_STRUCT__entry(
+ __field(__u16, vcpu)
+ __field(__u16, id)
+ __field(__u32, seq)
+ ),
+ TP_fast_assign(
+ __entry->vcpu = vcpu;
+ __entry->id = id;
+ __entry->seq = seq;
+ ),
+ TP_printk("vcpu %d %s seq %d",
+ __entry->vcpu,
+ trace_print_symbols_seq(p, __entry->id, kvmi_msg_id_symbol),
+ __entry->seq)
+);
+
+TRACE_EVENT(
+ kvmi_vcpu_reply,
+ TP_PROTO(__u16 vcpu, __u16 id, __u32 seq, __s32 err),
+ TP_ARGS(vcpu, id, seq, err),
+ TP_STRUCT__entry(
+ __field(__u16, vcpu)
+ __field(__u16, id)
+ __field(__u32, seq)
+ __field(__s32, err)
+ ),
+ TP_fast_assign(
+ __entry->vcpu = vcpu;
+ __entry->id = id;
+ __entry->seq = seq;
+ __entry->err = err;
+ ),
+ TP_printk("vcpu %d %s seq %d err %d",
+ __entry->vcpu,
+ trace_print_symbols_seq(p, __entry->id, kvmi_msg_id_symbol),
+ __entry->seq,
+ __entry->err)
+);
+
+TRACE_EVENT(
+ kvmi_event,
+ TP_PROTO(__u16 vcpu, __u32 id, __u32 seq),
+ TP_ARGS(vcpu, id, seq),
+ TP_STRUCT__entry(
+ __field(__u16, vcpu)
+ __field(__u32, id)
+ __field(__u32, seq)
+ ),
+ TP_fast_assign(
+ __entry->vcpu = vcpu;
+ __entry->id = id;
+ __entry->seq = seq;
+ ),
+ TP_printk("vcpu %d %s seq %d",
+ __entry->vcpu,
+ trace_print_symbols_seq(p, __entry->id, kvmi_event_symbol),
+ __entry->seq)
+);
+
+TRACE_EVENT(
+ kvmi_event_reply,
+ TP_PROTO(__u32 id, __u32 seq),
+ TP_ARGS(id, seq),
+ TP_STRUCT__entry(
+ __field(__u32, id)
+ __field(__u32, seq)
+ ),
+ TP_fast_assign(
+ __entry->id = id;
+ __entry->seq = seq;
+ ),
+ TP_printk("%s seq %d",
+ trace_print_symbols_seq(p, __entry->id, kvmi_event_symbol),
+ __entry->seq)
+);
+
+#define KVMI_ACCESS_PRINTK() ({ \
+ const char *saved_ptr = trace_seq_buffer_ptr(p); \
+ static const char * const access_str[] = { \
+ "---", "r--", "-w-", "rw-", "--x", "r-x", "-wx", "rwx" \
+ }; \
+ trace_seq_printf(p, "%s", access_str[__entry->access & 7]); \
+ saved_ptr; \
+})
+
+TRACE_EVENT(
+ kvmi_set_gfn_access,
+ TP_PROTO(__u64 gfn, __u8 access, __u32 bitmap, __u16 slot),
+ TP_ARGS(gfn, access, bitmap, slot),
+ TP_STRUCT__entry(
+ __field(__u64, gfn)
+ __field(__u8, access)
+ __field(__u32, bitmap)
+ __field(__u16, slot)
+ ),
+ TP_fast_assign(
+ __entry->gfn = gfn;
+ __entry->access = access;
+ __entry->bitmap = bitmap;
+ __entry->slot = slot;
+ ),
+ TP_printk("gfn %llx %s write bitmap %x slot %d",
+ __entry->gfn, KVMI_ACCESS_PRINTK(),
+ __entry->bitmap, __entry->slot)
+);
+
+DECLARE_EVENT_CLASS(
+ kvmi_event_send_template,
+ TP_PROTO(__u16 vcpu),
+ TP_ARGS(vcpu),
+ TP_STRUCT__entry(
+ __field(__u16, vcpu)
+ ),
+ TP_fast_assign(
+ __entry->vcpu = vcpu;
+ ),
+ TP_printk("vcpu %d",
+ __entry->vcpu
+ )
+);
+DECLARE_EVENT_CLASS(
+ kvmi_event_recv_template,
+ TP_PROTO(__u16 vcpu, __u32 action),
+ TP_ARGS(vcpu, action),
+ TP_STRUCT__entry(
+ __field(__u16, vcpu)
+ __field(__u32, action)
+ ),
+ TP_fast_assign(
+ __entry->vcpu = vcpu;
+ __entry->action = action;
+ ),
+ TP_printk("vcpu %d %s",
+ __entry->vcpu,
+ trace_print_symbols_seq(p, __entry->action,
+ kvmi_action_symbol)
+ )
+);
+
+TRACE_EVENT(
+ kvmi_event_cr_send,
+ TP_PROTO(__u16 vcpu, __u32 cr, __u64 old_value, __u64 new_value),
+ TP_ARGS(vcpu, cr, old_value, new_value),
+ TP_STRUCT__entry(
+ __field(__u16, vcpu)
+ __field(__u32, cr)
+ __field(__u64, old_value)
+ __field(__u64, new_value)
+ ),
+ TP_fast_assign(
+ __entry->vcpu = vcpu;
+ __entry->cr = cr;
+ __entry->old_value = old_value;
+ __entry->new_value = new_value;
+ ),
+ TP_printk("vcpu %d cr %x old_value %llx new_value %llx",
+ __entry->vcpu,
+ __entry->cr,
+ __entry->old_value,
+ __entry->new_value
+ )
+);
+TRACE_EVENT(
+ kvmi_event_cr_recv,
+ TP_PROTO(__u16 vcpu, __u32 action, __u64 new_value),
+ TP_ARGS(vcpu, action, new_value),
+ TP_STRUCT__entry(
+ __field(__u16, vcpu)
+ __field(__u32, action)
+ __field(__u64, new_value)
+ ),
+ TP_fast_assign(
+ __entry->vcpu = vcpu;
+ __entry->action = action;
+ __entry->new_value = new_value;
+ ),
+ TP_printk("vcpu %d %s new_value %llx",
+ __entry->vcpu,
+ trace_print_symbols_seq(p, __entry->action,
+ kvmi_action_symbol),
+ __entry->new_value
+ )
+);
+
+TRACE_EVENT(
+ kvmi_event_msr_send,
+ TP_PROTO(__u16 vcpu, __u32 msr, __u64 old_value, __u64 new_value),
+ TP_ARGS(vcpu, msr, old_value, new_value),
+ TP_STRUCT__entry(
+ __field(__u16, vcpu)
+ __field(__u32, msr)
+ __field(__u64, old_value)
+ __field(__u64, new_value)
+ ),
+ TP_fast_assign(
+ __entry->vcpu = vcpu;
+ __entry->msr = msr;
+ __entry->old_value = old_value;
+ __entry->new_value = new_value;
+ ),
+ TP_printk("vcpu %d msr %x old_value %llx new_value %llx",
+ __entry->vcpu,
+ __entry->msr,
+ __entry->old_value,
+ __entry->new_value
+ )
+);
+TRACE_EVENT(
+ kvmi_event_msr_recv,
+ TP_PROTO(__u16 vcpu, __u32 action, __u64 new_value),
+ TP_ARGS(vcpu, action, new_value),
+ TP_STRUCT__entry(
+ __field(__u16, vcpu)
+ __field(__u32, action)
+ __field(__u64, new_value)
+ ),
+ TP_fast_assign(
+ __entry->vcpu = vcpu;
+ __entry->action = action;
+ __entry->new_value = new_value;
+ ),
+ TP_printk("vcpu %d %s new_value %llx",
+ __entry->vcpu,
+ trace_print_symbols_seq(p, __entry->action,
+ kvmi_action_symbol),
+ __entry->new_value
+ )
+);
+
+DEFINE_EVENT(kvmi_event_send_template, kvmi_event_xsetbv_send,
+ TP_PROTO(__u16 vcpu),
+ TP_ARGS(vcpu)
+);
+DEFINE_EVENT(kvmi_event_recv_template, kvmi_event_xsetbv_recv,
+ TP_PROTO(__u16 vcpu, __u32 action),
+ TP_ARGS(vcpu, action)
+);
+
+TRACE_EVENT(
+ kvmi_event_bp_send,
+ TP_PROTO(__u16 vcpu, __u64 gpa, __u64 old_rip),
+ TP_ARGS(vcpu, gpa, old_rip),
+ TP_STRUCT__entry(
+ __field(__u16, vcpu)
+ __field(__u64, gpa)
+ __field(__u64, old_rip)
+ ),
+ TP_fast_assign(
+ __entry->vcpu = vcpu;
+ __entry->gpa = gpa;
+ __entry->old_rip = old_rip;
+ ),
+ TP_printk("vcpu %d gpa %llx rip %llx",
+ __entry->vcpu,
+ __entry->gpa,
+ __entry->old_rip
+ )
+);
+TRACE_EVENT(
+ kvmi_event_bp_recv,
+ TP_PROTO(__u16 vcpu, __u32 action, __u64 new_rip),
+ TP_ARGS(vcpu, action, new_rip),
+ TP_STRUCT__entry(
+ __field(__u16, vcpu)
+ __field(__u32, action)
+ __field(__u64, new_rip)
+ ),
+ TP_fast_assign(
+ __entry->vcpu = vcpu;
+ __entry->action = action;
+ __entry->new_rip = new_rip;
+ ),
+ TP_printk("vcpu %d %s rip %llx",
+ __entry->vcpu,
+ trace_print_symbols_seq(p, __entry->action,
+ kvmi_action_symbol),
+ __entry->new_rip
+ )
+);
+
+DEFINE_EVENT(kvmi_event_send_template, kvmi_event_hc_send,
+ TP_PROTO(__u16 vcpu),
+ TP_ARGS(vcpu)
+);
+DEFINE_EVENT(kvmi_event_recv_template, kvmi_event_hc_recv,
+ TP_PROTO(__u16 vcpu, __u32 action),
+ TP_ARGS(vcpu, action)
+);
+
+TRACE_EVENT(
+ kvmi_event_pf_send,
+ TP_PROTO(__u16 vcpu, __u64 gpa, __u64 gva, __u8 access, __u64 rip),
+ TP_ARGS(vcpu, gpa, gva, access, rip),
+ TP_STRUCT__entry(
+ __field(__u16, vcpu)
+ __field(__u64, gpa)
+ __field(__u64, gva)
+ __field(__u8, access)
+ __field(__u64, rip)
+ ),
+ TP_fast_assign(
+ __entry->vcpu = vcpu;
+ __entry->gpa = gpa;
+ __entry->gva = gva;
+ __entry->access = access;
+ __entry->rip = rip;
+ ),
+ TP_printk("vcpu %d gpa %llx %s gva %llx rip %llx",
+ __entry->vcpu,
+ __entry->gpa,
+ KVMI_ACCESS_PRINTK(),
+ __entry->gva,
+ __entry->rip
+ )
+);
+TRACE_EVENT(
+ kvmi_event_pf_recv,
+ TP_PROTO(__u16 vcpu, __u32 action, __u64 next_rip, size_t custom_data,
+ bool singlestep, bool ret),
+ TP_ARGS(vcpu, action, next_rip, custom_data, singlestep, ret),
+ TP_STRUCT__entry(
+ __field(__u16, vcpu)
+ __field(__u32, action)
+ __field(__u64, next_rip)
+ __field(size_t, custom_data)
+ __field(bool, singlestep)
+ __field(bool, ret)
+ ),
+ TP_fast_assign(
+ __entry->vcpu = vcpu;
+ __entry->action = action;
+ __entry->next_rip = next_rip;
+ __entry->custom_data = custom_data;
+ __entry->singlestep = singlestep;
+ __entry->ret = ret;
+ ),
+ TP_printk("vcpu %d %s rip %llx custom %zu %s",
+ __entry->vcpu,
+ trace_print_symbols_seq(p, __entry->action,
+ kvmi_action_symbol),
+ __entry->next_rip, __entry->custom_data,
+ (__entry->singlestep ? (__entry->ret ? "singlestep failed" :
+ "singlestep running")
+ : "")
+ )
+);
+
+TRACE_EVENT(
+ kvmi_event_trap_send,
+ TP_PROTO(__u16 vcpu, __u32 vector, __u8 nr, __u32 err, __u16 error_code,
+ __u64 cr2),
+ TP_ARGS(vcpu, vector, nr, err, error_code, cr2),
+ TP_STRUCT__entry(
+ __field(__u16, vcpu)
+ __field(__u32, vector)
+ __field(__u8, nr)
+ __field(__u32, err)
+ __field(__u16, error_code)
+ __field(__u64, cr2)
+ ),
+ TP_fast_assign(
+ __entry->vcpu = vcpu;
+ __entry->vector = vector;
+ __entry->nr = nr;
+ __entry->err = err;
+ __entry->error_code = error_code;
+ __entry->cr2 = cr2;
+ ),
+ TP_printk("vcpu %d vector %x/%x err %x/%x address %llx",
+ __entry->vcpu,
+ __entry->vector, __entry->nr,
+ __entry->err, __entry->error_code,
+ __entry->cr2
+ )
+);
+DEFINE_EVENT(kvmi_event_recv_template, kvmi_event_trap_recv,
+ TP_PROTO(__u16 vcpu, __u32 action),
+ TP_ARGS(vcpu, action)
+);
+
+TRACE_EVENT(
+ kvmi_event_desc_send,
+ TP_PROTO(__u16 vcpu, __u8 descriptor, __u8 write),
+ TP_ARGS(vcpu, descriptor, write),
+ TP_STRUCT__entry(
+ __field(__u16, vcpu)
+ __field(__u8, descriptor)
+ __field(__u8, write)
+ ),
+ TP_fast_assign(
+ __entry->vcpu = vcpu;
+ __entry->descriptor = descriptor;
+ __entry->write = write;
+ ),
+ TP_printk("vcpu %d %s %s",
+ __entry->vcpu,
+ __entry->write ? "write" : "read",
+ trace_print_symbols_seq(p, __entry->descriptor,
+ kvmi_descriptor_symbol)
+ )
+);
+DEFINE_EVENT(kvmi_event_recv_template, kvmi_event_desc_recv,
+ TP_PROTO(__u16 vcpu, __u32 action),
+ TP_ARGS(vcpu, action)
+);
+
+DEFINE_EVENT(kvmi_event_send_template, kvmi_event_create_vcpu_send,
+ TP_PROTO(__u16 vcpu),
+ TP_ARGS(vcpu)
+);
+DEFINE_EVENT(kvmi_event_recv_template, kvmi_event_create_vcpu_recv,
+ TP_PROTO(__u16 vcpu, __u32 action),
+ TP_ARGS(vcpu, action)
+);
+
+DEFINE_EVENT(kvmi_event_send_template, kvmi_event_pause_vcpu_send,
+ TP_PROTO(__u16 vcpu),
+ TP_ARGS(vcpu)
+);
+DEFINE_EVENT(kvmi_event_recv_template, kvmi_event_pause_vcpu_recv,
+ TP_PROTO(__u16 vcpu, __u32 action),
+ TP_ARGS(vcpu, action)
+);
+
+DEFINE_EVENT(kvmi_event_send_template, kvmi_event_singlestep_send,
+ TP_PROTO(__u16 vcpu),
+ TP_ARGS(vcpu)
+);
+DEFINE_EVENT(kvmi_event_recv_template, kvmi_event_singlestep_recv,
+ TP_PROTO(__u16 vcpu, __u32 action),
+ TP_ARGS(vcpu, action)
+);
+
+TRACE_EVENT(
+ kvmi_run_singlestep,
+ TP_PROTO(struct kvm_vcpu *vcpu, __u64 gpa, __u8 access, __u8 level,
+ size_t custom_data),
+ TP_ARGS(vcpu, gpa, access, level, custom_data),
+ TP_STRUCT__entry(
+ __field(__u16, vcpu_id)
+ __field(__u64, gpa)
+ __field(__u8, access)
+ __field(size_t, len)
+ __array(__u8, insn, 15)
+ __field(__u8, level)
+ __field(size_t, custom_data)
+ ),
+ TP_fast_assign(
+ __entry->vcpu_id = vcpu->vcpu_id;
+ __entry->gpa = gpa;
+ __entry->access = access;
+ __entry->len = min_t(size_t, 15,
+ vcpu->arch.emulate_ctxt.fetch.ptr
+ - vcpu->arch.emulate_ctxt.fetch.data);
+ memcpy(__entry->insn, vcpu->arch.emulate_ctxt.fetch.data, 15);
+ __entry->level = level;
+ __entry->custom_data = custom_data;
+ ),
+ TP_printk("vcpu %d gpa %llx %s insn %s level %x custom %zu",
+ __entry->vcpu_id,
+ __entry->gpa,
+ KVMI_ACCESS_PRINTK(),
+ __print_hex(__entry->insn, __entry->len),
+ __entry->level,
+ __entry->custom_data
+ )
+);
+
+TRACE_EVENT(
+ kvmi_stop_singlestep,
+ TP_PROTO(__u16 vcpu),
+ TP_ARGS(vcpu),
+ TP_STRUCT__entry(
+ __field(__u16, vcpu)
+ ),
+ TP_fast_assign(
+ __entry->vcpu = vcpu;
+ ),
+ TP_printk("vcpu %d", __entry->vcpu
+ )
+);
+
+TRACE_EVENT(
+ kvmi_mem_map,
+ TP_PROTO(struct kvm *kvm, gpa_t req_gpa, gpa_t map_gpa),
+ TP_ARGS(kvm, req_gpa, map_gpa),
+ TP_STRUCT__entry(
+ __field_struct(uuid_t, uuid)
+ __field(gpa_t, req_gpa)
+ __field(gpa_t, map_gpa)
+ ),
+ TP_fast_assign(
+ struct kvmi *ikvm = kvmi_get(kvm);
+
+ if (ikvm) {
+ memcpy(&__entry->uuid, &ikvm->uuid, sizeof(uuid_t));
+ kvmi_put(kvm);
+ } else
+ memset(&__entry->uuid, 0, sizeof(uuid_t));
+ __entry->req_gpa = req_gpa;
+ __entry->map_gpa = map_gpa;
+ ),
+ TP_printk("vm %pU req_gpa %llx map_gpa %llx",
+ &__entry->uuid,
+ __entry->req_gpa,
+ __entry->map_gpa
+ )
+);
+
+TRACE_EVENT(
+ kvmi_mem_unmap,
+ TP_PROTO(gpa_t map_gpa),
+ TP_ARGS(map_gpa),
+ TP_STRUCT__entry(
+ __field(gpa_t, map_gpa)
+ ),
+ TP_fast_assign(
+ __entry->map_gpa = map_gpa;
+ ),
+ TP_printk("map_gpa %llx",
+ __entry->map_gpa
+ )
+);
+
+#define EXS(x) { x##_VECTOR, "#" #x }
+
+#define kvm_trace_sym_exc \
+ EXS(DE), EXS(DB), EXS(BP), EXS(OF), EXS(BR), EXS(UD), EXS(NM), \
+ EXS(DF), EXS(TS), EXS(NP), EXS(SS), EXS(GP), EXS(PF), \
+ EXS(MF), EXS(AC), EXS(MC)
+
+TRACE_EVENT(
+ kvmi_cmd_inject_exception,
+ TP_PROTO(struct kvm_vcpu *vcpu, struct x86_exception *fault),
+ TP_ARGS(vcpu, fault),
+ TP_STRUCT__entry(
+ __field(__u16, vcpu_id)
+ __field(__u8, vector)
+ __field(__u64, address)
+ __field(__u16, error_code)
+ __field(bool, error_code_valid)
+ ),
+ TP_fast_assign(
+ __entry->vcpu_id = vcpu->vcpu_id;
+ __entry->vector = fault->vector;
+ __entry->address = fault->address;
+ __entry->error_code = fault->error_code;
+ __entry->error_code_valid = fault->error_code_valid;
+ ),
+ TP_printk("vcpu %d %s address %llx error %x",
+ __entry->vcpu_id,
+ __print_symbolic(__entry->vector, kvm_trace_sym_exc),
+ __entry->vector == PF_VECTOR ? __entry->address : 0,
+ __entry->error_code_valid ? __entry->error_code : 0
+ )
+);
+
+#endif /* _TRACE_KVMI_H */
+
+#include <trace/define_trace.h>
diff --git a/virt/kvm/kvmi.c b/virt/kvm/kvmi.c
index 157f3a401d64..ce28ca8c8d77 100644
--- a/virt/kvm/kvmi.c
+++ b/virt/kvm/kvmi.c
@@ -12,6 +12,9 @@
#include <linux/bitmap.h>
#include <linux/remote_mapping.h>
+#define CREATE_TRACE_POINTS
+#include <trace/events/kvmi.h>
+
#define MAX_PAUSE_REQUESTS 1001
static struct kmem_cache *msg_cache;
@@ -1284,6 +1287,8 @@ static void __kvmi_singlestep_event(struct kvm_vcpu *vcpu)
{
u32 action;
+ trace_kvmi_event_singlestep_send(vcpu->vcpu_id);
+
action = kvmi_send_singlestep(vcpu);
switch (action) {
case KVMI_EVENT_ACTION_CONTINUE:
@@ -1291,6 +1296,8 @@ static void __kvmi_singlestep_event(struct kvm_vcpu *vcpu)
default:
kvmi_handle_common_event_actions(vcpu, action, "SINGLESTEP");
}
+
+ trace_kvmi_event_singlestep_recv(vcpu->vcpu_id, action);
}
static void kvmi_singlestep_event(struct kvm_vcpu *vcpu)
@@ -1311,6 +1318,8 @@ static bool __kvmi_create_vcpu_event(struct kvm_vcpu *vcpu)
u32 action;
bool ret = false;
+ trace_kvmi_event_create_vcpu_send(vcpu->vcpu_id);
+
action = kvmi_msg_send_create_vcpu(vcpu);
switch (action) {
case KVMI_EVENT_ACTION_CONTINUE:
@@ -1320,6 +1329,8 @@ static bool __kvmi_create_vcpu_event(struct kvm_vcpu *vcpu)
kvmi_handle_common_event_actions(vcpu, action, "CREATE");
}
+ trace_kvmi_event_create_vcpu_recv(vcpu->vcpu_id, action);
+
return ret;
}
@@ -1345,6 +1356,8 @@ static bool __kvmi_pause_vcpu_event(struct kvm_vcpu *vcpu)
u32 action;
bool ret = false;
+ trace_kvmi_event_pause_vcpu_send(vcpu->vcpu_id);
+
action = kvmi_msg_send_pause_vcpu(vcpu);
switch (action) {
case KVMI_EVENT_ACTION_CONTINUE:
@@ -1354,6 +1367,8 @@ static bool __kvmi_pause_vcpu_event(struct kvm_vcpu *vcpu)
kvmi_handle_common_event_actions(vcpu, action, "PAUSE");
}
+ trace_kvmi_event_pause_vcpu_recv(vcpu->vcpu_id, action);
+
return ret;
}
@@ -1857,6 +1872,8 @@ void kvmi_stop_ss(struct kvm_vcpu *vcpu)
ivcpu->ss_owner = false;
+ trace_kvmi_stop_singlestep(vcpu->vcpu_id);
+
kvmi_singlestep_event(vcpu);
out:
@@ -1892,6 +1909,9 @@ static bool kvmi_run_ss(struct kvm_vcpu *vcpu, gpa_t gpa, u8 access)
gfn_t gfn = gpa_to_gfn(gpa);
int err;
+ trace_kvmi_run_singlestep(vcpu, gpa, access, ikvm->ss_level,
+ IVCPU(vcpu)->ctx_size);
+
kvmi_arch_start_single_step(vcpu);
err = write_custom_data(vcpu);
diff --git a/virt/kvm/kvmi_mem.c b/virt/kvm/kvmi_mem.c
index 6244add60062..a7a01646ea5c 100644
--- a/virt/kvm/kvmi_mem.c
+++ b/virt/kvm/kvmi_mem.c
@@ -23,6 +23,7 @@
#include <linux/remote_mapping.h>
#include <uapi/linux/kvmi.h>
+#include <trace/events/kvmi.h>
#include "kvmi_int.h"
@@ -221,6 +222,8 @@ int kvmi_host_mem_map(struct kvm_vcpu *vcpu, gva_t tkn_gva,
}
req_mm = target_kvm->mm;
+ trace_kvmi_mem_map(target_kvm, req_gpa, map_gpa);
+
/* translate source addresses */
req_gfn = gpa_to_gfn(req_gpa);
req_hva = gfn_to_hva_safe(target_kvm, req_gfn);
@@ -274,6 +277,8 @@ int kvmi_host_mem_unmap(struct kvm_vcpu *vcpu, gpa_t map_gpa)
kvm_debug("kvmi: unmapping request for map_gpa %016llx\n", map_gpa);
+ trace_kvmi_mem_unmap(map_gpa);
+
/* convert GPA -> HVA */
map_gfn = gpa_to_gfn(map_gpa);
map_hva = gfn_to_hva_safe(vcpu->kvm, map_gfn);
diff --git a/virt/kvm/kvmi_msg.c b/virt/kvm/kvmi_msg.c
index a5f87aafa237..bdb1e60906f9 100644
--- a/virt/kvm/kvmi_msg.c
+++ b/virt/kvm/kvmi_msg.c
@@ -8,6 +8,8 @@
#include <linux/net.h>
#include "kvmi_int.h"
+#include <trace/events/kvmi.h>
+
typedef int (*vcpu_reply_fct)(struct kvm_vcpu *vcpu,
const struct kvmi_msg_hdr *msg, int err,
const void *rpl, size_t rpl_size);
@@ -165,6 +167,8 @@ static int kvmi_msg_vm_reply(struct kvmi *ikvm,
const struct kvmi_msg_hdr *msg, int err,
const void *rpl, size_t rpl_size)
{
+ trace_kvmi_vm_reply(msg->id, msg->seq, err);
+
return kvmi_msg_reply(ikvm, msg, err, rpl, rpl_size);
}
@@ -202,6 +206,8 @@ int kvmi_msg_vcpu_reply(struct kvm_vcpu *vcpu,
const struct kvmi_msg_hdr *msg, int err,
const void *rpl, size_t rpl_size)
{
+ trace_kvmi_vcpu_reply(vcpu->vcpu_id, msg->id, msg->seq, err);
+
return kvmi_msg_reply(IKVM(vcpu->kvm), msg, err, rpl, rpl_size);
}
@@ -559,6 +565,8 @@ static int handle_event_reply(struct kvm_vcpu *vcpu,
struct kvmi_vcpu_reply *expected = &ivcpu->reply;
size_t useful, received, common;
+ trace_kvmi_event_reply(reply->event, msg->seq);
+
if (unlikely(msg->seq != expected->seq))
goto out;
@@ -883,6 +891,8 @@ static struct kvmi_msg_hdr *kvmi_msg_recv(struct kvmi *ikvm, bool *unsupported)
static int kvmi_msg_dispatch_vm_cmd(struct kvmi *ikvm,
const struct kvmi_msg_hdr *msg)
{
+ trace_kvmi_vm_command(msg->id, msg->seq);
+
return msg_vm[msg->id](ikvm, msg, msg + 1);
}
@@ -895,6 +905,8 @@ static int kvmi_msg_dispatch_vcpu_job(struct kvmi *ikvm,
struct kvm_vcpu *vcpu = NULL;
int err;
+ trace_kvmi_vcpu_command(cmd->vcpu, hdr->id, hdr->seq);
+
if (invalid_vcpu_hdr(cmd))
return -KVM_EINVAL;
@@ -1051,6 +1063,8 @@ int kvmi_send_event(struct kvm_vcpu *vcpu, u32 ev_id,
ivcpu->reply.size = rpl_size;
ivcpu->reply.error = -EINTR;
+ trace_kvmi_event(vcpu->vcpu_id, common.event, hdr.seq);
+
err = kvmi_sock_write(ikvm, vec, n, msg_size);
if (err)
goto out;
@@ -1091,6 +1105,8 @@ int kvmi_msg_send_unhook(struct kvmi *ikvm)
kvmi_setup_event_common(&common, KVMI_EVENT_UNHOOK, 0);
+ trace_kvmi_event(0, common.event, hdr.seq);
+
return kvmi_sock_write(ikvm, vec, n, msg_size);
}
next prev parent reply other threads:[~2019-08-09 16:05 UTC|newest]
Thread overview: 168+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-08-09 15:59 [RFC PATCH v6 00/92] VM introspection Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 01/92] kvm: introduce KVMI (VM introspection subsystem) Adalbert Lazăr
2019-08-12 20:20 ` Sean Christopherson
2019-08-13 9:11 ` Paolo Bonzini
2019-08-13 11:57 ` Adalbert Lazăr
[not found] ` <5d52a5ae.1c69fb81.5c260.1573SMTPIN_ADDED_BROKEN@mx.google.com>
2019-08-13 12:09 ` Paolo Bonzini
2019-08-13 15:01 ` Sean Christopherson
2019-08-13 21:03 ` Paolo Bonzini
2019-08-14 9:48 ` Adalbert Lazăr
[not found] ` <5d53d8d1.1c69fb81.7d32.0bedSMTPIN_ADDED_BROKEN@mx.google.com>
2019-08-14 10:37 ` Paolo Bonzini
2019-08-09 15:59 ` [RFC PATCH v6 02/92] kvm: introspection: add basic ioctls (hook/unhook) Adalbert Lazăr
2019-08-13 8:44 ` Paolo Bonzini
2019-08-13 14:24 ` Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 03/92] kvm: introspection: add permission access ioctls Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 04/92] kvm: introspection: add the read/dispatch message function Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 05/92] kvm: introspection: add KVMI_GET_VERSION Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 06/92] kvm: introspection: add KVMI_CONTROL_CMD_RESPONSE Adalbert Lazăr
2019-08-13 9:15 ` Paolo Bonzini
2019-08-13 17:08 ` Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 07/92] kvm: introspection: honor the reply option when handling the KVMI_GET_VERSION command Adalbert Lazăr
2019-08-13 9:16 ` Paolo Bonzini
2019-08-09 15:59 ` [RFC PATCH v6 08/92] kvm: introspection: add KVMI_CHECK_COMMAND and KVMI_CHECK_EVENT Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 09/92] kvm: introspection: add KVMI_GET_GUEST_INFO Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 10/92] kvm: introspection: add KVMI_CONTROL_VM_EVENTS Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 11/92] kvm: introspection: add vCPU related data Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 12/92] kvm: introspection: add a jobs list to every introspected vCPU Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 13/92] kvm: introspection: make the vCPU wait even when its jobs list is empty Adalbert Lazăr
2019-08-13 8:43 ` Paolo Bonzini
2019-08-13 14:19 ` Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 14/92] kvm: introspection: handle introspection commands before returning to guest Adalbert Lazăr
2019-08-13 8:26 ` Paolo Bonzini
2019-08-13 13:54 ` Adalbert Lazăr
[not found] ` <5d52c10e.1c69fb81.26904.fd34SMTPIN_ADDED_BROKEN@mx.google.com>
2019-08-13 14:45 ` Paolo Bonzini
2019-08-14 9:39 ` Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 15/92] kvm: introspection: handle vCPU related introspection commands Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 16/92] kvm: introspection: handle events and event replies Adalbert Lazăr
2019-08-13 8:55 ` Paolo Bonzini
2019-08-13 15:25 ` Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 17/92] kvm: introspection: introduce event actions Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 18/92] kvm: introspection: add KVMI_EVENT_UNHOOK Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 19/92] kvm: introspection: add KVMI_EVENT_CREATE_VCPU Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 20/92] kvm: introspection: add KVMI_GET_VCPU_INFO Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 21/92] kvm: page track: add track_create_slot() callback Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 22/92] kvm: x86: provide all page tracking hooks with the guest virtual address Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 23/92] kvm: page track: add support for preread, prewrite and preexec Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 24/92] kvm: x86: wire in the preread/prewrite/preexec page trackers Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 25/92] kvm: x86: intercept the write access on sidt and other emulated instructions Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 26/92] kvm: x86: add kvm_mmu_nested_pagefault() Adalbert Lazăr
2019-08-13 8:12 ` Paolo Bonzini
2019-08-09 15:59 ` [RFC PATCH v6 27/92] kvm: introspection: use page track Adalbert Lazăr
2019-08-13 9:06 ` Paolo Bonzini
2019-08-09 15:59 ` [RFC PATCH v6 28/92] kvm: x86: consult the page tracking from kvm_mmu_get_page() and __direct_map() Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 29/92] kvm: introspection: add KVMI_CONTROL_EVENTS Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 30/92] kvm: x86: add kvm_spt_fault() Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 31/92] kvm: introspection: add KVMI_EVENT_PF Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 32/92] kvm: introspection: add KVMI_GET_PAGE_ACCESS Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 33/92] kvm: introspection: add KVMI_SET_PAGE_ACCESS Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 34/92] Documentation: Introduce EPT based Subpage Protection Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 35/92] KVM: VMX: Add control flags for SPP enabling Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 36/92] KVM: VMX: Implement functions for SPPT paging setup Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 37/92] KVM: VMX: Introduce SPP access bitmap and operation functions Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 38/92] KVM: VMX: Add init/set/get functions for SPP Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 39/92] KVM: VMX: Introduce SPP user-space IOCTLs Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 40/92] KVM: VMX: Handle SPP induced vmexit and page fault Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 41/92] KVM: MMU: Enable Lazy mode SPPT setup Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 42/92] KVM: MMU: Handle host memory remapping and reclaim Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 43/92] kvm: introspection: add KVMI_CONTROL_SPP Adalbert Lazăr
2019-08-09 15:59 ` [RFC PATCH v6 44/92] kvm: introspection: extend the internal database of tracked pages with write_bitmap info Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 45/92] kvm: introspection: add KVMI_GET_PAGE_WRITE_BITMAP Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 46/92] kvm: introspection: add KVMI_SET_PAGE_WRITE_BITMAP Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 47/92] kvm: introspection: add KVMI_READ_PHYSICAL and KVMI_WRITE_PHYSICAL Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 48/92] kvm: add kvm_vcpu_kick_and_wait() Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 49/92] kvm: introspection: add KVMI_PAUSE_VCPU and KVMI_EVENT_PAUSE_VCPU Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 50/92] kvm: introspection: add KVMI_GET_REGISTERS Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 51/92] kvm: introspection: add KVMI_SET_REGISTERS Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 52/92] kvm: introspection: add KVMI_GET_CPUID Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 53/92] kvm: introspection: add KVMI_INJECT_EXCEPTION + KVMI_EVENT_TRAP Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 54/92] kvm: introspection: add KVMI_CONTROL_CR and KVMI_EVENT_CR Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 55/92] kvm: introspection: add KVMI_CONTROL_MSR and KVMI_EVENT_MSR Adalbert Lazăr
2019-08-12 21:05 ` Sean Christopherson
2019-08-15 6:36 ` Nicusor CITU
2019-08-19 18:36 ` Sean Christopherson
2019-08-20 8:44 ` Nicusor CITU
2019-08-20 11:43 ` Mihai Donțu
2019-08-21 15:18 ` Sean Christopherson
2019-08-19 18:52 ` Sean Christopherson
2019-08-09 16:00 ` [RFC PATCH v6 56/92] kvm: x86: block any attempt to disable MSR interception if tracked by introspection Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 57/92] kvm: introspection: add KVMI_GET_XSAVE Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 58/92] kvm: introspection: add KVMI_GET_MTRR_TYPE Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 59/92] kvm: introspection: add KVMI_EVENT_XSETBV Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 60/92] kvm: x86: add kvm_arch_vcpu_set_guest_debug() Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 61/92] kvm: introspection: add KVMI_EVENT_BREAKPOINT Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 62/92] kvm: introspection: add KVMI_EVENT_HYPERCALL Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 63/92] kvm: introspection: add KVMI_EVENT_DESCRIPTOR Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 64/92] kvm: introspection: add single-stepping Adalbert Lazăr
2019-08-12 20:50 ` Sean Christopherson
2019-08-13 12:51 ` Adalbert Lazăr
2019-08-14 12:36 ` Nicusor CITU
2019-08-14 12:53 ` Paolo Bonzini
2019-08-09 16:00 ` [RFC PATCH v6 65/92] kvm: introspection: add KVMI_EVENT_SINGLESTEP Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 66/92] kvm: introspection: add custom input when single-stepping a vCPU Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 67/92] kvm: introspection: use single stepping on unimplemented instructions Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 68/92] kvm: x86: emulate a guest page table walk on SPT violations due to A/D bit updates Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 69/92] kvm: x86: keep the page protected if tracked by the introspection tool Adalbert Lazăr
2019-09-10 14:26 ` Konrad Rzeszutek Wilk
2019-09-10 16:28 ` Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 70/92] kvm: x86: filter out access rights only when " Adalbert Lazăr
2019-08-13 9:08 ` Paolo Bonzini
2019-08-13 16:06 ` Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 71/92] mm: add support for remote mapping Adalbert Lazăr
2019-08-09 16:24 ` DANGER WILL ROBINSON, DANGER Matthew Wilcox
2019-08-13 9:29 ` Paolo Bonzini
2019-08-13 11:24 ` Matthew Wilcox
2019-08-13 12:02 ` Paolo Bonzini
2019-08-13 11:01 ` Adalbert Lazăr
2019-08-15 19:19 ` Jerome Glisse
2019-08-15 20:16 ` Jerome Glisse
2019-08-16 17:45 ` Jason Gunthorpe
2019-08-23 12:39 ` Mircea CIRJALIU - MELIU
2019-09-05 18:09 ` Jerome Glisse
2019-09-09 17:00 ` Paolo Bonzini
2019-09-10 7:49 ` Mircea CIRJALIU - MELIU
2019-10-02 19:27 ` Jerome Glisse
2019-10-02 13:46 ` Paolo Bonzini
2019-10-02 14:15 ` Jerome Glisse
2019-10-02 16:18 ` Paolo Bonzini
2019-10-02 17:04 ` Jerome Glisse
2019-10-02 20:10 ` Paolo Bonzini
2019-10-03 15:42 ` Jerome Glisse
2019-10-03 15:50 ` Paolo Bonzini
2019-10-03 16:42 ` Mircea CIRJALIU - MELIU
2019-10-03 18:31 ` Jerome Glisse
2019-10-03 19:38 ` Paolo Bonzini
2019-10-04 9:41 ` Mircea CIRJALIU - MELIU
2019-10-04 11:46 ` Paolo Bonzini
2019-10-03 16:36 ` Mircea CIRJALIU - MELIU
2019-08-09 16:00 ` [RFC PATCH v6 72/92] kvm: introspection: add memory map/unmap support on the guest side Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 73/92] kvm: introspection: use remote mapping Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 74/92] kvm: x86: do not unconditionally patch the hypercall instruction during emulation Adalbert Lazăr
2019-08-13 9:20 ` Paolo Bonzini
2019-08-14 12:07 ` Adalbert Lazăr
[not found] ` <5d53f965.1c69fb81.cd952.035bSMTPIN_ADDED_BROKEN@mx.google.com>
2019-08-14 12:33 ` Paolo Bonzini
2019-08-09 16:00 ` [RFC PATCH v6 75/92] kvm: x86: disable gpa_available optimization in emulator_read_write_onepage() Adalbert Lazăr
2019-08-13 8:47 ` Paolo Bonzini
2019-08-13 14:33 ` Adalbert Lazăr
[not found] ` <5d52ca22.1c69fb81.4ceb8.e90bSMTPIN_ADDED_BROKEN@mx.google.com>
2019-08-13 14:35 ` Paolo Bonzini
2019-08-09 16:00 ` [RFC PATCH v6 76/92] kvm: x86: disable EPT A/D bits if introspection is present Adalbert Lazăr
2019-08-13 9:18 ` Paolo Bonzini
2019-08-09 16:00 ` Adalbert Lazăr [this message]
2019-08-09 16:00 ` [RFC PATCH v6 78/92] kvm: x86: add tracepoints for interrupt and exception injections Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 79/92] kvm: x86: emulate movsd xmm, m64 Adalbert Lazăr
2019-08-13 9:17 ` Paolo Bonzini
2019-08-09 16:00 ` [RFC PATCH v6 80/92] kvm: x86: emulate movss xmm, m32 Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 81/92] kvm: x86: emulate movq xmm, m64 Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 82/92] kvm: x86: emulate movq r, xmm Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 83/92] kvm: x86: emulate movd xmm, m32 Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 84/92] kvm: x86: enable the half part of movss, movsd, movups Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 85/92] kvm: x86: emulate lfence Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 86/92] kvm: x86: emulate xorpd xmm2/m128, xmm1 Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 87/92] kvm: x86: emulate xorps xmm/m128, xmm Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 88/92] kvm: x86: emulate fst/fstp m64fp Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 89/92] kvm: x86: make lock cmpxchg r, r/m atomic Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 90/92] kvm: x86: emulate lock cmpxchg8b atomically Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 91/92] kvm: x86: emulate lock cmpxchg16b m128 Adalbert Lazăr
2019-08-09 16:00 ` [RFC PATCH v6 92/92] kvm: x86: fallback to the single-step on multipage CMPXCHG emulation Adalbert Lazăr
2019-08-12 18:23 ` [RFC PATCH v6 00/92] VM introspection Sean Christopherson
2019-08-12 21:40 ` Sean Christopherson
2019-08-13 9:34 ` Paolo Bonzini
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20190809160047.8319-78-alazar@bitdefender.com \
--to=alazar@bitdefender.com \
--cc=Zhang@kvack.org \
--cc=jan.kiszka@siemens.com \
--cc=konrad.wilk@oracle.com \
--cc=kvm@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=marian.c.rotariu@gmail.com \
--cc=mathieu.tarral@protonmail.com \
--cc=mcirjaliu@bitdefender.com \
--cc=mdontu@bitdefender.com \
--cc=ncitu@bitdefender.com \
--cc=patrick.colp@oracle.com \
--cc=pbonzini@redhat.com \
--cc=rkrcmar@redhat.com \
--cc=samuel.lauren@iki.fi \
--cc=stefanha@redhat.com \
--cc=tamas@tklengyel.com \
--cc=virtualization@lists.linux-foundation.org \
--cc=weijiang.yang@intel.com \
--cc=yu.c.zhang@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).