qemu-devel.nongnu.org archive mirror
 help / color / mirror / Atom feed
From: Paul Durrant <xadimgnik@gmail.com>
To: David Woodhouse <dwmw2@infradead.org>, qemu-devel@nongnu.org
Cc: "Paolo Bonzini" <pbonzini@redhat.com>,
	"Joao Martins" <joao.m.martins@oracle.com>,
	"Ankur Arora" <ankur.a.arora@oracle.com>,
	"Philippe Mathieu-Daudé" <philmd@linaro.org>,
	"Thomas Huth" <thuth@redhat.com>,
	"Alex Bennée" <alex.bennee@linaro.org>,
	"Juan Quintela" <quintela@redhat.com>,
	"Dr . David Alan Gilbert" <dgilbert@redhat.com>,
	"Claudio Fontana" <cfontana@suse.de>
Subject: Re: [RFC PATCH v2 17/22] i386/xen: handle VCPUOP_register_vcpu_time_info
Date: Mon, 12 Dec 2022 15:34:37 +0000	[thread overview]
Message-ID: <de303284-0604-ed40-bf8a-ec25a7b11776@xen.org> (raw)
In-Reply-To: <20221209095612.689243-18-dwmw2@infradead.org>

On 09/12/2022 09:56, David Woodhouse wrote:
> From: Joao Martins <joao.m.martins@oracle.com>
> 
> In order to support Linux vdso in Xen.
> 
> Signed-off-by: Joao Martins <joao.m.martins@oracle.com>
> Signed-off-by: David Woodhouse <dwmw@amazon.co.uk>
> ---
>   target/i386/cpu.h     |  1 +
>   target/i386/kvm/kvm.c |  9 ++++++
>   target/i386/machine.c |  4 ++-
>   target/i386/xen.c     | 70 ++++++++++++++++++++++++++++++++++++-------
>   4 files changed, 72 insertions(+), 12 deletions(-)
> 
> diff --git a/target/i386/cpu.h b/target/i386/cpu.h
> index 109b2e5669..96c2d0d5cb 100644
> --- a/target/i386/cpu.h
> +++ b/target/i386/cpu.h
> @@ -1790,6 +1790,7 @@ typedef struct CPUArchState {
>       struct kvm_nested_state *nested_state;
>       uint64_t xen_vcpu_info_gpa;
>       uint64_t xen_vcpu_info_default_gpa;
> +    uint64_t xen_vcpu_time_info_gpa;
>   #endif
>   #if defined(CONFIG_HVF)
>       HVFX86LazyFlags hvf_lflags;
> diff --git a/target/i386/kvm/kvm.c b/target/i386/kvm/kvm.c
> index fa45e2f99a..3f19fff21f 100644
> --- a/target/i386/kvm/kvm.c
> +++ b/target/i386/kvm/kvm.c
> @@ -1813,6 +1813,7 @@ int kvm_arch_init_vcpu(CPUState *cs)
>   
>       env->xen_vcpu_info_gpa = UINT64_MAX;
>       env->xen_vcpu_info_default_gpa = UINT64_MAX;
> +    env->xen_vcpu_time_info_gpa = UINT64_MAX;

Another few candidates for INVALID_GPA.

>   
>       xen_version = kvm_arch_xen_version(MACHINE(qdev_get_machine()));
>       if (xen_version) {
> @@ -4744,6 +4745,14 @@ int kvm_arch_put_registers(CPUState *cpu, int level)
>                   return ret;
>               }
>           }
> +
> +        gpa = x86_cpu->env.xen_vcpu_time_info_gpa;
> +        if (gpa != UINT64_MAX) {
> +            ret = kvm_xen_set_vcpu_attr(cpu, KVM_XEN_VCPU_ATTR_TYPE_VCPU_TIME_INFO, gpa);
> +            if (ret < 0) {
> +                return ret;
> +            }
> +        }
>       }
>   #endif
>   
> diff --git a/target/i386/machine.c b/target/i386/machine.c
> index 104cd6047c..9acef102a3 100644
> --- a/target/i386/machine.c
> +++ b/target/i386/machine.c
> @@ -1263,7 +1263,8 @@ static bool xen_vcpu_needed(void *opaque)
>       CPUX86State *env = &cpu->env;
>   
>       return (env->xen_vcpu_info_gpa != UINT64_MAX ||
> -            env->xen_vcpu_info_default_gpa != UINT64_MAX);
> +            env->xen_vcpu_info_default_gpa != UINT64_MAX ||
> +            env->xen_vcpu_time_info_gpa != UINT64_MAX);
>   }
>   
>   static const VMStateDescription vmstate_xen_vcpu = {
> @@ -1274,6 +1275,7 @@ static const VMStateDescription vmstate_xen_vcpu = {
>       .fields = (VMStateField[]) {
>           VMSTATE_UINT64(env.xen_vcpu_info_gpa, X86CPU),
>           VMSTATE_UINT64(env.xen_vcpu_info_default_gpa, X86CPU),
> +        VMSTATE_UINT64(env.xen_vcpu_time_info_gpa, X86CPU),
>           VMSTATE_END_OF_LIST()
>       }
>   };
> diff --git a/target/i386/xen.c b/target/i386/xen.c
> index cd816bb711..427729ab4d 100644
> --- a/target/i386/xen.c
> +++ b/target/i386/xen.c
> @@ -21,28 +21,41 @@
>   #include "standard-headers/xen/hvm/hvm_op.h"
>   #include "standard-headers/xen/vcpu.h"
>   
> +static bool kvm_gva_to_gpa(CPUState *cs, uint64_t gva, uint64_t *gpa,
> +                           size_t *len, bool is_write)
> +{
> +        struct kvm_translation tr = {
> +            .linear_address = gva,
> +        };
> +
> +        if (len) {
> +                *len = TARGET_PAGE_SIZE - (gva & ~TARGET_PAGE_MASK);
> +        }
> +
> +        if (kvm_vcpu_ioctl(cs, KVM_TRANSLATE, &tr) || !tr.valid ||
> +            (is_write && !tr.writeable)) {
> +            return false;
> +        }
> +        *gpa = tr.physical_address;
> +        return true;
> +}
> +
>   static int kvm_gva_rw(CPUState *cs, uint64_t gva, void *_buf, size_t sz,
>                         bool is_write)
>   {
>       uint8_t *buf = (uint8_t *)_buf;
>       size_t i = 0, len = 0;
> -    int ret;
>   
>       for (i = 0; i < sz; i+= len) {
> -        struct kvm_translation tr = {
> -            .linear_address = gva + i,
> -        };
> +        uint64_t gpa;
>   
> -        len = TARGET_PAGE_SIZE - (tr.linear_address & ~TARGET_PAGE_MASK);
> +        if (!kvm_gva_to_gpa(cs, gva + i, &gpa, &len, is_write)) {
> +                return -EFAULT;
> +        }
>           if (len > sz)
>               len = sz;
>   
> -        ret = kvm_vcpu_ioctl(cs, KVM_TRANSLATE, &tr);
> -        if (ret || !tr.valid || (is_write && !tr.writeable)) {
> -            return -EFAULT;
> -        }
> -
> -        cpu_physical_memory_rw(tr.physical_address, buf + i, len, is_write);
> +        cpu_physical_memory_rw(gpa, buf + i, len, is_write);
>       }
>   
>       return 0;
> @@ -166,6 +179,17 @@ static void do_set_vcpu_info_gpa(CPUState *cs, run_on_cpu_data data)
>                             env->xen_vcpu_info_gpa);
>   }
>   
> +static void do_set_vcpu_time_info_gpa(CPUState *cs, run_on_cpu_data data)
> +{
> +    X86CPU *cpu = X86_CPU(cs);
> +    CPUX86State *env = &cpu->env;
> +
> +    env->xen_vcpu_time_info_gpa = data.host_ulong;
> +
> +    kvm_xen_set_vcpu_attr(cs, KVM_XEN_VCPU_ATTR_TYPE_VCPU_TIME_INFO,
> +                          env->xen_vcpu_time_info_gpa);
> +}
> +
>   static int xen_set_shared_info(CPUState *cs, uint64_t gfn)
>   {
>       uint64_t gpa = gfn << TARGET_PAGE_BITS;
> @@ -258,6 +282,27 @@ static int vcpuop_register_vcpu_info(CPUState *cs, CPUState *target,
>       return 0;
>   }
>   
> +static int vcpuop_register_vcpu_time_info(CPUState *cs, CPUState *target,
> +                                          uint64_t arg)
> +{
> +    struct vcpu_register_time_memory_area tma;
> +    uint64_t gpa;
> +    size_t len;
> +
> +    if (kvm_copy_from_gva(cs, arg, &tma, sizeof(*tma.addr.v))) {
> +        return -EFAULT;
> +    }
> +
> +    if (!kvm_gva_to_gpa(cs, tma.addr.p, &gpa, &len, false) ||
> +        len < sizeof(tma)) {
> +        return -EFAULT;
> +    }

Xen stashes the GVA, not the GPA, and so it would be possible to 
register the same GVA on different vcpus to point at different areas of 
memory.

   Paul

> +
> +    async_run_on_cpu(target, do_set_vcpu_time_info_gpa,
> +                     RUN_ON_CPU_HOST_ULONG(gpa));
> +    return 0;
> +}
> +
>   static bool kvm_xen_hcall_vcpu_op(struct kvm_xen_exit *exit, X86CPU *cpu,
>                                     int cmd, int vcpu_id, uint64_t arg)
>   {
> @@ -266,6 +311,9 @@ static bool kvm_xen_hcall_vcpu_op(struct kvm_xen_exit *exit, X86CPU *cpu,
>       int err;
>   
>       switch (cmd) {
> +    case VCPUOP_register_vcpu_time_memory_area:
> +            err = vcpuop_register_vcpu_time_info(cs, dest, arg);
> +            break;
>       case VCPUOP_register_vcpu_info:
>               err = vcpuop_register_vcpu_info(cs, dest, arg);
>               break;



  reply	other threads:[~2022-12-12 15:35 UTC|newest]

Thread overview: 78+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2022-12-09  9:55 [RFC PATCH v2 00/22] Xen HVM support under KVM David Woodhouse
2022-12-09  9:55 ` [RFC PATCH v2 01/22] include: import xen public headers David Woodhouse
2022-12-12  9:17   ` Paul Durrant
2022-12-09  9:55 ` [RFC PATCH v2 02/22] xen: add CONFIG_XENFV_MACHINE and CONFIG_XEN_EMU options for Xen emulation David Woodhouse
2022-12-12  9:19   ` Paul Durrant
2022-12-12 17:07   ` Paolo Bonzini
2022-12-12 22:22     ` David Woodhouse
2022-12-13  0:39       ` Paolo Bonzini
2022-12-13  0:59         ` David Woodhouse
2022-12-13 22:32           ` Paolo Bonzini
2022-12-16  8:40             ` David Woodhouse
2022-12-09  9:55 ` [RFC PATCH v2 03/22] i386/xen: Add xen-version machine property and init KVM Xen support David Woodhouse
2022-12-12 12:48   ` Paul Durrant
2022-12-12 17:30   ` Paolo Bonzini
2022-12-12 17:55     ` Paul Durrant
2022-12-13  0:13     ` David Woodhouse
2023-01-17 13:49     ` David Woodhouse
2022-12-09  9:55 ` [RFC PATCH v2 04/22] i386/kvm: handle Xen HVM cpuid leaves David Woodhouse
2022-12-12 13:13   ` Paul Durrant
2022-12-13  9:47     ` David Woodhouse
2022-12-09  9:55 ` [RFC PATCH v2 05/22] xen-platform-pci: allow its creation with XEN_EMULATE mode David Woodhouse
2022-12-12 13:24   ` Paul Durrant
2022-12-12 22:07     ` David Woodhouse
2022-12-09  9:55 ` [RFC PATCH v2 06/22] hw/xen_backend: refactor xen_be_init() David Woodhouse
2022-12-12 13:27   ` Paul Durrant
2022-12-09  9:55 ` [RFC PATCH v2 07/22] pc_piix: handle XEN_EMULATE backend init David Woodhouse
2022-12-12 13:47   ` Paul Durrant
2022-12-12 14:50     ` David Woodhouse
2022-12-09  9:55 ` [RFC PATCH v2 08/22] xen_platform: exclude vfio-pci from the PCI platform unplug David Woodhouse
2022-12-12 13:52   ` Paul Durrant
2022-12-09  9:55 ` [RFC PATCH v2 09/22] pc_piix: allow xenfv machine with XEN_EMULATE David Woodhouse
2022-12-12 14:05   ` Paul Durrant
2022-12-09  9:56 ` [RFC PATCH v2 10/22] i386/xen: handle guest hypercalls David Woodhouse
2022-12-12 14:11   ` Paul Durrant
2022-12-12 14:17     ` David Woodhouse
2022-12-12 17:07   ` Paolo Bonzini
2022-12-09  9:56 ` [RFC PATCH v2 11/22] i386/xen: implement HYPERCALL_xen_version David Woodhouse
2022-12-12 14:17   ` Paul Durrant
2022-12-13  0:06     ` David Woodhouse
2022-12-09  9:56 ` [RFC PATCH v2 12/22] hw/xen: Add xen_overlay device for emulating shared xenheap pages David Woodhouse
2022-12-12 14:29   ` Paul Durrant
2022-12-12 17:14   ` Paolo Bonzini
2022-12-09  9:56 ` [RFC PATCH v2 13/22] i386/xen: implement HYPERVISOR_memory_op David Woodhouse
2022-12-12 14:38   ` Paul Durrant
2022-12-13  0:08     ` David Woodhouse
2022-12-09  9:56 ` [RFC PATCH v2 14/22] i386/xen: implement HYPERVISOR_hvm_op David Woodhouse
2022-12-12 14:41   ` Paul Durrant
2022-12-09  9:56 ` [RFC PATCH v2 15/22] i386/xen: implement HYPERVISOR_vcpu_op David Woodhouse
2022-12-12 14:51   ` Paul Durrant
2022-12-13  0:10     ` David Woodhouse
2022-12-09  9:56 ` [RFC PATCH v2 16/22] i386/xen: handle VCPUOP_register_vcpu_info David Woodhouse
2022-12-12 14:58   ` Paul Durrant
2022-12-13  0:13     ` David Woodhouse
2022-12-14 10:28       ` Paul Durrant
2022-12-14 11:04         ` David Woodhouse
2022-12-09  9:56 ` [RFC PATCH v2 17/22] i386/xen: handle VCPUOP_register_vcpu_time_info David Woodhouse
2022-12-12 15:34   ` Paul Durrant [this message]
2022-12-09  9:56 ` [RFC PATCH v2 18/22] i386/xen: handle VCPUOP_register_runstate_memory_area David Woodhouse
2022-12-12 15:38   ` Paul Durrant
2022-12-09  9:56 ` [RFC PATCH v2 19/22] i386/xen: implement HVMOP_set_evtchn_upcall_vector David Woodhouse
2022-12-12 15:52   ` Paul Durrant
2022-12-09  9:56 ` [RFC PATCH v2 20/22] i386/xen: HVMOP_set_param / HVM_PARAM_CALLBACK_IRQ David Woodhouse
2022-12-12 16:16   ` Paul Durrant
2022-12-12 16:26     ` David Woodhouse
2022-12-12 16:39       ` Paul Durrant
2022-12-15 20:54         ` David Woodhouse
2022-12-20 13:56           ` Paul Durrant
2022-12-20 16:27             ` David Woodhouse
2022-12-20 17:25               ` Paul Durrant
2022-12-20 17:29                 ` David Woodhouse
2022-12-28 10:45                   ` David Woodhouse
2022-12-21  1:41     ` David Woodhouse
2022-12-21  9:37       ` Paul Durrant
2022-12-21 12:16         ` David Woodhouse
2022-12-09  9:56 ` [RFC PATCH v2 21/22] i386/xen: implement HYPERVISOR_event_channel_op David Woodhouse
2022-12-12 16:23   ` Paul Durrant
2022-12-09  9:56 ` [RFC PATCH v2 22/22] i386/xen: implement HYPERVISOR_sched_op David Woodhouse
2022-12-12 16:37   ` Paul Durrant

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=de303284-0604-ed40-bf8a-ec25a7b11776@xen.org \
    --to=xadimgnik@gmail.com \
    --cc=alex.bennee@linaro.org \
    --cc=ankur.a.arora@oracle.com \
    --cc=cfontana@suse.de \
    --cc=dgilbert@redhat.com \
    --cc=dwmw2@infradead.org \
    --cc=joao.m.martins@oracle.com \
    --cc=pbonzini@redhat.com \
    --cc=philmd@linaro.org \
    --cc=qemu-devel@nongnu.org \
    --cc=quintela@redhat.com \
    --cc=thuth@redhat.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).