From: Janosch Frank <frankja@linux.ibm.com>
To: kvm@vger.kernel.org
Cc: frankja@linux.ibm.com, david@redhat.com, cohuck@redhat.com,
linux-s390@vger.kernel.org, imbrenda@linux.ibm.com,
thuth@redhat.com
Subject: [kvm-unit-tests PATCH v2 6/6] s390x: Add UV host test
Date: Mon, 10 May 2021 13:51:48 +0000 [thread overview]
Message-ID: <20210510135148.1904-7-frankja@linux.ibm.com> (raw)
In-Reply-To: <20210510135148.1904-1-frankja@linux.ibm.com>
Let's also test the UV host interfaces.
Signed-off-by: Janosch Frank <frankja@linux.ibm.com>
Reviewed-by: Claudio Imbrenda <imbrenda@linux.ibm.com>
---
s390x/Makefile | 1 +
s390x/uv-host.c | 480 ++++++++++++++++++++++++++++++++++++++++++++++++
2 files changed, 481 insertions(+)
create mode 100644 s390x/uv-host.c
diff --git a/s390x/Makefile b/s390x/Makefile
index bbf177fa..8de926ab 100644
--- a/s390x/Makefile
+++ b/s390x/Makefile
@@ -21,6 +21,7 @@ tests += $(TEST_DIR)/css.elf
tests += $(TEST_DIR)/uv-guest.elf
tests += $(TEST_DIR)/sie.elf
tests += $(TEST_DIR)/mvpg.elf
+tests += $(TEST_DIR)/uv-host.elf
tests_binary = $(patsubst %.elf,%.bin,$(tests))
ifneq ($(HOST_KEY_DOCUMENT),)
diff --git a/s390x/uv-host.c b/s390x/uv-host.c
new file mode 100644
index 00000000..2e4cba10
--- /dev/null
+++ b/s390x/uv-host.c
@@ -0,0 +1,480 @@
+/* SPDX-License-Identifier: GPL-2.0-only */
+/*
+ * Guest Ultravisor Call tests
+ *
+ * Copyright (c) 2021 IBM Corp
+ *
+ * Authors:
+ * Janosch Frank <frankja@linux.ibm.com>
+ */
+
+#include <libcflat.h>
+#include <alloc.h>
+#include <vmalloc.h>
+#include <sclp.h>
+#include <smp.h>
+#include <uv.h>
+#include <asm/page.h>
+#include <asm/sigp.h>
+#include <asm/pgtable.h>
+#include <asm/asm-offsets.h>
+#include <asm/interrupt.h>
+#include <asm/facility.h>
+#include <asm/uv.h>
+#include <asm-generic/barrier.h>
+
+static struct uv_cb_qui uvcb_qui;
+static struct uv_cb_init uvcb_init;
+static struct uv_cb_cgc uvcb_cgc;
+static struct uv_cb_csc uvcb_csc;
+
+extern int diag308_load_reset(u64 code);
+
+struct cmd_list{
+ const char *name;
+ uint16_t cmd;
+ uint16_t len;
+ int call_bit;
+};
+
+static void cpu_loop(void)
+{
+ for (;;) {}
+}
+
+static struct cmd_list cmds[] = {
+ { "init", UVC_CMD_INIT_UV, sizeof(struct uv_cb_init), BIT_UVC_CMD_INIT_UV },
+ { "create conf", UVC_CMD_CREATE_SEC_CONF, sizeof(struct uv_cb_cgc), BIT_UVC_CMD_CREATE_SEC_CONF },
+ { "destroy conf", UVC_CMD_DESTROY_SEC_CONF, sizeof(struct uv_cb_nodata), BIT_UVC_CMD_DESTROY_SEC_CONF },
+ { "create cpu", UVC_CMD_CREATE_SEC_CPU, sizeof(struct uv_cb_csc), BIT_UVC_CMD_CREATE_SEC_CPU },
+ { "destroy cpu", UVC_CMD_DESTROY_SEC_CPU, sizeof(struct uv_cb_nodata), BIT_UVC_CMD_DESTROY_SEC_CPU },
+ { "conv to", UVC_CMD_CONV_TO_SEC_STOR, sizeof(struct uv_cb_cts), BIT_UVC_CMD_CONV_TO_SEC_STOR },
+ { "conv from", UVC_CMD_CONV_FROM_SEC_STOR, sizeof(struct uv_cb_cfs), BIT_UVC_CMD_CONV_FROM_SEC_STOR },
+ { "set sec conf", UVC_CMD_SET_SEC_CONF_PARAMS, sizeof(struct uv_cb_ssc), BIT_UVC_CMD_SET_SEC_PARMS },
+ { "unpack", UVC_CMD_UNPACK_IMG, sizeof(struct uv_cb_unp), BIT_UVC_CMD_UNPACK_IMG },
+ { "verify", UVC_CMD_VERIFY_IMG, sizeof(struct uv_cb_nodata), BIT_UVC_CMD_VERIFY_IMG },
+ { "cpu reset", UVC_CMD_CPU_RESET, sizeof(struct uv_cb_nodata), BIT_UVC_CMD_CPU_RESET },
+ { "cpu initial reset", UVC_CMD_CPU_RESET_INITIAL, sizeof(struct uv_cb_nodata), BIT_UVC_CMD_CPU_RESET_INITIAL },
+ { "conf clear reset", UVC_CMD_PERF_CONF_CLEAR_RESET, sizeof(struct uv_cb_nodata), BIT_UVC_CMD_PREPARE_CLEAR_RESET },
+ { "cpu clear reset", UVC_CMD_CPU_RESET_CLEAR, sizeof(struct uv_cb_nodata), BIT_UVC_CMD_CPU_PERFORM_CLEAR_RESET },
+ { "cpu set state", UVC_CMD_CPU_SET_STATE, sizeof(struct uv_cb_cpu_set_state), BIT_UVC_CMD_CPU_SET_STATE },
+ { "pin shared", UVC_CMD_PIN_PAGE_SHARED, sizeof(struct uv_cb_cfs), BIT_UVC_CMD_PIN_PAGE_SHARED },
+ { "unpin shared", UVC_CMD_UNPIN_PAGE_SHARED, sizeof(struct uv_cb_cts), BIT_UVC_CMD_UNPIN_PAGE_SHARED },
+ { NULL, 0, 0 },
+};
+
+static void test_priv(void)
+{
+ struct uv_cb_header uvcb = {};
+ uint16_t pgm;
+ int i;
+
+ report_prefix_push("privileged");
+ for (i = 0; cmds[i].name; i++) {
+ expect_pgm_int();
+ uvcb.cmd = cmds[i].cmd;
+ uvcb.len = cmds[i].len;
+ enter_pstate();
+ uv_call(0, (uint64_t)&uvcb);
+ pgm = clear_pgm_int();
+ report(pgm == PGM_INT_CODE_PRIVILEGED_OPERATION, "%s", cmds[i].name);
+ }
+ report_prefix_pop();
+}
+
+static void test_config_destroy(void)
+{
+ int rc;
+ struct uv_cb_nodata uvcb = {
+ .header.cmd = UVC_CMD_DESTROY_SEC_CONF,
+ .header.len = sizeof(uvcb),
+ .handle = uvcb_cgc.guest_handle,
+ };
+
+ report_prefix_push("dsc");
+ uvcb.header.len -= 8;
+ rc = uv_call(0, (uint64_t)&uvcb);
+ report(rc == 1 && uvcb.header.rc == UVC_RC_INV_LEN,
+ "hdr invalid length");
+ uvcb.header.len += 8;
+
+ uvcb.handle += 1;
+ rc = uv_call(0, (uint64_t)&uvcb);
+ report(rc == 1 && uvcb.header.rc == UVC_RC_INV_GHANDLE, "invalid handle");
+ uvcb.handle -= 1;
+
+ rc = uv_call(0, (uint64_t)&uvcb);
+ report(rc == 0 && uvcb.header.rc == UVC_RC_EXECUTED, "success");
+ report_prefix_pop();
+}
+
+static void test_cpu_destroy(void)
+{
+ int rc;
+ struct uv_cb_nodata uvcb = {
+ .header.len = sizeof(uvcb),
+ .header.cmd = UVC_CMD_DESTROY_SEC_CPU,
+ .handle = uvcb_csc.cpu_handle,
+ };
+
+ report_prefix_push("dcpu");
+
+ uvcb.header.len -= 8;
+ rc = uv_call(0, (uint64_t)&uvcb);
+ report(rc == 1 && uvcb.header.rc == UVC_RC_INV_LEN,
+ "hdr invalid length");
+ uvcb.header.len += 8;
+
+ uvcb.handle += 1;
+ rc = uv_call(0, (uint64_t)&uvcb);
+ report(rc == 1 && uvcb.header.rc == UVC_RC_INV_CHANDLE, "invalid handle");
+ uvcb.handle -= 1;
+
+ rc = uv_call(0, (uint64_t)&uvcb);
+ report(rc == 0 && uvcb.header.rc == UVC_RC_EXECUTED, "success");
+
+ report_prefix_pop();
+}
+
+static void test_cpu_create(void)
+{
+ int rc;
+ unsigned long tmp;
+
+ report_prefix_push("csc");
+ uvcb_csc.header.len = sizeof(uvcb_csc);
+ uvcb_csc.header.cmd = UVC_CMD_CREATE_SEC_CPU;
+ uvcb_csc.guest_handle = uvcb_cgc.guest_handle;
+ uvcb_csc.stor_origin = (unsigned long)memalign(PAGE_SIZE, uvcb_qui.cpu_stor_len);
+ uvcb_csc.state_origin = (unsigned long)memalign(PAGE_SIZE, PAGE_SIZE);
+
+ uvcb_csc.header.len -= 8;
+ rc = uv_call(0, (uint64_t)&uvcb_csc);
+ report(uvcb_csc.header.rc == UVC_RC_INV_LEN && rc == 1 &&
+ !uvcb_csc.cpu_handle, "hdr invalid length");
+ uvcb_csc.header.len += 8;
+
+ uvcb_csc.guest_handle += 1;
+ rc = uv_call(0, (uint64_t)&uvcb_csc);
+ report(uvcb_csc.header.rc == UVC_RC_INV_GHANDLE && rc == 1,
+ "invalid guest handle");
+ uvcb_csc.guest_handle -= 1;
+
+ uvcb_csc.num = uvcb_qui.max_guest_cpus + 1;
+ rc = uv_call(0, (uint64_t)&uvcb_csc);
+ report(uvcb_csc.header.rc == 0x103 && rc == 1,
+ "invalid cpu #");
+ uvcb_csc.num = 0;
+
+ tmp = uvcb_csc.stor_origin;
+ uvcb_csc.stor_origin = get_max_ram_size() + PAGE_SIZE;
+ rc = uv_call(0, (uint64_t)&uvcb_csc);
+ report(uvcb_csc.header.rc == 0x105 && rc == 1,
+ "cpu stor inaccessible");
+ uvcb_csc.stor_origin = tmp;
+
+ tmp = uvcb_csc.stor_origin;
+ uvcb_csc.stor_origin = 0;
+ rc = uv_call(0, (uint64_t)&uvcb_csc);
+ report(uvcb_csc.header.rc == 0x106 && rc == 1,
+ "cpu stor in lowcore");
+ uvcb_csc.stor_origin = tmp;
+
+ tmp = uvcb_csc.state_origin;
+ uvcb_csc.state_origin = get_max_ram_size() + PAGE_SIZE;
+ rc = uv_call(0, (uint64_t)&uvcb_csc);
+ report(uvcb_csc.header.rc == 0x107 && rc == 1,
+ "SIE SD inaccessible");
+ uvcb_csc.state_origin = tmp;
+
+ rc = uv_call(0, (uint64_t)&uvcb_csc);
+ report(rc == 0 && uvcb_csc.header.rc == UVC_RC_EXECUTED &&
+ uvcb_csc.cpu_handle, "success");
+
+ tmp = uvcb_csc.stor_origin;
+ uvcb_csc.stor_origin = (unsigned long)memalign(PAGE_SIZE, uvcb_qui.cpu_stor_len);
+ rc = uv_call(0, (uint64_t)&uvcb_csc);
+ report(rc == 1 && uvcb_csc.header.rc == 0x104, "already defined");
+ uvcb_csc.stor_origin = tmp;
+ report_prefix_pop();
+}
+
+static void test_config_create(void)
+{
+ int rc;
+ unsigned long vsize, tmp;
+ static struct uv_cb_cgc uvcb;
+
+ uvcb_cgc.header.cmd = UVC_CMD_CREATE_SEC_CONF;
+ uvcb_cgc.header.len = sizeof(uvcb_cgc);
+ report_prefix_push("cgc");
+
+ uvcb_cgc.guest_stor_origin = 0;
+ uvcb_cgc.guest_stor_len = 42 * (1UL << 20);
+ vsize = uvcb_qui.conf_base_virt_stor_len +
+ ((uvcb_cgc.guest_stor_len / (1UL << 20)) * uvcb_qui.conf_virt_var_stor_len);
+
+ uvcb_cgc.conf_base_stor_origin = (uint64_t)memalign(PAGE_SIZE * 4, uvcb_qui.conf_base_phys_stor_len);
+ uvcb_cgc.conf_var_stor_origin = (uint64_t)memalign(PAGE_SIZE, vsize);
+ uvcb_cgc.guest_asce = (uint64_t)memalign(PAGE_SIZE, 4 * PAGE_SIZE) | ASCE_DT_SEGMENT | REGION_TABLE_LENGTH | ASCE_P;
+ uvcb_cgc.guest_sca = (uint64_t)memalign(PAGE_SIZE * 4, PAGE_SIZE * 4);
+
+ uvcb_cgc.header.len -= 8;
+ rc = uv_call(0, (uint64_t)&uvcb_cgc);
+ report(uvcb_cgc.header.rc == UVC_RC_INV_LEN && rc == 1 &&
+ !uvcb_cgc.guest_handle, "hdr invalid length");
+ uvcb_cgc.header.len += 8;
+
+ uvcb_cgc.guest_stor_origin = uvcb_qui.max_guest_stor_addr + (1UL << 20) * 2 + 1;
+ rc = uv_call(0, (uint64_t)&uvcb_cgc);
+ report(uvcb_cgc.header.rc == 0x101 && rc == 1,
+ "MSO > max guest addr");
+ uvcb_cgc.guest_stor_origin = 0;
+
+ uvcb_cgc.guest_stor_origin = uvcb_qui.max_guest_stor_addr - (1UL << 20);
+ rc = uv_call(0, (uint64_t)&uvcb_cgc);
+ report(uvcb_cgc.header.rc == 0x102 && rc == 1,
+ "MSO + MSL > max guest addr");
+ uvcb_cgc.guest_stor_origin = 0;
+
+ uvcb_cgc.guest_asce &= ~ASCE_P;
+ rc = uv_call(0, (uint64_t)&uvcb_cgc);
+ report(uvcb_cgc.header.rc == 0x105 && rc == 1,
+ "ASCE private bit missing");
+ uvcb_cgc.guest_asce |= ASCE_P;
+
+ uvcb_cgc.guest_asce |= 0x20;
+ rc = uv_call(0, (uint64_t)&uvcb_cgc);
+ report(uvcb_cgc.header.rc == 0x105 && rc == 1,
+ "ASCE bit 58 set");
+ uvcb_cgc.guest_asce &= ~0x20;
+
+ tmp = uvcb_cgc.conf_base_stor_origin;
+ uvcb_cgc.conf_base_stor_origin = get_max_ram_size() + 8;
+ rc = uv_call(0, (uint64_t)&uvcb_cgc);
+ report(uvcb_cgc.header.rc == 0x108 && rc == 1,
+ "base storage origin > available memory");
+ uvcb_cgc.conf_base_stor_origin = tmp;
+
+ tmp = uvcb_cgc.conf_base_stor_origin;
+ uvcb_cgc.conf_base_stor_origin = 0x1000;
+ rc = uv_call(0, (uint64_t)&uvcb_cgc);
+ report(uvcb_cgc.header.rc == 0x109 && rc == 1,
+ "base storage origin contains lowcore");
+ uvcb_cgc.conf_base_stor_origin = tmp;
+
+ if (smp_query_num_cpus() == 1) {
+ sigp_retry(1, SIGP_SET_PREFIX,
+ uvcb_cgc.conf_var_stor_origin + PAGE_SIZE, NULL);
+ rc = uv_call(0, (uint64_t)&uvcb_cgc);
+ report(uvcb_cgc.header.rc == 0x10e && rc == 1 &&
+ !uvcb_cgc.guest_handle, "variable storage area contains lowcore");
+ sigp_retry(1, SIGP_SET_PREFIX, 0x0, NULL);
+ }
+
+ tmp = uvcb_cgc.guest_sca;
+ uvcb_cgc.guest_sca = 0;
+ rc = uv_call(0, (uint64_t)&uvcb_cgc);
+ report(uvcb_cgc.header.rc == 0x10c && rc == 1,
+ "sca == 0");
+ uvcb_cgc.guest_sca = tmp;
+
+ tmp = uvcb_cgc.guest_sca;
+ uvcb_cgc.guest_sca = get_max_ram_size() + + PAGE_SIZE * 4;
+ rc = uv_call(0, (uint64_t)&uvcb_cgc);
+ report(uvcb_cgc.header.rc == 0x10d && rc == 1,
+ "sca inaccessible");
+ uvcb_cgc.guest_sca = tmp;
+
+ rc = uv_call(0, (uint64_t)&uvcb_cgc);
+ report(rc == 0 && uvcb_cgc.header.rc == UVC_RC_EXECUTED, "successful");
+
+ uvcb_cgc.header.rc = 0;
+ uvcb_cgc.header.rrc = 0;
+ tmp = uvcb_cgc.guest_handle;
+ uvcb_cgc.guest_handle = 0;
+ rc = uv_call(0, (uint64_t)&uvcb_cgc);
+ report(uvcb_cgc.header.rc >= 0x100 && rc == 1, "reuse uvcb");
+ uvcb_cgc.guest_handle = tmp;
+
+ /* Copy over most data from uvcb_cgc, so we have the ASCE that was used. */
+ memcpy(&uvcb, &uvcb_cgc, sizeof(uvcb));
+
+ /* Reset the header and handle */
+ uvcb.header.rc = 0;
+ uvcb.header.rrc = 0;
+ uvcb.guest_handle = 0;
+
+ /* Use new storage areas. */
+ uvcb.conf_base_stor_origin = (uint64_t)memalign(PAGE_SIZE * 4, uvcb_qui.conf_base_phys_stor_len);
+ uvcb.conf_var_stor_origin = (uint64_t)memalign(PAGE_SIZE, vsize);
+
+ rc = uv_call(0, (uint64_t)&uvcb);
+ report(uvcb.header.rc >= 0x104 && rc == 1 && !uvcb.guest_handle,
+ "reuse ASCE");
+ free((void *)uvcb.conf_base_stor_origin);
+ free((void *)uvcb.conf_var_stor_origin);
+
+ /* Missing: 106, 10a, a0b */
+ report_prefix_pop();
+}
+
+static void test_init(void)
+{
+ int rc;
+ uint64_t mem;
+ struct psw psw;
+
+ /* Donated storage needs to be over 2GB */
+ mem = (uint64_t)memalign(1UL << 31, uvcb_qui.uv_base_stor_len);
+
+ uvcb_init.header.len = sizeof(uvcb_init);
+ uvcb_init.header.cmd = UVC_CMD_INIT_UV;
+ uvcb_init.stor_origin = mem;
+ uvcb_init.stor_len = uvcb_qui.uv_base_stor_len;
+
+ report_prefix_push("init");
+ uvcb_init.header.len -= 8;
+ rc = uv_call(0, (uint64_t)&uvcb_init);
+ report(rc == 1 && uvcb_init.header.rc == UVC_RC_INV_LEN,
+ "hdr invalid length");
+ uvcb_init.header.len += 8;
+
+ uvcb_init.stor_len -= 8;
+ rc = uv_call(0, (uint64_t)&uvcb_init);
+ report(rc == 1 && uvcb_init.header.rc == 0x103,
+ "storage invalid length");
+ uvcb_init.stor_len += 8;
+
+ uvcb_init.stor_origin = get_max_ram_size() + 8;
+ rc = uv_call(0, (uint64_t)&uvcb_init);
+ report(rc == 1 && uvcb_init.header.rc == 0x104,
+ "storage origin invalid");
+ uvcb_init.stor_origin = mem;
+
+ uvcb_init.stor_origin = get_max_ram_size() - 8;
+ rc = uv_call(0, (uint64_t)&uvcb_init);
+ report(rc == 1 && uvcb_init.header.rc == 0x105,
+ "storage + length invalid");
+ uvcb_init.stor_origin = mem;
+
+ uvcb_init.stor_origin = 1UL << 30;
+ rc = uv_call(0, (uint64_t)&uvcb_init);
+ report(rc == 1 && uvcb_init.header.rc == 0x108,
+ "storage below 2GB");
+ uvcb_init.stor_origin = mem;
+
+ psw.mask = extract_psw_mask();
+ psw.addr = (unsigned long)cpu_loop;
+ smp_cpu_setup(1, psw);
+ rc = uv_call(0, (uint64_t)&uvcb_init);
+ report(rc == 1 && uvcb_init.header.rc == 0x102,
+ "too many running cpus");
+ smp_cpu_stop(1);
+
+ rc = uv_call(0, (uint64_t)&uvcb_init);
+ report(rc == 0 && uvcb_init.header.rc == UVC_RC_EXECUTED, "successful");
+
+ mem = (uint64_t)memalign(1UL << 31, uvcb_qui.uv_base_stor_len);
+ rc = uv_call(0, (uint64_t)&uvcb_init);
+ report(rc == 1 && uvcb_init.header.rc == 0x101, "double init");
+ free((void *)mem);
+
+ report_prefix_pop();
+}
+
+static void test_query(void)
+{
+ int i = 0;
+
+ uvcb_qui.header.cmd = UVC_CMD_QUI;
+ uvcb_qui.header.len = sizeof(uvcb_qui);
+
+ report_prefix_push("query");
+ uvcb_qui.header.len = 0xa0;
+ uv_call(0, (uint64_t)&uvcb_qui);
+ report(uvcb_qui.header.rc == UVC_RC_INV_LEN, "length");
+
+ uvcb_qui.header.len = 0xa8;
+ uv_call(0, (uint64_t)&uvcb_qui);
+ report(uvcb_qui.header.rc == 0x100, "insf length");
+
+ uvcb_qui.header.len = sizeof(uvcb_qui);
+ uv_call(0, (uint64_t)&uvcb_qui);
+ report(uvcb_qui.header.rc == UVC_RC_EXECUTED, "successful query");
+
+ for (i = 0; cmds[i].name; i++)
+ report(uv_query_test_call(cmds[i].call_bit), "%s", cmds[i].name);
+
+ report_prefix_pop();
+}
+
+static struct cmd_list invalid_cmds[] = {
+ { "bogus", 0x4242, sizeof(struct uv_cb_header), -1},
+ { "share", UVC_CMD_SET_SHARED_ACCESS, sizeof(struct uv_cb_share), BIT_UVC_CMD_SET_SHARED_ACCESS },
+ { "unshare", UVC_CMD_REMOVE_SHARED_ACCESS, sizeof(struct uv_cb_share), BIT_UVC_CMD_REMOVE_SHARED_ACCESS },
+ { NULL, 0, 0 },
+};
+
+static void test_invalid(void)
+{
+ struct uv_cb_header hdr = {};
+ int i, cc;
+
+ report_prefix_push("invalid");
+ for (i = 0; invalid_cmds[i].name; i++) {
+ hdr.cmd = invalid_cmds[i].cmd;
+ hdr.len = invalid_cmds[i].len;
+ cc = uv_call(0, (uint64_t)&hdr);
+ report(cc == 1 && hdr.rc == UVC_RC_INV_CMD &&
+ invalid_cmds[i].call_bit == -1 ? true : !uv_query_test_call(invalid_cmds[i].call_bit),
+ "%s", invalid_cmds[i].name);
+ }
+ report_prefix_pop();
+}
+
+static void test_clear(void)
+{
+ uint64_t *tmp = (void *)uvcb_init.stor_origin;
+
+ diag308_load_reset(1);
+ sclp_console_setup();
+ report(!*tmp, "memory cleared after reset 1");
+}
+
+static void setup_vmem(void)
+{
+ uint64_t asce, mask;
+
+ setup_mmu(get_max_ram_size());
+ asce = stctg(1);
+ lctlg(13, asce);
+ mask = extract_psw_mask() | 0x0000C00000000000UL;
+ load_psw_mask(mask);
+}
+
+int main(void)
+{
+ bool has_uvc = test_facility(158);
+
+ report_prefix_push("uvc");
+ if (!has_uvc) {
+ report_skip("Ultravisor call facility is not available");
+ goto done;
+ }
+
+ test_priv();
+ test_invalid();
+ test_query();
+ test_init();
+
+ setup_vmem();
+ test_config_create();
+ test_cpu_create();
+ test_cpu_destroy();
+ test_config_destroy();
+ test_clear();
+
+done:
+ return report_summary();
+}
--
2.30.2
next prev parent reply other threads:[~2021-05-10 13:57 UTC|newest]
Thread overview: 14+ messages / expand[flat|nested] mbox.gz Atom feed top
2021-05-10 13:51 [kvm-unit-tests PATCH v2 0/6] s390x: uv: Extend guest test and add host test Janosch Frank
2021-05-10 13:51 ` [kvm-unit-tests PATCH v2 1/6] s390x: uv-guest: Add invalid share location test Janosch Frank
2021-05-10 13:51 ` [kvm-unit-tests PATCH v2 2/6] s390x: Add more Ultravisor command structure definitions Janosch Frank
2021-05-11 16:02 ` Cornelia Huck
2021-05-10 13:51 ` [kvm-unit-tests PATCH v2 3/6] s390x: uv: Add UV lib Janosch Frank
2021-05-11 16:13 ` Cornelia Huck
2021-05-18 13:10 ` Janosch Frank
2021-05-10 13:51 ` [kvm-unit-tests PATCH v2 4/6] s390x: Test for share/unshare call support before using them Janosch Frank
2021-05-11 16:15 ` Cornelia Huck
2021-05-10 13:51 ` [kvm-unit-tests PATCH v2 5/6] s390x: uv-guest: Test invalid commands Janosch Frank
2021-05-11 16:22 ` Cornelia Huck
2021-05-11 17:12 ` Claudio Imbrenda
2021-05-10 13:51 ` Janosch Frank [this message]
2021-05-11 16:25 ` [kvm-unit-tests PATCH v2 6/6] s390x: Add UV host test Cornelia Huck
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20210510135148.1904-7-frankja@linux.ibm.com \
--to=frankja@linux.ibm.com \
--cc=cohuck@redhat.com \
--cc=david@redhat.com \
--cc=imbrenda@linux.ibm.com \
--cc=kvm@vger.kernel.org \
--cc=linux-s390@vger.kernel.org \
--cc=thuth@redhat.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).