linux-efi.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Tom Lendacky <thomas.lendacky@amd.com>
To: linux-arch@vger.kernel.org, linux-efi@vger.kernel.org,
	kvm@vger.kernel.org, linux-doc@vger.kernel.org, x86@kernel.org,
	kexec@lists.infradead.org, linux-kernel@vger.kernel.org,
	kasan-dev@googlegroups.com, linux-mm@kvack.org,
	iommu@lists.linux-foundation.org
Cc: "Rik van Riel" <riel@redhat.com>,
	"Radim Krčmář" <rkrcmar@redhat.com>,
	"Toshimitsu Kani" <toshi.kani@hpe.com>,
	"Arnd Bergmann" <arnd@arndb.de>,
	"Jonathan Corbet" <corbet@lwn.net>,
	"Matt Fleming" <matt@codeblueprint.co.uk>,
	"Michael S. Tsirkin" <mst@redhat.com>,
	"Joerg Roedel" <joro@8bytes.org>,
	"Konrad Rzeszutek Wilk" <konrad.wilk@oracle.com>,
	"Paolo Bonzini" <pbonzini@redhat.com>,
	"Larry Woodman" <lwoodman@redhat.com>,
	"Brijesh Singh" <brijesh.singh@amd.com>,
	"Ingo Molnar" <mingo@redhat.com>,
	"Borislav Petkov" <bp@alien8.de>,
	"Andy Lutomirski" <luto@kernel.org>,
	"H. Peter Anvin" <hpa@zytor.com>,
	"Andrey Ryabinin" <aryabinin@virtuozzo.com>,
	"Alexander Potapenko" <glider@google.com>,
	"Dave Young" <dyoung@redhat.com>,
	"Thomas Gleixner" <tglx@linutronix.de>,
	"Dmitry Vyukov" <dvyukov@google.com>
Subject: [PATCH v5 29/32] x86/mm: Add support to encrypt the kernel in-place
Date: Tue, 18 Apr 2017 16:21:49 -0500	[thread overview]
Message-ID: <20170418212149.10190.70894.stgit@tlendack-t1.amdoffice.net> (raw)
In-Reply-To: <20170418211612.10190.82788.stgit@tlendack-t1.amdoffice.net>

Add the support to encrypt the kernel in-place. This is done by creating
new page mappings for the kernel - a decrypted write-protected mapping
and an encrypted mapping. The kernel is encrypted by copying it through
a temporary buffer.

Signed-off-by: Tom Lendacky <thomas.lendacky@amd.com>
---
 arch/x86/include/asm/mem_encrypt.h |    6 +
 arch/x86/mm/Makefile               |    2 
 arch/x86/mm/mem_encrypt.c          |  262 ++++++++++++++++++++++++++++++++++++
 arch/x86/mm/mem_encrypt_boot.S     |  151 +++++++++++++++++++++
 4 files changed, 421 insertions(+)
 create mode 100644 arch/x86/mm/mem_encrypt_boot.S

diff --git a/arch/x86/include/asm/mem_encrypt.h b/arch/x86/include/asm/mem_encrypt.h
index b406df2..8f6f9b4 100644
--- a/arch/x86/include/asm/mem_encrypt.h
+++ b/arch/x86/include/asm/mem_encrypt.h
@@ -31,6 +31,12 @@ static inline u64 sme_dma_mask(void)
 	return ((u64)sme_me_mask << 1) - 1;
 }
 
+void sme_encrypt_execute(unsigned long encrypted_kernel_vaddr,
+			 unsigned long decrypted_kernel_vaddr,
+			 unsigned long kernel_len,
+			 unsigned long encryption_wa,
+			 unsigned long encryption_pgd);
+
 void __init sme_early_encrypt(resource_size_t paddr,
 			      unsigned long size);
 void __init sme_early_decrypt(resource_size_t paddr,
diff --git a/arch/x86/mm/Makefile b/arch/x86/mm/Makefile
index 9e13841..0633142 100644
--- a/arch/x86/mm/Makefile
+++ b/arch/x86/mm/Makefile
@@ -38,3 +38,5 @@ obj-$(CONFIG_NUMA_EMU)		+= numa_emulation.o
 obj-$(CONFIG_X86_INTEL_MPX)	+= mpx.o
 obj-$(CONFIG_X86_INTEL_MEMORY_PROTECTION_KEYS) += pkeys.o
 obj-$(CONFIG_RANDOMIZE_MEMORY) += kaslr.o
+
+obj-$(CONFIG_AMD_MEM_ENCRYPT)	+= mem_encrypt_boot.o
diff --git a/arch/x86/mm/mem_encrypt.c b/arch/x86/mm/mem_encrypt.c
index 30b07a3..0ff41a4 100644
--- a/arch/x86/mm/mem_encrypt.c
+++ b/arch/x86/mm/mem_encrypt.c
@@ -24,6 +24,7 @@
 #include <asm/setup.h>
 #include <asm/bootparam.h>
 #include <asm/cacheflush.h>
+#include <asm/sections.h>
 
 /*
  * Since SME related variables are set early in the boot process they must
@@ -216,8 +217,269 @@ void swiotlb_set_mem_attributes(void *vaddr, unsigned long size)
 	set_memory_decrypted((unsigned long)vaddr, size >> PAGE_SHIFT);
 }
 
+void __init sme_clear_pgd(pgd_t *pgd_base, unsigned long start,
+			  unsigned long end)
+{
+	unsigned long addr = start;
+	pgdval_t *pgd_p;
+
+	while (addr < end) {
+		unsigned long pgd_end;
+
+		pgd_end = (addr & PGDIR_MASK) + PGDIR_SIZE;
+		if (pgd_end > end)
+			pgd_end = end;
+
+		pgd_p = (pgdval_t *)pgd_base + pgd_index(addr);
+		*pgd_p = 0;
+
+		addr = pgd_end;
+	}
+}
+
+#define PGD_FLAGS	_KERNPG_TABLE_NOENC
+#define PUD_FLAGS	_KERNPG_TABLE_NOENC
+#define PMD_FLAGS	(__PAGE_KERNEL_LARGE_EXEC & ~_PAGE_GLOBAL)
+
+static void __init *sme_populate_pgd(pgd_t *pgd_base, void *pgtable_area,
+				     unsigned long vaddr, pmdval_t pmd_val)
+{
+	pgdval_t pgd, *pgd_p;
+	pudval_t pud, *pud_p;
+	pmdval_t pmd, *pmd_p;
+
+	pgd_p = (pgdval_t *)pgd_base + pgd_index(vaddr);
+	pgd = *pgd_p;
+	if (pgd) {
+		pud_p = (pudval_t *)(pgd & ~PTE_FLAGS_MASK);
+	} else {
+		pud_p = pgtable_area;
+		memset(pud_p, 0, sizeof(*pud_p) * PTRS_PER_PUD);
+		pgtable_area += sizeof(*pud_p) * PTRS_PER_PUD;
+
+		*pgd_p = (pgdval_t)pud_p + PGD_FLAGS;
+	}
+
+	pud_p += pud_index(vaddr);
+	pud = *pud_p;
+	if (pud) {
+		if (pud & _PAGE_PSE)
+			goto out;
+
+		pmd_p = (pmdval_t *)(pud & ~PTE_FLAGS_MASK);
+	} else {
+		pmd_p = pgtable_area;
+		memset(pmd_p, 0, sizeof(*pmd_p) * PTRS_PER_PMD);
+		pgtable_area += sizeof(*pmd_p) * PTRS_PER_PMD;
+
+		*pud_p = (pudval_t)pmd_p + PUD_FLAGS;
+	}
+
+	pmd_p += pmd_index(vaddr);
+	pmd = *pmd_p;
+	if (!pmd || !(pmd & _PAGE_PSE))
+		*pmd_p = pmd_val;
+
+out:
+	return pgtable_area;
+}
+
+static unsigned long __init sme_pgtable_calc(unsigned long len)
+{
+	unsigned long pud_tables, pmd_tables;
+	unsigned long total = 0;
+
+	/*
+	 * Perform a relatively simplistic calculation of the pagetable
+	 * entries that are needed. That mappings will be covered by 2MB
+	 * PMD entries so we can conservatively calculate the required
+	 * number of PUD and PMD structures needed to perform the mappings.
+	 * Incrementing the count for each covers the case where the
+	 * addresses cross entries.
+	 */
+	pud_tables = ALIGN(len, PGDIR_SIZE) / PGDIR_SIZE;
+	pud_tables++;
+	pmd_tables = ALIGN(len, PUD_SIZE) / PUD_SIZE;
+	pmd_tables++;
+
+	total += pud_tables * sizeof(pud_t) * PTRS_PER_PUD;
+	total += pmd_tables * sizeof(pmd_t) * PTRS_PER_PMD;
+
+	/*
+	 * Now calculate the added pagetable structures needed to populate
+	 * the new pagetables.
+	 */
+	pud_tables = ALIGN(total, PGDIR_SIZE) / PGDIR_SIZE;
+	pmd_tables = ALIGN(total, PUD_SIZE) / PUD_SIZE;
+
+	total += pud_tables * sizeof(pud_t) * PTRS_PER_PUD;
+	total += pmd_tables * sizeof(pmd_t) * PTRS_PER_PMD;
+
+	return total;
+}
+
 void __init sme_encrypt_kernel(void)
 {
+	pgd_t *pgd;
+	void *pgtable_area;
+	unsigned long kernel_start, kernel_end, kernel_len;
+	unsigned long workarea_start, workarea_end, workarea_len;
+	unsigned long execute_start, execute_end, execute_len;
+	unsigned long pgtable_area_len;
+	unsigned long decrypted_base;
+	unsigned long paddr, pmd_flags;
+
+	if (!sme_active())
+		return;
+
+	/*
+	 * Prepare for encrypting the kernel by building new pagetables with
+	 * the necessary attributes needed to encrypt the kernel in place.
+	 *
+	 *   One range of virtual addresses will map the memory occupied
+	 *   by the kernel as encrypted.
+	 *
+	 *   Another range of virtual addresses will map the memory occupied
+	 *   by the kernel as decrypted and write-protected.
+	 *
+	 *     The use of write-protect attribute will prevent any of the
+	 *     memory from being cached.
+	 */
+
+	/* Physical addresses gives us the identity mapped virtual addresses */
+	kernel_start = __pa_symbol(_text);
+	kernel_end = ALIGN(__pa_symbol(_end), PMD_PAGE_SIZE);
+	kernel_len = kernel_end - kernel_start;
+
+	/* Set the encryption workarea to be immediately after the kernel */
+	workarea_start = kernel_end;
+
+	/*
+	 * Calculate required number of workarea bytes needed:
+	 *   executable encryption area size:
+	 *     stack page (PAGE_SIZE)
+	 *     encryption routine page (PAGE_SIZE)
+	 *     intermediate copy buffer (PMD_PAGE_SIZE)
+	 *   pagetable structures for the encryption of the kernel
+	 *   pagetable structures for workarea (in case not currently mapped)
+	 */
+	execute_start = workarea_start;
+	execute_end = execute_start + (PAGE_SIZE * 2) + PMD_PAGE_SIZE;
+	execute_len = execute_end - execute_start;
+
+	/*
+	 * One PGD for both encrypted and decrypted mappings and a set of
+	 * PUDs and PMDs for each of the encrypted and decrypted mappings.
+	 */
+	pgtable_area_len = sizeof(pgd_t) * PTRS_PER_PGD;
+	pgtable_area_len += sme_pgtable_calc(execute_end - kernel_start) * 2;
+
+	/* PUDs and PMDs needed in the current pagetables for the workarea */
+	pgtable_area_len += sme_pgtable_calc(execute_len + pgtable_area_len);
+
+	/*
+	 * The total workarea includes the executable encryption area and
+	 * the pagetable area.
+	 */
+	workarea_len = execute_len + pgtable_area_len;
+	workarea_end = workarea_start + workarea_len;
+
+	/*
+	 * Set the address to the start of where newly created pagetable
+	 * structures (PGDs, PUDs and PMDs) will be allocated. New pagetable
+	 * structures are created when the workarea is added to the current
+	 * pagetables and when the new encrypted and decrypted kernel
+	 * mappings are populated.
+	 */
+	pgtable_area = (void *)execute_end;
+
+	/*
+	 * Make sure the current pagetable structure has entries for
+	 * addressing the workarea.
+	 */
+	pgd = (pgd_t *)native_read_cr3();
+	paddr = workarea_start;
+	while (paddr < workarea_end) {
+		pgtable_area = sme_populate_pgd(pgd, pgtable_area,
+						paddr,
+						paddr + PMD_FLAGS);
+
+		paddr += PMD_PAGE_SIZE;
+	}
+	native_write_cr3((unsigned long)pgd);
+
+	/*
+	 * A new pagetable structure is being built to allow for the kernel
+	 * to be encrypted. It starts with an empty PGD that will then be
+	 * populated with new PUDs and PMDs as the encrypted and decrypted
+	 * kernel mappings are created.
+	 */
+	pgd = pgtable_area;
+	memset(pgd, 0, sizeof(*pgd) * PTRS_PER_PGD);
+	pgtable_area += sizeof(*pgd) * PTRS_PER_PGD;
+
+	/* Add encrypted kernel (identity) mappings */
+	pmd_flags = PMD_FLAGS | _PAGE_ENC;
+	paddr = kernel_start;
+	while (paddr < kernel_end) {
+		pgtable_area = sme_populate_pgd(pgd, pgtable_area,
+						paddr,
+						paddr + pmd_flags);
+
+		paddr += PMD_PAGE_SIZE;
+	}
+
+	/*
+	 * A different PGD index/entry must be used to get different
+	 * pagetable entries for the decrypted mapping. Choose the next
+	 * PGD index and convert it to a virtual address to be used as
+	 * the base of the mapping.
+	 */
+	decrypted_base = (pgd_index(workarea_end) + 1) & (PTRS_PER_PGD - 1);
+	decrypted_base <<= PGDIR_SHIFT;
+
+	/* Add decrypted, write-protected kernel (non-identity) mappings */
+	pmd_flags = (PMD_FLAGS & ~_PAGE_CACHE_MASK) | (_PAGE_PAT | _PAGE_PWT);
+	paddr = kernel_start;
+	while (paddr < kernel_end) {
+		pgtable_area = sme_populate_pgd(pgd, pgtable_area,
+						paddr + decrypted_base,
+						paddr + pmd_flags);
+
+		paddr += PMD_PAGE_SIZE;
+	}
+
+	/* Add decrypted workarea mappings to both kernel mappings */
+	paddr = workarea_start;
+	while (paddr < workarea_end) {
+		pgtable_area = sme_populate_pgd(pgd, pgtable_area,
+						paddr,
+						paddr + PMD_FLAGS);
+
+		pgtable_area = sme_populate_pgd(pgd, pgtable_area,
+						paddr + decrypted_base,
+						paddr + PMD_FLAGS);
+
+		paddr += PMD_PAGE_SIZE;
+	}
+
+	/* Perform the encryption */
+	sme_encrypt_execute(kernel_start, kernel_start + decrypted_base,
+			    kernel_len, workarea_start, (unsigned long)pgd);
+
+	/*
+	 * At this point we are running encrypted.  Remove the mappings for
+	 * the decrypted areas - all that is needed for this is to remove
+	 * the PGD entry/entries.
+	 */
+	sme_clear_pgd(pgd, kernel_start + decrypted_base,
+		      kernel_end + decrypted_base);
+
+	sme_clear_pgd(pgd, workarea_start + decrypted_base,
+		      workarea_end + decrypted_base);
+
+	/* Flush the TLB - no globals so cr3 is enough */
+	native_write_cr3(native_read_cr3());
 }
 
 unsigned long __init sme_enable(void)
diff --git a/arch/x86/mm/mem_encrypt_boot.S b/arch/x86/mm/mem_encrypt_boot.S
new file mode 100644
index 0000000..fb58f9f
--- /dev/null
+++ b/arch/x86/mm/mem_encrypt_boot.S
@@ -0,0 +1,151 @@
+/*
+ * AMD Memory Encryption Support
+ *
+ * Copyright (C) 2016 Advanced Micro Devices, Inc.
+ *
+ * Author: Tom Lendacky <thomas.lendacky@amd.com>
+ *
+ * This program is free software; you can redistribute it and/or modify
+ * it under the terms of the GNU General Public License version 2 as
+ * published by the Free Software Foundation.
+ */
+
+#include <linux/linkage.h>
+#include <asm/pgtable.h>
+#include <asm/page.h>
+#include <asm/processor-flags.h>
+#include <asm/msr-index.h>
+
+	.text
+	.code64
+ENTRY(sme_encrypt_execute)
+
+	/*
+	 * Entry parameters:
+	 *   RDI - virtual address for the encrypted kernel mapping
+	 *   RSI - virtual address for the decrypted kernel mapping
+	 *   RDX - length of kernel
+	 *   RCX - virtual address of the encryption workarea, including:
+	 *     - stack page (PAGE_SIZE)
+	 *     - encryption routine page (PAGE_SIZE)
+	 *     - intermediate copy buffer (PMD_PAGE_SIZE)
+	 *    R8 - physcial address of the pagetables to use for encryption
+	 */
+
+	push	%rbp
+	push	%r12
+
+	/* Set up a one page stack in the non-encrypted memory area */
+	movq	%rsp, %rbp		/* Save current stack pointer */
+	movq	%rcx, %rax		/* Workarea stack page */
+	movq	%rax, %rsp		/* Set new stack pointer */
+	addq	$PAGE_SIZE, %rsp	/* Stack grows from the bottom */
+	addq	$PAGE_SIZE, %rax	/* Workarea encryption routine */
+
+	movq	%rdi, %r10		/* Encrypted kernel */
+	movq	%rsi, %r11		/* Decrypted kernel */
+	movq	%rdx, %r12		/* Kernel length */
+
+	/* Copy encryption routine into the workarea */
+	movq	%rax, %rdi		/* Workarea encryption routine */
+	leaq	.Lenc_start(%rip), %rsi	/* Encryption routine */
+	movq	$(.Lenc_stop - .Lenc_start), %rcx	/* Encryption routine length */
+	rep	movsb
+
+	/* Setup registers for call */
+	movq	%r10, %rdi		/* Encrypted kernel */
+	movq	%r11, %rsi		/* Decrypted kernel */
+	movq	%r8, %rdx		/* Pagetables used for encryption */
+	movq	%r12, %rcx		/* Kernel length */
+	movq	%rax, %r8		/* Workarea encryption routine */
+	addq	$PAGE_SIZE, %r8		/* Workarea intermediate copy buffer */
+
+	call	*%rax			/* Call the encryption routine */
+
+	movq	%rbp, %rsp		/* Restore original stack pointer */
+
+	pop	%r12
+	pop	%rbp
+
+	ret
+ENDPROC(sme_encrypt_execute)
+
+.Lenc_start:
+ENTRY(sme_enc_routine)
+/*
+ * Routine used to encrypt kernel.
+ *   This routine must be run outside of the kernel proper since
+ *   the kernel will be encrypted during the process. So this
+ *   routine is defined here and then copied to an area outside
+ *   of the kernel where it will remain and run decrypted
+ *   during execution.
+ *
+ *   On entry the registers must be:
+ *     RDI - virtual address for the encrypted kernel mapping
+ *     RSI - virtual address for the decrypted kernel mapping
+ *     RDX - address of the pagetables to use for encryption
+ *     RCX - length of kernel
+ *      R8 - intermediate copy buffer
+ *
+ *     RAX - points to this routine
+ *
+ * The kernel will be encrypted by copying from the non-encrypted
+ * kernel space to an intermediate buffer and then copying from the
+ * intermediate buffer back to the encrypted kernel space. The physical
+ * addresses of the two kernel space mappings are the same which
+ * results in the kernel being encrypted "in place".
+ */
+	/* Enable the new page tables */
+	mov	%rdx, %cr3
+
+	/* Flush any global TLBs */
+	mov	%cr4, %rdx
+	andq	$~X86_CR4_PGE, %rdx
+	mov	%rdx, %cr4
+	orq	$X86_CR4_PGE, %rdx
+	mov	%rdx, %cr4
+
+	/* Set the PAT register PA5 entry to write-protect */
+	push	%rcx
+	movl	$MSR_IA32_CR_PAT, %ecx
+	rdmsr
+	push	%rdx			/* Save original PAT value */
+	andl	$0xffff00ff, %edx	/* Clear PA5 */
+	orl	$0x00000500, %edx	/* Set PA5 to WP */
+	wrmsr
+	pop	%rdx			/* RDX contains original PAT value */
+	pop	%rcx
+
+	movq	%rcx, %r9		/* Save kernel length */
+	movq	%rdi, %r10		/* Save encrypted kernel address */
+	movq	%rsi, %r11		/* Save decrypted kernel address */
+
+	wbinvd				/* Invalidate any cache entries */
+
+	/* Copy/encrypt 2MB at a time */
+1:
+	movq	%r11, %rsi		/* Source - decrypted kernel */
+	movq	%r8, %rdi		/* Dest   - intermediate copy buffer */
+	movq	$PMD_PAGE_SIZE, %rcx	/* 2MB length */
+	rep	movsb
+
+	movq	%r8, %rsi		/* Source - intermediate copy buffer */
+	movq	%r10, %rdi		/* Dest   - encrypted kernel */
+	movq	$PMD_PAGE_SIZE, %rcx	/* 2MB length */
+	rep	movsb
+
+	addq	$PMD_PAGE_SIZE, %r11
+	addq	$PMD_PAGE_SIZE, %r10
+	subq	$PMD_PAGE_SIZE, %r9	/* Kernel length decrement */
+	jnz	1b			/* Kernel length not zero? */
+
+	/* Restore PAT register */
+	push	%rdx			/* Save original PAT value */
+	movl	$MSR_IA32_CR_PAT, %ecx
+	rdmsr
+	pop	%rdx			/* Restore original PAT value */
+	wrmsr
+
+	ret
+ENDPROC(sme_enc_routine)
+.Lenc_stop:

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

  parent reply	other threads:[~2017-04-18 21:21 UTC|newest]

Thread overview: 126+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2017-04-18 21:16 [PATCH v5 00/32] x86: Secure Memory Encryption (AMD) Tom Lendacky
2017-04-18 21:16 ` [PATCH v5 01/32] x86: Documentation for AMD Secure Memory Encryption (SME) Tom Lendacky
     [not found]   ` <20170418211625.10190.52568.stgit-qCXWGYdRb2BnqfbPTmsdiZQ+2ll4COg0XqFh9Ls21Oc@public.gmane.org>
2017-04-19  9:02     ` Borislav Petkov
2017-04-19 14:23       ` Tom Lendacky
2017-04-19 15:38         ` Borislav Petkov
     [not found]   ` <20170419090224.frmv2jhwfwoxvdie-fF5Pk5pvG8Y@public.gmane.org>
2017-04-19  9:52     ` David Howells
2017-04-18 21:16 ` [PATCH v5 02/32] x86/mm/pat: Set write-protect cache mode for full PAT support Tom Lendacky
2017-04-18 21:16 ` [PATCH v5 03/32] x86, mpparse, x86/acpi, x86/PCI, SFI: Use memremap for RAM mappings Tom Lendacky
2017-04-18 21:17 ` [PATCH v5 04/32] x86/CPU/AMD: Add the Secure Memory Encryption CPU feature Tom Lendacky
2017-04-18 21:17 ` [PATCH v5 05/32] x86/CPU/AMD: Handle SME reduction in physical address size Tom Lendacky
2017-04-20 16:59   ` Borislav Petkov
2017-04-20 17:29     ` Tom Lendacky
2017-04-20 18:52       ` Borislav Petkov
2017-04-18 21:17 ` [PATCH v5 07/32] x86/mm: Add support to enable SME in early boot processing Tom Lendacky
     [not found]   ` <20170418211735.10190.29562.stgit-qCXWGYdRb2BnqfbPTmsdiZQ+2ll4COg0XqFh9Ls21Oc@public.gmane.org>
2017-04-21 14:55     ` Borislav Petkov
     [not found]       ` <20170421145555.v3xeaijv3vjclsos-fF5Pk5pvG8Y@public.gmane.org>
2017-04-21 21:40         ` Tom Lendacky
2017-04-18 21:17 ` [PATCH v5 08/32] x86/mm: Simplify p[g4um]d_page() macros Tom Lendacky
2017-04-18 21:17 ` [PATCH v5 09/32] x86/mm: Provide general kernel support for memory encryption Tom Lendacky
2017-04-21 21:52   ` Dave Hansen
2017-04-24 15:53     ` Tom Lendacky
     [not found]       ` <9fc79e28-ad64-1c2f-4c46-a4efcdd550b0-5C7GfCeVMHo@public.gmane.org>
2017-04-24 15:57         ` Dave Hansen
     [not found]           ` <67926f62-a068-6114-92ee-39bc08488b32-ral2JQCrhuEAvxtiuMwx3w@public.gmane.org>
2017-04-24 16:10             ` Tom Lendacky
2017-04-27 16:12   ` Borislav Petkov
2017-05-04 14:34     ` Tom Lendacky
2017-05-04 17:01       ` Borislav Petkov
2017-04-18 21:18 ` [PATCH v5 10/32] x86/mm: Extend early_memremap() support with additional attrs Tom Lendacky
2017-04-18 21:18 ` [PATCH v5 11/32] x86/mm: Add support for early encrypt/decrypt of memory Tom Lendacky
2017-04-18 21:18 ` [PATCH v5 12/32] x86/mm: Insure that boot memory areas are mapped properly Tom Lendacky
     [not found]   ` <20170418211822.10190.67435.stgit-qCXWGYdRb2BnqfbPTmsdiZQ+2ll4COg0XqFh9Ls21Oc@public.gmane.org>
2017-05-04 10:16     ` Borislav Petkov
2017-05-04 14:39       ` Tom Lendacky
2017-04-18 21:18 ` [PATCH v5 14/32] efi: Add an EFI table address match function Tom Lendacky
2017-05-15 18:09   ` Borislav Petkov
2017-05-16 21:53     ` Tom Lendacky
2017-04-18 21:19 ` [PATCH v5 15/32] efi: Update efi_mem_type() to return an error rather than 0 Tom Lendacky
     [not found]   ` <20170418211900.10190.98158.stgit-qCXWGYdRb2BnqfbPTmsdiZQ+2ll4COg0XqFh9Ls21Oc@public.gmane.org>
2017-05-07 17:18     ` Borislav Petkov
2017-05-08 13:20       ` Tom Lendacky
2017-04-18 21:19 ` [PATCH v5 16/32] x86/efi: Update EFI pagetable creation to work with SME Tom Lendacky
2017-04-18 21:19 ` [PATCH v5 18/32] x86, mpparse: Use memremap to map the mpf and mpc data Tom Lendacky
2017-05-16  8:36   ` Borislav Petkov
2017-05-17 20:26     ` Tom Lendacky
     [not found]       ` <6e155b8f-b691-2ee0-8977-969aaab6199a-5C7GfCeVMHo@public.gmane.org>
2017-05-18  9:03         ` Borislav Petkov
2017-04-18 21:19 ` [PATCH v5 20/32] x86/mm: Add support for changing the memory encryption attribute Tom Lendacky
     [not found] ` <20170418211612.10190.82788.stgit-qCXWGYdRb2BnqfbPTmsdiZQ+2ll4COg0XqFh9Ls21Oc@public.gmane.org>
2017-04-18 21:17   ` [PATCH v5 06/32] x86/mm: Add Secure Memory Encryption (SME) support Tom Lendacky
2017-04-27 15:46     ` Borislav Petkov
2017-05-04 14:24       ` Tom Lendacky
     [not found]         ` <d9d9f10a-0ce5-53e8-41f5-f8690dbd7362-5C7GfCeVMHo@public.gmane.org>
2017-05-04 14:36           ` Borislav Petkov
2017-05-16 19:28             ` Tom Lendacky
2017-05-17  7:05               ` Borislav Petkov
2017-04-18 21:18   ` [PATCH v5 13/32] x86/boot/e820: Add support to determine the E820 type of an address Tom Lendacky
     [not found]     ` <20170418211831.10190.80158.stgit-qCXWGYdRb2BnqfbPTmsdiZQ+2ll4COg0XqFh9Ls21Oc@public.gmane.org>
2017-05-05 17:11       ` Borislav Petkov
2017-05-06  7:48         ` Ard Biesheuvel
2017-04-18 21:19   ` [PATCH v5 17/32] x86/mm: Add support to access boot related data in the clear Tom Lendacky
2017-05-15 18:35     ` Borislav Petkov
     [not found]       ` <20170515183517.mb4k2gp2qobbuvtm-fF5Pk5pvG8Y@public.gmane.org>
2017-05-17 18:54         ` Tom Lendacky
2017-05-18  9:02           ` Borislav Petkov
     [not found]             ` <20170518090212.kebstmnjv4h3cjf2-fF5Pk5pvG8Y@public.gmane.org>
2017-05-19 20:50               ` Tom Lendacky
2017-05-21  7:16                 ` Borislav Petkov
2017-05-30 16:46                   ` Tom Lendacky
2017-05-31 11:31                     ` Borislav Petkov
2017-05-18 19:50       ` Matt Fleming
2017-05-26 16:22         ` Tom Lendacky
     [not found]           ` <4c2ef3ba-2940-3330-d362-5b2b0d812c6f-5C7GfCeVMHo@public.gmane.org>
2017-05-26 16:35             ` Borislav Petkov
     [not found]               ` <20170526163517.nrweesvse24dszkz-fF5Pk5pvG8Y@public.gmane.org>
2017-05-30 17:47                 ` Tom Lendacky
2017-04-18 21:19   ` [PATCH v5 19/32] x86/mm: Add support to access persistent memory " Tom Lendacky
2017-05-16 14:04     ` Borislav Petkov
2017-05-19 19:52       ` Tom Lendacky
2017-04-18 21:19   ` [PATCH v5 21/32] x86, realmode: Decrypt trampoline area if memory encryption is active Tom Lendacky
2017-04-18 21:20   ` [PATCH v5 22/32] x86, swiotlb: DMA support for memory encryption Tom Lendacky
2017-05-16 14:27     ` Borislav Petkov
2017-05-19 19:54       ` Tom Lendacky
2017-04-18 21:20   ` [PATCH v5 23/32] swiotlb: Add warnings for use of bounce buffers with SME Tom Lendacky
2017-05-16 14:52     ` Borislav Petkov
2017-05-19 19:55       ` Tom Lendacky
2017-04-18 21:20   ` [PATCH v5 24/32] iommu/amd: Disable AMD IOMMU if memory encryption is active Tom Lendacky
2017-04-18 21:20   ` [PATCH v5 25/32] x86, realmode: Check for memory encryption on the APs Tom Lendacky
2017-04-18 21:20   ` [PATCH v5 26/32] x86, drm, fbdev: Do not specify encrypted memory for video mappings Tom Lendacky
2017-05-16 17:35     ` Borislav Petkov
2017-05-30 20:07       ` Tom Lendacky
2017-04-18 21:22   ` [PATCH v5 31/32] x86: Add sysfs support for Secure Memory Encryption Tom Lendacky
2017-04-21 21:55     ` Dave Hansen
2017-04-27  7:25       ` Dave Young
2017-04-27 15:52         ` Dave Hansen
     [not found]           ` <1f034974-20e6-b5e9-e6ff-434b634e1522-ral2JQCrhuEAvxtiuMwx3w@public.gmane.org>
2017-04-28  5:32             ` Dave Young
2017-05-04 14:17           ` Tom Lendacky
     [not found]         ` <20170427072547.GB15297-0VdLhd/A9Pl+NNSt+8eSiB/sF2h8X+2i0E9HWUfgJXw@public.gmane.org>
2017-05-04 14:13           ` Tom Lendacky
     [not found]     ` <20170418212212.10190.73484.stgit-qCXWGYdRb2BnqfbPTmsdiZQ+2ll4COg0XqFh9Ls21Oc@public.gmane.org>
2017-05-18 17:01       ` Borislav Petkov
     [not found]         ` <20170518170153.eqiyat5s6q3yeejl-fF5Pk5pvG8Y@public.gmane.org>
2017-05-26  2:49           ` Dave Young
2017-05-26  5:04             ` Xunlei Pang
2017-05-26 15:47               ` Tom Lendacky
2017-04-18 21:21 ` [PATCH v5 27/32] kvm: x86: svm: Enable Secure Memory Encryption within KVM Tom Lendacky
2017-04-18 21:21 ` [PATCH v5 28/32] x86/mm, kexec: Allow kexec to be used with SME Tom Lendacky
     [not found]   ` <20170418212121.10190.94885.stgit-qCXWGYdRb2BnqfbPTmsdiZQ+2ll4COg0XqFh9Ls21Oc@public.gmane.org>
2017-05-17 19:17     ` Borislav Petkov
2017-05-19 20:45       ` Tom Lendacky
     [not found]         ` <1b74e0e6-3dda-f638-461b-f73af9904360-5C7GfCeVMHo@public.gmane.org>
2017-05-19 20:58           ` Borislav Petkov
     [not found]             ` <20170519205836.3wvl3nztqyyouje3-fF5Pk5pvG8Y@public.gmane.org>
2017-05-19 21:07               ` Tom Lendacky
     [not found]                 ` <5ef96f3a-6ebd-1d4d-7ac9-05dbed45d998-5C7GfCeVMHo@public.gmane.org>
2017-05-19 21:28                   ` Borislav Petkov
2017-05-19 21:38                     ` Tom Lendacky
2017-05-26  4:17     ` Xunlei Pang
     [not found]       ` <5927AC6E.8080209-H+wXaHxf7aLQT0dZR+AlfA@public.gmane.org>
2017-05-27  2:17         ` Dave Young
2017-05-30 17:46         ` Tom Lendacky
     [not found]           ` <de4d2efc-6636-4120-98d9-7fdf4707f68d-5C7GfCeVMHo@public.gmane.org>
2017-05-31 10:01             ` Borislav Petkov
2017-05-31 15:03           ` Xunlei Pang
2017-05-31 15:48             ` Borislav Petkov
2017-04-18 21:21 ` Tom Lendacky [this message]
     [not found]   ` <20170418212149.10190.70894.stgit-qCXWGYdRb2BnqfbPTmsdiZQ+2ll4COg0XqFh9Ls21Oc@public.gmane.org>
2017-05-18 12:46     ` [PATCH v5 29/32] x86/mm: Add support to encrypt the kernel in-place Borislav Petkov
2017-05-25 22:24       ` Tom Lendacky
2017-05-26 16:25         ` Borislav Petkov
2017-05-30 16:39           ` Tom Lendacky
2017-05-31  9:51             ` Borislav Petkov
     [not found]               ` <20170531095148.pba6ju6im4qxbwfg-fF5Pk5pvG8Y@public.gmane.org>
2017-05-31 13:12                 ` Tom Lendacky
2017-04-18 21:22 ` [PATCH v5 30/32] x86/boot: Add early cmdline parsing for options with arguments Tom Lendacky
2017-04-18 21:22 ` [PATCH v5 32/32] x86/mm: Add support to make use of Secure Memory Encryption Tom Lendacky
2017-04-21 18:56   ` Tom Lendacky
2017-05-19 11:30     ` Borislav Petkov
2017-05-19 20:16       ` Josh Poimboeuf
2017-05-19 20:29         ` Borislav Petkov
2017-05-30 15:48         ` Tom Lendacky
     [not found]           ` <1ac40d18-a8b2-94eb-35ed-c30768667be8-5C7GfCeVMHo@public.gmane.org>
2017-05-31  9:15             ` Borislav Petkov
2017-05-30 15:46       ` Tom Lendacky
2017-05-19 11:27   ` Borislav Petkov
2017-05-30 14:38     ` Tom Lendacky
2017-05-30 14:55       ` Borislav Petkov
2017-05-30 15:37         ` Tom Lendacky
2017-05-31  8:49           ` Borislav Petkov
2017-05-31 13:37             ` Tom Lendacky
2017-05-31 14:12               ` Borislav Petkov

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20170418212149.10190.70894.stgit@tlendack-t1.amdoffice.net \
    --to=thomas.lendacky@amd.com \
    --cc=arnd@arndb.de \
    --cc=aryabinin@virtuozzo.com \
    --cc=bp@alien8.de \
    --cc=brijesh.singh@amd.com \
    --cc=corbet@lwn.net \
    --cc=dvyukov@google.com \
    --cc=dyoung@redhat.com \
    --cc=glider@google.com \
    --cc=hpa@zytor.com \
    --cc=iommu@lists.linux-foundation.org \
    --cc=joro@8bytes.org \
    --cc=kasan-dev@googlegroups.com \
    --cc=kexec@lists.infradead.org \
    --cc=konrad.wilk@oracle.com \
    --cc=kvm@vger.kernel.org \
    --cc=linux-arch@vger.kernel.org \
    --cc=linux-doc@vger.kernel.org \
    --cc=linux-efi@vger.kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=luto@kernel.org \
    --cc=lwoodman@redhat.com \
    --cc=matt@codeblueprint.co.uk \
    --cc=mingo@redhat.com \
    --cc=mst@redhat.com \
    --cc=pbonzini@redhat.com \
    --cc=riel@redhat.com \
    --cc=rkrcmar@redhat.com \
    --cc=tglx@linutronix.de \
    --cc=toshi.kani@hpe.com \
    --cc=x86@kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).