All of lore.kernel.org
 help / color / mirror / Atom feed
From: Tom Lendacky <thomas.lendacky@amd.com>
To: linux-arch@vger.kernel.org, linux-efi@vger.kernel.org,
	kvm@vger.kernel.org, linux-doc@vger.kernel.org, x86@kernel.org,
	kexec@lists.infradead.org, linux-kernel@vger.kernel.org,
	kasan-dev@googlegroups.com, xen-devel@lists.xen.org,
	linux-mm@kvack.org, iommu@lists.linux-foundation.org
Cc: "Brijesh Singh" <brijesh.singh@amd.com>,
	"Toshimitsu Kani" <toshi.kani@hpe.com>,
	"Radim Krčmář" <rkrcmar@redhat.com>,
	"Matt Fleming" <matt@codeblueprint.co.uk>,
	"Alexander Potapenko" <glider@google.com>,
	"H. Peter Anvin" <hpa@zytor.com>,
	"Larry Woodman" <lwoodman@redhat.com>,
	"Jonathan Corbet" <corbet@lwn.net>,
	"Joerg Roedel" <joro@8bytes.org>,
	"Michael S. Tsirkin" <mst@redhat.com>,
	"Ingo Molnar" <mingo@redhat.com>,
	"Andrey Ryabinin" <aryabinin@virtuozzo.com>,
	"Dave Young" <dyoung@redhat.com>,
	"Rik van Riel" <riel@redhat.com>, "Arnd Bergmann" <arnd@arndb.de>,
	"Konrad Rzeszutek Wilk" <konrad.wilk@oracle.com>,
	"Borislav Petkov" <bp@alien8.de>,
	"Andy Lutomirski" <luto@kernel.org>,
	"Boris Ostrovsky" <boris.ostrovsky@oracle.com>,
	"Dmitry Vyukov" <dvyukov@google.com>,
	"Juergen Gross" <jgross@suse.com>,
	"Thomas Gleixner" <tglx@linutronix.de>,
	"Paolo Bonzini" <pbonzini@redhat.com>
Subject: [PATCH v7 27/36] iommu/amd: Allow the AMD IOMMU to work with memory encryption
Date: Fri, 16 Jun 2017 13:54:59 -0500	[thread overview]
Message-ID: <20170616185459.18967.72790.stgit@tlendack-t1.amdoffice.net> (raw)
In-Reply-To: <20170616184947.18967.84890.stgit@tlendack-t1.amdoffice.net>

The IOMMU is programmed with physical addresses for the various tables
and buffers that are used to communicate between the device and the
driver. When the driver allocates this memory it is encrypted. In order
for the IOMMU to access the memory as encrypted the encryption mask needs
to be included in these physical addresses during configuration.

The PTE entries created by the IOMMU should also include the encryption
mask so that when the device behind the IOMMU performs a DMA, the DMA
will be performed to encrypted memory.

Signed-off-by: Tom Lendacky <thomas.lendacky@amd.com>
---
 drivers/iommu/amd_iommu.c       |   30 ++++++++++++++++--------------
 drivers/iommu/amd_iommu_init.c  |   34 ++++++++++++++++++++++++++++------
 drivers/iommu/amd_iommu_proto.h |   10 ++++++++++
 drivers/iommu/amd_iommu_types.h |    2 +-
 4 files changed, 55 insertions(+), 21 deletions(-)

diff --git a/drivers/iommu/amd_iommu.c b/drivers/iommu/amd_iommu.c
index 63cacf5..912008c 100644
--- a/drivers/iommu/amd_iommu.c
+++ b/drivers/iommu/amd_iommu.c
@@ -544,7 +544,7 @@ static void dump_dte_entry(u16 devid)
 
 static void dump_command(unsigned long phys_addr)
 {
-	struct iommu_cmd *cmd = phys_to_virt(phys_addr);
+	struct iommu_cmd *cmd = iommu_phys_to_virt(phys_addr);
 	int i;
 
 	for (i = 0; i < 4; ++i)
@@ -865,11 +865,13 @@ static void copy_cmd_to_buffer(struct amd_iommu *iommu,
 
 static void build_completion_wait(struct iommu_cmd *cmd, u64 address)
 {
+	u64 paddr = iommu_virt_to_phys((void *)address);
+
 	WARN_ON(address & 0x7ULL);
 
 	memset(cmd, 0, sizeof(*cmd));
-	cmd->data[0] = lower_32_bits(__pa(address)) | CMD_COMPL_WAIT_STORE_MASK;
-	cmd->data[1] = upper_32_bits(__pa(address));
+	cmd->data[0] = lower_32_bits(paddr) | CMD_COMPL_WAIT_STORE_MASK;
+	cmd->data[1] = upper_32_bits(paddr);
 	cmd->data[2] = 1;
 	CMD_SET_TYPE(cmd, CMD_COMPL_WAIT);
 }
@@ -1328,7 +1330,7 @@ static bool increase_address_space(struct protection_domain *domain,
 		return false;
 
 	*pte             = PM_LEVEL_PDE(domain->mode,
-					virt_to_phys(domain->pt_root));
+					iommu_virt_to_phys(domain->pt_root));
 	domain->pt_root  = pte;
 	domain->mode    += 1;
 	domain->updated  = true;
@@ -1365,7 +1367,7 @@ static u64 *alloc_pte(struct protection_domain *domain,
 			if (!page)
 				return NULL;
 
-			__npte = PM_LEVEL_PDE(level, virt_to_phys(page));
+			__npte = PM_LEVEL_PDE(level, iommu_virt_to_phys(page));
 
 			/* pte could have been changed somewhere. */
 			if (cmpxchg64(pte, __pte, __npte) != __pte) {
@@ -1481,10 +1483,10 @@ static int iommu_map_page(struct protection_domain *dom,
 			return -EBUSY;
 
 	if (count > 1) {
-		__pte = PAGE_SIZE_PTE(phys_addr, page_size);
+		__pte = PAGE_SIZE_PTE(__sme_set(phys_addr), page_size);
 		__pte |= PM_LEVEL_ENC(7) | IOMMU_PTE_P | IOMMU_PTE_FC;
 	} else
-		__pte = phys_addr | IOMMU_PTE_P | IOMMU_PTE_FC;
+		__pte = __sme_set(phys_addr) | IOMMU_PTE_P | IOMMU_PTE_FC;
 
 	if (prot & IOMMU_PROT_IR)
 		__pte |= IOMMU_PTE_IR;
@@ -1700,7 +1702,7 @@ static void free_gcr3_tbl_level1(u64 *tbl)
 		if (!(tbl[i] & GCR3_VALID))
 			continue;
 
-		ptr = __va(tbl[i] & PAGE_MASK);
+		ptr = iommu_phys_to_virt(tbl[i] & PAGE_MASK);
 
 		free_page((unsigned long)ptr);
 	}
@@ -1715,7 +1717,7 @@ static void free_gcr3_tbl_level2(u64 *tbl)
 		if (!(tbl[i] & GCR3_VALID))
 			continue;
 
-		ptr = __va(tbl[i] & PAGE_MASK);
+		ptr = iommu_phys_to_virt(tbl[i] & PAGE_MASK);
 
 		free_gcr3_tbl_level1(ptr);
 	}
@@ -1807,7 +1809,7 @@ static void set_dte_entry(u16 devid, struct protection_domain *domain, bool ats)
 	u64 flags = 0;
 
 	if (domain->mode != PAGE_MODE_NONE)
-		pte_root = virt_to_phys(domain->pt_root);
+		pte_root = iommu_virt_to_phys(domain->pt_root);
 
 	pte_root |= (domain->mode & DEV_ENTRY_MODE_MASK)
 		    << DEV_ENTRY_MODE_SHIFT;
@@ -1819,7 +1821,7 @@ static void set_dte_entry(u16 devid, struct protection_domain *domain, bool ats)
 		flags |= DTE_FLAG_IOTLB;
 
 	if (domain->flags & PD_IOMMUV2_MASK) {
-		u64 gcr3 = __pa(domain->gcr3_tbl);
+		u64 gcr3 = iommu_virt_to_phys(domain->gcr3_tbl);
 		u64 glx  = domain->glx;
 		u64 tmp;
 
@@ -3470,10 +3472,10 @@ static u64 *__get_gcr3_pte(u64 *root, int level, int pasid, bool alloc)
 			if (root == NULL)
 				return NULL;
 
-			*pte = __pa(root) | GCR3_VALID;
+			*pte = iommu_virt_to_phys(root) | GCR3_VALID;
 		}
 
-		root = __va(*pte & PAGE_MASK);
+		root = iommu_phys_to_virt(*pte & PAGE_MASK);
 
 		level -= 1;
 	}
@@ -3652,7 +3654,7 @@ static void set_dte_irq_entry(u16 devid, struct irq_remap_table *table)
 
 	dte	= amd_iommu_dev_table[devid].data[2];
 	dte	&= ~DTE_IRQ_PHYS_ADDR_MASK;
-	dte	|= virt_to_phys(table->table);
+	dte	|= iommu_virt_to_phys(table->table);
 	dte	|= DTE_IRQ_REMAP_INTCTL;
 	dte	|= DTE_IRQ_TABLE_LEN;
 	dte	|= DTE_IRQ_REMAP_ENABLE;
diff --git a/drivers/iommu/amd_iommu_init.c b/drivers/iommu/amd_iommu_init.c
index 5a11328..26862f1 100644
--- a/drivers/iommu/amd_iommu_init.c
+++ b/drivers/iommu/amd_iommu_init.c
@@ -29,6 +29,7 @@
 #include <linux/export.h>
 #include <linux/iommu.h>
 #include <linux/kmemleak.h>
+#include <linux/mem_encrypt.h>
 #include <asm/pci-direct.h>
 #include <asm/iommu.h>
 #include <asm/gart.h>
@@ -346,7 +347,7 @@ static void iommu_set_device_table(struct amd_iommu *iommu)
 
 	BUG_ON(iommu->mmio_base == NULL);
 
-	entry = virt_to_phys(amd_iommu_dev_table);
+	entry = iommu_virt_to_phys(amd_iommu_dev_table);
 	entry |= (dev_table_size >> 12) - 1;
 	memcpy_toio(iommu->mmio_base + MMIO_DEV_TABLE_OFFSET,
 			&entry, sizeof(entry));
@@ -602,7 +603,7 @@ static void iommu_enable_command_buffer(struct amd_iommu *iommu)
 
 	BUG_ON(iommu->cmd_buf == NULL);
 
-	entry = (u64)virt_to_phys(iommu->cmd_buf);
+	entry = iommu_virt_to_phys(iommu->cmd_buf);
 	entry |= MMIO_CMD_SIZE_512;
 
 	memcpy_toio(iommu->mmio_base + MMIO_CMD_BUF_OFFSET,
@@ -631,7 +632,7 @@ static void iommu_enable_event_buffer(struct amd_iommu *iommu)
 
 	BUG_ON(iommu->evt_buf == NULL);
 
-	entry = (u64)virt_to_phys(iommu->evt_buf) | EVT_LEN_MASK;
+	entry = iommu_virt_to_phys(iommu->evt_buf) | EVT_LEN_MASK;
 
 	memcpy_toio(iommu->mmio_base + MMIO_EVT_BUF_OFFSET,
 		    &entry, sizeof(entry));
@@ -664,7 +665,7 @@ static void iommu_enable_ppr_log(struct amd_iommu *iommu)
 	if (iommu->ppr_log == NULL)
 		return;
 
-	entry = (u64)virt_to_phys(iommu->ppr_log) | PPR_LOG_SIZE_512;
+	entry = iommu_virt_to_phys(iommu->ppr_log) | PPR_LOG_SIZE_512;
 
 	memcpy_toio(iommu->mmio_base + MMIO_PPR_LOG_OFFSET,
 		    &entry, sizeof(entry));
@@ -744,10 +745,10 @@ static int iommu_init_ga_log(struct amd_iommu *iommu)
 	if (!iommu->ga_log_tail)
 		goto err_out;
 
-	entry = (u64)virt_to_phys(iommu->ga_log) | GA_LOG_SIZE_512;
+	entry = iommu_virt_to_phys(iommu->ga_log) | GA_LOG_SIZE_512;
 	memcpy_toio(iommu->mmio_base + MMIO_GA_LOG_BASE_OFFSET,
 		    &entry, sizeof(entry));
-	entry = ((u64)virt_to_phys(iommu->ga_log) & 0xFFFFFFFFFFFFFULL) & ~7ULL;
+	entry = (iommu_virt_to_phys(iommu->ga_log) & 0xFFFFFFFFFFFFFULL) & ~7ULL;
 	memcpy_toio(iommu->mmio_base + MMIO_GA_LOG_TAIL_OFFSET,
 		    &entry, sizeof(entry));
 	writel(0x00, iommu->mmio_base + MMIO_GA_HEAD_OFFSET);
@@ -2535,6 +2536,24 @@ static int __init amd_iommu_init(void)
 	return ret;
 }
 
+static bool amd_iommu_supports_sme(void)
+{
+	if (!sme_active() || (boot_cpu_data.x86 != 0x17))
+		return true;
+
+	/* For Fam17h, a specific level of support is required */
+	if (boot_cpu_data.microcode >= 0x08001205)
+		return true;
+
+	if ((boot_cpu_data.microcode >= 0x08001126) &&
+	    (boot_cpu_data.microcode <= 0x080011ff))
+		return true;
+
+	pr_notice("AMD-Vi: IOMMU not currently supported when SME is active\n");
+
+	return false;
+}
+
 /****************************************************************************
  *
  * Early detect code. This code runs at IOMMU detection time in the DMA
@@ -2552,6 +2571,9 @@ int __init amd_iommu_detect(void)
 	if (amd_iommu_disabled)
 		return -ENODEV;
 
+	if (!amd_iommu_supports_sme())
+		return -ENODEV;
+
 	ret = iommu_go_to_state(IOMMU_IVRS_DETECTED);
 	if (ret)
 		return ret;
diff --git a/drivers/iommu/amd_iommu_proto.h b/drivers/iommu/amd_iommu_proto.h
index 466260f..3f12fb2 100644
--- a/drivers/iommu/amd_iommu_proto.h
+++ b/drivers/iommu/amd_iommu_proto.h
@@ -87,4 +87,14 @@ static inline bool iommu_feature(struct amd_iommu *iommu, u64 f)
 	return !!(iommu->features & f);
 }
 
+static inline u64 iommu_virt_to_phys(void *vaddr)
+{
+	return (u64)__sme_set(virt_to_phys(vaddr));
+}
+
+static inline void *iommu_phys_to_virt(unsigned long paddr)
+{
+	return phys_to_virt(__sme_clr(paddr));
+}
+
 #endif /* _ASM_X86_AMD_IOMMU_PROTO_H  */
diff --git a/drivers/iommu/amd_iommu_types.h b/drivers/iommu/amd_iommu_types.h
index 4de8f41..3ce587d 100644
--- a/drivers/iommu/amd_iommu_types.h
+++ b/drivers/iommu/amd_iommu_types.h
@@ -343,7 +343,7 @@
 
 #define IOMMU_PAGE_MASK (((1ULL << 52) - 1) & ~0xfffULL)
 #define IOMMU_PTE_PRESENT(pte) ((pte) & IOMMU_PTE_P)
-#define IOMMU_PTE_PAGE(pte) (phys_to_virt((pte) & IOMMU_PAGE_MASK))
+#define IOMMU_PTE_PAGE(pte) (iommu_phys_to_virt((pte) & IOMMU_PAGE_MASK))
 #define IOMMU_PTE_MODE(pte) (((pte) >> 9) & 0x07)
 
 #define IOMMU_PROT_MASK 0x03

WARNING: multiple messages have this Message-ID (diff)
From: Tom Lendacky <thomas.lendacky@amd.com>
To: linux-arch@vger.kernel.org, linux-efi@vger.kernel.org,
	kvm@vger.kernel.org, linux-doc@vger.kernel.org, x86@kernel.org,
	kexec@lists.infradead.org, linux-kernel@vger.kernel.org,
	kasan-dev@googlegroups.com, xen-devel@lists.xen.org,
	linux-mm@kvack.org, iommu@lists.linux-foundation.org
Cc: "Brijesh Singh" <brijesh.singh@amd.com>,
	"Toshimitsu Kani" <toshi.kani@hpe.com>,
	"Radim Krčmář" <rkrcmar@redhat.com>,
	"Matt Fleming" <matt@codeblueprint.co.uk>,
	"Alexander Potapenko" <glider@google.com>,
	"H. Peter Anvin" <hpa@zytor.com>,
	"Larry Woodman" <lwoodman@redhat.com>,
	"Jonathan Corbet" <corbet@lwn.net>,
	"Joerg Roedel" <joro@8bytes.org>,
	"Michael S. Tsirkin" <mst@redhat.com>,
	"Ingo Molnar" <mingo@redhat.com>,
	"Andrey Ryabinin" <aryabinin@virtuozzo.com>,
	"Dave Young" <dyoung@redhat.com>,
	"Rik van Riel" <riel@redhat.com>, "Arnd Bergmann" <arnd@arndb.de>,
	"Konrad Rzeszutek Wilk" <konrad.wilk@oracle.com>,
	"Borislav Petkov" <bp@alien8.de>,
	"Andy Lutomirski" <luto@kernel.org>,
	"Boris Ostrovsky" <boris.ostrovsky@oracle.com>,
	"Dmitry Vyukov" <dvyukov@google.com>,
	"Juergen Gross" <jgross@suse.com>,
	"Thomas Gleixner" <tglx@linutronix.de>,
	"Paolo Bonzini" <pbonzini@redhat.com>
Subject: [PATCH v7 27/36] iommu/amd: Allow the AMD IOMMU to work with memory encryption
Date: Fri, 16 Jun 2017 13:54:59 -0500	[thread overview]
Message-ID: <20170616185459.18967.72790.stgit@tlendack-t1.amdoffice.net> (raw)
In-Reply-To: <20170616184947.18967.84890.stgit@tlendack-t1.amdoffice.net>

The IOMMU is programmed with physical addresses for the various tables
and buffers that are used to communicate between the device and the
driver. When the driver allocates this memory it is encrypted. In order
for the IOMMU to access the memory as encrypted the encryption mask needs
to be included in these physical addresses during configuration.

The PTE entries created by the IOMMU should also include the encryption
mask so that when the device behind the IOMMU performs a DMA, the DMA
will be performed to encrypted memory.

Signed-off-by: Tom Lendacky <thomas.lendacky@amd.com>
---
 drivers/iommu/amd_iommu.c       |   30 ++++++++++++++++--------------
 drivers/iommu/amd_iommu_init.c  |   34 ++++++++++++++++++++++++++++------
 drivers/iommu/amd_iommu_proto.h |   10 ++++++++++
 drivers/iommu/amd_iommu_types.h |    2 +-
 4 files changed, 55 insertions(+), 21 deletions(-)

diff --git a/drivers/iommu/amd_iommu.c b/drivers/iommu/amd_iommu.c
index 63cacf5..912008c 100644
--- a/drivers/iommu/amd_iommu.c
+++ b/drivers/iommu/amd_iommu.c
@@ -544,7 +544,7 @@ static void dump_dte_entry(u16 devid)
 
 static void dump_command(unsigned long phys_addr)
 {
-	struct iommu_cmd *cmd = phys_to_virt(phys_addr);
+	struct iommu_cmd *cmd = iommu_phys_to_virt(phys_addr);
 	int i;
 
 	for (i = 0; i < 4; ++i)
@@ -865,11 +865,13 @@ static void copy_cmd_to_buffer(struct amd_iommu *iommu,
 
 static void build_completion_wait(struct iommu_cmd *cmd, u64 address)
 {
+	u64 paddr = iommu_virt_to_phys((void *)address);
+
 	WARN_ON(address & 0x7ULL);
 
 	memset(cmd, 0, sizeof(*cmd));
-	cmd->data[0] = lower_32_bits(__pa(address)) | CMD_COMPL_WAIT_STORE_MASK;
-	cmd->data[1] = upper_32_bits(__pa(address));
+	cmd->data[0] = lower_32_bits(paddr) | CMD_COMPL_WAIT_STORE_MASK;
+	cmd->data[1] = upper_32_bits(paddr);
 	cmd->data[2] = 1;
 	CMD_SET_TYPE(cmd, CMD_COMPL_WAIT);
 }
@@ -1328,7 +1330,7 @@ static bool increase_address_space(struct protection_domain *domain,
 		return false;
 
 	*pte             = PM_LEVEL_PDE(domain->mode,
-					virt_to_phys(domain->pt_root));
+					iommu_virt_to_phys(domain->pt_root));
 	domain->pt_root  = pte;
 	domain->mode    += 1;
 	domain->updated  = true;
@@ -1365,7 +1367,7 @@ static u64 *alloc_pte(struct protection_domain *domain,
 			if (!page)
 				return NULL;
 
-			__npte = PM_LEVEL_PDE(level, virt_to_phys(page));
+			__npte = PM_LEVEL_PDE(level, iommu_virt_to_phys(page));
 
 			/* pte could have been changed somewhere. */
 			if (cmpxchg64(pte, __pte, __npte) != __pte) {
@@ -1481,10 +1483,10 @@ static int iommu_map_page(struct protection_domain *dom,
 			return -EBUSY;
 
 	if (count > 1) {
-		__pte = PAGE_SIZE_PTE(phys_addr, page_size);
+		__pte = PAGE_SIZE_PTE(__sme_set(phys_addr), page_size);
 		__pte |= PM_LEVEL_ENC(7) | IOMMU_PTE_P | IOMMU_PTE_FC;
 	} else
-		__pte = phys_addr | IOMMU_PTE_P | IOMMU_PTE_FC;
+		__pte = __sme_set(phys_addr) | IOMMU_PTE_P | IOMMU_PTE_FC;
 
 	if (prot & IOMMU_PROT_IR)
 		__pte |= IOMMU_PTE_IR;
@@ -1700,7 +1702,7 @@ static void free_gcr3_tbl_level1(u64 *tbl)
 		if (!(tbl[i] & GCR3_VALID))
 			continue;
 
-		ptr = __va(tbl[i] & PAGE_MASK);
+		ptr = iommu_phys_to_virt(tbl[i] & PAGE_MASK);
 
 		free_page((unsigned long)ptr);
 	}
@@ -1715,7 +1717,7 @@ static void free_gcr3_tbl_level2(u64 *tbl)
 		if (!(tbl[i] & GCR3_VALID))
 			continue;
 
-		ptr = __va(tbl[i] & PAGE_MASK);
+		ptr = iommu_phys_to_virt(tbl[i] & PAGE_MASK);
 
 		free_gcr3_tbl_level1(ptr);
 	}
@@ -1807,7 +1809,7 @@ static void set_dte_entry(u16 devid, struct protection_domain *domain, bool ats)
 	u64 flags = 0;
 
 	if (domain->mode != PAGE_MODE_NONE)
-		pte_root = virt_to_phys(domain->pt_root);
+		pte_root = iommu_virt_to_phys(domain->pt_root);
 
 	pte_root |= (domain->mode & DEV_ENTRY_MODE_MASK)
 		    << DEV_ENTRY_MODE_SHIFT;
@@ -1819,7 +1821,7 @@ static void set_dte_entry(u16 devid, struct protection_domain *domain, bool ats)
 		flags |= DTE_FLAG_IOTLB;
 
 	if (domain->flags & PD_IOMMUV2_MASK) {
-		u64 gcr3 = __pa(domain->gcr3_tbl);
+		u64 gcr3 = iommu_virt_to_phys(domain->gcr3_tbl);
 		u64 glx  = domain->glx;
 		u64 tmp;
 
@@ -3470,10 +3472,10 @@ static u64 *__get_gcr3_pte(u64 *root, int level, int pasid, bool alloc)
 			if (root == NULL)
 				return NULL;
 
-			*pte = __pa(root) | GCR3_VALID;
+			*pte = iommu_virt_to_phys(root) | GCR3_VALID;
 		}
 
-		root = __va(*pte & PAGE_MASK);
+		root = iommu_phys_to_virt(*pte & PAGE_MASK);
 
 		level -= 1;
 	}
@@ -3652,7 +3654,7 @@ static void set_dte_irq_entry(u16 devid, struct irq_remap_table *table)
 
 	dte	= amd_iommu_dev_table[devid].data[2];
 	dte	&= ~DTE_IRQ_PHYS_ADDR_MASK;
-	dte	|= virt_to_phys(table->table);
+	dte	|= iommu_virt_to_phys(table->table);
 	dte	|= DTE_IRQ_REMAP_INTCTL;
 	dte	|= DTE_IRQ_TABLE_LEN;
 	dte	|= DTE_IRQ_REMAP_ENABLE;
diff --git a/drivers/iommu/amd_iommu_init.c b/drivers/iommu/amd_iommu_init.c
index 5a11328..26862f1 100644
--- a/drivers/iommu/amd_iommu_init.c
+++ b/drivers/iommu/amd_iommu_init.c
@@ -29,6 +29,7 @@
 #include <linux/export.h>
 #include <linux/iommu.h>
 #include <linux/kmemleak.h>
+#include <linux/mem_encrypt.h>
 #include <asm/pci-direct.h>
 #include <asm/iommu.h>
 #include <asm/gart.h>
@@ -346,7 +347,7 @@ static void iommu_set_device_table(struct amd_iommu *iommu)
 
 	BUG_ON(iommu->mmio_base == NULL);
 
-	entry = virt_to_phys(amd_iommu_dev_table);
+	entry = iommu_virt_to_phys(amd_iommu_dev_table);
 	entry |= (dev_table_size >> 12) - 1;
 	memcpy_toio(iommu->mmio_base + MMIO_DEV_TABLE_OFFSET,
 			&entry, sizeof(entry));
@@ -602,7 +603,7 @@ static void iommu_enable_command_buffer(struct amd_iommu *iommu)
 
 	BUG_ON(iommu->cmd_buf == NULL);
 
-	entry = (u64)virt_to_phys(iommu->cmd_buf);
+	entry = iommu_virt_to_phys(iommu->cmd_buf);
 	entry |= MMIO_CMD_SIZE_512;
 
 	memcpy_toio(iommu->mmio_base + MMIO_CMD_BUF_OFFSET,
@@ -631,7 +632,7 @@ static void iommu_enable_event_buffer(struct amd_iommu *iommu)
 
 	BUG_ON(iommu->evt_buf == NULL);
 
-	entry = (u64)virt_to_phys(iommu->evt_buf) | EVT_LEN_MASK;
+	entry = iommu_virt_to_phys(iommu->evt_buf) | EVT_LEN_MASK;
 
 	memcpy_toio(iommu->mmio_base + MMIO_EVT_BUF_OFFSET,
 		    &entry, sizeof(entry));
@@ -664,7 +665,7 @@ static void iommu_enable_ppr_log(struct amd_iommu *iommu)
 	if (iommu->ppr_log == NULL)
 		return;
 
-	entry = (u64)virt_to_phys(iommu->ppr_log) | PPR_LOG_SIZE_512;
+	entry = iommu_virt_to_phys(iommu->ppr_log) | PPR_LOG_SIZE_512;
 
 	memcpy_toio(iommu->mmio_base + MMIO_PPR_LOG_OFFSET,
 		    &entry, sizeof(entry));
@@ -744,10 +745,10 @@ static int iommu_init_ga_log(struct amd_iommu *iommu)
 	if (!iommu->ga_log_tail)
 		goto err_out;
 
-	entry = (u64)virt_to_phys(iommu->ga_log) | GA_LOG_SIZE_512;
+	entry = iommu_virt_to_phys(iommu->ga_log) | GA_LOG_SIZE_512;
 	memcpy_toio(iommu->mmio_base + MMIO_GA_LOG_BASE_OFFSET,
 		    &entry, sizeof(entry));
-	entry = ((u64)virt_to_phys(iommu->ga_log) & 0xFFFFFFFFFFFFFULL) & ~7ULL;
+	entry = (iommu_virt_to_phys(iommu->ga_log) & 0xFFFFFFFFFFFFFULL) & ~7ULL;
 	memcpy_toio(iommu->mmio_base + MMIO_GA_LOG_TAIL_OFFSET,
 		    &entry, sizeof(entry));
 	writel(0x00, iommu->mmio_base + MMIO_GA_HEAD_OFFSET);
@@ -2535,6 +2536,24 @@ static int __init amd_iommu_init(void)
 	return ret;
 }
 
+static bool amd_iommu_supports_sme(void)
+{
+	if (!sme_active() || (boot_cpu_data.x86 != 0x17))
+		return true;
+
+	/* For Fam17h, a specific level of support is required */
+	if (boot_cpu_data.microcode >= 0x08001205)
+		return true;
+
+	if ((boot_cpu_data.microcode >= 0x08001126) &&
+	    (boot_cpu_data.microcode <= 0x080011ff))
+		return true;
+
+	pr_notice("AMD-Vi: IOMMU not currently supported when SME is active\n");
+
+	return false;
+}
+
 /****************************************************************************
  *
  * Early detect code. This code runs at IOMMU detection time in the DMA
@@ -2552,6 +2571,9 @@ int __init amd_iommu_detect(void)
 	if (amd_iommu_disabled)
 		return -ENODEV;
 
+	if (!amd_iommu_supports_sme())
+		return -ENODEV;
+
 	ret = iommu_go_to_state(IOMMU_IVRS_DETECTED);
 	if (ret)
 		return ret;
diff --git a/drivers/iommu/amd_iommu_proto.h b/drivers/iommu/amd_iommu_proto.h
index 466260f..3f12fb2 100644
--- a/drivers/iommu/amd_iommu_proto.h
+++ b/drivers/iommu/amd_iommu_proto.h
@@ -87,4 +87,14 @@ static inline bool iommu_feature(struct amd_iommu *iommu, u64 f)
 	return !!(iommu->features & f);
 }
 
+static inline u64 iommu_virt_to_phys(void *vaddr)
+{
+	return (u64)__sme_set(virt_to_phys(vaddr));
+}
+
+static inline void *iommu_phys_to_virt(unsigned long paddr)
+{
+	return phys_to_virt(__sme_clr(paddr));
+}
+
 #endif /* _ASM_X86_AMD_IOMMU_PROTO_H  */
diff --git a/drivers/iommu/amd_iommu_types.h b/drivers/iommu/amd_iommu_types.h
index 4de8f41..3ce587d 100644
--- a/drivers/iommu/amd_iommu_types.h
+++ b/drivers/iommu/amd_iommu_types.h
@@ -343,7 +343,7 @@
 
 #define IOMMU_PAGE_MASK (((1ULL << 52) - 1) & ~0xfffULL)
 #define IOMMU_PTE_PRESENT(pte) ((pte) & IOMMU_PTE_P)
-#define IOMMU_PTE_PAGE(pte) (phys_to_virt((pte) & IOMMU_PAGE_MASK))
+#define IOMMU_PTE_PAGE(pte) (iommu_phys_to_virt((pte) & IOMMU_PAGE_MASK))
 #define IOMMU_PTE_MODE(pte) (((pte) >> 9) & 0x07)
 
 #define IOMMU_PROT_MASK 0x03

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

WARNING: multiple messages have this Message-ID (diff)
From: Tom Lendacky <thomas.lendacky@amd.com>
To: linux-arch@vger.kernel.org, linux-efi@vger.kernel.org,
	kvm@vger.kernel.org, linux-doc@vger.kernel.org, x86@kernel.org,
	kexec@lists.infradead.org, linux-kernel@vger.kernel.org,
	kasan-dev@googlegroups.com, xen-devel@lists.xen.org,
	linux-mm@kvack.org, iommu@lists.linux-foundation.org
Cc: "Brijesh Singh" <brijesh.singh@amd.com>,
	"Toshimitsu Kani" <toshi.kani@hpe.com>,
	"Michael S. Tsirkin" <mst@redhat.com>,
	"Matt Fleming" <matt@codeblueprint.co.uk>,
	"Alexander Potapenko" <glider@google.com>,
	"H. Peter Anvin" <hpa@zytor.com>,
	"Boris Ostrovsky" <boris.ostrovsky@oracle.com>,
	"Jonathan Corbet" <corbet@lwn.net>,
	"Joerg Roedel" <joro@8bytes.org>,
	"Radim Krčmář" <rkrcmar@redhat.com>,
	"Larry Woodman" <lwoodman@redhat.com>,
	"Ingo Molnar" <mingo@redhat.com>,
	"Andrey Ryabinin" <aryabinin@virtuozzo.com>,
	"Dave Young" <dyoung@redhat.com>,
	"Rik van Riel" <riel@redhat.com>, "Arnd Bergmann" <arnd@arndb.de>,
	"Konrad Rzeszutek Wilk" <konrad.wilk@oracle.com>,
	"Borislav Petkov" <bp@alien8.de>,
	"Andy Lutomirski" <luto@kernel.org>,
	"Thomas Gleixner" <tglx@linutronix.de>,
	"Dmitry Vyukov" <dvyukov@google.com>,
	"Juergen Gross" <jgross@suse.com>,
	"Paolo Bonzini" <pbonzini@redhat.com>
Subject: [PATCH v7 27/36] iommu/amd: Allow the AMD IOMMU to work with memory encryption
Date: Fri, 16 Jun 2017 13:54:59 -0500	[thread overview]
Message-ID: <20170616185459.18967.72790.stgit@tlendack-t1.amdoffice.net> (raw)
In-Reply-To: <20170616184947.18967.84890.stgit@tlendack-t1.amdoffice.net>

The IOMMU is programmed with physical addresses for the various tables
and buffers that are used to communicate between the device and the
driver. When the driver allocates this memory it is encrypted. In order
for the IOMMU to access the memory as encrypted the encryption mask needs
to be included in these physical addresses during configuration.

The PTE entries created by the IOMMU should also include the encryption
mask so that when the device behind the IOMMU performs a DMA, the DMA
will be performed to encrypted memory.

Signed-off-by: Tom Lendacky <thomas.lendacky@amd.com>
---
 drivers/iommu/amd_iommu.c       |   30 ++++++++++++++++--------------
 drivers/iommu/amd_iommu_init.c  |   34 ++++++++++++++++++++++++++++------
 drivers/iommu/amd_iommu_proto.h |   10 ++++++++++
 drivers/iommu/amd_iommu_types.h |    2 +-
 4 files changed, 55 insertions(+), 21 deletions(-)

diff --git a/drivers/iommu/amd_iommu.c b/drivers/iommu/amd_iommu.c
index 63cacf5..912008c 100644
--- a/drivers/iommu/amd_iommu.c
+++ b/drivers/iommu/amd_iommu.c
@@ -544,7 +544,7 @@ static void dump_dte_entry(u16 devid)
 
 static void dump_command(unsigned long phys_addr)
 {
-	struct iommu_cmd *cmd = phys_to_virt(phys_addr);
+	struct iommu_cmd *cmd = iommu_phys_to_virt(phys_addr);
 	int i;
 
 	for (i = 0; i < 4; ++i)
@@ -865,11 +865,13 @@ static void copy_cmd_to_buffer(struct amd_iommu *iommu,
 
 static void build_completion_wait(struct iommu_cmd *cmd, u64 address)
 {
+	u64 paddr = iommu_virt_to_phys((void *)address);
+
 	WARN_ON(address & 0x7ULL);
 
 	memset(cmd, 0, sizeof(*cmd));
-	cmd->data[0] = lower_32_bits(__pa(address)) | CMD_COMPL_WAIT_STORE_MASK;
-	cmd->data[1] = upper_32_bits(__pa(address));
+	cmd->data[0] = lower_32_bits(paddr) | CMD_COMPL_WAIT_STORE_MASK;
+	cmd->data[1] = upper_32_bits(paddr);
 	cmd->data[2] = 1;
 	CMD_SET_TYPE(cmd, CMD_COMPL_WAIT);
 }
@@ -1328,7 +1330,7 @@ static bool increase_address_space(struct protection_domain *domain,
 		return false;
 
 	*pte             = PM_LEVEL_PDE(domain->mode,
-					virt_to_phys(domain->pt_root));
+					iommu_virt_to_phys(domain->pt_root));
 	domain->pt_root  = pte;
 	domain->mode    += 1;
 	domain->updated  = true;
@@ -1365,7 +1367,7 @@ static u64 *alloc_pte(struct protection_domain *domain,
 			if (!page)
 				return NULL;
 
-			__npte = PM_LEVEL_PDE(level, virt_to_phys(page));
+			__npte = PM_LEVEL_PDE(level, iommu_virt_to_phys(page));
 
 			/* pte could have been changed somewhere. */
 			if (cmpxchg64(pte, __pte, __npte) != __pte) {
@@ -1481,10 +1483,10 @@ static int iommu_map_page(struct protection_domain *dom,
 			return -EBUSY;
 
 	if (count > 1) {
-		__pte = PAGE_SIZE_PTE(phys_addr, page_size);
+		__pte = PAGE_SIZE_PTE(__sme_set(phys_addr), page_size);
 		__pte |= PM_LEVEL_ENC(7) | IOMMU_PTE_P | IOMMU_PTE_FC;
 	} else
-		__pte = phys_addr | IOMMU_PTE_P | IOMMU_PTE_FC;
+		__pte = __sme_set(phys_addr) | IOMMU_PTE_P | IOMMU_PTE_FC;
 
 	if (prot & IOMMU_PROT_IR)
 		__pte |= IOMMU_PTE_IR;
@@ -1700,7 +1702,7 @@ static void free_gcr3_tbl_level1(u64 *tbl)
 		if (!(tbl[i] & GCR3_VALID))
 			continue;
 
-		ptr = __va(tbl[i] & PAGE_MASK);
+		ptr = iommu_phys_to_virt(tbl[i] & PAGE_MASK);
 
 		free_page((unsigned long)ptr);
 	}
@@ -1715,7 +1717,7 @@ static void free_gcr3_tbl_level2(u64 *tbl)
 		if (!(tbl[i] & GCR3_VALID))
 			continue;
 
-		ptr = __va(tbl[i] & PAGE_MASK);
+		ptr = iommu_phys_to_virt(tbl[i] & PAGE_MASK);
 
 		free_gcr3_tbl_level1(ptr);
 	}
@@ -1807,7 +1809,7 @@ static void set_dte_entry(u16 devid, struct protection_domain *domain, bool ats)
 	u64 flags = 0;
 
 	if (domain->mode != PAGE_MODE_NONE)
-		pte_root = virt_to_phys(domain->pt_root);
+		pte_root = iommu_virt_to_phys(domain->pt_root);
 
 	pte_root |= (domain->mode & DEV_ENTRY_MODE_MASK)
 		    << DEV_ENTRY_MODE_SHIFT;
@@ -1819,7 +1821,7 @@ static void set_dte_entry(u16 devid, struct protection_domain *domain, bool ats)
 		flags |= DTE_FLAG_IOTLB;
 
 	if (domain->flags & PD_IOMMUV2_MASK) {
-		u64 gcr3 = __pa(domain->gcr3_tbl);
+		u64 gcr3 = iommu_virt_to_phys(domain->gcr3_tbl);
 		u64 glx  = domain->glx;
 		u64 tmp;
 
@@ -3470,10 +3472,10 @@ static u64 *__get_gcr3_pte(u64 *root, int level, int pasid, bool alloc)
 			if (root == NULL)
 				return NULL;
 
-			*pte = __pa(root) | GCR3_VALID;
+			*pte = iommu_virt_to_phys(root) | GCR3_VALID;
 		}
 
-		root = __va(*pte & PAGE_MASK);
+		root = iommu_phys_to_virt(*pte & PAGE_MASK);
 
 		level -= 1;
 	}
@@ -3652,7 +3654,7 @@ static void set_dte_irq_entry(u16 devid, struct irq_remap_table *table)
 
 	dte	= amd_iommu_dev_table[devid].data[2];
 	dte	&= ~DTE_IRQ_PHYS_ADDR_MASK;
-	dte	|= virt_to_phys(table->table);
+	dte	|= iommu_virt_to_phys(table->table);
 	dte	|= DTE_IRQ_REMAP_INTCTL;
 	dte	|= DTE_IRQ_TABLE_LEN;
 	dte	|= DTE_IRQ_REMAP_ENABLE;
diff --git a/drivers/iommu/amd_iommu_init.c b/drivers/iommu/amd_iommu_init.c
index 5a11328..26862f1 100644
--- a/drivers/iommu/amd_iommu_init.c
+++ b/drivers/iommu/amd_iommu_init.c
@@ -29,6 +29,7 @@
 #include <linux/export.h>
 #include <linux/iommu.h>
 #include <linux/kmemleak.h>
+#include <linux/mem_encrypt.h>
 #include <asm/pci-direct.h>
 #include <asm/iommu.h>
 #include <asm/gart.h>
@@ -346,7 +347,7 @@ static void iommu_set_device_table(struct amd_iommu *iommu)
 
 	BUG_ON(iommu->mmio_base == NULL);
 
-	entry = virt_to_phys(amd_iommu_dev_table);
+	entry = iommu_virt_to_phys(amd_iommu_dev_table);
 	entry |= (dev_table_size >> 12) - 1;
 	memcpy_toio(iommu->mmio_base + MMIO_DEV_TABLE_OFFSET,
 			&entry, sizeof(entry));
@@ -602,7 +603,7 @@ static void iommu_enable_command_buffer(struct amd_iommu *iommu)
 
 	BUG_ON(iommu->cmd_buf == NULL);
 
-	entry = (u64)virt_to_phys(iommu->cmd_buf);
+	entry = iommu_virt_to_phys(iommu->cmd_buf);
 	entry |= MMIO_CMD_SIZE_512;
 
 	memcpy_toio(iommu->mmio_base + MMIO_CMD_BUF_OFFSET,
@@ -631,7 +632,7 @@ static void iommu_enable_event_buffer(struct amd_iommu *iommu)
 
 	BUG_ON(iommu->evt_buf == NULL);
 
-	entry = (u64)virt_to_phys(iommu->evt_buf) | EVT_LEN_MASK;
+	entry = iommu_virt_to_phys(iommu->evt_buf) | EVT_LEN_MASK;
 
 	memcpy_toio(iommu->mmio_base + MMIO_EVT_BUF_OFFSET,
 		    &entry, sizeof(entry));
@@ -664,7 +665,7 @@ static void iommu_enable_ppr_log(struct amd_iommu *iommu)
 	if (iommu->ppr_log == NULL)
 		return;
 
-	entry = (u64)virt_to_phys(iommu->ppr_log) | PPR_LOG_SIZE_512;
+	entry = iommu_virt_to_phys(iommu->ppr_log) | PPR_LOG_SIZE_512;
 
 	memcpy_toio(iommu->mmio_base + MMIO_PPR_LOG_OFFSET,
 		    &entry, sizeof(entry));
@@ -744,10 +745,10 @@ static int iommu_init_ga_log(struct amd_iommu *iommu)
 	if (!iommu->ga_log_tail)
 		goto err_out;
 
-	entry = (u64)virt_to_phys(iommu->ga_log) | GA_LOG_SIZE_512;
+	entry = iommu_virt_to_phys(iommu->ga_log) | GA_LOG_SIZE_512;
 	memcpy_toio(iommu->mmio_base + MMIO_GA_LOG_BASE_OFFSET,
 		    &entry, sizeof(entry));
-	entry = ((u64)virt_to_phys(iommu->ga_log) & 0xFFFFFFFFFFFFFULL) & ~7ULL;
+	entry = (iommu_virt_to_phys(iommu->ga_log) & 0xFFFFFFFFFFFFFULL) & ~7ULL;
 	memcpy_toio(iommu->mmio_base + MMIO_GA_LOG_TAIL_OFFSET,
 		    &entry, sizeof(entry));
 	writel(0x00, iommu->mmio_base + MMIO_GA_HEAD_OFFSET);
@@ -2535,6 +2536,24 @@ static int __init amd_iommu_init(void)
 	return ret;
 }
 
+static bool amd_iommu_supports_sme(void)
+{
+	if (!sme_active() || (boot_cpu_data.x86 != 0x17))
+		return true;
+
+	/* For Fam17h, a specific level of support is required */
+	if (boot_cpu_data.microcode >= 0x08001205)
+		return true;
+
+	if ((boot_cpu_data.microcode >= 0x08001126) &&
+	    (boot_cpu_data.microcode <= 0x080011ff))
+		return true;
+
+	pr_notice("AMD-Vi: IOMMU not currently supported when SME is active\n");
+
+	return false;
+}
+
 /****************************************************************************
  *
  * Early detect code. This code runs at IOMMU detection time in the DMA
@@ -2552,6 +2571,9 @@ int __init amd_iommu_detect(void)
 	if (amd_iommu_disabled)
 		return -ENODEV;
 
+	if (!amd_iommu_supports_sme())
+		return -ENODEV;
+
 	ret = iommu_go_to_state(IOMMU_IVRS_DETECTED);
 	if (ret)
 		return ret;
diff --git a/drivers/iommu/amd_iommu_proto.h b/drivers/iommu/amd_iommu_proto.h
index 466260f..3f12fb2 100644
--- a/drivers/iommu/amd_iommu_proto.h
+++ b/drivers/iommu/amd_iommu_proto.h
@@ -87,4 +87,14 @@ static inline bool iommu_feature(struct amd_iommu *iommu, u64 f)
 	return !!(iommu->features & f);
 }
 
+static inline u64 iommu_virt_to_phys(void *vaddr)
+{
+	return (u64)__sme_set(virt_to_phys(vaddr));
+}
+
+static inline void *iommu_phys_to_virt(unsigned long paddr)
+{
+	return phys_to_virt(__sme_clr(paddr));
+}
+
 #endif /* _ASM_X86_AMD_IOMMU_PROTO_H  */
diff --git a/drivers/iommu/amd_iommu_types.h b/drivers/iommu/amd_iommu_types.h
index 4de8f41..3ce587d 100644
--- a/drivers/iommu/amd_iommu_types.h
+++ b/drivers/iommu/amd_iommu_types.h
@@ -343,7 +343,7 @@
 
 #define IOMMU_PAGE_MASK (((1ULL << 52) - 1) & ~0xfffULL)
 #define IOMMU_PTE_PRESENT(pte) ((pte) & IOMMU_PTE_P)
-#define IOMMU_PTE_PAGE(pte) (phys_to_virt((pte) & IOMMU_PAGE_MASK))
+#define IOMMU_PTE_PAGE(pte) (iommu_phys_to_virt((pte) & IOMMU_PAGE_MASK))
 #define IOMMU_PTE_MODE(pte) (((pte) >> 9) & 0x07)
 
 #define IOMMU_PROT_MASK 0x03


_______________________________________________
kexec mailing list
kexec@lists.infradead.org
http://lists.infradead.org/mailman/listinfo/kexec

  parent reply	other threads:[~2017-06-16 18:55 UTC|newest]

Thread overview: 384+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2017-06-16 18:49 [PATCH v7 00/36] x86: Secure Memory Encryption (AMD) Tom Lendacky
2017-06-16 18:49 ` Tom Lendacky
2017-06-16 18:49 ` Tom Lendacky
2017-06-16 18:49 ` [PATCH v7 01/36] x86: Document AMD Secure Memory Encryption (SME) Tom Lendacky
2017-06-16 18:49 ` Tom Lendacky
2017-06-16 18:49   ` Tom Lendacky
2017-06-16 18:49   ` Tom Lendacky
2017-06-16 18:50 ` [PATCH v7 02/36] x86/mm/pat: Set write-protect cache mode for full PAT support Tom Lendacky
2017-06-16 18:50   ` Tom Lendacky
2017-06-16 18:50   ` Tom Lendacky
2017-06-16 18:50 ` Tom Lendacky
2017-06-16 18:50 ` [PATCH v7 03/36] x86, mpparse, x86/acpi, x86/PCI, x86/dmi, SFI: Use memremap for RAM mappings Tom Lendacky
2017-06-16 18:50   ` Tom Lendacky
2017-06-16 18:50   ` Tom Lendacky
2017-06-16 18:50   ` Tom Lendacky
2017-06-20  7:21   ` Borislav Petkov
2017-06-20  7:21   ` Borislav Petkov
2017-06-20  7:21     ` Borislav Petkov
2017-06-20  7:21     ` Borislav Petkov
2017-06-20  7:21     ` Borislav Petkov
2017-06-16 18:50 ` Tom Lendacky
2017-06-16 18:50 ` [PATCH v7 04/36] x86/CPU/AMD: Add the Secure Memory Encryption CPU feature Tom Lendacky
2017-06-16 18:50   ` Tom Lendacky
2017-06-16 18:50   ` Tom Lendacky
2017-06-16 18:50   ` Tom Lendacky
2017-06-16 18:50 ` Tom Lendacky
2017-06-16 18:50 ` [PATCH v7 05/36] x86/CPU/AMD: Handle SME reduction in physical address size Tom Lendacky
2017-06-16 18:50   ` Tom Lendacky
2017-06-16 18:50   ` Tom Lendacky
2017-06-16 18:50 ` Tom Lendacky
2017-06-16 18:50 ` [PATCH v7 06/36] x86/mm: Add Secure Memory Encryption (SME) support Tom Lendacky
2017-06-16 18:50 ` Tom Lendacky
2017-06-16 18:50   ` Tom Lendacky
2017-06-16 18:50   ` Tom Lendacky
2017-06-16 18:50   ` Tom Lendacky
2017-06-20 20:49   ` Thomas Gleixner
2017-06-20 20:49     ` Thomas Gleixner
2017-06-20 20:49     ` Thomas Gleixner
2017-06-20 20:49     ` Thomas Gleixner
2017-06-21 13:26     ` Tom Lendacky
2017-06-21 13:26       ` Tom Lendacky
2017-06-21 13:26       ` Tom Lendacky
2017-06-21 13:26       ` Tom Lendacky
2017-06-21 13:26     ` Tom Lendacky
2017-06-20 20:49   ` Thomas Gleixner
2017-06-16 18:51 ` [PATCH v7 07/36] x86/mm: Don't use phys_to_virt in ioremap() if SME is active Tom Lendacky
2017-06-16 18:51   ` Tom Lendacky
2017-06-16 18:51   ` Tom Lendacky
2017-06-20 20:55   ` Thomas Gleixner
2017-06-20 20:55   ` Thomas Gleixner
2017-06-20 20:55     ` Thomas Gleixner
2017-06-20 20:55     ` Thomas Gleixner
2017-06-20 20:55     ` Thomas Gleixner
2017-06-21 13:52     ` Tom Lendacky
2017-06-21 13:52       ` Tom Lendacky
2017-06-21 13:52       ` Tom Lendacky
2017-06-21 13:52       ` Tom Lendacky
2017-06-21 13:52     ` Tom Lendacky
2017-06-21  7:37   ` Thomas Gleixner
2017-06-21  7:37     ` Thomas Gleixner
2017-06-21  7:37     ` Thomas Gleixner
2017-06-21  7:37     ` Thomas Gleixner
2017-06-21 13:54     ` Tom Lendacky
2017-06-21 13:54     ` Tom Lendacky
2017-06-21 13:54       ` Tom Lendacky
2017-06-21 13:54       ` Tom Lendacky
2017-06-21 13:54       ` Tom Lendacky
2017-06-21  7:37   ` Thomas Gleixner
2017-06-16 18:51 ` Tom Lendacky
2017-06-16 18:51 ` [PATCH v7 08/36] x86/mm: Add support to enable SME in early boot processing Tom Lendacky
2017-06-16 18:51   ` Tom Lendacky
2017-06-16 18:51   ` Tom Lendacky
2017-06-16 18:51   ` Tom Lendacky
2017-06-20  7:38   ` Borislav Petkov
2017-06-20  7:38     ` Borislav Petkov
2017-06-20  7:38     ` Borislav Petkov
2017-06-20  7:38     ` Borislav Petkov
2017-06-20 15:52     ` Tom Lendacky
2017-06-20 15:52     ` Tom Lendacky
2017-06-20 15:52       ` Tom Lendacky
2017-06-20 15:52       ` Tom Lendacky
2017-06-20 15:52       ` Tom Lendacky
2017-06-20  7:38   ` Borislav Petkov
2017-06-21  7:16   ` Thomas Gleixner
2017-06-21  7:16   ` Thomas Gleixner
2017-06-21  7:16     ` Thomas Gleixner
2017-06-21  7:16     ` Thomas Gleixner
2017-06-21  7:16     ` Thomas Gleixner
2017-06-21 15:14     ` Tom Lendacky
2017-06-21 15:14       ` Tom Lendacky
2017-06-21 15:14       ` Tom Lendacky
2017-06-21 15:14       ` Tom Lendacky
2017-06-21 15:38       ` Thomas Gleixner
2017-06-21 15:38         ` Thomas Gleixner
2017-06-21 15:38         ` Thomas Gleixner
2017-06-21 15:38         ` Thomas Gleixner
2017-06-21 18:30         ` Tom Lendacky
2017-06-21 18:30           ` Tom Lendacky
2017-06-21 18:30           ` Tom Lendacky
2017-06-21 18:30           ` Tom Lendacky
2017-06-21 18:52           ` Thomas Gleixner
2017-06-21 18:52           ` Thomas Gleixner
2017-06-21 18:52             ` Thomas Gleixner
2017-06-21 18:52             ` Thomas Gleixner
2017-06-21 18:52             ` Thomas Gleixner
2017-06-21 18:30         ` Tom Lendacky
2017-06-21 15:38       ` Thomas Gleixner
2017-06-21 15:14     ` Tom Lendacky
2017-06-16 18:51 ` Tom Lendacky
2017-06-16 18:51 ` [PATCH v7 09/36] x86/mm: Simplify p[gum]d_page() macros Tom Lendacky
2017-06-16 18:51 ` Tom Lendacky
2017-06-16 18:51   ` Tom Lendacky
2017-06-16 18:51   ` Tom Lendacky
2017-06-16 18:51   ` Tom Lendacky
2017-06-16 18:51 ` [PATCH v7 10/36] x86/mm: Provide general kernel support for memory encryption Tom Lendacky
2017-06-16 18:51   ` Tom Lendacky
2017-06-16 18:51   ` Tom Lendacky
2017-06-16 18:51   ` Tom Lendacky
2017-06-21  7:18   ` Thomas Gleixner
2017-06-21  7:18   ` Thomas Gleixner
2017-06-21  7:18     ` Thomas Gleixner
2017-06-21  7:18     ` Thomas Gleixner
2017-06-21  7:18     ` Thomas Gleixner
2017-06-21  8:23     ` Borislav Petkov
2017-06-21  8:23       ` Borislav Petkov
2017-06-21  8:23       ` Borislav Petkov
2017-06-21  8:23       ` Borislav Petkov
2017-06-21  8:23     ` Borislav Petkov
2017-06-16 18:51 ` Tom Lendacky
2017-06-16 18:51 ` [PATCH v7 11/36] x86/mm: Add SME support for read_cr3_pa() Tom Lendacky
2017-06-16 18:51   ` Tom Lendacky
2017-06-16 18:51   ` Tom Lendacky
2017-06-20  8:25   ` Borislav Petkov
2017-06-20  8:25     ` Borislav Petkov
2017-06-20  8:25     ` Borislav Petkov
2017-06-20  8:25     ` Borislav Petkov
2017-06-20  8:25   ` Borislav Petkov
2017-06-20 16:17   ` Andy Lutomirski
2017-06-20 16:17   ` Andy Lutomirski
2017-06-20 16:17     ` Andy Lutomirski
2017-06-20 16:17     ` Andy Lutomirski
2017-06-20 16:17     ` Andy Lutomirski
2017-06-20 16:23     ` Tom Lendacky
2017-06-20 16:23       ` Tom Lendacky
2017-06-20 16:23       ` Tom Lendacky
2017-06-20 16:23       ` Tom Lendacky
2017-06-20 16:23     ` Tom Lendacky
2017-06-16 18:51 ` Tom Lendacky
2017-06-16 18:52 ` [PATCH v7 12/36] x86/mm: Extend early_memremap() support with additional attrs Tom Lendacky
2017-06-16 18:52   ` Tom Lendacky
2017-06-16 18:52   ` Tom Lendacky
2017-06-16 18:52 ` Tom Lendacky
2017-06-16 18:52 ` [PATCH v7 13/36] x86/mm: Add support for early encrypt/decrypt of memory Tom Lendacky
2017-06-16 18:52 ` Tom Lendacky
2017-06-16 18:52   ` Tom Lendacky
2017-06-16 18:52   ` Tom Lendacky
2017-06-16 18:52   ` Tom Lendacky
2017-06-16 18:52 ` [PATCH v7 14/36] x86/mm: Insure that boot memory areas are mapped properly Tom Lendacky
2017-06-16 18:52 ` Tom Lendacky
2017-06-16 18:52   ` Tom Lendacky
2017-06-16 18:52   ` Tom Lendacky
2017-06-16 18:52   ` Tom Lendacky
2017-06-20 15:30   ` Borislav Petkov
2017-06-20 15:30   ` Borislav Petkov
2017-06-20 15:30     ` Borislav Petkov
2017-06-20 15:30     ` Borislav Petkov
2017-06-20 15:30     ` Borislav Petkov
2017-06-16 18:52 ` [PATCH v7 15/36] x86/boot/e820: Add support to determine the E820 type of an address Tom Lendacky
2017-06-16 18:52 ` Tom Lendacky
2017-06-16 18:52   ` Tom Lendacky
2017-06-16 18:52   ` Tom Lendacky
2017-06-16 18:52 ` [PATCH v7 16/36] efi: Add an EFI table address match function Tom Lendacky
2017-06-16 18:52   ` Tom Lendacky
2017-06-16 18:52   ` Tom Lendacky
2017-06-22 10:57   ` Matt Fleming
2017-06-22 10:57     ` Matt Fleming
2017-06-22 10:57     ` Matt Fleming
2017-06-22 10:57     ` Matt Fleming
2017-06-22 10:57   ` Matt Fleming
2017-06-16 18:52 ` Tom Lendacky
2017-06-16 18:53 ` [PATCH v7 17/36] efi: Update efi_mem_type() to return an error rather than 0 Tom Lendacky
2017-06-16 18:53 ` Tom Lendacky
2017-06-16 18:53   ` Tom Lendacky
2017-06-16 18:53   ` Tom Lendacky
2017-06-16 18:53   ` Tom Lendacky
2017-06-22 10:03   ` Matt Fleming
2017-06-22 10:03     ` Matt Fleming
2017-06-22 10:03     ` Matt Fleming
2017-06-22 10:03     ` Matt Fleming
2017-06-22 10:03   ` Matt Fleming
2017-06-16 18:53 ` [PATCH v7 18/36] x86/efi: Update EFI pagetable creation to work with SME Tom Lendacky
2017-06-16 18:53   ` Tom Lendacky
2017-06-16 18:53   ` Tom Lendacky
2017-06-16 18:53   ` Tom Lendacky
2017-06-22 10:58   ` Matt Fleming
2017-06-22 10:58   ` Matt Fleming
2017-06-22 10:58     ` Matt Fleming
2017-06-22 10:58     ` Matt Fleming
2017-06-22 10:58     ` Matt Fleming
2017-06-16 18:53 ` Tom Lendacky
2017-06-16 18:53 ` [PATCH v7 19/36] x86/mm: Add support to access boot related data in the clear Tom Lendacky
2017-06-16 18:53 ` Tom Lendacky
2017-06-16 18:53   ` Tom Lendacky
2017-06-16 18:53   ` Tom Lendacky
2017-06-16 18:53   ` Tom Lendacky
2017-06-20 20:39   ` Borislav Petkov
2017-06-20 20:39   ` Borislav Petkov
2017-06-20 20:39     ` Borislav Petkov
2017-06-20 20:39     ` Borislav Petkov
2017-06-20 20:39     ` Borislav Petkov
2017-06-22 11:04   ` Matt Fleming
2017-06-22 11:04     ` Matt Fleming
2017-06-22 11:04     ` Matt Fleming
2017-06-22 11:04     ` Matt Fleming
2017-06-22 11:04   ` Matt Fleming
2017-06-16 18:53 ` [PATCH v7 20/36] x86, mpparse: Use memremap to map the mpf and mpc data Tom Lendacky
2017-06-16 18:53   ` Tom Lendacky
2017-06-16 18:53   ` Tom Lendacky
2017-06-21  8:39   ` Borislav Petkov
2017-06-21  8:39     ` Borislav Petkov
2017-06-21  8:39     ` Borislav Petkov
2017-06-21  8:39     ` Borislav Petkov
2017-06-21  8:39   ` Borislav Petkov
2017-06-16 18:53 ` Tom Lendacky
2017-06-16 18:53 ` [PATCH v7 21/36] x86/mm: Add support to access persistent memory in the clear Tom Lendacky
2017-06-16 18:53   ` Tom Lendacky
2017-06-16 18:53   ` Tom Lendacky
2017-06-16 18:53 ` Tom Lendacky
2017-06-16 18:54 ` [PATCH v7 22/36] x86/mm: Add support for changing the memory encryption attribute Tom Lendacky
2017-06-16 18:54 ` Tom Lendacky
2017-06-16 18:54   ` Tom Lendacky
2017-06-16 18:54   ` Tom Lendacky
2017-06-16 18:54 ` [PATCH v7 23/36] x86, realmode: Decrypt trampoline area if memory encryption is active Tom Lendacky
2017-06-16 18:54   ` Tom Lendacky
2017-06-16 18:54   ` Tom Lendacky
2017-06-21  9:43   ` Borislav Petkov
2017-06-21  9:43     ` Borislav Petkov
2017-06-21  9:43     ` Borislav Petkov
2017-06-21  9:43     ` Borislav Petkov
2017-06-21  9:43   ` Borislav Petkov
2017-06-16 18:54 ` Tom Lendacky
2017-06-16 18:54 ` [PATCH v7 24/36] x86, swiotlb: Add memory encryption support Tom Lendacky
2017-06-16 18:54   ` Tom Lendacky
2017-06-16 18:54   ` Tom Lendacky
2017-06-16 18:54   ` Tom Lendacky
2017-06-21  9:54   ` Borislav Petkov
2017-06-21  9:54   ` Borislav Petkov
2017-06-21  9:54     ` Borislav Petkov
2017-06-21  9:54     ` Borislav Petkov
2017-06-21  9:54     ` Borislav Petkov
2017-06-16 18:54 ` Tom Lendacky
2017-06-16 18:54 ` [PATCH v7 25/36] swiotlb: Add warnings for use of bounce buffers with SME Tom Lendacky
2017-06-16 18:54   ` Tom Lendacky
2017-06-16 18:54   ` Tom Lendacky
2017-06-16 18:54   ` Tom Lendacky
2017-06-21 10:50   ` Borislav Petkov
2017-06-21 10:50   ` Borislav Petkov
2017-06-21 10:50     ` Borislav Petkov
2017-06-21 10:50     ` Borislav Petkov
2017-06-21 10:50     ` Borislav Petkov
2017-06-21 15:37     ` Tom Lendacky
2017-06-21 15:37     ` Tom Lendacky
2017-06-21 15:37       ` Tom Lendacky
2017-06-21 15:37       ` Tom Lendacky
2017-06-21 15:37       ` Tom Lendacky
2017-06-16 18:54 ` Tom Lendacky
2017-06-16 18:54 ` [PATCH v7 26/36] x86/CPU/AMD: Make the microcode level available earlier in the boot Tom Lendacky
2017-06-16 18:54   ` Tom Lendacky
2017-06-16 18:54   ` Tom Lendacky
2017-06-22  9:40   ` Borislav Petkov
2017-06-22  9:40   ` Borislav Petkov
2017-06-22  9:40     ` Borislav Petkov
2017-06-22  9:40     ` Borislav Petkov
2017-06-22  9:40     ` Borislav Petkov
2017-06-16 18:54 ` Tom Lendacky
2017-06-16 18:54 ` [PATCH v7 27/36] iommu/amd: Allow the AMD IOMMU to work with memory encryption Tom Lendacky
2017-06-16 18:54 ` Tom Lendacky [this message]
2017-06-16 18:54   ` Tom Lendacky
2017-06-16 18:54   ` Tom Lendacky
2017-06-22 10:56   ` Borislav Petkov
2017-06-22 10:56   ` Borislav Petkov
2017-06-22 10:56     ` Borislav Petkov
2017-06-22 10:56     ` Borislav Petkov
2017-06-22 10:56     ` Borislav Petkov
2017-06-22 16:48     ` Tom Lendacky
2017-06-22 16:48     ` Tom Lendacky
2017-06-22 16:48       ` Tom Lendacky
2017-06-22 16:48       ` Tom Lendacky
2017-06-22 16:48       ` Tom Lendacky
2017-06-16 18:55 ` [PATCH v7 28/36] x86, realmode: Check for memory encryption on the APs Tom Lendacky
2017-06-16 18:55 ` Tom Lendacky
2017-06-16 18:55   ` Tom Lendacky
2017-06-16 18:55   ` Tom Lendacky
2017-06-16 18:55 ` [PATCH v7 29/36] x86, drm, fbdev: Do not specify encrypted memory for video mappings Tom Lendacky
2017-06-16 18:55   ` Tom Lendacky
2017-06-16 18:55   ` Tom Lendacky
2017-06-17 14:08   ` kbuild test robot
2017-06-17 14:08   ` kbuild test robot
2017-06-17 14:08     ` kbuild test robot
2017-06-17 14:08     ` kbuild test robot
2017-06-16 18:55 ` Tom Lendacky
2017-06-16 18:55 ` [PATCH v7 30/36] kvm: x86: svm: Support Secure Memory Encryption within KVM Tom Lendacky
2017-06-16 18:55   ` Tom Lendacky
2017-06-16 18:55   ` Tom Lendacky
2017-06-16 18:55 ` Tom Lendacky
2017-06-16 18:55 ` [PATCH v7 31/36] x86/mm, kexec: Allow kexec to be used with SME Tom Lendacky
2017-06-16 18:55 ` Tom Lendacky
2017-06-16 18:55   ` Tom Lendacky
2017-06-16 18:55   ` Tom Lendacky
2017-06-16 18:55   ` Tom Lendacky
2017-06-23  8:56   ` Borislav Petkov
2017-06-23  8:56   ` Borislav Petkov
2017-06-23  8:56     ` Borislav Petkov
2017-06-23  8:56     ` Borislav Petkov
2017-06-23  8:56     ` Borislav Petkov
2017-06-16 18:55 ` [PATCH v7 32/36] xen/x86: Remove SME feature in PV guests Tom Lendacky
2017-06-16 18:55 ` Tom Lendacky
2017-06-16 18:55   ` Tom Lendacky
2017-06-16 18:55   ` Tom Lendacky
2017-06-16 18:55   ` Tom Lendacky
2017-06-17 10:40   ` Juergen Gross
2017-06-17 10:40     ` Juergen Gross
2017-06-17 10:40     ` Juergen Gross
2017-06-17 10:40     ` Juergen Gross
2017-06-17 10:40   ` Juergen Gross
2017-06-23  9:09   ` Borislav Petkov
2017-06-23  9:09   ` Borislav Petkov
2017-06-23  9:09     ` Borislav Petkov
2017-06-23  9:09     ` Borislav Petkov
2017-06-23  9:09     ` Borislav Petkov
2017-06-16 18:56 ` [PATCH v7 33/36] x86/mm: Use proper encryption attributes with /dev/mem Tom Lendacky
2017-06-16 18:56 ` Tom Lendacky
2017-06-16 18:56   ` Tom Lendacky
2017-06-16 18:56   ` Tom Lendacky
2017-06-16 18:56   ` Tom Lendacky
2017-06-23  9:32   ` Borislav Petkov
2017-06-23  9:32   ` Borislav Petkov
2017-06-23  9:32     ` Borislav Petkov
2017-06-23  9:32     ` Borislav Petkov
2017-06-23  9:32     ` Borislav Petkov
2017-06-16 18:56 ` [PATCH v7 34/36] x86/mm: Add support to encrypt the kernel in-place Tom Lendacky
2017-06-16 18:56   ` Tom Lendacky
2017-06-16 18:56   ` Tom Lendacky
2017-06-16 18:56   ` Tom Lendacky
2017-06-23 10:00   ` Borislav Petkov
2017-06-23 10:00     ` Borislav Petkov
2017-06-23 10:00     ` Borislav Petkov
2017-06-23 10:00     ` Borislav Petkov
2017-06-23 17:44     ` Tom Lendacky
2017-06-23 17:44     ` Tom Lendacky
2017-06-23 17:44       ` Tom Lendacky
2017-06-23 17:44       ` Tom Lendacky
2017-06-23 17:44       ` Tom Lendacky
2017-06-26 15:45       ` Borislav Petkov
2017-06-26 15:45       ` Borislav Petkov
2017-06-26 15:45         ` Borislav Petkov
2017-06-26 15:45         ` Borislav Petkov
2017-06-26 15:45         ` Borislav Petkov
2017-06-26 16:34         ` Tom Lendacky
2017-06-26 16:34           ` Tom Lendacky
2017-06-26 16:34           ` Tom Lendacky
2017-06-26 16:34           ` Tom Lendacky
2017-06-26 16:34         ` Tom Lendacky
2017-06-23 10:00   ` Borislav Petkov
2017-06-16 18:56 ` Tom Lendacky
2017-06-16 18:56 ` [PATCH v7 35/36] x86/boot: Add early cmdline parsing for options with arguments Tom Lendacky
2017-06-16 18:56   ` Tom Lendacky
2017-06-16 18:56   ` Tom Lendacky
2017-06-23 11:57   ` Borislav Petkov
2017-06-23 11:57   ` Borislav Petkov
2017-06-23 11:57     ` Borislav Petkov
2017-06-23 11:57     ` Borislav Petkov
2017-06-23 11:57     ` Borislav Petkov
2017-06-16 18:56 ` Tom Lendacky
2017-06-16 18:56 ` [PATCH v7 36/36] x86/mm: Add support to make use of Secure Memory Encryption Tom Lendacky
2017-06-16 18:56 ` Tom Lendacky
2017-06-16 18:56   ` Tom Lendacky
2017-06-16 18:56   ` Tom Lendacky
2017-06-16 18:56   ` Tom Lendacky
2017-06-23 17:39   ` Borislav Petkov
2017-06-23 17:39     ` Borislav Petkov
2017-06-23 17:39     ` Borislav Petkov
2017-06-23 17:39     ` Borislav Petkov
2017-06-23 17:39   ` Borislav Petkov

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20170616185459.18967.72790.stgit@tlendack-t1.amdoffice.net \
    --to=thomas.lendacky@amd.com \
    --cc=arnd@arndb.de \
    --cc=aryabinin@virtuozzo.com \
    --cc=boris.ostrovsky@oracle.com \
    --cc=bp@alien8.de \
    --cc=brijesh.singh@amd.com \
    --cc=corbet@lwn.net \
    --cc=dvyukov@google.com \
    --cc=dyoung@redhat.com \
    --cc=glider@google.com \
    --cc=hpa@zytor.com \
    --cc=iommu@lists.linux-foundation.org \
    --cc=jgross@suse.com \
    --cc=joro@8bytes.org \
    --cc=kasan-dev@googlegroups.com \
    --cc=kexec@lists.infradead.org \
    --cc=konrad.wilk@oracle.com \
    --cc=kvm@vger.kernel.org \
    --cc=linux-arch@vger.kernel.org \
    --cc=linux-doc@vger.kernel.org \
    --cc=linux-efi@vger.kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=luto@kernel.org \
    --cc=lwoodman@redhat.com \
    --cc=matt@codeblueprint.co.uk \
    --cc=mingo@redhat.com \
    --cc=mst@redhat.com \
    --cc=pbonzini@redhat.com \
    --cc=riel@redhat.com \
    --cc=rkrcmar@redhat.com \
    --cc=tglx@linutronix.de \
    --cc=toshi.kani@hpe.com \
    --cc=x86@kernel.org \
    --cc=xen-devel@lists.xen.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.