From: Lu Baolu <baolu.lu@linux.intel.com> To: David Woodhouse <dwmw2@infradead.org>, Joerg Roedel <joro@8bytes.org>, Bjorn Helgaas <bhelgaas@google.com>, Christoph Hellwig <hch@lst.de> Cc: ashok.raj@intel.com, jacob.jun.pan@intel.com, alan.cox@intel.com, kevin.tian@intel.com, mika.westerberg@linux.intel.com, Ingo Molnar <mingo@redhat.com>, Greg Kroah-Hartman <gregkh@linuxfoundation.org>, pengfei.xu@intel.com, Konrad Rzeszutek Wilk <konrad.wilk@oracle.com>, Marek Szyprowski <m.szyprowski@samsung.com>, Robin Murphy <robin.murphy@arm.com>, Jonathan Corbet <corbet@lwn.net>, Boris Ostrovsky <boris.ostrovsky@oracle.com>, Juergen Gross <jgross@suse.com>, Stefano Stabellini <sstabellini@kernel.org>, Steven Rostedt <rostedt@goodmis.org>, iommu@lists.linux-foundation.org, linux-kernel@vger.kernel.org, Lu Baolu <baolu.lu@linux.intel.com>, Jacob Pan <jacob.jun.pan@linux.intel.com>, Alan Cox <alan@linux.intel.com>, Mika Westerberg <mika.westerberg@intel.com> Subject: [PATCH v6 5/8] iommu: Add bounce page APIs Date: Tue, 30 Jul 2019 12:52:26 +0800 [thread overview] Message-ID: <20190730045229.3826-6-baolu.lu@linux.intel.com> (raw) In-Reply-To: <20190730045229.3826-1-baolu.lu@linux.intel.com> IOMMU hardware always use paging for DMA remapping. The minimum mapped window is a page size. The device drivers may map buffers not filling whole IOMMU window. It allows device to access to possibly unrelated memory and various malicious devices can exploit this to perform DMA attack. This introduces the bouce buffer mechanism for DMA buffers which doesn't fill a minimal IOMMU page. It could be used by various vendor specific IOMMU drivers as long as the DMA domain is managed by the generic IOMMU layer. Below APIs are added: * iommu_bounce_map(dev, addr, paddr, size, dir, attrs) - Map a buffer start at DMA address @addr in bounce page manner. For buffer parts that doesn't cross a whole minimal IOMMU page, the bounce page policy is applied. A bounce page mapped by swiotlb will be used as the DMA target in the IOMMU page table. Otherwise, the physical address @paddr is mapped instead. * iommu_bounce_unmap(dev, addr, size, dir, attrs) - Unmap the buffer mapped with iommu_bounce_map(). The bounce page will be torn down after the bounced data get synced. * iommu_bounce_sync(dev, addr, size, dir, target) - Synce the bounced data in case the bounce mapped buffer is reused. The whole APIs are included within a kernel option IOMMU_BOUNCE_PAGE. It's useful for cases where bounce page doesn't needed, for example, embedded cases. Cc: Ashok Raj <ashok.raj@intel.com> Cc: Jacob Pan <jacob.jun.pan@linux.intel.com> Cc: Kevin Tian <kevin.tian@intel.com> Cc: Alan Cox <alan@linux.intel.com> Cc: Mika Westerberg <mika.westerberg@intel.com> Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com> --- drivers/iommu/Kconfig | 13 +++++ drivers/iommu/iommu.c | 118 ++++++++++++++++++++++++++++++++++++++++++ include/linux/iommu.h | 35 +++++++++++++ 3 files changed, 166 insertions(+) diff --git a/drivers/iommu/Kconfig b/drivers/iommu/Kconfig index e15cdcd8cb3c..d7f2e09cbcf2 100644 --- a/drivers/iommu/Kconfig +++ b/drivers/iommu/Kconfig @@ -86,6 +86,19 @@ config IOMMU_DEFAULT_PASSTHROUGH If unsure, say N here. +config IOMMU_BOUNCE_PAGE + bool "Use bounce page for untrusted devices" + depends on IOMMU_API && SWIOTLB + help + IOMMU hardware always use paging for DMA remapping. The minimum + mapped window is a page size. The device drivers may map buffers + not filling whole IOMMU window. This allows device to access to + possibly unrelated memory and malicious device can exploit this + to perform a DMA attack. Select this to use a bounce page for the + buffer which doesn't fill a whole IOMU page. + + If unsure, say N here. + config OF_IOMMU def_bool y depends on OF && IOMMU_API diff --git a/drivers/iommu/iommu.c b/drivers/iommu/iommu.c index 0c674d80c37f..fe3815186d72 100644 --- a/drivers/iommu/iommu.c +++ b/drivers/iommu/iommu.c @@ -2468,3 +2468,121 @@ int iommu_sva_get_pasid(struct iommu_sva *handle) return ops->sva_get_pasid(handle); } EXPORT_SYMBOL_GPL(iommu_sva_get_pasid); + +#ifdef CONFIG_IOMMU_BOUNCE_PAGE + +/* + * Bounce buffer support for external devices: + * + * IOMMU hardware always use paging for DMA remapping. The minimum mapped + * window is a page size. The device drivers may map buffers not filling + * whole IOMMU window. This allows device to access to possibly unrelated + * memory and malicious device can exploit this to perform a DMA attack. + * Use bounce pages for the buffer which doesn't fill whole IOMMU pages. + */ + +static inline size_t +get_aligned_size(struct iommu_domain *domain, size_t size) +{ + return ALIGN(size, 1 << __ffs(domain->pgsize_bitmap)); +} + +dma_addr_t iommu_bounce_map(struct device *dev, dma_addr_t iova, + phys_addr_t paddr, size_t size, + enum dma_data_direction dir, + unsigned long attrs) +{ + struct iommu_domain *domain; + unsigned int min_pagesz; + phys_addr_t tlb_addr; + size_t aligned_size; + int prot = 0; + int ret; + + domain = iommu_get_dma_domain(dev); + if (!domain) + return DMA_MAPPING_ERROR; + + if (dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL) + prot |= IOMMU_READ; + if (dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL) + prot |= IOMMU_WRITE; + + aligned_size = get_aligned_size(domain, size); + min_pagesz = 1 << __ffs(domain->pgsize_bitmap); + + /* + * If both the physical buffer start address and size are + * page aligned, we don't need to use a bounce page. + */ + if (!IS_ALIGNED(paddr | size, min_pagesz)) { + tlb_addr = swiotlb_tbl_map_single(dev, + __phys_to_dma(dev, io_tlb_start), + paddr, size, aligned_size, dir, attrs); + if (tlb_addr == DMA_MAPPING_ERROR) + return DMA_MAPPING_ERROR; + } else { + tlb_addr = paddr; + } + + ret = iommu_map(domain, iova, tlb_addr, aligned_size, prot); + if (ret) { + if (is_swiotlb_buffer(tlb_addr)) + swiotlb_tbl_unmap_single(dev, tlb_addr, size, + aligned_size, dir, attrs); + + return DMA_MAPPING_ERROR; + } + + return iova; +} +EXPORT_SYMBOL_GPL(iommu_bounce_map); + +static inline phys_addr_t +iova_to_tlb_addr(struct iommu_domain *domain, dma_addr_t addr) +{ + if (unlikely(!domain->ops || !domain->ops->iova_to_phys)) + return 0; + + return domain->ops->iova_to_phys(domain, addr); +} + +void iommu_bounce_unmap(struct device *dev, dma_addr_t iova, size_t size, + enum dma_data_direction dir, unsigned long attrs) +{ + struct iommu_domain *domain; + phys_addr_t tlb_addr; + size_t aligned_size; + + domain = iommu_get_dma_domain(dev); + if (WARN_ON(!domain)) + return; + + aligned_size = get_aligned_size(domain, size); + tlb_addr = iova_to_tlb_addr(domain, iova); + if (WARN_ON(!tlb_addr)) + return; + + iommu_unmap(domain, iova, aligned_size); + if (is_swiotlb_buffer(tlb_addr)) + swiotlb_tbl_unmap_single(dev, tlb_addr, size, + aligned_size, dir, attrs); +} +EXPORT_SYMBOL_GPL(iommu_bounce_unmap); + +void iommu_bounce_sync(struct device *dev, dma_addr_t addr, size_t size, + enum dma_data_direction dir, enum dma_sync_target target) +{ + struct iommu_domain *domain; + phys_addr_t tlb_addr; + + domain = iommu_get_dma_domain(dev); + if (WARN_ON(!domain)) + return; + + tlb_addr = iova_to_tlb_addr(domain, addr); + if (is_swiotlb_buffer(tlb_addr)) + swiotlb_tbl_sync_single(dev, tlb_addr, size, dir, target); +} +EXPORT_SYMBOL_GPL(iommu_bounce_sync); +#endif /* CONFIG_IOMMU_BOUNCE_PAGE */ diff --git a/include/linux/iommu.h b/include/linux/iommu.h index fdc355ccc570..5569b84cc9be 100644 --- a/include/linux/iommu.h +++ b/include/linux/iommu.h @@ -14,6 +14,8 @@ #include <linux/err.h> #include <linux/of.h> #include <uapi/linux/iommu.h> +#include <linux/swiotlb.h> +#include <linux/dma-direct.h> #define IOMMU_READ (1 << 0) #define IOMMU_WRITE (1 << 1) @@ -560,6 +562,39 @@ int iommu_sva_set_ops(struct iommu_sva *handle, const struct iommu_sva_ops *ops); int iommu_sva_get_pasid(struct iommu_sva *handle); +#ifdef CONFIG_IOMMU_BOUNCE_PAGE +dma_addr_t iommu_bounce_map(struct device *dev, dma_addr_t iova, + phys_addr_t paddr, size_t size, + enum dma_data_direction dir, + unsigned long attrs); +void iommu_bounce_unmap(struct device *dev, dma_addr_t iova, size_t size, + enum dma_data_direction dir, unsigned long attrs); +void iommu_bounce_sync(struct device *dev, dma_addr_t addr, size_t size, + enum dma_data_direction dir, + enum dma_sync_target target); +#else +static inline +dma_addr_t iommu_bounce_map(struct device *dev, dma_addr_t iova, + phys_addr_t paddr, size_t size, + enum dma_data_direction dir, + unsigned long attrs) +{ + return DMA_MAPPING_ERROR; +} + +static inline +void iommu_bounce_unmap(struct device *dev, dma_addr_t iova, size_t size, + enum dma_data_direction dir, unsigned long attrs) +{ +} + +static inline +void iommu_bounce_sync(struct device *dev, dma_addr_t addr, size_t size, + enum dma_data_direction dir, enum dma_sync_target target) +{ +} +#endif /* CONFIG_IOMMU_BOUNCE_PAGE */ + #else /* CONFIG_IOMMU_API */ struct iommu_ops {}; -- 2.17.1
WARNING: multiple messages have this Message-ID (diff)
From: Lu Baolu <baolu.lu@linux.intel.com> To: David Woodhouse <dwmw2@infradead.org>, Joerg Roedel <joro@8bytes.org>, Bjorn Helgaas <bhelgaas@google.com>, Christoph Hellwig <hch@lst.de> Cc: alan.cox@intel.com, Stefano Stabellini <sstabellini@kernel.org>, ashok.raj@intel.com, Jonathan Corbet <corbet@lwn.net>, pengfei.xu@intel.com, Ingo Molnar <mingo@redhat.com>, kevin.tian@intel.com, Konrad Rzeszutek Wilk <konrad.wilk@oracle.com>, Steven Rostedt <rostedt@goodmis.org>, Boris Ostrovsky <boris.ostrovsky@oracle.com>, mika.westerberg@linux.intel.com, Alan Cox <alan@linux.intel.com>, Juergen Gross <jgross@suse.com>, Greg Kroah-Hartman <gregkh@linuxfoundation.org>, Mika Westerberg <mika.westerberg@intel.com>, linux-kernel@vger.kernel.org, iommu@lists.linux-foundation.org, jacob.jun.pan@intel.com, Robin Murphy <robin.murphy@arm.com> Subject: [PATCH v6 5/8] iommu: Add bounce page APIs Date: Tue, 30 Jul 2019 12:52:26 +0800 [thread overview] Message-ID: <20190730045229.3826-6-baolu.lu@linux.intel.com> (raw) In-Reply-To: <20190730045229.3826-1-baolu.lu@linux.intel.com> IOMMU hardware always use paging for DMA remapping. The minimum mapped window is a page size. The device drivers may map buffers not filling whole IOMMU window. It allows device to access to possibly unrelated memory and various malicious devices can exploit this to perform DMA attack. This introduces the bouce buffer mechanism for DMA buffers which doesn't fill a minimal IOMMU page. It could be used by various vendor specific IOMMU drivers as long as the DMA domain is managed by the generic IOMMU layer. Below APIs are added: * iommu_bounce_map(dev, addr, paddr, size, dir, attrs) - Map a buffer start at DMA address @addr in bounce page manner. For buffer parts that doesn't cross a whole minimal IOMMU page, the bounce page policy is applied. A bounce page mapped by swiotlb will be used as the DMA target in the IOMMU page table. Otherwise, the physical address @paddr is mapped instead. * iommu_bounce_unmap(dev, addr, size, dir, attrs) - Unmap the buffer mapped with iommu_bounce_map(). The bounce page will be torn down after the bounced data get synced. * iommu_bounce_sync(dev, addr, size, dir, target) - Synce the bounced data in case the bounce mapped buffer is reused. The whole APIs are included within a kernel option IOMMU_BOUNCE_PAGE. It's useful for cases where bounce page doesn't needed, for example, embedded cases. Cc: Ashok Raj <ashok.raj@intel.com> Cc: Jacob Pan <jacob.jun.pan@linux.intel.com> Cc: Kevin Tian <kevin.tian@intel.com> Cc: Alan Cox <alan@linux.intel.com> Cc: Mika Westerberg <mika.westerberg@intel.com> Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com> --- drivers/iommu/Kconfig | 13 +++++ drivers/iommu/iommu.c | 118 ++++++++++++++++++++++++++++++++++++++++++ include/linux/iommu.h | 35 +++++++++++++ 3 files changed, 166 insertions(+) diff --git a/drivers/iommu/Kconfig b/drivers/iommu/Kconfig index e15cdcd8cb3c..d7f2e09cbcf2 100644 --- a/drivers/iommu/Kconfig +++ b/drivers/iommu/Kconfig @@ -86,6 +86,19 @@ config IOMMU_DEFAULT_PASSTHROUGH If unsure, say N here. +config IOMMU_BOUNCE_PAGE + bool "Use bounce page for untrusted devices" + depends on IOMMU_API && SWIOTLB + help + IOMMU hardware always use paging for DMA remapping. The minimum + mapped window is a page size. The device drivers may map buffers + not filling whole IOMMU window. This allows device to access to + possibly unrelated memory and malicious device can exploit this + to perform a DMA attack. Select this to use a bounce page for the + buffer which doesn't fill a whole IOMU page. + + If unsure, say N here. + config OF_IOMMU def_bool y depends on OF && IOMMU_API diff --git a/drivers/iommu/iommu.c b/drivers/iommu/iommu.c index 0c674d80c37f..fe3815186d72 100644 --- a/drivers/iommu/iommu.c +++ b/drivers/iommu/iommu.c @@ -2468,3 +2468,121 @@ int iommu_sva_get_pasid(struct iommu_sva *handle) return ops->sva_get_pasid(handle); } EXPORT_SYMBOL_GPL(iommu_sva_get_pasid); + +#ifdef CONFIG_IOMMU_BOUNCE_PAGE + +/* + * Bounce buffer support for external devices: + * + * IOMMU hardware always use paging for DMA remapping. The minimum mapped + * window is a page size. The device drivers may map buffers not filling + * whole IOMMU window. This allows device to access to possibly unrelated + * memory and malicious device can exploit this to perform a DMA attack. + * Use bounce pages for the buffer which doesn't fill whole IOMMU pages. + */ + +static inline size_t +get_aligned_size(struct iommu_domain *domain, size_t size) +{ + return ALIGN(size, 1 << __ffs(domain->pgsize_bitmap)); +} + +dma_addr_t iommu_bounce_map(struct device *dev, dma_addr_t iova, + phys_addr_t paddr, size_t size, + enum dma_data_direction dir, + unsigned long attrs) +{ + struct iommu_domain *domain; + unsigned int min_pagesz; + phys_addr_t tlb_addr; + size_t aligned_size; + int prot = 0; + int ret; + + domain = iommu_get_dma_domain(dev); + if (!domain) + return DMA_MAPPING_ERROR; + + if (dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL) + prot |= IOMMU_READ; + if (dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL) + prot |= IOMMU_WRITE; + + aligned_size = get_aligned_size(domain, size); + min_pagesz = 1 << __ffs(domain->pgsize_bitmap); + + /* + * If both the physical buffer start address and size are + * page aligned, we don't need to use a bounce page. + */ + if (!IS_ALIGNED(paddr | size, min_pagesz)) { + tlb_addr = swiotlb_tbl_map_single(dev, + __phys_to_dma(dev, io_tlb_start), + paddr, size, aligned_size, dir, attrs); + if (tlb_addr == DMA_MAPPING_ERROR) + return DMA_MAPPING_ERROR; + } else { + tlb_addr = paddr; + } + + ret = iommu_map(domain, iova, tlb_addr, aligned_size, prot); + if (ret) { + if (is_swiotlb_buffer(tlb_addr)) + swiotlb_tbl_unmap_single(dev, tlb_addr, size, + aligned_size, dir, attrs); + + return DMA_MAPPING_ERROR; + } + + return iova; +} +EXPORT_SYMBOL_GPL(iommu_bounce_map); + +static inline phys_addr_t +iova_to_tlb_addr(struct iommu_domain *domain, dma_addr_t addr) +{ + if (unlikely(!domain->ops || !domain->ops->iova_to_phys)) + return 0; + + return domain->ops->iova_to_phys(domain, addr); +} + +void iommu_bounce_unmap(struct device *dev, dma_addr_t iova, size_t size, + enum dma_data_direction dir, unsigned long attrs) +{ + struct iommu_domain *domain; + phys_addr_t tlb_addr; + size_t aligned_size; + + domain = iommu_get_dma_domain(dev); + if (WARN_ON(!domain)) + return; + + aligned_size = get_aligned_size(domain, size); + tlb_addr = iova_to_tlb_addr(domain, iova); + if (WARN_ON(!tlb_addr)) + return; + + iommu_unmap(domain, iova, aligned_size); + if (is_swiotlb_buffer(tlb_addr)) + swiotlb_tbl_unmap_single(dev, tlb_addr, size, + aligned_size, dir, attrs); +} +EXPORT_SYMBOL_GPL(iommu_bounce_unmap); + +void iommu_bounce_sync(struct device *dev, dma_addr_t addr, size_t size, + enum dma_data_direction dir, enum dma_sync_target target) +{ + struct iommu_domain *domain; + phys_addr_t tlb_addr; + + domain = iommu_get_dma_domain(dev); + if (WARN_ON(!domain)) + return; + + tlb_addr = iova_to_tlb_addr(domain, addr); + if (is_swiotlb_buffer(tlb_addr)) + swiotlb_tbl_sync_single(dev, tlb_addr, size, dir, target); +} +EXPORT_SYMBOL_GPL(iommu_bounce_sync); +#endif /* CONFIG_IOMMU_BOUNCE_PAGE */ diff --git a/include/linux/iommu.h b/include/linux/iommu.h index fdc355ccc570..5569b84cc9be 100644 --- a/include/linux/iommu.h +++ b/include/linux/iommu.h @@ -14,6 +14,8 @@ #include <linux/err.h> #include <linux/of.h> #include <uapi/linux/iommu.h> +#include <linux/swiotlb.h> +#include <linux/dma-direct.h> #define IOMMU_READ (1 << 0) #define IOMMU_WRITE (1 << 1) @@ -560,6 +562,39 @@ int iommu_sva_set_ops(struct iommu_sva *handle, const struct iommu_sva_ops *ops); int iommu_sva_get_pasid(struct iommu_sva *handle); +#ifdef CONFIG_IOMMU_BOUNCE_PAGE +dma_addr_t iommu_bounce_map(struct device *dev, dma_addr_t iova, + phys_addr_t paddr, size_t size, + enum dma_data_direction dir, + unsigned long attrs); +void iommu_bounce_unmap(struct device *dev, dma_addr_t iova, size_t size, + enum dma_data_direction dir, unsigned long attrs); +void iommu_bounce_sync(struct device *dev, dma_addr_t addr, size_t size, + enum dma_data_direction dir, + enum dma_sync_target target); +#else +static inline +dma_addr_t iommu_bounce_map(struct device *dev, dma_addr_t iova, + phys_addr_t paddr, size_t size, + enum dma_data_direction dir, + unsigned long attrs) +{ + return DMA_MAPPING_ERROR; +} + +static inline +void iommu_bounce_unmap(struct device *dev, dma_addr_t iova, size_t size, + enum dma_data_direction dir, unsigned long attrs) +{ +} + +static inline +void iommu_bounce_sync(struct device *dev, dma_addr_t addr, size_t size, + enum dma_data_direction dir, enum dma_sync_target target) +{ +} +#endif /* CONFIG_IOMMU_BOUNCE_PAGE */ + #else /* CONFIG_IOMMU_API */ struct iommu_ops {}; -- 2.17.1 _______________________________________________ iommu mailing list iommu@lists.linux-foundation.org https://lists.linuxfoundation.org/mailman/listinfo/iommu
next prev parent reply other threads:[~2019-07-30 4:53 UTC|newest] Thread overview: 30+ messages / expand[flat|nested] mbox.gz Atom feed top 2019-07-30 4:52 [PATCH v6 0/8] iommu: Bounce page for untrusted devices Lu Baolu 2019-07-30 4:52 ` Lu Baolu 2019-07-30 4:52 ` [PATCH v6 1/8] iommu/vt-d: Don't switch off swiotlb if use direct dma Lu Baolu 2019-07-30 4:52 ` Lu Baolu 2019-07-30 4:52 ` [PATCH v6 2/8] PCI: Add dev_is_untrusted helper Lu Baolu 2019-07-30 4:52 ` Lu Baolu 2019-07-30 4:52 ` [PATCH v6 3/8] swiotlb: Split size parameter to map/unmap APIs Lu Baolu 2019-07-30 4:52 ` Lu Baolu 2019-07-30 4:52 ` [PATCH v6 4/8] swiotlb: Zero out bounce buffer for untrusted device Lu Baolu 2019-07-30 4:52 ` Lu Baolu 2019-07-30 4:52 ` Lu Baolu [this message] 2019-07-30 4:52 ` [PATCH v6 5/8] iommu: Add bounce page APIs Lu Baolu 2019-08-14 8:38 ` Joerg Roedel 2019-08-14 8:38 ` Joerg Roedel 2019-08-15 6:15 ` Lu Baolu 2019-08-15 6:15 ` Lu Baolu 2019-08-15 15:48 ` Joerg Roedel 2019-08-15 15:48 ` Joerg Roedel 2019-08-16 2:45 ` Lu Baolu 2019-08-16 2:45 ` Lu Baolu 2019-08-16 4:46 ` Christoph Hellwig 2019-08-16 4:46 ` Christoph Hellwig 2019-08-18 3:07 ` Lu Baolu 2019-08-18 3:07 ` Lu Baolu 2019-07-30 4:52 ` [PATCH v6 6/8] iommu/vt-d: Check whether device requires bounce buffer Lu Baolu 2019-07-30 4:52 ` Lu Baolu 2019-07-30 4:52 ` [PATCH v6 7/8] iommu/vt-d: Add trace events for device dma map/unmap Lu Baolu 2019-07-30 4:52 ` Lu Baolu 2019-07-30 4:52 ` [PATCH v6 8/8] iommu/vt-d: Use bounce buffer for untrusted devices Lu Baolu 2019-07-30 4:52 ` Lu Baolu
Reply instructions: You may reply publicly to this message via plain-text email using any one of the following methods: * Save the following mbox file, import it into your mail client, and reply-to-all from there: mbox Avoid top-posting and favor interleaved quoting: https://en.wikipedia.org/wiki/Posting_style#Interleaved_style * Reply using the --to, --cc, and --in-reply-to switches of git-send-email(1): git send-email \ --in-reply-to=20190730045229.3826-6-baolu.lu@linux.intel.com \ --to=baolu.lu@linux.intel.com \ --cc=alan.cox@intel.com \ --cc=alan@linux.intel.com \ --cc=ashok.raj@intel.com \ --cc=bhelgaas@google.com \ --cc=boris.ostrovsky@oracle.com \ --cc=corbet@lwn.net \ --cc=dwmw2@infradead.org \ --cc=gregkh@linuxfoundation.org \ --cc=hch@lst.de \ --cc=iommu@lists.linux-foundation.org \ --cc=jacob.jun.pan@intel.com \ --cc=jacob.jun.pan@linux.intel.com \ --cc=jgross@suse.com \ --cc=joro@8bytes.org \ --cc=kevin.tian@intel.com \ --cc=konrad.wilk@oracle.com \ --cc=linux-kernel@vger.kernel.org \ --cc=m.szyprowski@samsung.com \ --cc=mika.westerberg@intel.com \ --cc=mika.westerberg@linux.intel.com \ --cc=mingo@redhat.com \ --cc=pengfei.xu@intel.com \ --cc=robin.murphy@arm.com \ --cc=rostedt@goodmis.org \ --cc=sstabellini@kernel.org \ /path/to/YOUR_REPLY https://kernel.org/pub/software/scm/git/docs/git-send-email.html * If your mail client supports setting the In-Reply-To header via mailto: links, try the mailto: linkBe sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes, see mirroring instructions on how to clone and mirror all data and code used by this external index.