From: Dave Jiang <dave.jiang@intel.com>
To: vkoul@kernel.org, megha.dey@linux.intel.com, maz@kernel.org,
bhelgaas@google.com, rafael@kernel.org,
gregkh@linuxfoundation.org, tglx@linutronix.de, hpa@zytor.com,
alex.williamson@redhat.com, jacob.jun.pan@intel.com,
ashok.raj@intel.com, jgg@mellanox.com, yi.l.liu@intel.com,
baolu.lu@intel.com, kevin.tian@intel.com,
sanjay.k.kumar@intel.com, tony.luck@intel.com,
jing.lin@intel.com, dan.j.williams@intel.com,
kwankhede@nvidia.com, eric.auger@redhat.com, parav@mellanox.com
Cc: dmaengine@vger.kernel.org, linux-kernel@vger.kernel.org,
x86@kernel.org, linux-pci@vger.kernel.org, kvm@vger.kernel.org
Subject: [PATCH RFC 03/15] drivers/base: Allocate/free platform-msi interrupts by group
Date: Tue, 21 Apr 2020 16:34:05 -0700 [thread overview]
Message-ID: <158751204550.36773.459505651659406529.stgit@djiang5-desk3.ch.intel.com> (raw)
In-Reply-To: <158751095889.36773.6009825070990637468.stgit@djiang5-desk3.ch.intel.com>
From: Megha Dey <megha.dey@linux.intel.com>
This is a preparatory patch to introduce Interrupt Message Store (IMS).
Dynamic allocation is IMS vectors is a requirement for devices which
support Scalable I/o virtualization. A driver can allocate and free
vectors not just once during probe (as was the case with the MSI/MSI-X)
but also in the post probe phase where actual demand is available.
Thus, introduce an API, platform_msi_domain_alloc_irqs_group() which
drivers using IMS would be able to call multiple times. The vectors
allocated each time this API is called are associated with a group ID,
starting from 1. To free the vectors associated with a particular group,
platform_msi_domain_free_irqs_group() API can be called.
The existing drivers using platform-msi infrastructure will continue to
use the existing alloc (platform_msi_domain_alloc_irqs) and free
(platform_msi_domain_free_irqs) APIs and are assigned a default group 0.
Signed-off-by: Megha Dey <megha.dey@linux.intel.com>
---
drivers/base/platform-msi.c | 131 ++++++++++++++++++++++++++++++++-----------
include/linux/device.h | 1
include/linux/msi.h | 47 +++++++++++----
kernel/irq/msi.c | 43 +++++++++++---
4 files changed, 169 insertions(+), 53 deletions(-)
diff --git a/drivers/base/platform-msi.c b/drivers/base/platform-msi.c
index b25c52f734dc..2696aa75983b 100644
--- a/drivers/base/platform-msi.c
+++ b/drivers/base/platform-msi.c
@@ -106,16 +106,28 @@ static void platform_msi_update_chip_ops(struct msi_domain_info *info)
info->flags &= ~MSI_FLAG_LEVEL_CAPABLE;
}
-static void platform_msi_free_descs(struct device *dev, int base, int nvec)
+static void platform_msi_free_descs(struct device *dev, int base, int nvec,
+ unsigned int group)
{
struct msi_desc *desc, *tmp;
-
- list_for_each_entry_safe(desc, tmp, dev_to_platform_msi_list(dev),
- list) {
- if (desc->platform.msi_index >= base &&
- desc->platform.msi_index < (base + nvec)) {
- list_del(&desc->list);
- free_msi_entry(desc);
+ struct platform_msi_group_entry *platform_msi_group,
+ *tmp_platform_msi_group;
+
+ list_for_each_entry_safe(platform_msi_group, tmp_platform_msi_group,
+ dev_to_platform_msi_group_list(dev),
+ group_list) {
+ if (platform_msi_group->group_id == group) {
+ list_for_each_entry_safe(desc, tmp,
+ &platform_msi_group->entry_list,
+ list) {
+ if (desc->platform.msi_index >= base &&
+ desc->platform.msi_index < (base + nvec)) {
+ list_del(&desc->list);
+ free_msi_entry(desc);
+ }
+ }
+ list_del(&platform_msi_group->group_list);
+ kfree(platform_msi_group);
}
}
}
@@ -128,8 +140,8 @@ static int platform_msi_alloc_descs_with_irq(struct device *dev, int virq,
struct msi_desc *desc;
int i, base = 0;
- if (!list_empty(dev_to_platform_msi_list(dev))) {
- desc = list_last_entry(dev_to_platform_msi_list(dev),
+ if (!list_empty(platform_msi_current_group_entry_list(dev))) {
+ desc = list_last_entry(platform_msi_current_group_entry_list(dev),
struct msi_desc, list);
base = desc->platform.msi_index + 1;
}
@@ -143,12 +155,13 @@ static int platform_msi_alloc_descs_with_irq(struct device *dev, int virq,
desc->platform.msi_index = base + i;
desc->irq = virq ? virq + i : 0;
- list_add_tail(&desc->list, dev_to_platform_msi_list(dev));
+ list_add_tail(&desc->list,
+ platform_msi_current_group_entry_list(dev));
}
if (i != nvec) {
/* Clean up the mess */
- platform_msi_free_descs(dev, base, nvec);
+ platform_msi_free_descs(dev, base, nvec, dev->group_id);
return -ENOMEM;
}
@@ -214,7 +227,7 @@ platform_msi_alloc_priv_data(struct device *dev, unsigned int nvec,
}
/* Already had a helping of MSI? Greed... */
- if (!list_empty(dev_to_platform_msi_list(dev)))
+ if (!list_empty(platform_msi_current_group_entry_list(dev)))
return ERR_PTR(-EBUSY);
datap = kzalloc(sizeof(*datap), GFP_KERNEL);
@@ -253,11 +266,36 @@ static void platform_msi_free_priv_data(struct platform_msi_priv_data *data)
int platform_msi_domain_alloc_irqs(struct device *dev, unsigned int nvec,
const struct platform_msi_ops *platform_ops)
{
+ return platform_msi_domain_alloc_irqs_group(dev, nvec, platform_ops,
+ NULL);
+}
+EXPORT_SYMBOL_GPL(platform_msi_domain_alloc_irqs);
+
+int platform_msi_domain_alloc_irqs_group(struct device *dev, unsigned int nvec,
+ const struct platform_msi_ops *platform_ops,
+ unsigned int *group_id)
+{
+ struct platform_msi_group_entry *platform_msi_group;
struct platform_msi_priv_data *priv_data;
int err;
dev->platform_msi_type = GEN_PLAT_MSI;
+ if (group_id)
+ *group_id = ++dev->group_id;
+
+ platform_msi_group = kzalloc(sizeof(*platform_msi_group), GFP_KERNEL);
+ if (!platform_msi_group) {
+ err = -ENOMEM;
+ goto out_platform_msi_group;
+ }
+
+ INIT_LIST_HEAD(&platform_msi_group->group_list);
+ INIT_LIST_HEAD(&platform_msi_group->entry_list);
+ platform_msi_group->group_id = dev->group_id;
+ list_add_tail(&platform_msi_group->group_list,
+ dev_to_platform_msi_group_list(dev));
+
priv_data = platform_msi_alloc_priv_data(dev, nvec, platform_ops);
if (IS_ERR(priv_data))
return PTR_ERR(priv_data);
@@ -273,13 +311,14 @@ int platform_msi_domain_alloc_irqs(struct device *dev, unsigned int nvec,
return 0;
out_free_desc:
- platform_msi_free_descs(dev, 0, nvec);
+ platform_msi_free_descs(dev, 0, nvec, dev->group_id);
out_free_priv_data:
platform_msi_free_priv_data(priv_data);
-
+out_platform_msi_group:
+ kfree(platform_msi_group);
return err;
}
-EXPORT_SYMBOL_GPL(platform_msi_domain_alloc_irqs);
+EXPORT_SYMBOL_GPL(platform_msi_domain_alloc_irqs_group);
/**
* platform_msi_domain_free_irqs - Free MSI interrupts for @dev
@@ -287,17 +326,30 @@ EXPORT_SYMBOL_GPL(platform_msi_domain_alloc_irqs);
*/
void platform_msi_domain_free_irqs(struct device *dev)
{
- if (!list_empty(dev_to_platform_msi_list(dev))) {
- struct msi_desc *desc;
+ platform_msi_domain_free_irqs_group(dev, 0);
+}
+EXPORT_SYMBOL_GPL(platform_msi_domain_free_irqs);
- desc = first_platform_msi_entry(dev);
- platform_msi_free_priv_data(desc->platform.msi_priv_data);
+void platform_msi_domain_free_irqs_group(struct device *dev, unsigned int group)
+{
+ struct platform_msi_group_entry *platform_msi_group;
+
+ list_for_each_entry(platform_msi_group,
+ dev_to_platform_msi_group_list((dev)), group_list) {
+ if (platform_msi_group->group_id == group) {
+ if (!list_empty(&platform_msi_group->entry_list)) {
+ struct msi_desc *desc;
+
+ desc = list_first_entry(&(platform_msi_group)->entry_list,
+ struct msi_desc, list);
+ platform_msi_free_priv_data(desc->platform.msi_priv_data);
+ }
+ }
}
-
- msi_domain_free_irqs(dev->msi_domain, dev);
- platform_msi_free_descs(dev, 0, MAX_DEV_MSIS);
+ msi_domain_free_irqs_group(dev->msi_domain, dev, group);
+ platform_msi_free_descs(dev, 0, MAX_DEV_MSIS, group);
}
-EXPORT_SYMBOL_GPL(platform_msi_domain_free_irqs);
+EXPORT_SYMBOL_GPL(platform_msi_domain_free_irqs_group);
/**
* platform_msi_get_host_data - Query the private data associated with
@@ -373,15 +425,28 @@ void platform_msi_domain_free(struct irq_domain *domain, unsigned int virq,
{
struct platform_msi_priv_data *data = domain->host_data;
struct msi_desc *desc, *tmp;
- for_each_platform_msi_entry_safe(desc, tmp, data->dev) {
- if (WARN_ON(!desc->irq || desc->nvec_used != 1))
- return;
- if (!(desc->irq >= virq && desc->irq < (virq + nvec)))
- continue;
-
- irq_domain_free_irqs_common(domain, desc->irq, 1);
- list_del(&desc->list);
- free_msi_entry(desc);
+ struct platform_msi_group_entry *platform_msi_group,
+ *tmp_platform_msi_group;
+
+ list_for_each_entry_safe(platform_msi_group, tmp_platform_msi_group,
+ dev_to_platform_msi_group_list(data->dev),
+ group_list) {
+ if (platform_msi_group->group_id == data->dev->group_id) {
+ list_for_each_entry_safe(desc, tmp,
+ &platform_msi_group->entry_list,
+ list) {
+ if (WARN_ON(!desc->irq || desc->nvec_used != 1))
+ return;
+ if (!(desc->irq >= virq && desc->irq < (virq + nvec)))
+ continue;
+
+ irq_domain_free_irqs_common(domain, desc->irq, 1);
+ list_del(&desc->list);
+ free_msi_entry(desc);
+ }
+ list_del(&platform_msi_group->group_list);
+ kfree(platform_msi_group);
+ }
}
}
diff --git a/include/linux/device.h b/include/linux/device.h
index cbcecb14584e..f6700b85eb95 100644
--- a/include/linux/device.h
+++ b/include/linux/device.h
@@ -624,6 +624,7 @@ struct device {
defined(CONFIG_ARCH_HAS_SYNC_DMA_FOR_CPU_ALL)
bool dma_coherent:1;
#endif
+ unsigned int group_id;
};
static inline struct device *kobj_to_dev(struct kobject *kobj)
diff --git a/include/linux/msi.h b/include/linux/msi.h
index 9c15b7403694..3890b143b04d 100644
--- a/include/linux/msi.h
+++ b/include/linux/msi.h
@@ -135,6 +135,12 @@ enum platform_msi_type {
GEN_PLAT_MSI = 1,
};
+struct platform_msi_group_entry {
+ unsigned int group_id;
+ struct list_head group_list;
+ struct list_head entry_list;
+};
+
/* Helpers to hide struct msi_desc implementation details */
#define msi_desc_to_dev(desc) ((desc)->dev)
#define dev_to_msi_list(dev) (&(dev)->msi_list)
@@ -145,21 +151,31 @@ enum platform_msi_type {
#define for_each_msi_entry_safe(desc, tmp, dev) \
list_for_each_entry_safe((desc), (tmp), dev_to_msi_list((dev)), list)
-#define dev_to_platform_msi_list(dev) (&(dev)->platform_msi_list)
-#define first_platform_msi_entry(dev) \
- list_first_entry(dev_to_platform_msi_list((dev)), struct msi_desc, list)
-#define for_each_platform_msi_entry(desc, dev) \
- list_for_each_entry((desc), dev_to_platform_msi_list((dev)), list)
-#define for_each_platform_msi_entry_safe(desc, tmp, dev) \
- list_for_each_entry_safe((desc), (tmp), dev_to_platform_msi_list((dev)), list)
+#define dev_to_platform_msi_group_list(dev) (&(dev)->platform_msi_list)
+
+#define first_platform_msi_group_entry(dev) \
+ list_first_entry(dev_to_platform_msi_group_list((dev)), \
+ struct platform_msi_group_entry, group_list)
-#define first_msi_entry_common(dev) \
- list_first_entry_select((dev)->platform_msi_type, dev_to_platform_msi_list((dev)), \
+#define platform_msi_current_group_entry_list(dev) \
+ (&((list_last_entry(dev_to_platform_msi_group_list((dev)), \
+ struct platform_msi_group_entry, \
+ group_list))->entry_list))
+
+#define first_msi_entry_current_group(dev) \
+ list_first_entry_select((dev)->platform_msi_type, \
+ platform_msi_current_group_entry_list((dev)), \
dev_to_msi_list((dev)), struct msi_desc, list)
-#define for_each_msi_entry_common(desc, dev) \
- list_for_each_entry_select((dev)->platform_msi_type, desc, dev_to_platform_msi_list((dev)), \
- dev_to_msi_list((dev)), list) \
+#define for_each_msi_entry_current_group(desc, dev) \
+ list_for_each_entry_select((dev)->platform_msi_type, desc, \
+ platform_msi_current_group_entry_list((dev)),\
+ dev_to_msi_list((dev)), list)
+
+#define for_each_platform_msi_entry_in_group(desc, platform_msi_group, group, dev) \
+ list_for_each_entry((platform_msi_group), dev_to_platform_msi_group_list((dev)), group_list) \
+ if (((platform_msi_group)->group_id) == (group)) \
+ list_for_each_entry((desc), (&(platform_msi_group)->entry_list), list)
#ifdef CONFIG_IRQ_MSI_IOMMU
static inline const void *msi_desc_get_iommu_cookie(struct msi_desc *desc)
@@ -363,6 +379,8 @@ struct irq_domain *msi_create_irq_domain(struct fwnode_handle *fwnode,
int msi_domain_alloc_irqs(struct irq_domain *domain, struct device *dev,
int nvec);
void msi_domain_free_irqs(struct irq_domain *domain, struct device *dev);
+void msi_domain_free_irqs_group(struct irq_domain *domain,
+ struct device *dev, unsigned int group);
struct msi_domain_info *msi_get_domain_info(struct irq_domain *domain);
struct irq_domain *platform_msi_create_irq_domain(struct fwnode_handle *fwnode,
@@ -371,6 +389,11 @@ struct irq_domain *platform_msi_create_irq_domain(struct fwnode_handle *fwnode,
int platform_msi_domain_alloc_irqs(struct device *dev, unsigned int nvec,
const struct platform_msi_ops *platform_ops);
void platform_msi_domain_free_irqs(struct device *dev);
+int platform_msi_domain_alloc_irqs_group(struct device *dev, unsigned int nvec,
+ const struct platform_msi_ops *platform_ops,
+ unsigned int *group_id);
+void platform_msi_domain_free_irqs_group(struct device *dev,
+ unsigned int group_id);
/* When an MSI domain is used as an intermediate domain */
int msi_domain_prepare_irqs(struct irq_domain *domain, struct device *dev,
diff --git a/kernel/irq/msi.c b/kernel/irq/msi.c
index bc5f9e32387f..899ade394ec8 100644
--- a/kernel/irq/msi.c
+++ b/kernel/irq/msi.c
@@ -320,7 +320,7 @@ int msi_domain_populate_irqs(struct irq_domain *domain, struct device *dev,
struct msi_desc *desc;
int ret = 0;
- for_each_msi_entry_common(desc, dev) {
+ for_each_msi_entry_current_group(desc, dev) {
/* Don't even try the multi-MSI brain damage. */
if (WARN_ON(!desc->irq || desc->nvec_used != 1)) {
ret = -EINVAL;
@@ -342,7 +342,7 @@ int msi_domain_populate_irqs(struct irq_domain *domain, struct device *dev,
if (ret) {
/* Mop up the damage */
- for_each_msi_entry_common(desc, dev) {
+ for_each_msi_entry_current_group(desc, dev) {
if (!(desc->irq >= virq && desc->irq < (virq + nvec)))
continue;
@@ -383,7 +383,7 @@ static bool msi_check_reservation_mode(struct irq_domain *domain,
* Checking the first MSI descriptor is sufficient. MSIX supports
* masking and MSI does so when the maskbit is set.
*/
- desc = first_msi_entry_common(dev);
+ desc = first_msi_entry_current_group(dev);
return desc->msi_attrib.is_msix || desc->msi_attrib.maskbit;
}
@@ -411,7 +411,7 @@ int msi_domain_alloc_irqs(struct irq_domain *domain, struct device *dev,
if (ret)
return ret;
- for_each_msi_entry_common(desc, dev) {
+ for_each_msi_entry_current_group(desc, dev) {
ops->set_desc(&arg, desc);
virq = __irq_domain_alloc_irqs(domain, -1, desc->nvec_used,
@@ -437,7 +437,7 @@ int msi_domain_alloc_irqs(struct irq_domain *domain, struct device *dev,
can_reserve = msi_check_reservation_mode(domain, info, dev);
- for_each_msi_entry_common(desc, dev) {
+ for_each_msi_entry_current_group(desc, dev) {
virq = desc->irq;
if (desc->nvec_used == 1)
dev_dbg(dev, "irq %d for MSI\n", virq);
@@ -468,7 +468,7 @@ int msi_domain_alloc_irqs(struct irq_domain *domain, struct device *dev,
* so request_irq() will assign the final vector.
*/
if (can_reserve) {
- for_each_msi_entry_common(desc, dev) {
+ for_each_msi_entry_current_group(desc, dev) {
irq_data = irq_domain_get_irq_data(domain, desc->irq);
irqd_clr_activated(irq_data);
}
@@ -476,7 +476,7 @@ int msi_domain_alloc_irqs(struct irq_domain *domain, struct device *dev,
return 0;
cleanup:
- for_each_msi_entry_common(desc, dev) {
+ for_each_msi_entry_current_group(desc, dev) {
struct irq_data *irqd;
if (desc->irq == virq)
@@ -500,7 +500,34 @@ void msi_domain_free_irqs(struct irq_domain *domain, struct device *dev)
{
struct msi_desc *desc;
- for_each_msi_entry_common(desc, dev) {
+ for_each_msi_entry_current_group(desc, dev) {
+ /*
+ * We might have failed to allocate an MSI early
+ * enough that there is no IRQ associated to this
+ * entry. If that's the case, don't do anything.
+ */
+ if (desc->irq) {
+ irq_domain_free_irqs(desc->irq, desc->nvec_used);
+ desc->irq = 0;
+ }
+ }
+}
+
+/**
+ * msi_domain_free_irqs_group - Free interrupts from a MSI interrupt @domain
+ * associated to @dev from a particular group
+ * @domain: The domain to managing the interrupts
+ * @dev: Pointer to device struct of the device for which the interrupts
+ * are free
+ * @group: group from which interrupts are to be freed
+ */
+void msi_domain_free_irqs_group(struct irq_domain *domain,
+ struct device *dev, unsigned int group)
+{
+ struct msi_desc *desc;
+ struct platform_msi_group_entry *platform_msi_group;
+
+ for_each_platform_msi_entry_in_group(desc, platform_msi_group, group, dev) {
/*
* We might have failed to allocate an MSI early
* enough that there is no IRQ associated to this
next prev parent reply other threads:[~2020-04-21 23:34 UTC|newest]
Thread overview: 89+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-04-21 23:33 [PATCH RFC 00/15] Add VFIO mediated device support and IMS support for the idxd driver Dave Jiang
2020-04-21 23:33 ` [PATCH RFC 01/15] drivers/base: Introduce platform_msi_ops Dave Jiang
2020-04-26 7:01 ` Greg KH
2020-04-27 21:38 ` Dave Jiang
2020-04-28 7:34 ` Greg KH
2020-04-21 23:33 ` [PATCH RFC 02/15] drivers/base: Introduce a new platform-msi list Dave Jiang
2020-04-25 21:13 ` Thomas Gleixner
2020-05-04 0:08 ` Dey, Megha
2020-04-21 23:34 ` Dave Jiang [this message]
2020-04-25 21:23 ` [PATCH RFC 03/15] drivers/base: Allocate/free platform-msi interrupts by group Thomas Gleixner
2020-05-04 0:08 ` Dey, Megha
2020-04-21 23:34 ` [PATCH RFC 04/15] drivers/base: Add support for a new IMS irq domain Dave Jiang
2020-04-23 20:11 ` Jason Gunthorpe
2020-05-01 22:30 ` Dey, Megha
2020-05-03 22:25 ` Jason Gunthorpe
2020-05-03 22:40 ` Dey, Megha
2020-05-03 22:46 ` Jason Gunthorpe
2020-05-04 0:25 ` Dey, Megha
2020-05-04 12:14 ` Jason Gunthorpe
2020-05-06 10:27 ` Tian, Kevin
2020-04-25 21:38 ` Thomas Gleixner
2020-05-04 0:11 ` Dey, Megha
2020-04-21 23:34 ` [PATCH RFC 05/15] ims-msi: Add mask/unmask routines Dave Jiang
2020-04-25 21:49 ` Thomas Gleixner
2020-05-04 0:16 ` Dey, Megha
2020-04-21 23:34 ` [PATCH RFC 06/15] ims-msi: Enable IMS interrupts Dave Jiang
2020-04-25 22:13 ` Thomas Gleixner
2020-05-04 0:17 ` Dey, Megha
2020-04-21 23:34 ` [PATCH RFC 07/15] Documentation: Interrupt Message store Dave Jiang
2020-04-23 20:04 ` Jason Gunthorpe
2020-05-01 22:32 ` Dey, Megha
2020-05-03 22:28 ` Jason Gunthorpe
2020-05-03 22:41 ` Dey, Megha
2020-04-21 23:34 ` [PATCH RFC 08/15] vfio/mdev: Add a member for iommu domain in mdev_device Dave Jiang
2020-04-21 23:34 ` [PATCH RFC 09/15] vfio/type1: Save domain when attach domain to mdev Dave Jiang
2020-04-21 23:34 ` [PATCH RFC 10/15] dmaengine: idxd: add config support for readonly devices Dave Jiang
2020-04-21 23:34 ` [PATCH RFC 11/15] dmaengine: idxd: add IMS support in base driver Dave Jiang
2020-04-21 23:35 ` [PATCH RFC 12/15] dmaengine: idxd: add device support functions in prep for mdev Dave Jiang
2020-04-21 23:35 ` [PATCH RFC 13/15] dmaengine: idxd: add support for VFIO mediated device Dave Jiang
2020-04-21 23:35 ` [PATCH RFC 14/15] dmaengine: idxd: add error notification from host driver to " Dave Jiang
2020-04-21 23:35 ` [PATCH RFC 15/15] dmaengine: idxd: add ABI documentation for mediated device support Dave Jiang
2020-04-21 23:54 ` [PATCH RFC 00/15] Add VFIO mediated device support and IMS support for the idxd driver Jason Gunthorpe
2020-04-22 0:53 ` Tian, Kevin
2020-04-22 11:50 ` Jason Gunthorpe
2020-04-22 21:14 ` Raj, Ashok
2020-04-23 19:12 ` Jason Gunthorpe
2020-04-24 3:27 ` Tian, Kevin
2020-04-24 12:44 ` Jason Gunthorpe
2020-04-24 16:25 ` Tian, Kevin
2020-04-24 18:12 ` Jason Gunthorpe
2020-04-26 5:18 ` Tian, Kevin
2020-04-26 19:13 ` Jason Gunthorpe
2020-04-27 3:43 ` Alex Williamson
2020-04-27 11:58 ` Jason Gunthorpe
2020-04-27 13:19 ` Alex Williamson
2020-04-27 13:22 ` Jason Gunthorpe
2020-04-27 14:18 ` Alex Williamson
2020-04-27 14:25 ` Jason Gunthorpe
2020-04-27 15:41 ` Alex Williamson
2020-04-27 16:16 ` Jason Gunthorpe
2020-04-27 16:25 ` Dave Jiang
2020-04-27 21:56 ` Jason Gunthorpe
2020-04-29 9:42 ` Tian, Kevin
2020-05-08 20:47 ` Raj, Ashok
2020-05-08 23:16 ` Jason Gunthorpe
2020-05-08 23:52 ` Dave Jiang
2020-05-09 0:09 ` Raj, Ashok
2020-05-09 12:21 ` Jason Gunthorpe
2020-05-13 2:29 ` Jason Wang
2020-05-13 8:30 ` Tian, Kevin
2020-05-13 12:40 ` Jason Gunthorpe
2020-04-27 12:13 ` Tian, Kevin
2020-04-27 12:55 ` Jason Gunthorpe
2020-04-22 21:24 ` Dan Williams
2020-04-23 19:17 ` Dan Williams
2020-04-23 19:49 ` Jason Gunthorpe
2020-05-01 22:31 ` Dey, Megha
2020-05-03 22:21 ` Jason Gunthorpe
2020-05-03 22:32 ` Dey, Megha
2020-04-23 19:18 ` Jason Gunthorpe
2020-05-01 22:31 ` Dey, Megha
2020-05-03 22:22 ` Jason Gunthorpe
2020-05-03 22:31 ` Dey, Megha
2020-05-03 22:36 ` Jason Gunthorpe
2020-05-04 0:20 ` Dey, Megha
2020-04-22 23:04 ` Dey, Megha
2020-04-23 19:44 ` Jason Gunthorpe
2020-05-01 22:32 ` Dey, Megha
2020-04-24 6:31 ` Jason Wang
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=158751204550.36773.459505651659406529.stgit@djiang5-desk3.ch.intel.com \
--to=dave.jiang@intel.com \
--cc=alex.williamson@redhat.com \
--cc=ashok.raj@intel.com \
--cc=baolu.lu@intel.com \
--cc=bhelgaas@google.com \
--cc=dan.j.williams@intel.com \
--cc=dmaengine@vger.kernel.org \
--cc=eric.auger@redhat.com \
--cc=gregkh@linuxfoundation.org \
--cc=hpa@zytor.com \
--cc=jacob.jun.pan@intel.com \
--cc=jgg@mellanox.com \
--cc=jing.lin@intel.com \
--cc=kevin.tian@intel.com \
--cc=kvm@vger.kernel.org \
--cc=kwankhede@nvidia.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-pci@vger.kernel.org \
--cc=maz@kernel.org \
--cc=megha.dey@linux.intel.com \
--cc=parav@mellanox.com \
--cc=rafael@kernel.org \
--cc=sanjay.k.kumar@intel.com \
--cc=tglx@linutronix.de \
--cc=tony.luck@intel.com \
--cc=vkoul@kernel.org \
--cc=x86@kernel.org \
--cc=yi.l.liu@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).