From: Dimitri Sivanich <sivanich@sgi.com>
To: Jiang Liu <jiang.liu@linux.intel.com>
Cc: Bjorn Helgaas <bhelgaas@google.com>,
Benjamin Herrenschmidt <benh@kernel.crashing.org>,
Thomas Gleixner <tglx@linutronix.de>,
Ingo Molnar <mingo@redhat.com>, "H. Peter Anvin" <hpa@zytor.com>,
"Rafael J. Wysocki" <rjw@rjwysocki.net>,
Randy Dunlap <rdunlap@infradead.org>,
Yinghai Lu <yinghai@kernel.org>, Borislav Petkov <bp@alien8.de>,
x86@kernel.org, Konrad Rzeszutek Wilk <konrad.wilk@oracle.com>,
Andrew Morton <akpm@linux-foundation.org>,
Tony Luck <tony.luck@intel.com>, Joerg Roedel <joro@8bytes.org>,
Greg Kroah-Hartman <gregkh@linuxfoundation.org>,
linux-kernel@vger.kernel.org, linux-pci@vger.kernel.org,
linux-acpi@vger.kernel.org, Russ Anderson <rja@sgi.com>,
Nathan Zimmer <nzimmer@sgi.com>,
Dimitri Sivanich <sivanich@sgi.com>
Subject: Re: [Patch Part2 v6 22/27] x86, uv: Use hierarchy irqdomain to manage UV interrupts
Date: Mon, 15 Dec 2014 15:37:35 -0600 [thread overview]
Message-ID: <20141215213735.GA8149@sgi.com> (raw)
In-Reply-To: <1416894816-23245-23-git-send-email-jiang.liu@linux.intel.com>
Was this patch ever tested on a UV system?
Also, adding some SGI folks to the CC list, since there were none listed before.
On Tue, Nov 25, 2014 at 01:53:31PM +0800, Jiang Liu wrote:
> Enhance UV code to support hierarchy irqdomain, it helps to make
> the architecture more clear.
>
> We should construct hwirq based on mmr_blade and mmr_offset, but
> mmr_offset is type of unsigned long, it may exceed the range of
> irq_hw_number_t. So help about the way to construct hwirq based
> on mmr_blade and mmr_offset is welcomed!
>
> Signed-off-by: Jiang Liu <jiang.liu@linux.intel.com>
> ---
> arch/x86/include/asm/hw_irq.h | 9 ++
> arch/x86/platform/uv/uv_irq.c | 287 ++++++++++++++++-------------------------
> 2 files changed, 117 insertions(+), 179 deletions(-)
>
> diff --git a/arch/x86/include/asm/hw_irq.h b/arch/x86/include/asm/hw_irq.h
> index 46dec7e37829..bbf90fe2a224 100644
> --- a/arch/x86/include/asm/hw_irq.h
> +++ b/arch/x86/include/asm/hw_irq.h
> @@ -123,6 +123,7 @@ enum irq_alloc_type {
> X86_IRQ_ALLOC_TYPE_MSI,
> X86_IRQ_ALLOC_TYPE_MSIX,
> X86_IRQ_ALLOC_TYPE_DMAR,
> + X86_IRQ_ALLOC_TYPE_UV,
> };
>
> struct irq_alloc_info {
> @@ -169,6 +170,14 @@ struct irq_alloc_info {
> void *ht_update;
> };
> #endif
> +#ifdef CONFIG_X86_UV
> + struct {
> + int uv_limit;
> + int uv_blade;
> + unsigned long uv_offset;
> + char *uv_name;
> + };
> +#endif
> };
> };
>
> diff --git a/arch/x86/platform/uv/uv_irq.c b/arch/x86/platform/uv/uv_irq.c
> index 474912d03f40..c237ed34a498 100644
> --- a/arch/x86/platform/uv/uv_irq.c
> +++ b/arch/x86/platform/uv/uv_irq.c
> @@ -19,17 +19,31 @@
> #include <asm/uv/uv_hub.h>
>
> /* MMR offset and pnode of hub sourcing interrupts for a given irq */
> -struct uv_irq_2_mmr_pnode{
> - struct rb_node list;
> +struct uv_irq_2_mmr_pnode {
> unsigned long offset;
> int pnode;
> - int irq;
> };
>
> -static DEFINE_SPINLOCK(uv_irq_lock);
> -static struct rb_root uv_irq_root;
> +static void uv_program_mmr(struct irq_cfg *cfg, struct uv_irq_2_mmr_pnode *info)
> +{
> + unsigned long mmr_value;
> + struct uv_IO_APIC_route_entry *entry;
> +
> + BUILD_BUG_ON(sizeof(struct uv_IO_APIC_route_entry) !=
> + sizeof(unsigned long));
> +
> + mmr_value = 0;
> + entry = (struct uv_IO_APIC_route_entry *)&mmr_value;
> + entry->vector = cfg->vector;
> + entry->delivery_mode = apic->irq_delivery_mode;
> + entry->dest_mode = apic->irq_dest_mode;
> + entry->polarity = 0;
> + entry->trigger = 0;
> + entry->mask = 0;
> + entry->dest = cfg->dest_apicid;
>
> -static int uv_set_irq_affinity(struct irq_data *, const struct cpumask *, bool);
> + uv_write_global_mmr64(info->pnode, info->offset, mmr_value);
> +}
>
> static void uv_noop(struct irq_data *data) { }
>
> @@ -38,6 +52,24 @@ static void uv_ack_apic(struct irq_data *data)
> ack_APIC_irq();
> }
>
> +static int
> +uv_set_irq_affinity(struct irq_data *data, const struct cpumask *mask,
> + bool force)
> +{
> + struct irq_data *parent = data->parent_data;
> + struct irq_cfg *cfg = irqd_cfg(data);
> + int ret;
> +
> + ret = parent->chip->irq_set_affinity(parent, mask, force);
> + if (ret >= 0) {
> + uv_program_mmr(cfg, data->chip_data);
> + if (cfg->move_in_progress)
> + send_cleanup_vector(cfg);
> + }
> +
> + return ret;
> +}
> +
> static struct irq_chip uv_irq_chip = {
> .name = "UV-CORE",
> .irq_mask = uv_noop,
> @@ -46,179 +78,99 @@ static struct irq_chip uv_irq_chip = {
> .irq_set_affinity = uv_set_irq_affinity,
> };
>
> -/*
> - * Add offset and pnode information of the hub sourcing interrupts to the
> - * rb tree for a specific irq.
> - */
> -static int uv_set_irq_2_mmr_info(int irq, unsigned long offset, unsigned blade)
> +static int uv_domain_alloc(struct irq_domain *domain, unsigned int virq,
> + unsigned int nr_irqs, void *arg)
> {
> - struct rb_node **link = &uv_irq_root.rb_node;
> - struct rb_node *parent = NULL;
> - struct uv_irq_2_mmr_pnode *n;
> - struct uv_irq_2_mmr_pnode *e;
> - unsigned long irqflags;
> -
> - n = kmalloc_node(sizeof(struct uv_irq_2_mmr_pnode), GFP_KERNEL,
> - uv_blade_to_memory_nid(blade));
> - if (!n)
> + struct uv_irq_2_mmr_pnode *chip_data;
> + struct irq_alloc_info *info = arg;
> + struct irq_data *irq_data = irq_domain_get_irq_data(domain, virq);
> + int ret;
> +
> + if (nr_irqs > 1 || !info || info->type != X86_IRQ_ALLOC_TYPE_UV)
> + return -EINVAL;
> +
> + chip_data = kmalloc_node(sizeof(*chip_data), GFP_KERNEL,
> + irq_data->node);
> + if (!chip_data)
> return -ENOMEM;
>
> - n->irq = irq;
> - n->offset = offset;
> - n->pnode = uv_blade_to_pnode(blade);
> - spin_lock_irqsave(&uv_irq_lock, irqflags);
> - /* Find the right place in the rbtree: */
> - while (*link) {
> - parent = *link;
> - e = rb_entry(parent, struct uv_irq_2_mmr_pnode, list);
> -
> - if (unlikely(irq == e->irq)) {
> - /* irq entry exists */
> - e->pnode = uv_blade_to_pnode(blade);
> - e->offset = offset;
> - spin_unlock_irqrestore(&uv_irq_lock, irqflags);
> - kfree(n);
> - return 0;
> - }
> -
> - if (irq < e->irq)
> - link = &(*link)->rb_left;
> + ret = irq_domain_alloc_irqs_parent(domain, virq, nr_irqs, arg);
> + if (ret >= 0) {
> + if (info->uv_limit == UV_AFFINITY_CPU)
> + irq_set_status_flags(virq, IRQ_NO_BALANCING);
> else
> - link = &(*link)->rb_right;
> + irq_set_status_flags(virq, IRQ_MOVE_PCNTXT);
> +
> + chip_data->pnode = uv_blade_to_pnode(info->uv_blade);
> + chip_data->offset = info->uv_offset;
> + irq_domain_set_info(domain, virq, virq, &uv_irq_chip, chip_data,
> + handle_percpu_irq, NULL, info->uv_name);
> + } else {
> + kfree(chip_data);
> }
>
> - /* Insert the node into the rbtree. */
> - rb_link_node(&n->list, parent, link);
> - rb_insert_color(&n->list, &uv_irq_root);
> -
> - spin_unlock_irqrestore(&uv_irq_lock, irqflags);
> - return 0;
> + return ret;
> }
>
> -/* Retrieve offset and pnode information from the rb tree for a specific irq */
> -int uv_irq_2_mmr_info(int irq, unsigned long *offset, int *pnode)
> +static void uv_domain_free(struct irq_domain *domain, unsigned int virq,
> + unsigned int nr_irqs)
> {
> - struct uv_irq_2_mmr_pnode *e;
> - struct rb_node *n;
> - unsigned long irqflags;
> -
> - spin_lock_irqsave(&uv_irq_lock, irqflags);
> - n = uv_irq_root.rb_node;
> - while (n) {
> - e = rb_entry(n, struct uv_irq_2_mmr_pnode, list);
> -
> - if (e->irq == irq) {
> - *offset = e->offset;
> - *pnode = e->pnode;
> - spin_unlock_irqrestore(&uv_irq_lock, irqflags);
> - return 0;
> - }
> -
> - if (irq < e->irq)
> - n = n->rb_left;
> - else
> - n = n->rb_right;
> - }
> - spin_unlock_irqrestore(&uv_irq_lock, irqflags);
> - return -1;
> + struct irq_data *irq_data = irq_domain_get_irq_data(domain, virq);
> +
> + BUG_ON(nr_irqs != 1);
> + kfree(irq_data->chip_data);
> + irq_clear_status_flags(virq, IRQ_MOVE_PCNTXT);
> + irq_clear_status_flags(virq, IRQ_NO_BALANCING);
> + irq_domain_free_irqs_top(domain, virq, nr_irqs);
> }
>
> /*
> * Re-target the irq to the specified CPU and enable the specified MMR located
> * on the specified blade to allow the sending of MSIs to the specified CPU.
> */
> -static int
> -arch_enable_uv_irq(char *irq_name, unsigned int irq, int cpu, int mmr_blade,
> - unsigned long mmr_offset, int limit)
> +static void uv_domain_activate(struct irq_domain *domain,
> + struct irq_data *irq_data)
> {
> - struct irq_cfg *cfg = irq_cfg(irq);
> - unsigned long mmr_value;
> - struct uv_IO_APIC_route_entry *entry;
> - int mmr_pnode;
> -
> - BUILD_BUG_ON(sizeof(struct uv_IO_APIC_route_entry) !=
> - sizeof(unsigned long));
> -
> - if (limit == UV_AFFINITY_CPU)
> - irq_set_status_flags(irq, IRQ_NO_BALANCING);
> - else
> - irq_set_status_flags(irq, IRQ_MOVE_PCNTXT);
> -
> - irq_set_chip_and_handler_name(irq, &uv_irq_chip, handle_percpu_irq,
> - irq_name);
> -
> - mmr_value = 0;
> - entry = (struct uv_IO_APIC_route_entry *)&mmr_value;
> - entry->vector = cfg->vector;
> - entry->delivery_mode = apic->irq_delivery_mode;
> - entry->dest_mode = apic->irq_dest_mode;
> - entry->polarity = 0;
> - entry->trigger = 0;
> - entry->mask = 0;
> - entry->dest = cfg->dest_apicid;
> -
> - mmr_pnode = uv_blade_to_pnode(mmr_blade);
> - uv_write_global_mmr64(mmr_pnode, mmr_offset, mmr_value);
> -
> - if (cfg->move_in_progress)
> - send_cleanup_vector(cfg);
> -
> - return irq;
> + uv_program_mmr(irqd_cfg(irq_data), irq_data->chip_data);
> }
>
> /*
> * Disable the specified MMR located on the specified blade so that MSIs are
> * longer allowed to be sent.
> */
> -static void arch_disable_uv_irq(int mmr_pnode, unsigned long mmr_offset)
> +static void uv_domain_deactivate(struct irq_domain *domain,
> + struct irq_data *irq_data)
> {
> unsigned long mmr_value;
> struct uv_IO_APIC_route_entry *entry;
>
> - BUILD_BUG_ON(sizeof(struct uv_IO_APIC_route_entry) !=
> - sizeof(unsigned long));
> -
> mmr_value = 0;
> entry = (struct uv_IO_APIC_route_entry *)&mmr_value;
> entry->mask = 1;
> -
> - uv_write_global_mmr64(mmr_pnode, mmr_offset, mmr_value);
> + uv_program_mmr(irqd_cfg(irq_data), irq_data->chip_data);
> }
>
> -static int
> -uv_set_irq_affinity(struct irq_data *data, const struct cpumask *mask,
> - bool force)
> -{
> - struct irq_cfg *cfg = irqd_cfg(data);
> - unsigned int dest;
> - unsigned long mmr_value, mmr_offset;
> - struct uv_IO_APIC_route_entry *entry;
> - int mmr_pnode;
> -
> - if (apic_set_affinity(data, mask, &dest))
> - return -1;
> -
> - mmr_value = 0;
> - entry = (struct uv_IO_APIC_route_entry *)&mmr_value;
> -
> - entry->vector = cfg->vector;
> - entry->delivery_mode = apic->irq_delivery_mode;
> - entry->dest_mode = apic->irq_dest_mode;
> - entry->polarity = 0;
> - entry->trigger = 0;
> - entry->mask = 0;
> - entry->dest = dest;
> -
> - /* Get previously stored MMR and pnode of hub sourcing interrupts */
> - if (uv_irq_2_mmr_info(data->irq, &mmr_offset, &mmr_pnode))
> - return -1;
> -
> - uv_write_global_mmr64(mmr_pnode, mmr_offset, mmr_value);
> +static struct irq_domain_ops uv_domain_ops = {
> + .alloc = uv_domain_alloc,
> + .free = uv_domain_free,
> + .activate = uv_domain_activate,
> + .deactivate = uv_domain_deactivate,
> +};
>
> - if (cfg->move_in_progress)
> - send_cleanup_vector(cfg);
> +static struct irq_domain *uv_get_irq_domain(void)
> +{
> + static struct irq_domain *uv_domain;
> + static DEFINE_MUTEX(uv_lock);
> +
> + mutex_lock(&uv_lock);
> + if (uv_domain == NULL) {
> + uv_domain = irq_domain_add_tree(NULL, &uv_domain_ops, NULL);
> + if (uv_domain)
> + uv_domain->parent = x86_vector_domain;
> + }
> + mutex_unlock(&uv_lock);
>
> - return IRQ_SET_MASK_OK_NOCOPY;
> + return uv_domain;
> }
>
> /*
> @@ -229,23 +181,20 @@ uv_set_irq_affinity(struct irq_data *data, const struct cpumask *mask,
> int uv_setup_irq(char *irq_name, int cpu, int mmr_blade,
> unsigned long mmr_offset, int limit)
> {
> - int ret, irq;
> struct irq_alloc_info info;
> + struct irq_domain *domain = uv_get_irq_domain();
> +
> + if (!domain)
> + return -ENOMEM;
>
> init_irq_alloc_info(&info, cpumask_of(cpu));
> - irq = irq_domain_alloc_irqs(NULL, 1, uv_blade_to_memory_nid(mmr_blade),
> - &info);
> - if (irq <= 0)
> - return -EBUSY;
> -
> - ret = arch_enable_uv_irq(irq_name, irq, cpu, mmr_blade, mmr_offset,
> - limit);
> - if (ret == irq)
> - uv_set_irq_2_mmr_info(irq, mmr_offset, mmr_blade);
> - else
> - irq_domain_free_irqs(irq, 1);
> + info.uv_limit = limit;
> + info.uv_blade = mmr_blade;
> + info.uv_offset = mmr_offset;
> + info.uv_name = irq_name;
>
> - return ret;
> + return irq_domain_alloc_irqs(domain, 1,
> + uv_blade_to_memory_nid(mmr_blade), &info);
> }
> EXPORT_SYMBOL_GPL(uv_setup_irq);
>
> @@ -258,26 +207,6 @@ EXPORT_SYMBOL_GPL(uv_setup_irq);
> */
> void uv_teardown_irq(unsigned int irq)
> {
> - struct uv_irq_2_mmr_pnode *e;
> - struct rb_node *n;
> - unsigned long irqflags;
> -
> - spin_lock_irqsave(&uv_irq_lock, irqflags);
> - n = uv_irq_root.rb_node;
> - while (n) {
> - e = rb_entry(n, struct uv_irq_2_mmr_pnode, list);
> - if (e->irq == irq) {
> - arch_disable_uv_irq(e->pnode, e->offset);
> - rb_erase(n, &uv_irq_root);
> - kfree(e);
> - break;
> - }
> - if (irq < e->irq)
> - n = n->rb_left;
> - else
> - n = n->rb_right;
> - }
> - spin_unlock_irqrestore(&uv_irq_lock, irqflags);
> irq_domain_free_irqs(irq, 1);
> }
> EXPORT_SYMBOL_GPL(uv_teardown_irq);
> --
> 1.7.10.4
>
> --
> To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at http://vger.kernel.org/majordomo-info.html
> Please read the FAQ at http://www.tux.org/lkml/
next prev parent reply other threads:[~2014-12-15 21:37 UTC|newest]
Thread overview: 62+ messages / expand[flat|nested] mbox.gz Atom feed top
2014-11-25 5:53 [Patch Part2 v6 00/27] Enable hierarchy irqdomian on x86 platforms Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 01/27] x86, irq: Save destination CPU ID in irq_cfg Jiang Liu
2014-11-26 23:05 ` [tip:x86/apic] " tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 02/27] x86, irq: Use hierarchy irqdomain to manage CPU interrupt vectors Jiang Liu
2014-11-26 23:06 ` [tip:x86/apic] " tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 03/27] x86, hpet: Use new irqdomain interfaces to allocate/free IRQ Jiang Liu
2014-11-26 23:06 ` [tip:x86/apic] " tip-bot for Jiang Liu
2015-04-24 15:44 ` [tip:x86/apic] x86/hpet: Use new irqdomain interfaces to allocate /free IRQ tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 04/27] x86, MSI: Use new irqdomain interfaces to allocate/free IRQ Jiang Liu
2014-11-26 23:06 ` [tip:x86/apic] x86, MSI: Use new irqdomain interfaces to allocate /free IRQ tip-bot for Jiang Liu
2015-04-24 15:44 ` [tip:x86/apic] x86/MSI: Use new irqdomain interfaces to allocate/ free IRQ tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 05/27] x86, uv: Use new irqdomain interfaces to allocate/free IRQ Jiang Liu
2014-11-26 23:06 ` [tip:x86/apic] x86, uv: Use new irqdomain interfaces to allocate/ free IRQ tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 06/27] x86, htirq: Use new irqdomain interfaces to allocate/free IRQ Jiang Liu
2014-11-26 23:07 ` [tip:x86/apic] " tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 07/27] x86, dmar: " Jiang Liu
2014-11-26 23:07 ` [tip:x86/apic] " tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 08/27] x86: irq_remapping: Introduce new interfaces to support hierarchy irqdomain Jiang Liu
2014-11-26 23:07 ` [tip:x86/apic] " tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 09/27] iommu/vt-d: Change prototypes to prepare for enabling " Jiang Liu
2014-11-26 23:07 ` [tip:x86/apic] " tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 10/27] iommu/vt-d: Enhance Intel IR driver to suppport " Jiang Liu
2014-11-26 23:08 ` [tip:x86/apic] " tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 11/27] iommu/amd: Enhance AMD " Jiang Liu
2014-11-26 23:08 ` [tip:x86/apic] " tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 12/27] x86, hpet: Enhance HPET IRQ to support " Jiang Liu
2014-11-26 23:08 ` [tip:x86/apic] " tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 13/27] x86, PCI, MSI: Use hierarchy irqdomain to manage MSI interrupts Jiang Liu
2014-11-26 23:09 ` [tip:x86/apic] " tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 14/27] x86, irq: Directly call native_compose_msi_msg() for DMAR IRQ Jiang Liu
2014-11-26 23:09 ` [tip:x86/apic] " tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 15/27] iommu/vt-d: Clean up unused MSI related code Jiang Liu
2014-11-26 23:09 ` [tip:x86/apic] " tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 16/27] iommu/amd: " Jiang Liu
2014-11-26 23:09 ` [tip:x86/apic] " tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 17/27] x86: irq_remapping: " Jiang Liu
2014-11-26 23:10 ` [tip:x86/apic] " tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 18/27] x86, irq: Clean up unused MSI related code and interfaces Jiang Liu
2014-11-26 23:10 ` [tip:x86/apic] " tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 19/27] iommu/vt-d: Refine the interfaces to create IRQ for DMAR unit Jiang Liu
2014-11-26 23:10 ` [tip:x86/apic] " tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 20/27] x86, irq: Use hierarchy irqdomain to manage DMAR interrupts Jiang Liu
2014-11-26 23:10 ` [tip:x86/apic] " tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 21/27] x86, htirq: Use hierarchy irqdomain to manage Hypertransport interrupts Jiang Liu
2014-11-26 23:11 ` [tip:x86/apic] " tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 22/27] x86, uv: Use hierarchy irqdomain to manage UV interrupts Jiang Liu
2014-11-26 23:11 ` [tip:x86/apic] " tip-bot for Jiang Liu
2014-12-15 21:37 ` Dimitri Sivanich [this message]
2014-12-16 17:29 ` [Patch Part2 v6 22/27] " Dimitri Sivanich
2014-12-17 2:41 ` Jiang Liu
2014-12-17 16:45 ` Russ Anderson
2014-12-23 7:18 ` Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 23/27] x86, irq: Normalize x86 irq_chip name Jiang Liu
2014-11-26 23:11 ` [tip:x86/apic] " tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 24/27] x86, PCI/MSI: Simplify the way to deal with remapped MSI interrupts Jiang Liu
2014-11-26 23:11 ` [tip:x86/apic] " tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 25/27] PCI/MSI: Replace msi_update_msg() with irq_chip_compose_msi_msg() Jiang Liu
2014-11-26 23:12 ` [tip:x86/apic] " tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 26/27] x86, irq: Implement irq_chip.irq_write_msi_msg for MSI/DMAR/HPET irq_chips Jiang Liu
2014-11-26 23:12 ` [tip:x86/apic] " tip-bot for Jiang Liu
2014-11-25 5:53 ` [Patch Part2 v6 27/27] x86, irq: Simplify MSI/DMAR/HPET implementation by using common code Jiang Liu
2014-11-26 23:12 ` [tip:x86/apic] x86, irq: Simplify MSI/DMAR/ HPET " tip-bot for Jiang Liu
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20141215213735.GA8149@sgi.com \
--to=sivanich@sgi.com \
--cc=akpm@linux-foundation.org \
--cc=benh@kernel.crashing.org \
--cc=bhelgaas@google.com \
--cc=bp@alien8.de \
--cc=gregkh@linuxfoundation.org \
--cc=hpa@zytor.com \
--cc=jiang.liu@linux.intel.com \
--cc=joro@8bytes.org \
--cc=konrad.wilk@oracle.com \
--cc=linux-acpi@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-pci@vger.kernel.org \
--cc=mingo@redhat.com \
--cc=nzimmer@sgi.com \
--cc=rdunlap@infradead.org \
--cc=rja@sgi.com \
--cc=rjw@rjwysocki.net \
--cc=tglx@linutronix.de \
--cc=tony.luck@intel.com \
--cc=x86@kernel.org \
--cc=yinghai@kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).