Was this patch ever tested on a UV system?

Also, adding some SGI folks to the CC list, since there were none listed before.

On Tue, Nov 25, 2014 at 01:53:31PM +0800, Jiang Liu wrote:
> Enhance UV code to support hierarchy irqdomain, it helps to make
> the architecture more clear.
> 
> We should construct hwirq based on mmr_blade and mmr_offset, but
> mmr_offset is type of unsigned long, it may exceed the range of
> irq_hw_number_t. So help about the way to construct hwirq based
> on mmr_blade and mmr_offset is welcomed!
> 
> Signed-off-by: Jiang Liu <jiang....@linux.intel.com>
> ---
>  arch/x86/include/asm/hw_irq.h |    9 ++
>  arch/x86/platform/uv/uv_irq.c |  287 
> ++++++++++++++++-------------------------
>  2 files changed, 117 insertions(+), 179 deletions(-)
> 
> diff --git a/arch/x86/include/asm/hw_irq.h b/arch/x86/include/asm/hw_irq.h
> index 46dec7e37829..bbf90fe2a224 100644
> --- a/arch/x86/include/asm/hw_irq.h
> +++ b/arch/x86/include/asm/hw_irq.h
> @@ -123,6 +123,7 @@ enum irq_alloc_type {
>       X86_IRQ_ALLOC_TYPE_MSI,
>       X86_IRQ_ALLOC_TYPE_MSIX,
>       X86_IRQ_ALLOC_TYPE_DMAR,
> +     X86_IRQ_ALLOC_TYPE_UV,
>  };
>  
>  struct irq_alloc_info {
> @@ -169,6 +170,14 @@ struct irq_alloc_info {
>                       void            *ht_update;
>               };
>  #endif
> +#ifdef       CONFIG_X86_UV
> +             struct {
> +                     int             uv_limit;
> +                     int             uv_blade;
> +                     unsigned long   uv_offset;
> +                     char            *uv_name;
> +             };
> +#endif
>       };
>  };
>  
> diff --git a/arch/x86/platform/uv/uv_irq.c b/arch/x86/platform/uv/uv_irq.c
> index 474912d03f40..c237ed34a498 100644
> --- a/arch/x86/platform/uv/uv_irq.c
> +++ b/arch/x86/platform/uv/uv_irq.c
> @@ -19,17 +19,31 @@
>  #include <asm/uv/uv_hub.h>
>  
>  /* MMR offset and pnode of hub sourcing interrupts for a given irq */
> -struct uv_irq_2_mmr_pnode{
> -     struct rb_node          list;
> +struct uv_irq_2_mmr_pnode {
>       unsigned long           offset;
>       int                     pnode;
> -     int                     irq;
>  };
>  
> -static DEFINE_SPINLOCK(uv_irq_lock);
> -static struct rb_root                uv_irq_root;
> +static void uv_program_mmr(struct irq_cfg *cfg, struct uv_irq_2_mmr_pnode 
> *info)
> +{
> +     unsigned long mmr_value;
> +     struct uv_IO_APIC_route_entry *entry;
> +
> +     BUILD_BUG_ON(sizeof(struct uv_IO_APIC_route_entry) !=
> +                  sizeof(unsigned long));
> +
> +     mmr_value = 0;
> +     entry = (struct uv_IO_APIC_route_entry *)&mmr_value;
> +     entry->vector           = cfg->vector;
> +     entry->delivery_mode    = apic->irq_delivery_mode;
> +     entry->dest_mode        = apic->irq_dest_mode;
> +     entry->polarity         = 0;
> +     entry->trigger          = 0;
> +     entry->mask             = 0;
> +     entry->dest             = cfg->dest_apicid;
>  
> -static int uv_set_irq_affinity(struct irq_data *, const struct cpumask *, 
> bool);
> +     uv_write_global_mmr64(info->pnode, info->offset, mmr_value);
> +}
>  
>  static void uv_noop(struct irq_data *data) { }
>  
> @@ -38,6 +52,24 @@ static void uv_ack_apic(struct irq_data *data)
>       ack_APIC_irq();
>  }
>  
> +static int
> +uv_set_irq_affinity(struct irq_data *data, const struct cpumask *mask,
> +                 bool force)
> +{
> +     struct irq_data *parent = data->parent_data;
> +     struct irq_cfg *cfg = irqd_cfg(data);
> +     int ret;
> +
> +     ret = parent->chip->irq_set_affinity(parent, mask, force);
> +     if (ret >= 0) {
> +             uv_program_mmr(cfg, data->chip_data);
> +             if (cfg->move_in_progress)
> +                     send_cleanup_vector(cfg);
> +     }
> +
> +     return ret;
> +}
> +
>  static struct irq_chip uv_irq_chip = {
>       .name                   = "UV-CORE",
>       .irq_mask               = uv_noop,
> @@ -46,179 +78,99 @@ static struct irq_chip uv_irq_chip = {
>       .irq_set_affinity       = uv_set_irq_affinity,
>  };
>  
> -/*
> - * Add offset and pnode information of the hub sourcing interrupts to the
> - * rb tree for a specific irq.
> - */
> -static int uv_set_irq_2_mmr_info(int irq, unsigned long offset, unsigned 
> blade)
> +static int uv_domain_alloc(struct irq_domain *domain, unsigned int virq,
> +                        unsigned int nr_irqs, void *arg)
>  {
> -     struct rb_node **link = &uv_irq_root.rb_node;
> -     struct rb_node *parent = NULL;
> -     struct uv_irq_2_mmr_pnode *n;
> -     struct uv_irq_2_mmr_pnode *e;
> -     unsigned long irqflags;
> -
> -     n = kmalloc_node(sizeof(struct uv_irq_2_mmr_pnode), GFP_KERNEL,
> -                             uv_blade_to_memory_nid(blade));
> -     if (!n)
> +     struct uv_irq_2_mmr_pnode *chip_data;
> +     struct irq_alloc_info *info = arg;
> +     struct irq_data *irq_data = irq_domain_get_irq_data(domain, virq);
> +     int ret;
> +
> +     if (nr_irqs > 1 || !info || info->type != X86_IRQ_ALLOC_TYPE_UV)
> +             return -EINVAL;
> +
> +     chip_data = kmalloc_node(sizeof(*chip_data), GFP_KERNEL,
> +                              irq_data->node);
> +     if (!chip_data)
>               return -ENOMEM;
>  
> -     n->irq = irq;
> -     n->offset = offset;
> -     n->pnode = uv_blade_to_pnode(blade);
> -     spin_lock_irqsave(&uv_irq_lock, irqflags);
> -     /* Find the right place in the rbtree: */
> -     while (*link) {
> -             parent = *link;
> -             e = rb_entry(parent, struct uv_irq_2_mmr_pnode, list);
> -
> -             if (unlikely(irq == e->irq)) {
> -                     /* irq entry exists */
> -                     e->pnode = uv_blade_to_pnode(blade);
> -                     e->offset = offset;
> -                     spin_unlock_irqrestore(&uv_irq_lock, irqflags);
> -                     kfree(n);
> -                     return 0;
> -             }
> -
> -             if (irq < e->irq)
> -                     link = &(*link)->rb_left;
> +     ret = irq_domain_alloc_irqs_parent(domain, virq, nr_irqs, arg);
> +     if (ret >= 0) {
> +             if (info->uv_limit == UV_AFFINITY_CPU)
> +                     irq_set_status_flags(virq, IRQ_NO_BALANCING);
>               else
> -                     link = &(*link)->rb_right;
> +                     irq_set_status_flags(virq, IRQ_MOVE_PCNTXT);
> +
> +             chip_data->pnode = uv_blade_to_pnode(info->uv_blade);
> +             chip_data->offset = info->uv_offset;
> +             irq_domain_set_info(domain, virq, virq, &uv_irq_chip, chip_data,
> +                                 handle_percpu_irq, NULL, info->uv_name);
> +     } else {
> +             kfree(chip_data);
>       }
>  
> -     /* Insert the node into the rbtree. */
> -     rb_link_node(&n->list, parent, link);
> -     rb_insert_color(&n->list, &uv_irq_root);
> -
> -     spin_unlock_irqrestore(&uv_irq_lock, irqflags);
> -     return 0;
> +     return ret;
>  }
>  
> -/* Retrieve offset and pnode information from the rb tree for a specific irq 
> */
> -int uv_irq_2_mmr_info(int irq, unsigned long *offset, int *pnode)
> +static void uv_domain_free(struct irq_domain *domain, unsigned int virq,
> +                        unsigned int nr_irqs)
>  {
> -     struct uv_irq_2_mmr_pnode *e;
> -     struct rb_node *n;
> -     unsigned long irqflags;
> -
> -     spin_lock_irqsave(&uv_irq_lock, irqflags);
> -     n = uv_irq_root.rb_node;
> -     while (n) {
> -             e = rb_entry(n, struct uv_irq_2_mmr_pnode, list);
> -
> -             if (e->irq == irq) {
> -                     *offset = e->offset;
> -                     *pnode = e->pnode;
> -                     spin_unlock_irqrestore(&uv_irq_lock, irqflags);
> -                     return 0;
> -             }
> -
> -             if (irq < e->irq)
> -                     n = n->rb_left;
> -             else
> -                     n = n->rb_right;
> -     }
> -     spin_unlock_irqrestore(&uv_irq_lock, irqflags);
> -     return -1;
> +     struct irq_data *irq_data = irq_domain_get_irq_data(domain, virq);
> +
> +     BUG_ON(nr_irqs != 1);
> +     kfree(irq_data->chip_data);
> +     irq_clear_status_flags(virq, IRQ_MOVE_PCNTXT);
> +     irq_clear_status_flags(virq, IRQ_NO_BALANCING);
> +     irq_domain_free_irqs_top(domain, virq, nr_irqs);
>  }
>  
>  /*
>   * Re-target the irq to the specified CPU and enable the specified MMR 
> located
>   * on the specified blade to allow the sending of MSIs to the specified CPU.
>   */
> -static int
> -arch_enable_uv_irq(char *irq_name, unsigned int irq, int cpu, int mmr_blade,
> -                    unsigned long mmr_offset, int limit)
> +static void uv_domain_activate(struct irq_domain *domain,
> +                            struct irq_data *irq_data)
>  {
> -     struct irq_cfg *cfg = irq_cfg(irq);
> -     unsigned long mmr_value;
> -     struct uv_IO_APIC_route_entry *entry;
> -     int mmr_pnode;
> -
> -     BUILD_BUG_ON(sizeof(struct uv_IO_APIC_route_entry) !=
> -                     sizeof(unsigned long));
> -
> -     if (limit == UV_AFFINITY_CPU)
> -             irq_set_status_flags(irq, IRQ_NO_BALANCING);
> -     else
> -             irq_set_status_flags(irq, IRQ_MOVE_PCNTXT);
> -
> -     irq_set_chip_and_handler_name(irq, &uv_irq_chip, handle_percpu_irq,
> -                                   irq_name);
> -
> -     mmr_value = 0;
> -     entry = (struct uv_IO_APIC_route_entry *)&mmr_value;
> -     entry->vector           = cfg->vector;
> -     entry->delivery_mode    = apic->irq_delivery_mode;
> -     entry->dest_mode        = apic->irq_dest_mode;
> -     entry->polarity         = 0;
> -     entry->trigger          = 0;
> -     entry->mask             = 0;
> -     entry->dest             = cfg->dest_apicid;
> -
> -     mmr_pnode = uv_blade_to_pnode(mmr_blade);
> -     uv_write_global_mmr64(mmr_pnode, mmr_offset, mmr_value);
> -
> -     if (cfg->move_in_progress)
> -             send_cleanup_vector(cfg);
> -
> -     return irq;
> +     uv_program_mmr(irqd_cfg(irq_data), irq_data->chip_data);
>  }
>  
>  /*
>   * Disable the specified MMR located on the specified blade so that MSIs are
>   * longer allowed to be sent.
>   */
> -static void arch_disable_uv_irq(int mmr_pnode, unsigned long mmr_offset)
> +static void uv_domain_deactivate(struct irq_domain *domain,
> +                              struct irq_data *irq_data)
>  {
>       unsigned long mmr_value;
>       struct uv_IO_APIC_route_entry *entry;
>  
> -     BUILD_BUG_ON(sizeof(struct uv_IO_APIC_route_entry) !=
> -                     sizeof(unsigned long));
> -
>       mmr_value = 0;
>       entry = (struct uv_IO_APIC_route_entry *)&mmr_value;
>       entry->mask = 1;
> -
> -     uv_write_global_mmr64(mmr_pnode, mmr_offset, mmr_value);
> +     uv_program_mmr(irqd_cfg(irq_data), irq_data->chip_data);
>  }
>  
> -static int
> -uv_set_irq_affinity(struct irq_data *data, const struct cpumask *mask,
> -                 bool force)
> -{
> -     struct irq_cfg *cfg = irqd_cfg(data);
> -     unsigned int dest;
> -     unsigned long mmr_value, mmr_offset;
> -     struct uv_IO_APIC_route_entry *entry;
> -     int mmr_pnode;
> -
> -     if (apic_set_affinity(data, mask, &dest))
> -             return -1;
> -
> -     mmr_value = 0;
> -     entry = (struct uv_IO_APIC_route_entry *)&mmr_value;
> -
> -     entry->vector           = cfg->vector;
> -     entry->delivery_mode    = apic->irq_delivery_mode;
> -     entry->dest_mode        = apic->irq_dest_mode;
> -     entry->polarity         = 0;
> -     entry->trigger          = 0;
> -     entry->mask             = 0;
> -     entry->dest             = dest;
> -
> -     /* Get previously stored MMR and pnode of hub sourcing interrupts */
> -     if (uv_irq_2_mmr_info(data->irq, &mmr_offset, &mmr_pnode))
> -             return -1;
> -
> -     uv_write_global_mmr64(mmr_pnode, mmr_offset, mmr_value);
> +static struct irq_domain_ops uv_domain_ops = {
> +     .alloc = uv_domain_alloc,
> +     .free = uv_domain_free,
> +     .activate = uv_domain_activate,
> +     .deactivate = uv_domain_deactivate,
> +};
>  
> -     if (cfg->move_in_progress)
> -             send_cleanup_vector(cfg);
> +static struct irq_domain *uv_get_irq_domain(void)
> +{
> +     static struct irq_domain *uv_domain;
> +     static DEFINE_MUTEX(uv_lock);
> +
> +     mutex_lock(&uv_lock);
> +     if (uv_domain == NULL) {
> +             uv_domain = irq_domain_add_tree(NULL, &uv_domain_ops, NULL);
> +             if (uv_domain)
> +                     uv_domain->parent = x86_vector_domain;
> +     }
> +     mutex_unlock(&uv_lock);
>  
> -     return IRQ_SET_MASK_OK_NOCOPY;
> +     return uv_domain;
>  }
>  
>  /*
> @@ -229,23 +181,20 @@ uv_set_irq_affinity(struct irq_data *data, const struct 
> cpumask *mask,
>  int uv_setup_irq(char *irq_name, int cpu, int mmr_blade,
>                unsigned long mmr_offset, int limit)
>  {
> -     int ret, irq;
>       struct irq_alloc_info info;
> +     struct irq_domain *domain = uv_get_irq_domain();
> +
> +     if (!domain)
> +             return -ENOMEM;
>  
>       init_irq_alloc_info(&info, cpumask_of(cpu));
> -     irq = irq_domain_alloc_irqs(NULL, 1, uv_blade_to_memory_nid(mmr_blade),
> -                                 &info);
> -     if (irq <= 0)
> -             return -EBUSY;
> -
> -     ret = arch_enable_uv_irq(irq_name, irq, cpu, mmr_blade, mmr_offset,
> -             limit);
> -     if (ret == irq)
> -             uv_set_irq_2_mmr_info(irq, mmr_offset, mmr_blade);
> -     else
> -             irq_domain_free_irqs(irq, 1);
> +     info.uv_limit = limit;
> +     info.uv_blade = mmr_blade;
> +     info.uv_offset = mmr_offset;
> +     info.uv_name = irq_name;
>  
> -     return ret;
> +     return irq_domain_alloc_irqs(domain, 1,
> +                                  uv_blade_to_memory_nid(mmr_blade), &info);
>  }
>  EXPORT_SYMBOL_GPL(uv_setup_irq);
>  
> @@ -258,26 +207,6 @@ EXPORT_SYMBOL_GPL(uv_setup_irq);
>   */
>  void uv_teardown_irq(unsigned int irq)
>  {
> -     struct uv_irq_2_mmr_pnode *e;
> -     struct rb_node *n;
> -     unsigned long irqflags;
> -
> -     spin_lock_irqsave(&uv_irq_lock, irqflags);
> -     n = uv_irq_root.rb_node;
> -     while (n) {
> -             e = rb_entry(n, struct uv_irq_2_mmr_pnode, list);
> -             if (e->irq == irq) {
> -                     arch_disable_uv_irq(e->pnode, e->offset);
> -                     rb_erase(n, &uv_irq_root);
> -                     kfree(e);
> -                     break;
> -             }
> -             if (irq < e->irq)
> -                     n = n->rb_left;
> -             else
> -                     n = n->rb_right;
> -     }
> -     spin_unlock_irqrestore(&uv_irq_lock, irqflags);
>       irq_domain_free_irqs(irq, 1);
>  }
>  EXPORT_SYMBOL_GPL(uv_teardown_irq);
> -- 
> 1.7.10.4
> 
> --
> To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
> the body of a message to majord...@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> Please read the FAQ at  http://www.tux.org/lkml/
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to