forked from Minki/linux
Merge branch 'irq/for-x86' into irq/core
Get the infrastructure patches which are required for x86/apic into core
This commit is contained in:
commit
f05218651b
@ -126,13 +126,21 @@ struct msi_desc;
|
||||
struct irq_domain;
|
||||
|
||||
/**
|
||||
* struct irq_data - per irq and irq chip data passed down to chip functions
|
||||
* struct irq_common_data - per irq data shared by all irqchips
|
||||
* @state_use_accessors: status information for irq chip functions.
|
||||
* Use accessor functions to deal with it
|
||||
*/
|
||||
struct irq_common_data {
|
||||
unsigned int state_use_accessors;
|
||||
};
|
||||
|
||||
/**
|
||||
* struct irq_data - per irq chip data passed down to chip functions
|
||||
* @mask: precomputed bitmask for accessing the chip registers
|
||||
* @irq: interrupt number
|
||||
* @hwirq: hardware interrupt number, local to the interrupt domain
|
||||
* @node: node index useful for balancing
|
||||
* @state_use_accessors: status information for irq chip functions.
|
||||
* Use accessor functions to deal with it
|
||||
* @common: point to data shared by all irqchips
|
||||
* @chip: low level interrupt hardware access
|
||||
* @domain: Interrupt translation domain; responsible for mapping
|
||||
* between hwirq number and linux irq number.
|
||||
@ -153,7 +161,7 @@ struct irq_data {
|
||||
unsigned int irq;
|
||||
unsigned long hwirq;
|
||||
unsigned int node;
|
||||
unsigned int state_use_accessors;
|
||||
struct irq_common_data *common;
|
||||
struct irq_chip *chip;
|
||||
struct irq_domain *domain;
|
||||
#ifdef CONFIG_IRQ_DOMAIN_HIERARCHY
|
||||
@ -166,7 +174,7 @@ struct irq_data {
|
||||
};
|
||||
|
||||
/*
|
||||
* Bit masks for irq_data.state
|
||||
* Bit masks for irq_common_data.state_use_accessors
|
||||
*
|
||||
* IRQD_TRIGGER_MASK - Mask for the trigger type bits
|
||||
* IRQD_SETAFFINITY_PENDING - Affinity setting is pending
|
||||
@ -198,34 +206,36 @@ enum {
|
||||
IRQD_WAKEUP_ARMED = (1 << 19),
|
||||
};
|
||||
|
||||
#define __irqd_to_state(d) ((d)->common->state_use_accessors)
|
||||
|
||||
static inline bool irqd_is_setaffinity_pending(struct irq_data *d)
|
||||
{
|
||||
return d->state_use_accessors & IRQD_SETAFFINITY_PENDING;
|
||||
return __irqd_to_state(d) & IRQD_SETAFFINITY_PENDING;
|
||||
}
|
||||
|
||||
static inline bool irqd_is_per_cpu(struct irq_data *d)
|
||||
{
|
||||
return d->state_use_accessors & IRQD_PER_CPU;
|
||||
return __irqd_to_state(d) & IRQD_PER_CPU;
|
||||
}
|
||||
|
||||
static inline bool irqd_can_balance(struct irq_data *d)
|
||||
{
|
||||
return !(d->state_use_accessors & (IRQD_PER_CPU | IRQD_NO_BALANCING));
|
||||
return !(__irqd_to_state(d) & (IRQD_PER_CPU | IRQD_NO_BALANCING));
|
||||
}
|
||||
|
||||
static inline bool irqd_affinity_was_set(struct irq_data *d)
|
||||
{
|
||||
return d->state_use_accessors & IRQD_AFFINITY_SET;
|
||||
return __irqd_to_state(d) & IRQD_AFFINITY_SET;
|
||||
}
|
||||
|
||||
static inline void irqd_mark_affinity_was_set(struct irq_data *d)
|
||||
{
|
||||
d->state_use_accessors |= IRQD_AFFINITY_SET;
|
||||
__irqd_to_state(d) |= IRQD_AFFINITY_SET;
|
||||
}
|
||||
|
||||
static inline u32 irqd_get_trigger_type(struct irq_data *d)
|
||||
{
|
||||
return d->state_use_accessors & IRQD_TRIGGER_MASK;
|
||||
return __irqd_to_state(d) & IRQD_TRIGGER_MASK;
|
||||
}
|
||||
|
||||
/*
|
||||
@ -233,43 +243,43 @@ static inline u32 irqd_get_trigger_type(struct irq_data *d)
|
||||
*/
|
||||
static inline void irqd_set_trigger_type(struct irq_data *d, u32 type)
|
||||
{
|
||||
d->state_use_accessors &= ~IRQD_TRIGGER_MASK;
|
||||
d->state_use_accessors |= type & IRQD_TRIGGER_MASK;
|
||||
__irqd_to_state(d) &= ~IRQD_TRIGGER_MASK;
|
||||
__irqd_to_state(d) |= type & IRQD_TRIGGER_MASK;
|
||||
}
|
||||
|
||||
static inline bool irqd_is_level_type(struct irq_data *d)
|
||||
{
|
||||
return d->state_use_accessors & IRQD_LEVEL;
|
||||
return __irqd_to_state(d) & IRQD_LEVEL;
|
||||
}
|
||||
|
||||
static inline bool irqd_is_wakeup_set(struct irq_data *d)
|
||||
{
|
||||
return d->state_use_accessors & IRQD_WAKEUP_STATE;
|
||||
return __irqd_to_state(d) & IRQD_WAKEUP_STATE;
|
||||
}
|
||||
|
||||
static inline bool irqd_can_move_in_process_context(struct irq_data *d)
|
||||
{
|
||||
return d->state_use_accessors & IRQD_MOVE_PCNTXT;
|
||||
return __irqd_to_state(d) & IRQD_MOVE_PCNTXT;
|
||||
}
|
||||
|
||||
static inline bool irqd_irq_disabled(struct irq_data *d)
|
||||
{
|
||||
return d->state_use_accessors & IRQD_IRQ_DISABLED;
|
||||
return __irqd_to_state(d) & IRQD_IRQ_DISABLED;
|
||||
}
|
||||
|
||||
static inline bool irqd_irq_masked(struct irq_data *d)
|
||||
{
|
||||
return d->state_use_accessors & IRQD_IRQ_MASKED;
|
||||
return __irqd_to_state(d) & IRQD_IRQ_MASKED;
|
||||
}
|
||||
|
||||
static inline bool irqd_irq_inprogress(struct irq_data *d)
|
||||
{
|
||||
return d->state_use_accessors & IRQD_IRQ_INPROGRESS;
|
||||
return __irqd_to_state(d) & IRQD_IRQ_INPROGRESS;
|
||||
}
|
||||
|
||||
static inline bool irqd_is_wakeup_armed(struct irq_data *d)
|
||||
{
|
||||
return d->state_use_accessors & IRQD_WAKEUP_ARMED;
|
||||
return __irqd_to_state(d) & IRQD_WAKEUP_ARMED;
|
||||
}
|
||||
|
||||
|
||||
@ -280,12 +290,12 @@ static inline bool irqd_is_wakeup_armed(struct irq_data *d)
|
||||
*/
|
||||
static inline void irqd_set_chained_irq_inprogress(struct irq_data *d)
|
||||
{
|
||||
d->state_use_accessors |= IRQD_IRQ_INPROGRESS;
|
||||
__irqd_to_state(d) |= IRQD_IRQ_INPROGRESS;
|
||||
}
|
||||
|
||||
static inline void irqd_clr_chained_irq_inprogress(struct irq_data *d)
|
||||
{
|
||||
d->state_use_accessors &= ~IRQD_IRQ_INPROGRESS;
|
||||
__irqd_to_state(d) &= ~IRQD_IRQ_INPROGRESS;
|
||||
}
|
||||
|
||||
static inline irq_hw_number_t irqd_to_hwirq(struct irq_data *d)
|
||||
@ -641,6 +651,23 @@ static inline u32 irq_get_trigger_type(unsigned int irq)
|
||||
return d ? irqd_get_trigger_type(d) : 0;
|
||||
}
|
||||
|
||||
static inline int irq_data_get_node(struct irq_data *d)
|
||||
{
|
||||
return d->node;
|
||||
}
|
||||
|
||||
static inline struct cpumask *irq_get_affinity_mask(int irq)
|
||||
{
|
||||
struct irq_data *d = irq_get_irq_data(irq);
|
||||
|
||||
return d ? d->affinity : NULL;
|
||||
}
|
||||
|
||||
static inline struct cpumask *irq_data_get_affinity_mask(struct irq_data *d)
|
||||
{
|
||||
return d->affinity;
|
||||
}
|
||||
|
||||
unsigned int arch_dynirq_lower_bound(unsigned int from);
|
||||
|
||||
int __irq_alloc_descs(int irq, unsigned int from, unsigned int cnt, int node,
|
||||
|
@ -17,7 +17,7 @@ struct pt_regs;
|
||||
|
||||
/**
|
||||
* struct irq_desc - interrupt descriptor
|
||||
* @irq_data: per irq and chip data passed down to chip functions
|
||||
* @irq_common_data: per irq and chip data passed down to chip functions
|
||||
* @kstat_irqs: irq stats per cpu
|
||||
* @handle_irq: highlevel irq-events handler
|
||||
* @preflow_handler: handler called before the flow handler (currently used by sparc)
|
||||
@ -47,6 +47,7 @@ struct pt_regs;
|
||||
* @name: flow handler name for /proc/interrupts output
|
||||
*/
|
||||
struct irq_desc {
|
||||
struct irq_common_data irq_common_data;
|
||||
struct irq_data irq_data;
|
||||
unsigned int __percpu *kstat_irqs;
|
||||
irq_flow_handler_t handle_irq;
|
||||
@ -93,6 +94,15 @@ struct irq_desc {
|
||||
extern struct irq_desc irq_desc[NR_IRQS];
|
||||
#endif
|
||||
|
||||
static inline struct irq_desc *irq_data_to_desc(struct irq_data *data)
|
||||
{
|
||||
#ifdef CONFIG_IRQ_DOMAIN_HIERARCHY
|
||||
return irq_to_desc(data->irq);
|
||||
#else
|
||||
return container_of(data, struct irq_desc, irq_data);
|
||||
#endif
|
||||
}
|
||||
|
||||
static inline struct irq_data *irq_desc_get_irq_data(struct irq_desc *desc)
|
||||
{
|
||||
return &desc->irq_data;
|
||||
|
@ -59,8 +59,6 @@ enum {
|
||||
#include "debug.h"
|
||||
#include "settings.h"
|
||||
|
||||
#define irq_data_to_desc(data) container_of(data, struct irq_desc, irq_data)
|
||||
|
||||
extern int __irq_set_trigger(struct irq_desc *desc, unsigned int irq,
|
||||
unsigned long flags);
|
||||
extern void __disable_irq(struct irq_desc *desc, unsigned int irq);
|
||||
@ -170,27 +168,27 @@ irq_put_desc_unlock(struct irq_desc *desc, unsigned long flags)
|
||||
*/
|
||||
static inline void irqd_set_move_pending(struct irq_data *d)
|
||||
{
|
||||
d->state_use_accessors |= IRQD_SETAFFINITY_PENDING;
|
||||
__irqd_to_state(d) |= IRQD_SETAFFINITY_PENDING;
|
||||
}
|
||||
|
||||
static inline void irqd_clr_move_pending(struct irq_data *d)
|
||||
{
|
||||
d->state_use_accessors &= ~IRQD_SETAFFINITY_PENDING;
|
||||
__irqd_to_state(d) &= ~IRQD_SETAFFINITY_PENDING;
|
||||
}
|
||||
|
||||
static inline void irqd_clear(struct irq_data *d, unsigned int mask)
|
||||
{
|
||||
d->state_use_accessors &= ~mask;
|
||||
__irqd_to_state(d) &= ~mask;
|
||||
}
|
||||
|
||||
static inline void irqd_set(struct irq_data *d, unsigned int mask)
|
||||
{
|
||||
d->state_use_accessors |= mask;
|
||||
__irqd_to_state(d) |= mask;
|
||||
}
|
||||
|
||||
static inline bool irqd_has_set(struct irq_data *d, unsigned int mask)
|
||||
{
|
||||
return d->state_use_accessors & mask;
|
||||
return __irqd_to_state(d) & mask;
|
||||
}
|
||||
|
||||
static inline void kstat_incr_irqs_this_cpu(unsigned int irq, struct irq_desc *desc)
|
||||
@ -199,6 +197,11 @@ static inline void kstat_incr_irqs_this_cpu(unsigned int irq, struct irq_desc *d
|
||||
__this_cpu_inc(kstat.irqs_sum);
|
||||
}
|
||||
|
||||
static inline int irq_desc_get_node(struct irq_desc *desc)
|
||||
{
|
||||
return irq_data_get_node(&desc->irq_data);
|
||||
}
|
||||
|
||||
#ifdef CONFIG_PM_SLEEP
|
||||
bool irq_pm_check_wakeup(struct irq_desc *desc);
|
||||
void irq_pm_install_action(struct irq_desc *desc, struct irqaction *action);
|
||||
|
@ -59,16 +59,10 @@ static void desc_smp_init(struct irq_desc *desc, int node)
|
||||
#endif
|
||||
}
|
||||
|
||||
static inline int desc_node(struct irq_desc *desc)
|
||||
{
|
||||
return desc->irq_data.node;
|
||||
}
|
||||
|
||||
#else
|
||||
static inline int
|
||||
alloc_masks(struct irq_desc *desc, gfp_t gfp, int node) { return 0; }
|
||||
static inline void desc_smp_init(struct irq_desc *desc, int node) { }
|
||||
static inline int desc_node(struct irq_desc *desc) { return 0; }
|
||||
#endif
|
||||
|
||||
static void desc_set_defaults(unsigned int irq, struct irq_desc *desc, int node,
|
||||
@ -76,6 +70,7 @@ static void desc_set_defaults(unsigned int irq, struct irq_desc *desc, int node,
|
||||
{
|
||||
int cpu;
|
||||
|
||||
desc->irq_data.common = &desc->irq_common_data;
|
||||
desc->irq_data.irq = irq;
|
||||
desc->irq_data.chip = &no_irq_chip;
|
||||
desc->irq_data.chip_data = NULL;
|
||||
@ -299,7 +294,7 @@ static void free_desc(unsigned int irq)
|
||||
unsigned long flags;
|
||||
|
||||
raw_spin_lock_irqsave(&desc->lock, flags);
|
||||
desc_set_defaults(irq, desc, desc_node(desc), NULL);
|
||||
desc_set_defaults(irq, desc, irq_desc_get_node(desc), NULL);
|
||||
raw_spin_unlock_irqrestore(&desc->lock, flags);
|
||||
}
|
||||
|
||||
|
@ -830,10 +830,12 @@ static struct irq_data *irq_domain_insert_irq_data(struct irq_domain *domain,
|
||||
{
|
||||
struct irq_data *irq_data;
|
||||
|
||||
irq_data = kzalloc_node(sizeof(*irq_data), GFP_KERNEL, child->node);
|
||||
irq_data = kzalloc_node(sizeof(*irq_data), GFP_KERNEL,
|
||||
irq_data_get_node(child));
|
||||
if (irq_data) {
|
||||
child->parent_data = irq_data;
|
||||
irq_data->irq = child->irq;
|
||||
irq_data->common = child->common;
|
||||
irq_data->node = child->node;
|
||||
irq_data->domain = domain;
|
||||
}
|
||||
|
@ -363,7 +363,7 @@ static int
|
||||
setup_affinity(unsigned int irq, struct irq_desc *desc, struct cpumask *mask)
|
||||
{
|
||||
struct cpumask *set = irq_default_affinity;
|
||||
int node = desc->irq_data.node;
|
||||
int node = irq_desc_get_node(desc);
|
||||
|
||||
/* Excludes PER_CPU and NO_BALANCE interrupts */
|
||||
if (!irq_can_set_affinity(irq))
|
||||
|
@ -7,21 +7,21 @@
|
||||
void irq_move_masked_irq(struct irq_data *idata)
|
||||
{
|
||||
struct irq_desc *desc = irq_data_to_desc(idata);
|
||||
struct irq_chip *chip = idata->chip;
|
||||
struct irq_chip *chip = desc->irq_data.chip;
|
||||
|
||||
if (likely(!irqd_is_setaffinity_pending(&desc->irq_data)))
|
||||
return;
|
||||
|
||||
irqd_clr_move_pending(&desc->irq_data);
|
||||
|
||||
/*
|
||||
* Paranoia: cpu-local interrupts shouldn't be calling in here anyway.
|
||||
*/
|
||||
if (!irqd_can_balance(&desc->irq_data)) {
|
||||
if (irqd_is_per_cpu(&desc->irq_data)) {
|
||||
WARN_ON(1);
|
||||
return;
|
||||
}
|
||||
|
||||
irqd_clr_move_pending(&desc->irq_data);
|
||||
|
||||
if (unlikely(cpumask_empty(desc->pending_mask)))
|
||||
return;
|
||||
|
||||
@ -52,6 +52,13 @@ void irq_move_irq(struct irq_data *idata)
|
||||
{
|
||||
bool masked;
|
||||
|
||||
/*
|
||||
* Get top level irq_data when CONFIG_IRQ_DOMAIN_HIERARCHY is enabled,
|
||||
* and it should be optimized away when CONFIG_IRQ_DOMAIN_HIERARCHY is
|
||||
* disabled. So we avoid an "#ifdef CONFIG_IRQ_DOMAIN_HIERARCHY" here.
|
||||
*/
|
||||
idata = irq_desc_get_irq_data(irq_data_to_desc(idata));
|
||||
|
||||
if (likely(!irqd_is_setaffinity_pending(idata)))
|
||||
return;
|
||||
|
||||
|
@ -241,7 +241,7 @@ static int irq_node_proc_show(struct seq_file *m, void *v)
|
||||
{
|
||||
struct irq_desc *desc = irq_to_desc((long) m->private);
|
||||
|
||||
seq_printf(m, "%d\n", desc->irq_data.node);
|
||||
seq_printf(m, "%d\n", irq_desc_get_node(desc));
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user