提交 d585d060 编写于 作者: T Thomas Gleixner

intr_remap: Simplify the code further

Having irq_2_iommu in struct irq_cfg allows further simplifications.
Signed-off-by: NThomas Gleixner <tglx@linutronix.de>
Reviewed-by: NIngo Molnar <mingo@elte.hu>
Acked-by: NSuresh Siddha <suresh.b.siddha@intel.com>
Cc: David Woodhouse <dwmw2@infradead.org>
Cc: Jesse Barnes <jbarnes@virtuousgeek.org>
上级 349d6767
...@@ -46,58 +46,31 @@ static __init int setup_intremap(char *str) ...@@ -46,58 +46,31 @@ static __init int setup_intremap(char *str)
} }
early_param("intremap", setup_intremap); early_param("intremap", setup_intremap);
static DEFINE_SPINLOCK(irq_2_ir_lock);
static struct irq_2_iommu *irq_2_iommu(unsigned int irq) static struct irq_2_iommu *irq_2_iommu(unsigned int irq)
{ {
struct irq_cfg *cfg = get_irq_chip_data(irq); struct irq_cfg *cfg = get_irq_chip_data(irq);
return cfg ? &cfg->irq_2_iommu : NULL; return cfg ? &cfg->irq_2_iommu : NULL;
} }
static struct irq_2_iommu *irq_2_iommu_alloc(unsigned int irq)
{
return irq_2_iommu(irq);
}
static void irq_2_iommu_free(unsigned int irq)
{
}
static DEFINE_SPINLOCK(irq_2_ir_lock);
static struct irq_2_iommu *valid_irq_2_iommu(unsigned int irq)
{
struct irq_2_iommu *irq_iommu;
irq_iommu = irq_2_iommu(irq);
if (!irq_iommu)
return NULL;
if (!irq_iommu->iommu)
return NULL;
return irq_iommu;
}
int irq_remapped(int irq) int irq_remapped(int irq)
{ {
return valid_irq_2_iommu(irq) != NULL; struct irq_2_iommu *irq_iommu = irq_2_iommu(irq);
return irq_iommu ? irq_iommu->iommu != NULL : 0;
} }
int get_irte(int irq, struct irte *entry) int get_irte(int irq, struct irte *entry)
{ {
int index; struct irq_2_iommu *irq_iommu = irq_2_iommu(irq);
struct irq_2_iommu *irq_iommu;
unsigned long flags; unsigned long flags;
int index;
if (!entry) if (!entry || !irq_iommu)
return -1; return -1;
spin_lock_irqsave(&irq_2_ir_lock, flags); spin_lock_irqsave(&irq_2_ir_lock, flags);
irq_iommu = valid_irq_2_iommu(irq);
if (!irq_iommu) {
spin_unlock_irqrestore(&irq_2_ir_lock, flags);
return -1;
}
index = irq_iommu->irte_index + irq_iommu->sub_handle; index = irq_iommu->irte_index + irq_iommu->sub_handle;
*entry = *(irq_iommu->iommu->ir_table->base + index); *entry = *(irq_iommu->iommu->ir_table->base + index);
...@@ -109,20 +82,14 @@ int get_irte(int irq, struct irte *entry) ...@@ -109,20 +82,14 @@ int get_irte(int irq, struct irte *entry)
int alloc_irte(struct intel_iommu *iommu, int irq, u16 count) int alloc_irte(struct intel_iommu *iommu, int irq, u16 count)
{ {
struct ir_table *table = iommu->ir_table; struct ir_table *table = iommu->ir_table;
struct irq_2_iommu *irq_iommu; struct irq_2_iommu *irq_iommu = irq_2_iommu(irq);
u16 index, start_index; u16 index, start_index;
unsigned int mask = 0; unsigned int mask = 0;
unsigned long flags; unsigned long flags;
int i; int i;
if (!count) if (!count || !irq_iommu)
return -1;
#ifndef CONFIG_SPARSE_IRQ
/* protect irq_2_iommu_alloc later */
if (irq >= nr_irqs)
return -1; return -1;
#endif
/* /*
* start the IRTE search from index 0. * start the IRTE search from index 0.
...@@ -163,13 +130,6 @@ int alloc_irte(struct intel_iommu *iommu, int irq, u16 count) ...@@ -163,13 +130,6 @@ int alloc_irte(struct intel_iommu *iommu, int irq, u16 count)
for (i = index; i < index + count; i++) for (i = index; i < index + count; i++)
table->base[i].present = 1; table->base[i].present = 1;
irq_iommu = irq_2_iommu_alloc(irq);
if (!irq_iommu) {
spin_unlock_irqrestore(&irq_2_ir_lock, flags);
printk(KERN_ERR "can't allocate irq_2_iommu\n");
return -1;
}
irq_iommu->iommu = iommu; irq_iommu->iommu = iommu;
irq_iommu->irte_index = index; irq_iommu->irte_index = index;
irq_iommu->sub_handle = 0; irq_iommu->sub_handle = 0;
...@@ -193,17 +153,14 @@ static int qi_flush_iec(struct intel_iommu *iommu, int index, int mask) ...@@ -193,17 +153,14 @@ static int qi_flush_iec(struct intel_iommu *iommu, int index, int mask)
int map_irq_to_irte_handle(int irq, u16 *sub_handle) int map_irq_to_irte_handle(int irq, u16 *sub_handle)
{ {
int index; struct irq_2_iommu *irq_iommu = irq_2_iommu(irq);
struct irq_2_iommu *irq_iommu;
unsigned long flags; unsigned long flags;
int index;
spin_lock_irqsave(&irq_2_ir_lock, flags); if (!irq_iommu)
irq_iommu = valid_irq_2_iommu(irq);
if (!irq_iommu) {
spin_unlock_irqrestore(&irq_2_ir_lock, flags);
return -1; return -1;
}
spin_lock_irqsave(&irq_2_ir_lock, flags);
*sub_handle = irq_iommu->sub_handle; *sub_handle = irq_iommu->sub_handle;
index = irq_iommu->irte_index; index = irq_iommu->irte_index;
spin_unlock_irqrestore(&irq_2_ir_lock, flags); spin_unlock_irqrestore(&irq_2_ir_lock, flags);
...@@ -212,18 +169,13 @@ int map_irq_to_irte_handle(int irq, u16 *sub_handle) ...@@ -212,18 +169,13 @@ int map_irq_to_irte_handle(int irq, u16 *sub_handle)
int set_irte_irq(int irq, struct intel_iommu *iommu, u16 index, u16 subhandle) int set_irte_irq(int irq, struct intel_iommu *iommu, u16 index, u16 subhandle)
{ {
struct irq_2_iommu *irq_iommu; struct irq_2_iommu *irq_iommu = irq_2_iommu(irq);
unsigned long flags; unsigned long flags;
spin_lock_irqsave(&irq_2_ir_lock, flags); if (!irq_iommu)
irq_iommu = irq_2_iommu_alloc(irq);
if (!irq_iommu) {
spin_unlock_irqrestore(&irq_2_ir_lock, flags);
printk(KERN_ERR "can't allocate irq_2_iommu\n");
return -1; return -1;
}
spin_lock_irqsave(&irq_2_ir_lock, flags);
irq_iommu->iommu = iommu; irq_iommu->iommu = iommu;
irq_iommu->irte_index = index; irq_iommu->irte_index = index;
...@@ -237,19 +189,16 @@ int set_irte_irq(int irq, struct intel_iommu *iommu, u16 index, u16 subhandle) ...@@ -237,19 +189,16 @@ int set_irte_irq(int irq, struct intel_iommu *iommu, u16 index, u16 subhandle)
int modify_irte(int irq, struct irte *irte_modified) int modify_irte(int irq, struct irte *irte_modified)
{ {
int rc; struct irq_2_iommu *irq_iommu = irq_2_iommu(irq);
int index;
struct irte *irte;
struct intel_iommu *iommu; struct intel_iommu *iommu;
struct irq_2_iommu *irq_iommu;
unsigned long flags; unsigned long flags;
struct irte *irte;
int rc, index;
spin_lock_irqsave(&irq_2_ir_lock, flags); if (!irq_iommu)
irq_iommu = valid_irq_2_iommu(irq);
if (!irq_iommu) {
spin_unlock_irqrestore(&irq_2_ir_lock, flags);
return -1; return -1;
}
spin_lock_irqsave(&irq_2_ir_lock, flags);
iommu = irq_iommu->iommu; iommu = irq_iommu->iommu;
...@@ -322,16 +271,14 @@ static int clear_entries(struct irq_2_iommu *irq_iommu) ...@@ -322,16 +271,14 @@ static int clear_entries(struct irq_2_iommu *irq_iommu)
int free_irte(int irq) int free_irte(int irq)
{ {
int rc = 0; struct irq_2_iommu *irq_iommu = irq_2_iommu(irq);
struct irq_2_iommu *irq_iommu;
unsigned long flags; unsigned long flags;
int rc;
spin_lock_irqsave(&irq_2_ir_lock, flags); if (!irq_iommu)
irq_iommu = valid_irq_2_iommu(irq);
if (!irq_iommu) {
spin_unlock_irqrestore(&irq_2_ir_lock, flags);
return -1; return -1;
}
spin_lock_irqsave(&irq_2_ir_lock, flags);
rc = clear_entries(irq_iommu); rc = clear_entries(irq_iommu);
...@@ -342,8 +289,6 @@ int free_irte(int irq) ...@@ -342,8 +289,6 @@ int free_irte(int irq)
spin_unlock_irqrestore(&irq_2_ir_lock, flags); spin_unlock_irqrestore(&irq_2_ir_lock, flags);
irq_2_iommu_free(irq);
return rc; return rc;
} }
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册