Commit a33a5d2d authored by Thomas Gleixner's avatar Thomas Gleixner

genirq/generic_pending: Do not lose pending affinity update

The generic pending interrupt mechanism moves interrupts from the interrupt
handler on the original target CPU to the new destination CPU. This is
required for x86 and ia64 due to the way the interrupt delivery and
acknowledge works if the interrupts are not remapped.

However that update can fail for various reasons. Some of them are valid
reasons to discard the pending update, but the case, when the previous move
has not been fully cleaned up is not a legit reason to fail.

Check the return value of irq_do_set_affinity() for -EBUSY, which indicates
a pending cleanup, and rearm the pending move in the irq dexcriptor so it's
tried again when the next interrupt arrives.

Fixes: 996c5912 ("x86/irq: Plug vector cleanup race")
Signed-off-by: default avatarThomas Gleixner <tglx@linutronix.de>
Tested-by: default avatarSong Liu <songliubraving@fb.com>
Cc: Joerg Roedel <jroedel@suse.de>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Song Liu <liu.song.a23@gmail.com>
Cc: Dmitry Safonov <0x7f454c46@gmail.com>
Cc: stable@vger.kernel.org
Cc: Mike Travis <mike.travis@hpe.com>
Cc: Borislav Petkov <bp@alien8.de>
Cc: Tariq Toukan <tariqt@mellanox.com>
Link: https://lkml.kernel.org/r/20180604162224.386544292@linutronix.de
parent 80ae7b1a
...@@ -38,17 +38,18 @@ bool irq_fixup_move_pending(struct irq_desc *desc, bool force_clear) ...@@ -38,17 +38,18 @@ bool irq_fixup_move_pending(struct irq_desc *desc, bool force_clear)
void irq_move_masked_irq(struct irq_data *idata) void irq_move_masked_irq(struct irq_data *idata)
{ {
struct irq_desc *desc = irq_data_to_desc(idata); struct irq_desc *desc = irq_data_to_desc(idata);
struct irq_chip *chip = desc->irq_data.chip; struct irq_data *data = &desc->irq_data;
struct irq_chip *chip = data->chip;
if (likely(!irqd_is_setaffinity_pending(&desc->irq_data))) if (likely(!irqd_is_setaffinity_pending(data)))
return; return;
irqd_clr_move_pending(&desc->irq_data); irqd_clr_move_pending(data);
/* /*
* Paranoia: cpu-local interrupts shouldn't be calling in here anyway. * Paranoia: cpu-local interrupts shouldn't be calling in here anyway.
*/ */
if (irqd_is_per_cpu(&desc->irq_data)) { if (irqd_is_per_cpu(data)) {
WARN_ON(1); WARN_ON(1);
return; return;
} }
...@@ -73,9 +74,20 @@ void irq_move_masked_irq(struct irq_data *idata) ...@@ -73,9 +74,20 @@ void irq_move_masked_irq(struct irq_data *idata)
* For correct operation this depends on the caller * For correct operation this depends on the caller
* masking the irqs. * masking the irqs.
*/ */
if (cpumask_any_and(desc->pending_mask, cpu_online_mask) < nr_cpu_ids) if (cpumask_any_and(desc->pending_mask, cpu_online_mask) < nr_cpu_ids) {
irq_do_set_affinity(&desc->irq_data, desc->pending_mask, false); int ret;
ret = irq_do_set_affinity(data, desc->pending_mask, false);
/*
* If the there is a cleanup pending in the underlying
* vector management, reschedule the move for the next
* interrupt. Leave desc->pending_mask intact.
*/
if (ret == -EBUSY) {
irqd_set_move_pending(data);
return;
}
}
cpumask_clear(desc->pending_mask); cpumask_clear(desc->pending_mask);
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment