|
@@ -192,7 +192,7 @@ int irq_do_set_affinity(struct irq_data *data, const struct cpumask *mask,
|
|
|
switch (ret) {
|
|
|
case IRQ_SET_MASK_OK:
|
|
|
case IRQ_SET_MASK_OK_DONE:
|
|
|
- cpumask_copy(data->affinity, mask);
|
|
|
+ cpumask_copy(desc->irq_common_data.affinity, mask);
|
|
|
case IRQ_SET_MASK_OK_NOCOPY:
|
|
|
irq_set_thread_affinity(desc);
|
|
|
ret = 0;
|
|
@@ -304,7 +304,7 @@ static void irq_affinity_notify(struct work_struct *work)
|
|
|
if (irq_move_pending(&desc->irq_data))
|
|
|
irq_get_pending(cpumask, desc);
|
|
|
else
|
|
|
- cpumask_copy(cpumask, desc->irq_data.affinity);
|
|
|
+ cpumask_copy(cpumask, desc->irq_common_data.affinity);
|
|
|
raw_spin_unlock_irqrestore(&desc->lock, flags);
|
|
|
|
|
|
notify->notify(notify, cpumask);
|
|
@@ -375,9 +375,9 @@ static int setup_affinity(struct irq_desc *desc, struct cpumask *mask)
|
|
|
* one of the targets is online.
|
|
|
*/
|
|
|
if (irqd_has_set(&desc->irq_data, IRQD_AFFINITY_SET)) {
|
|
|
- if (cpumask_intersects(desc->irq_data.affinity,
|
|
|
+ if (cpumask_intersects(desc->irq_common_data.affinity,
|
|
|
cpu_online_mask))
|
|
|
- set = desc->irq_data.affinity;
|
|
|
+ set = desc->irq_common_data.affinity;
|
|
|
else
|
|
|
irqd_clear(&desc->irq_data, IRQD_AFFINITY_SET);
|
|
|
}
|
|
@@ -829,8 +829,8 @@ irq_thread_check_affinity(struct irq_desc *desc, struct irqaction *action)
|
|
|
* This code is triggered unconditionally. Check the affinity
|
|
|
* mask pointer. For CPU_MASK_OFFSTACK=n this is optimized out.
|
|
|
*/
|
|
|
- if (desc->irq_data.affinity)
|
|
|
- cpumask_copy(mask, desc->irq_data.affinity);
|
|
|
+ if (desc->irq_common_data.affinity)
|
|
|
+ cpumask_copy(mask, desc->irq_common_data.affinity);
|
|
|
else
|
|
|
valid = false;
|
|
|
raw_spin_unlock_irq(&desc->lock);
|