|
@@ -67,8 +67,8 @@ void arch_spin_lock_wait(arch_spinlock_t *lp)
|
|
|
|
|
|
/* Pass the virtual CPU to the lock holder if it is not running */
|
|
/* Pass the virtual CPU to the lock holder if it is not running */
|
|
owner = arch_load_niai4(&lp->lock);
|
|
owner = arch_load_niai4(&lp->lock);
|
|
- if (owner && arch_vcpu_is_preempted(~owner))
|
|
|
|
- smp_yield_cpu(~owner);
|
|
|
|
|
|
+ if (owner && arch_vcpu_is_preempted(owner - 1))
|
|
|
|
+ smp_yield_cpu(owner - 1);
|
|
|
|
|
|
count = spin_retry;
|
|
count = spin_retry;
|
|
while (1) {
|
|
while (1) {
|
|
@@ -87,8 +87,8 @@ void arch_spin_lock_wait(arch_spinlock_t *lp)
|
|
* yield the CPU unconditionally. For LPAR rely on the
|
|
* yield the CPU unconditionally. For LPAR rely on the
|
|
* sense running status.
|
|
* sense running status.
|
|
*/
|
|
*/
|
|
- if (!MACHINE_IS_LPAR || arch_vcpu_is_preempted(~owner))
|
|
|
|
- smp_yield_cpu(~owner);
|
|
|
|
|
|
+ if (!MACHINE_IS_LPAR || arch_vcpu_is_preempted(owner - 1))
|
|
|
|
+ smp_yield_cpu(owner - 1);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
EXPORT_SYMBOL(arch_spin_lock_wait);
|
|
EXPORT_SYMBOL(arch_spin_lock_wait);
|
|
@@ -102,8 +102,8 @@ void arch_spin_lock_wait_flags(arch_spinlock_t *lp, unsigned long flags)
|
|
|
|
|
|
/* Pass the virtual CPU to the lock holder if it is not running */
|
|
/* Pass the virtual CPU to the lock holder if it is not running */
|
|
owner = arch_load_niai4(&lp->lock);
|
|
owner = arch_load_niai4(&lp->lock);
|
|
- if (owner && arch_vcpu_is_preempted(~owner))
|
|
|
|
- smp_yield_cpu(~owner);
|
|
|
|
|
|
+ if (owner && arch_vcpu_is_preempted(owner - 1))
|
|
|
|
+ smp_yield_cpu(owner - 1);
|
|
|
|
|
|
count = spin_retry;
|
|
count = spin_retry;
|
|
while (1) {
|
|
while (1) {
|
|
@@ -124,8 +124,8 @@ void arch_spin_lock_wait_flags(arch_spinlock_t *lp, unsigned long flags)
|
|
* yield the CPU unconditionally. For LPAR rely on the
|
|
* yield the CPU unconditionally. For LPAR rely on the
|
|
* sense running status.
|
|
* sense running status.
|
|
*/
|
|
*/
|
|
- if (!MACHINE_IS_LPAR || arch_vcpu_is_preempted(~owner))
|
|
|
|
- smp_yield_cpu(~owner);
|
|
|
|
|
|
+ if (!MACHINE_IS_LPAR || arch_vcpu_is_preempted(owner - 1))
|
|
|
|
+ smp_yield_cpu(owner - 1);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
EXPORT_SYMBOL(arch_spin_lock_wait_flags);
|
|
EXPORT_SYMBOL(arch_spin_lock_wait_flags);
|
|
@@ -158,8 +158,8 @@ void _raw_read_lock_wait(arch_rwlock_t *rw)
|
|
owner = 0;
|
|
owner = 0;
|
|
while (1) {
|
|
while (1) {
|
|
if (count-- <= 0) {
|
|
if (count-- <= 0) {
|
|
- if (owner && arch_vcpu_is_preempted(~owner))
|
|
|
|
- smp_yield_cpu(~owner);
|
|
|
|
|
|
+ if (owner && arch_vcpu_is_preempted(owner - 1))
|
|
|
|
+ smp_yield_cpu(owner - 1);
|
|
count = spin_retry;
|
|
count = spin_retry;
|
|
}
|
|
}
|
|
old = ACCESS_ONCE(rw->lock);
|
|
old = ACCESS_ONCE(rw->lock);
|
|
@@ -198,8 +198,8 @@ void _raw_write_lock_wait(arch_rwlock_t *rw, int prev)
|
|
owner = 0;
|
|
owner = 0;
|
|
while (1) {
|
|
while (1) {
|
|
if (count-- <= 0) {
|
|
if (count-- <= 0) {
|
|
- if (owner && arch_vcpu_is_preempted(~owner))
|
|
|
|
- smp_yield_cpu(~owner);
|
|
|
|
|
|
+ if (owner && arch_vcpu_is_preempted(owner - 1))
|
|
|
|
+ smp_yield_cpu(owner - 1);
|
|
count = spin_retry;
|
|
count = spin_retry;
|
|
}
|
|
}
|
|
old = ACCESS_ONCE(rw->lock);
|
|
old = ACCESS_ONCE(rw->lock);
|
|
@@ -226,8 +226,8 @@ void _raw_write_lock_wait(arch_rwlock_t *rw)
|
|
owner = 0;
|
|
owner = 0;
|
|
while (1) {
|
|
while (1) {
|
|
if (count-- <= 0) {
|
|
if (count-- <= 0) {
|
|
- if (owner && arch_vcpu_is_preempted(~owner))
|
|
|
|
- smp_yield_cpu(~owner);
|
|
|
|
|
|
+ if (owner && arch_vcpu_is_preempted(owner - 1))
|
|
|
|
+ smp_yield_cpu(owner - 1);
|
|
count = spin_retry;
|
|
count = spin_retry;
|
|
}
|
|
}
|
|
old = ACCESS_ONCE(rw->lock);
|
|
old = ACCESS_ONCE(rw->lock);
|
|
@@ -265,8 +265,8 @@ void arch_lock_relax(int cpu)
|
|
{
|
|
{
|
|
if (!cpu)
|
|
if (!cpu)
|
|
return;
|
|
return;
|
|
- if (MACHINE_IS_LPAR && !arch_vcpu_is_preempted(~cpu))
|
|
|
|
|
|
+ if (MACHINE_IS_LPAR && !arch_vcpu_is_preempted(cpu - 1))
|
|
return;
|
|
return;
|
|
- smp_yield_cpu(~cpu);
|
|
|
|
|
|
+ smp_yield_cpu(cpu - 1);
|
|
}
|
|
}
|
|
EXPORT_SYMBOL(arch_lock_relax);
|
|
EXPORT_SYMBOL(arch_lock_relax);
|