|
@@ -225,7 +225,7 @@ struct rw_semaphore __sched *rwsem_down_read_failed(struct rw_semaphore *sem)
|
|
|
long count, adjustment = -RWSEM_ACTIVE_READ_BIAS;
|
|
|
struct rwsem_waiter waiter;
|
|
|
struct task_struct *tsk = current;
|
|
|
- WAKE_Q(wake_q);
|
|
|
+ DEFINE_WAKE_Q(wake_q);
|
|
|
|
|
|
waiter.task = tsk;
|
|
|
waiter.type = RWSEM_WAITING_FOR_READ;
|
|
@@ -461,7 +461,7 @@ __rwsem_down_write_failed_common(struct rw_semaphore *sem, int state)
|
|
|
bool waiting = true; /* any queued threads before us */
|
|
|
struct rwsem_waiter waiter;
|
|
|
struct rw_semaphore *ret = sem;
|
|
|
- WAKE_Q(wake_q);
|
|
|
+ DEFINE_WAKE_Q(wake_q);
|
|
|
|
|
|
/* undo write bias from down_write operation, stop active locking */
|
|
|
count = atomic_long_sub_return(RWSEM_ACTIVE_WRITE_BIAS, &sem->count);
|
|
@@ -495,7 +495,7 @@ __rwsem_down_write_failed_common(struct rw_semaphore *sem, int state)
|
|
|
* wake any read locks that were queued ahead of us.
|
|
|
*/
|
|
|
if (count > RWSEM_WAITING_BIAS) {
|
|
|
- WAKE_Q(wake_q);
|
|
|
+ DEFINE_WAKE_Q(wake_q);
|
|
|
|
|
|
__rwsem_mark_wake(sem, RWSEM_WAKE_READERS, &wake_q);
|
|
|
/*
|
|
@@ -571,7 +571,7 @@ __visible
|
|
|
struct rw_semaphore *rwsem_wake(struct rw_semaphore *sem)
|
|
|
{
|
|
|
unsigned long flags;
|
|
|
- WAKE_Q(wake_q);
|
|
|
+ DEFINE_WAKE_Q(wake_q);
|
|
|
|
|
|
/*
|
|
|
* If a spinner is present, it is not necessary to do the wakeup.
|
|
@@ -625,7 +625,7 @@ __visible
|
|
|
struct rw_semaphore *rwsem_downgrade_wake(struct rw_semaphore *sem)
|
|
|
{
|
|
|
unsigned long flags;
|
|
|
- WAKE_Q(wake_q);
|
|
|
+ DEFINE_WAKE_Q(wake_q);
|
|
|
|
|
|
raw_spin_lock_irqsave(&sem->wait_lock, flags);
|
|
|
|