|
@@ -433,12 +433,13 @@ static inline bool rwsem_has_spinner(struct rw_semaphore *sem)
|
|
|
/*
|
|
|
* Wait until we successfully acquire the write lock
|
|
|
*/
|
|
|
-__visible
|
|
|
-struct rw_semaphore __sched *rwsem_down_write_failed(struct rw_semaphore *sem)
|
|
|
+static inline struct rw_semaphore *
|
|
|
+__rwsem_down_write_failed_common(struct rw_semaphore *sem, int state)
|
|
|
{
|
|
|
long count;
|
|
|
bool waiting = true; /* any queued threads before us */
|
|
|
struct rwsem_waiter waiter;
|
|
|
+ struct rw_semaphore *ret = sem;
|
|
|
|
|
|
/* undo write bias from down_write operation, stop active locking */
|
|
|
count = rwsem_atomic_update(-RWSEM_ACTIVE_WRITE_BIAS, sem);
|
|
@@ -478,7 +479,7 @@ struct rw_semaphore __sched *rwsem_down_write_failed(struct rw_semaphore *sem)
|
|
|
count = rwsem_atomic_update(RWSEM_WAITING_BIAS, sem);
|
|
|
|
|
|
/* wait until we successfully acquire the lock */
|
|
|
- set_current_state(TASK_UNINTERRUPTIBLE);
|
|
|
+ set_current_state(state);
|
|
|
while (true) {
|
|
|
if (rwsem_try_write_lock(count, sem))
|
|
|
break;
|
|
@@ -486,21 +487,39 @@ struct rw_semaphore __sched *rwsem_down_write_failed(struct rw_semaphore *sem)
|
|
|
|
|
|
/* Block until there are no active lockers. */
|
|
|
do {
|
|
|
+ if (signal_pending_state(state, current)) {
|
|
|
+ raw_spin_lock_irq(&sem->wait_lock);
|
|
|
+ ret = ERR_PTR(-EINTR);
|
|
|
+ goto out;
|
|
|
+ }
|
|
|
schedule();
|
|
|
- set_current_state(TASK_UNINTERRUPTIBLE);
|
|
|
+ set_current_state(state);
|
|
|
} while ((count = sem->count) & RWSEM_ACTIVE_MASK);
|
|
|
|
|
|
raw_spin_lock_irq(&sem->wait_lock);
|
|
|
}
|
|
|
+out:
|
|
|
__set_current_state(TASK_RUNNING);
|
|
|
-
|
|
|
list_del(&waiter.list);
|
|
|
raw_spin_unlock_irq(&sem->wait_lock);
|
|
|
|
|
|
- return sem;
|
|
|
+ return ret;
|
|
|
+}
|
|
|
+
|
|
|
+__visible struct rw_semaphore * __sched
|
|
|
+rwsem_down_write_failed(struct rw_semaphore *sem)
|
|
|
+{
|
|
|
+ return __rwsem_down_write_failed_common(sem, TASK_UNINTERRUPTIBLE);
|
|
|
}
|
|
|
EXPORT_SYMBOL(rwsem_down_write_failed);
|
|
|
|
|
|
+__visible struct rw_semaphore * __sched
|
|
|
+rwsem_down_write_failed_killable(struct rw_semaphore *sem)
|
|
|
+{
|
|
|
+ return __rwsem_down_write_failed_common(sem, TASK_KILLABLE);
|
|
|
+}
|
|
|
+EXPORT_SYMBOL(rwsem_down_write_failed_killable);
|
|
|
+
|
|
|
/*
|
|
|
* handle waking up a waiter on the semaphore
|
|
|
* - up_read/up_write has decremented the active part of count if we come here
|