|
@@ -68,25 +68,17 @@ static void blk_mq_sched_mark_restart_hctx(struct blk_mq_hw_ctx *hctx)
|
|
|
set_bit(BLK_MQ_S_SCHED_RESTART, &hctx->state);
|
|
|
}
|
|
|
|
|
|
-static bool blk_mq_sched_restart_hctx(struct blk_mq_hw_ctx *hctx)
|
|
|
+void blk_mq_sched_restart(struct blk_mq_hw_ctx *hctx)
|
|
|
{
|
|
|
if (!test_bit(BLK_MQ_S_SCHED_RESTART, &hctx->state))
|
|
|
- return false;
|
|
|
-
|
|
|
- if (hctx->flags & BLK_MQ_F_TAG_SHARED) {
|
|
|
- struct request_queue *q = hctx->queue;
|
|
|
+ return;
|
|
|
|
|
|
- if (test_and_clear_bit(BLK_MQ_S_SCHED_RESTART, &hctx->state))
|
|
|
- atomic_dec(&q->shared_hctx_restart);
|
|
|
- } else
|
|
|
- clear_bit(BLK_MQ_S_SCHED_RESTART, &hctx->state);
|
|
|
+ clear_bit(BLK_MQ_S_SCHED_RESTART, &hctx->state);
|
|
|
|
|
|
if (blk_mq_hctx_has_pending(hctx)) {
|
|
|
blk_mq_run_hw_queue(hctx, true);
|
|
|
- return true;
|
|
|
+ return;
|
|
|
}
|
|
|
-
|
|
|
- return false;
|
|
|
}
|
|
|
|
|
|
/* return true if hctx need to run again */
|
|
@@ -385,68 +377,6 @@ static bool blk_mq_sched_bypass_insert(struct blk_mq_hw_ctx *hctx,
|
|
|
return true;
|
|
|
}
|
|
|
|
|
|
-/**
|
|
|
- * list_for_each_entry_rcu_rr - iterate in a round-robin fashion over rcu list
|
|
|
- * @pos: loop cursor.
|
|
|
- * @skip: the list element that will not be examined. Iteration starts at
|
|
|
- * @skip->next.
|
|
|
- * @head: head of the list to examine. This list must have at least one
|
|
|
- * element, namely @skip.
|
|
|
- * @member: name of the list_head structure within typeof(*pos).
|
|
|
- */
|
|
|
-#define list_for_each_entry_rcu_rr(pos, skip, head, member) \
|
|
|
- for ((pos) = (skip); \
|
|
|
- (pos = (pos)->member.next != (head) ? list_entry_rcu( \
|
|
|
- (pos)->member.next, typeof(*pos), member) : \
|
|
|
- list_entry_rcu((pos)->member.next->next, typeof(*pos), member)), \
|
|
|
- (pos) != (skip); )
|
|
|
-
|
|
|
-/*
|
|
|
- * Called after a driver tag has been freed to check whether a hctx needs to
|
|
|
- * be restarted. Restarts @hctx if its tag set is not shared. Restarts hardware
|
|
|
- * queues in a round-robin fashion if the tag set of @hctx is shared with other
|
|
|
- * hardware queues.
|
|
|
- */
|
|
|
-void blk_mq_sched_restart(struct blk_mq_hw_ctx *const hctx)
|
|
|
-{
|
|
|
- struct blk_mq_tags *const tags = hctx->tags;
|
|
|
- struct blk_mq_tag_set *const set = hctx->queue->tag_set;
|
|
|
- struct request_queue *const queue = hctx->queue, *q;
|
|
|
- struct blk_mq_hw_ctx *hctx2;
|
|
|
- unsigned int i, j;
|
|
|
-
|
|
|
- if (set->flags & BLK_MQ_F_TAG_SHARED) {
|
|
|
- /*
|
|
|
- * If this is 0, then we know that no hardware queues
|
|
|
- * have RESTART marked. We're done.
|
|
|
- */
|
|
|
- if (!atomic_read(&queue->shared_hctx_restart))
|
|
|
- return;
|
|
|
-
|
|
|
- rcu_read_lock();
|
|
|
- list_for_each_entry_rcu_rr(q, queue, &set->tag_list,
|
|
|
- tag_set_list) {
|
|
|
- queue_for_each_hw_ctx(q, hctx2, i)
|
|
|
- if (hctx2->tags == tags &&
|
|
|
- blk_mq_sched_restart_hctx(hctx2))
|
|
|
- goto done;
|
|
|
- }
|
|
|
- j = hctx->queue_num + 1;
|
|
|
- for (i = 0; i < queue->nr_hw_queues; i++, j++) {
|
|
|
- if (j == queue->nr_hw_queues)
|
|
|
- j = 0;
|
|
|
- hctx2 = queue->queue_hw_ctx[j];
|
|
|
- if (hctx2->tags == tags &&
|
|
|
- blk_mq_sched_restart_hctx(hctx2))
|
|
|
- break;
|
|
|
- }
|
|
|
-done:
|
|
|
- rcu_read_unlock();
|
|
|
- } else {
|
|
|
- blk_mq_sched_restart_hctx(hctx);
|
|
|
- }
|
|
|
-}
|
|
|
-
|
|
|
/*
|
|
|
* Add flush/fua to the queue. If we fail getting a driver tag, then
|
|
|
* punt to the requeue list. Requeue will re-invoke us from a context
|