|
@@ -1811,7 +1811,6 @@ static void blk_mq_map_swqueue(struct request_queue *q)
|
|
|
|
|
|
|
|
hctx = q->mq_ops->map_queue(q, i);
|
|
hctx = q->mq_ops->map_queue(q, i);
|
|
|
cpumask_set_cpu(i, hctx->cpumask);
|
|
cpumask_set_cpu(i, hctx->cpumask);
|
|
|
- cpumask_set_cpu(i, hctx->tags->cpumask);
|
|
|
|
|
ctx->index_hw = hctx->nr_ctx;
|
|
ctx->index_hw = hctx->nr_ctx;
|
|
|
hctx->ctxs[hctx->nr_ctx++] = ctx;
|
|
hctx->ctxs[hctx->nr_ctx++] = ctx;
|
|
|
}
|
|
}
|
|
@@ -1851,6 +1850,14 @@ static void blk_mq_map_swqueue(struct request_queue *q)
|
|
|
hctx->next_cpu = cpumask_first(hctx->cpumask);
|
|
hctx->next_cpu = cpumask_first(hctx->cpumask);
|
|
|
hctx->next_cpu_batch = BLK_MQ_CPU_WORK_BATCH;
|
|
hctx->next_cpu_batch = BLK_MQ_CPU_WORK_BATCH;
|
|
|
}
|
|
}
|
|
|
|
|
+
|
|
|
|
|
+ queue_for_each_ctx(q, ctx, i) {
|
|
|
|
|
+ if (!cpu_online(i))
|
|
|
|
|
+ continue;
|
|
|
|
|
+
|
|
|
|
|
+ hctx = q->mq_ops->map_queue(q, i);
|
|
|
|
|
+ cpumask_set_cpu(i, hctx->tags->cpumask);
|
|
|
|
|
+ }
|
|
|
}
|
|
}
|
|
|
|
|
|
|
|
static void blk_mq_update_tag_set_depth(struct blk_mq_tag_set *set)
|
|
static void blk_mq_update_tag_set_depth(struct blk_mq_tag_set *set)
|