|
@@ -204,7 +204,7 @@ static void blk_mq_rq_ctx_init(struct request_queue *q, struct blk_mq_ctx *ctx,
|
|
|
rq->end_io_data = NULL;
|
|
|
rq->next_rq = NULL;
|
|
|
|
|
|
- ctx->rq_dispatched[rw_is_sync(op | op_flags)]++;
|
|
|
+ ctx->rq_dispatched[rw_is_sync(op, op_flags)]++;
|
|
|
}
|
|
|
|
|
|
static struct request *
|
|
@@ -1178,7 +1178,7 @@ static struct request *blk_mq_map_request(struct request_queue *q,
|
|
|
ctx = blk_mq_get_ctx(q);
|
|
|
hctx = q->mq_ops->map_queue(q, ctx->cpu);
|
|
|
|
|
|
- if (rw_is_sync(bio->bi_rw))
|
|
|
+ if (rw_is_sync(bio_op(bio), bio->bi_rw))
|
|
|
op_flags |= REQ_SYNC;
|
|
|
|
|
|
trace_block_getrq(q, bio, op);
|
|
@@ -1246,7 +1246,7 @@ static int blk_mq_direct_issue_request(struct request *rq, blk_qc_t *cookie)
|
|
|
*/
|
|
|
static blk_qc_t blk_mq_make_request(struct request_queue *q, struct bio *bio)
|
|
|
{
|
|
|
- const int is_sync = rw_is_sync(bio->bi_rw);
|
|
|
+ const int is_sync = rw_is_sync(bio_op(bio), bio->bi_rw);
|
|
|
const int is_flush_fua = bio->bi_rw & (REQ_FLUSH | REQ_FUA);
|
|
|
struct blk_map_ctx data;
|
|
|
struct request *rq;
|
|
@@ -1343,7 +1343,7 @@ done:
|
|
|
*/
|
|
|
static blk_qc_t blk_sq_make_request(struct request_queue *q, struct bio *bio)
|
|
|
{
|
|
|
- const int is_sync = rw_is_sync(bio->bi_rw);
|
|
|
+ const int is_sync = rw_is_sync(bio_op(bio), bio->bi_rw);
|
|
|
const int is_flush_fua = bio->bi_rw & (REQ_FLUSH | REQ_FUA);
|
|
|
struct blk_plug *plug;
|
|
|
unsigned int request_count = 0;
|