|
@@ -911,15 +911,6 @@ static inline struct cfq_data *cic_to_cfqd(struct cfq_io_cq *cic)
|
|
|
return cic->icq.q->elevator->elevator_data;
|
|
|
}
|
|
|
|
|
|
-/*
|
|
|
- * We regard a request as SYNC, if it's either a read or has the SYNC bit
|
|
|
- * set (in which case it could also be direct WRITE).
|
|
|
- */
|
|
|
-static inline bool cfq_bio_sync(struct bio *bio)
|
|
|
-{
|
|
|
- return bio_data_dir(bio) == READ || (bio->bi_opf & REQ_SYNC);
|
|
|
-}
|
|
|
-
|
|
|
/*
|
|
|
* scheduler run of queue, if there are requests pending and no one in the
|
|
|
* driver that will restart queueing
|
|
@@ -2490,7 +2481,7 @@ cfq_find_rq_fmerge(struct cfq_data *cfqd, struct bio *bio)
|
|
|
if (!cic)
|
|
|
return NULL;
|
|
|
|
|
|
- cfqq = cic_to_cfqq(cic, cfq_bio_sync(bio));
|
|
|
+ cfqq = cic_to_cfqq(cic, op_is_sync(bio->bi_opf));
|
|
|
if (cfqq)
|
|
|
return elv_rb_find(&cfqq->sort_list, bio_end_sector(bio));
|
|
|
|
|
@@ -2604,13 +2595,14 @@ static int cfq_allow_bio_merge(struct request_queue *q, struct request *rq,
|
|
|
struct bio *bio)
|
|
|
{
|
|
|
struct cfq_data *cfqd = q->elevator->elevator_data;
|
|
|
+ bool is_sync = op_is_sync(bio->bi_opf);
|
|
|
struct cfq_io_cq *cic;
|
|
|
struct cfq_queue *cfqq;
|
|
|
|
|
|
/*
|
|
|
* Disallow merge of a sync bio into an async request.
|
|
|
*/
|
|
|
- if (cfq_bio_sync(bio) && !rq_is_sync(rq))
|
|
|
+ if (is_sync && !rq_is_sync(rq))
|
|
|
return false;
|
|
|
|
|
|
/*
|
|
@@ -2621,7 +2613,7 @@ static int cfq_allow_bio_merge(struct request_queue *q, struct request *rq,
|
|
|
if (!cic)
|
|
|
return false;
|
|
|
|
|
|
- cfqq = cic_to_cfqq(cic, cfq_bio_sync(bio));
|
|
|
+ cfqq = cic_to_cfqq(cic, is_sync);
|
|
|
return cfqq == RQ_CFQQ(rq);
|
|
|
}
|
|
|
|