|
@@ -505,6 +505,7 @@ struct request_queue {
|
|
|
#define QUEUE_FLAG_WC 23 /* Write back caching */
|
|
|
#define QUEUE_FLAG_FUA 24 /* device supports FUA writes */
|
|
|
#define QUEUE_FLAG_FLUSH_NQ 25 /* flush not queueuable */
|
|
|
+#define QUEUE_FLAG_DAX 26 /* device supports DAX */
|
|
|
|
|
|
#define QUEUE_FLAG_DEFAULT ((1 << QUEUE_FLAG_IO_STAT) | \
|
|
|
(1 << QUEUE_FLAG_STACKABLE) | \
|
|
@@ -594,6 +595,7 @@ static inline void queue_flag_clear(unsigned int flag, struct request_queue *q)
|
|
|
#define blk_queue_discard(q) test_bit(QUEUE_FLAG_DISCARD, &(q)->queue_flags)
|
|
|
#define blk_queue_secdiscard(q) (blk_queue_discard(q) && \
|
|
|
test_bit(QUEUE_FLAG_SECDISCARD, &(q)->queue_flags))
|
|
|
+#define blk_queue_dax(q) test_bit(QUEUE_FLAG_DAX, &(q)->queue_flags)
|
|
|
|
|
|
#define blk_noretry_request(rq) \
|
|
|
((rq)->cmd_flags & (REQ_FAILFAST_DEV|REQ_FAILFAST_TRANSPORT| \
|