|
@@ -1114,15 +1114,18 @@ static void drbd_setup_queue_param(struct drbd_device *device, unsigned int max_
|
|
|
struct request_queue * const q = device->rq_queue;
|
|
|
unsigned int max_hw_sectors = max_bio_size >> 9;
|
|
|
unsigned int max_segments = 0;
|
|
|
+ struct request_queue *b = NULL;
|
|
|
|
|
|
if (get_ldev_if_state(device, D_ATTACHING)) {
|
|
|
- struct request_queue * const b = device->ldev->backing_bdev->bd_disk->queue;
|
|
|
+ b = device->ldev->backing_bdev->bd_disk->queue;
|
|
|
|
|
|
max_hw_sectors = min(queue_max_hw_sectors(b), max_bio_size >> 9);
|
|
|
rcu_read_lock();
|
|
|
max_segments = rcu_dereference(device->ldev->disk_conf)->max_bio_bvecs;
|
|
|
rcu_read_unlock();
|
|
|
- put_ldev(device);
|
|
|
+
|
|
|
+ blk_set_stacking_limits(&q->limits);
|
|
|
+ blk_queue_max_write_same_sectors(q, 0);
|
|
|
}
|
|
|
|
|
|
blk_queue_logical_block_size(q, 512);
|
|
@@ -1131,14 +1134,11 @@ static void drbd_setup_queue_param(struct drbd_device *device, unsigned int max_
|
|
|
blk_queue_max_segments(q, max_segments ? max_segments : BLK_MAX_SEGMENTS);
|
|
|
blk_queue_segment_boundary(q, PAGE_CACHE_SIZE-1);
|
|
|
|
|
|
- if (get_ldev_if_state(device, D_ATTACHING)) {
|
|
|
- struct request_queue * const b = device->ldev->backing_bdev->bd_disk->queue;
|
|
|
+ if (b) {
|
|
|
struct drbd_connection *connection = first_peer_device(device)->connection;
|
|
|
|
|
|
if (blk_queue_discard(b) &&
|
|
|
(connection->cstate < C_CONNECTED || connection->agreed_features & FF_TRIM)) {
|
|
|
- /* inherit from backing queue */
|
|
|
- q->limits.discard_zeroes_data = 1;
|
|
|
/* For now, don't allow more than one activity log extent worth of data
|
|
|
* to be discarded in one go. We may need to rework drbd_al_begin_io()
|
|
|
* to allow for even larger discard ranges */
|