|
@@ -2295,12 +2295,16 @@ repeat:
|
|
|
if (mddev_is_clustered(mddev)) {
|
|
|
if (test_and_clear_bit(MD_CHANGE_DEVS, &mddev->flags))
|
|
|
force_change = 1;
|
|
|
+ if (test_and_clear_bit(MD_CHANGE_CLEAN, &mddev->flags))
|
|
|
+ nospares = 1;
|
|
|
ret = md_cluster_ops->metadata_update_start(mddev);
|
|
|
/* Has someone else has updated the sb */
|
|
|
if (!does_sb_need_changing(mddev)) {
|
|
|
if (ret == 0)
|
|
|
md_cluster_ops->metadata_update_cancel(mddev);
|
|
|
- clear_bit(MD_CHANGE_PENDING, &mddev->flags);
|
|
|
+ bit_clear_unless(&mddev->flags, BIT(MD_CHANGE_PENDING),
|
|
|
+ BIT(MD_CHANGE_DEVS) |
|
|
|
+ BIT(MD_CHANGE_CLEAN));
|
|
|
return;
|
|
|
}
|
|
|
}
|
|
@@ -2434,15 +2438,11 @@ repeat:
|
|
|
if (mddev_is_clustered(mddev) && ret == 0)
|
|
|
md_cluster_ops->metadata_update_finish(mddev);
|
|
|
|
|
|
- spin_lock(&mddev->lock);
|
|
|
if (mddev->in_sync != sync_req ||
|
|
|
- test_bit(MD_CHANGE_DEVS, &mddev->flags)) {
|
|
|
+ !bit_clear_unless(&mddev->flags, BIT(MD_CHANGE_PENDING),
|
|
|
+ BIT(MD_CHANGE_DEVS) | BIT(MD_CHANGE_CLEAN)))
|
|
|
/* have to write it out again */
|
|
|
- spin_unlock(&mddev->lock);
|
|
|
goto repeat;
|
|
|
- }
|
|
|
- clear_bit(MD_CHANGE_PENDING, &mddev->flags);
|
|
|
- spin_unlock(&mddev->lock);
|
|
|
wake_up(&mddev->sb_wait);
|
|
|
if (test_bit(MD_RECOVERY_RUNNING, &mddev->recovery))
|
|
|
sysfs_notify(&mddev->kobj, NULL, "sync_completed");
|
|
@@ -8147,18 +8147,18 @@ void md_do_sync(struct md_thread *thread)
|
|
|
}
|
|
|
}
|
|
|
skip:
|
|
|
- set_bit(MD_CHANGE_DEVS, &mddev->flags);
|
|
|
-
|
|
|
if (mddev_is_clustered(mddev) &&
|
|
|
ret == 0) {
|
|
|
/* set CHANGE_PENDING here since maybe another
|
|
|
* update is needed, so other nodes are informed */
|
|
|
- set_bit(MD_CHANGE_PENDING, &mddev->flags);
|
|
|
+ set_mask_bits(&mddev->flags, 0,
|
|
|
+ BIT(MD_CHANGE_PENDING) | BIT(MD_CHANGE_DEVS));
|
|
|
md_wakeup_thread(mddev->thread);
|
|
|
wait_event(mddev->sb_wait,
|
|
|
!test_bit(MD_CHANGE_PENDING, &mddev->flags));
|
|
|
md_cluster_ops->resync_finish(mddev);
|
|
|
- }
|
|
|
+ } else
|
|
|
+ set_bit(MD_CHANGE_DEVS, &mddev->flags);
|
|
|
|
|
|
spin_lock(&mddev->lock);
|
|
|
if (!test_bit(MD_RECOVERY_INTR, &mddev->recovery)) {
|
|
@@ -8550,6 +8550,7 @@ EXPORT_SYMBOL(md_finish_reshape);
|
|
|
int rdev_set_badblocks(struct md_rdev *rdev, sector_t s, int sectors,
|
|
|
int is_new)
|
|
|
{
|
|
|
+ struct mddev *mddev = rdev->mddev;
|
|
|
int rv;
|
|
|
if (is_new)
|
|
|
s += rdev->new_data_offset;
|
|
@@ -8559,8 +8560,8 @@ int rdev_set_badblocks(struct md_rdev *rdev, sector_t s, int sectors,
|
|
|
if (rv == 0) {
|
|
|
/* Make sure they get written out promptly */
|
|
|
sysfs_notify_dirent_safe(rdev->sysfs_state);
|
|
|
- set_bit(MD_CHANGE_CLEAN, &rdev->mddev->flags);
|
|
|
- set_bit(MD_CHANGE_PENDING, &rdev->mddev->flags);
|
|
|
+ set_mask_bits(&mddev->flags, 0,
|
|
|
+ BIT(MD_CHANGE_CLEAN) | BIT(MD_CHANGE_PENDING));
|
|
|
md_wakeup_thread(rdev->mddev->thread);
|
|
|
return 1;
|
|
|
} else
|