@@ -439,7 +439,7 @@ static bool rs_is_reshapable(struct raid_set *rs)
439
439
/* Return true, if raid set in @rs is recovering */
440
440
static bool rs_is_recovering (struct raid_set * rs )
441
441
{
442
- return rs -> md .recovery_cp < rs -> md .dev_sectors ;
442
+ return rs -> md .resync_offset < rs -> md .dev_sectors ;
443
443
}
444
444
445
445
/* Return true, if raid set in @rs is reshaping */
@@ -769,7 +769,7 @@ static struct raid_set *raid_set_alloc(struct dm_target *ti, struct raid_type *r
769
769
rs -> md .layout = raid_type -> algorithm ;
770
770
rs -> md .new_layout = rs -> md .layout ;
771
771
rs -> md .delta_disks = 0 ;
772
- rs -> md .recovery_cp = MaxSector ;
772
+ rs -> md .resync_offset = MaxSector ;
773
773
774
774
for (i = 0 ; i < raid_devs ; i ++ )
775
775
md_rdev_init (& rs -> dev [i ].rdev );
@@ -913,7 +913,7 @@ static int parse_dev_params(struct raid_set *rs, struct dm_arg_set *as)
913
913
rs -> md .external = 0 ;
914
914
rs -> md .persistent = 1 ;
915
915
rs -> md .major_version = 2 ;
916
- } else if (rebuild && !rs -> md .recovery_cp ) {
916
+ } else if (rebuild && !rs -> md .resync_offset ) {
917
917
/*
918
918
* Without metadata, we will not be able to tell if the array
919
919
* is in-sync or not - we must assume it is not. Therefore,
@@ -1696,20 +1696,20 @@ static void rs_setup_recovery(struct raid_set *rs, sector_t dev_sectors)
1696
1696
{
1697
1697
/* raid0 does not recover */
1698
1698
if (rs_is_raid0 (rs ))
1699
- rs -> md .recovery_cp = MaxSector ;
1699
+ rs -> md .resync_offset = MaxSector ;
1700
1700
/*
1701
1701
* A raid6 set has to be recovered either
1702
1702
* completely or for the grown part to
1703
1703
* ensure proper parity and Q-Syndrome
1704
1704
*/
1705
1705
else if (rs_is_raid6 (rs ))
1706
- rs -> md .recovery_cp = dev_sectors ;
1706
+ rs -> md .resync_offset = dev_sectors ;
1707
1707
/*
1708
1708
* Other raid set types may skip recovery
1709
1709
* depending on the 'nosync' flag.
1710
1710
*/
1711
1711
else
1712
- rs -> md .recovery_cp = test_bit (__CTR_FLAG_NOSYNC , & rs -> ctr_flags )
1712
+ rs -> md .resync_offset = test_bit (__CTR_FLAG_NOSYNC , & rs -> ctr_flags )
1713
1713
? MaxSector : dev_sectors ;
1714
1714
}
1715
1715
@@ -2144,7 +2144,7 @@ static void super_sync(struct mddev *mddev, struct md_rdev *rdev)
2144
2144
sb -> events = cpu_to_le64 (mddev -> events );
2145
2145
2146
2146
sb -> disk_recovery_offset = cpu_to_le64 (rdev -> recovery_offset );
2147
- sb -> array_resync_offset = cpu_to_le64 (mddev -> recovery_cp );
2147
+ sb -> array_resync_offset = cpu_to_le64 (mddev -> resync_offset );
2148
2148
2149
2149
sb -> level = cpu_to_le32 (mddev -> level );
2150
2150
sb -> layout = cpu_to_le32 (mddev -> layout );
@@ -2335,18 +2335,18 @@ static int super_init_validation(struct raid_set *rs, struct md_rdev *rdev)
2335
2335
}
2336
2336
2337
2337
if (!test_bit (__CTR_FLAG_NOSYNC , & rs -> ctr_flags ))
2338
- mddev -> recovery_cp = le64_to_cpu (sb -> array_resync_offset );
2338
+ mddev -> resync_offset = le64_to_cpu (sb -> array_resync_offset );
2339
2339
2340
2340
/*
2341
2341
* During load, we set FirstUse if a new superblock was written.
2342
2342
* There are two reasons we might not have a superblock:
2343
2343
* 1) The raid set is brand new - in which case, all of the
2344
2344
* devices must have their In_sync bit set. Also,
2345
- * recovery_cp must be 0, unless forced.
2345
+ * resync_offset must be 0, unless forced.
2346
2346
* 2) This is a new device being added to an old raid set
2347
2347
* and the new device needs to be rebuilt - in which
2348
2348
* case the In_sync bit will /not/ be set and
2349
- * recovery_cp must be MaxSector.
2349
+ * resync_offset must be MaxSector.
2350
2350
* 3) This is/are a new device(s) being added to an old
2351
2351
* raid set during takeover to a higher raid level
2352
2352
* to provide capacity for redundancy or during reshape
@@ -2391,8 +2391,8 @@ static int super_init_validation(struct raid_set *rs, struct md_rdev *rdev)
2391
2391
new_devs > 1 ? "s" : "" );
2392
2392
return - EINVAL ;
2393
2393
} else if (!test_bit (__CTR_FLAG_REBUILD , & rs -> ctr_flags ) && rs_is_recovering (rs )) {
2394
- DMERR ("'rebuild' specified while raid set is not in-sync (recovery_cp =%llu)" ,
2395
- (unsigned long long ) mddev -> recovery_cp );
2394
+ DMERR ("'rebuild' specified while raid set is not in-sync (resync_offset =%llu)" ,
2395
+ (unsigned long long ) mddev -> resync_offset );
2396
2396
return - EINVAL ;
2397
2397
} else if (rs_is_reshaping (rs )) {
2398
2398
DMERR ("'rebuild' specified while raid set is being reshaped (reshape_position=%llu)" ,
@@ -2697,11 +2697,11 @@ static int rs_adjust_data_offsets(struct raid_set *rs)
2697
2697
}
2698
2698
out :
2699
2699
/*
2700
- * Raise recovery_cp in case data_offset != 0 to
2700
+ * Raise resync_offset in case data_offset != 0 to
2701
2701
* avoid false recovery positives in the constructor.
2702
2702
*/
2703
- if (rs -> md .recovery_cp < rs -> md .dev_sectors )
2704
- rs -> md .recovery_cp += rs -> dev [0 ].rdev .data_offset ;
2703
+ if (rs -> md .resync_offset < rs -> md .dev_sectors )
2704
+ rs -> md .resync_offset += rs -> dev [0 ].rdev .data_offset ;
2705
2705
2706
2706
/* Adjust data offsets on all rdevs but on any raid4/5/6 journal device */
2707
2707
rdev_for_each (rdev , & rs -> md ) {
@@ -2756,15 +2756,15 @@ static int rs_setup_takeover(struct raid_set *rs)
2756
2756
}
2757
2757
2758
2758
clear_bit (MD_ARRAY_FIRST_USE , & mddev -> flags );
2759
- mddev -> recovery_cp = MaxSector ;
2759
+ mddev -> resync_offset = MaxSector ;
2760
2760
2761
2761
while (d -- ) {
2762
2762
rdev = & rs -> dev [d ].rdev ;
2763
2763
2764
2764
if (test_bit (d , (void * ) rs -> rebuild_disks )) {
2765
2765
clear_bit (In_sync , & rdev -> flags );
2766
2766
clear_bit (Faulty , & rdev -> flags );
2767
- mddev -> recovery_cp = rdev -> recovery_offset = 0 ;
2767
+ mddev -> resync_offset = rdev -> recovery_offset = 0 ;
2768
2768
/* Bitmap has to be created when we do an "up" takeover */
2769
2769
set_bit (MD_ARRAY_FIRST_USE , & mddev -> flags );
2770
2770
}
@@ -3222,7 +3222,7 @@ static int raid_ctr(struct dm_target *ti, unsigned int argc, char **argv)
3222
3222
if (r )
3223
3223
goto bad ;
3224
3224
3225
- rs_setup_recovery (rs , rs -> md .recovery_cp < rs -> md .dev_sectors ? rs -> md .recovery_cp : rs -> md .dev_sectors );
3225
+ rs_setup_recovery (rs , rs -> md .resync_offset < rs -> md .dev_sectors ? rs -> md .resync_offset : rs -> md .dev_sectors );
3226
3226
} else {
3227
3227
/* This is no size change or it is shrinking, update size and record in superblocks */
3228
3228
r = rs_set_dev_and_array_sectors (rs , rs -> ti -> len , false);
@@ -3446,7 +3446,7 @@ static sector_t rs_get_progress(struct raid_set *rs, unsigned long recovery,
3446
3446
3447
3447
} else {
3448
3448
if (state == st_idle && !test_bit (MD_RECOVERY_INTR , & recovery ))
3449
- r = mddev -> recovery_cp ;
3449
+ r = mddev -> resync_offset ;
3450
3450
else
3451
3451
r = mddev -> curr_resync_completed ;
3452
3452
@@ -4074,9 +4074,9 @@ static int raid_preresume(struct dm_target *ti)
4074
4074
}
4075
4075
4076
4076
/* Check for any resize/reshape on @rs and adjust/initiate */
4077
- if (mddev -> recovery_cp && mddev -> recovery_cp < MaxSector ) {
4077
+ if (mddev -> resync_offset && mddev -> resync_offset < MaxSector ) {
4078
4078
set_bit (MD_RECOVERY_REQUESTED , & mddev -> recovery );
4079
- mddev -> resync_min = mddev -> recovery_cp ;
4079
+ mddev -> resync_min = mddev -> resync_offset ;
4080
4080
if (test_bit (RT_FLAG_RS_GROW , & rs -> runtime_flags ))
4081
4081
mddev -> resync_max_sectors = mddev -> dev_sectors ;
4082
4082
}
0 commit comments