Commit cc94015a authored by NeilBrown's avatar NeilBrown

md/raid5: move stripe_head_state and more code into handle_stripe.

By defining the 'stripe_head_state' in 'handle_stripe', we can move
some common code out of handle_stripe[56]() and into handle_stripe.

The means that all accesses for stripe_head_state in handle_stripe[56]
need to be 's->' instead of 's.', but the compiler should inline
those functions and just use a direct stack reference, and future
patches while hoist most of this code up into handle_stripe()
so we will revert to "s.".
Signed-off-by: default avatarNeilBrown <neilb@suse.de>
Reviewed-by: default avatarNamhyung Kim <namhyung@gmail.com>
parent c5709ef6
...@@ -2999,24 +2999,13 @@ static void handle_stripe_expansion(raid5_conf_t *conf, struct stripe_head *sh, ...@@ -2999,24 +2999,13 @@ static void handle_stripe_expansion(raid5_conf_t *conf, struct stripe_head *sh,
* *
*/ */
static void handle_stripe5(struct stripe_head *sh) static void handle_stripe5(struct stripe_head *sh, struct stripe_head_state *s)
{ {
raid5_conf_t *conf = sh->raid_conf; raid5_conf_t *conf = sh->raid_conf;
int disks = sh->disks, i; int disks = sh->disks, i;
struct stripe_head_state s;
struct r5dev *dev; struct r5dev *dev;
int prexor; int prexor;
memset(&s, 0, sizeof(s));
pr_debug("handling stripe %llu, state=%#lx cnt=%d, pd_idx=%d check:%d "
"reconstruct:%d\n", (unsigned long long)sh->sector, sh->state,
atomic_read(&sh->count), sh->pd_idx, sh->check_state,
sh->reconstruct_state);
s.syncing = test_bit(STRIPE_SYNCING, &sh->state);
s.expanding = test_bit(STRIPE_EXPAND_SOURCE, &sh->state);
s.expanded = test_bit(STRIPE_EXPAND_READY, &sh->state);
/* Now to look around and see what can be done */ /* Now to look around and see what can be done */
rcu_read_lock(); rcu_read_lock();
spin_lock_irq(&conf->device_lock); spin_lock_irq(&conf->device_lock);
...@@ -3039,25 +3028,28 @@ static void handle_stripe5(struct stripe_head *sh) ...@@ -3039,25 +3028,28 @@ static void handle_stripe5(struct stripe_head *sh)
set_bit(R5_Wantfill, &dev->flags); set_bit(R5_Wantfill, &dev->flags);
/* now count some things */ /* now count some things */
if (test_bit(R5_LOCKED, &dev->flags)) s.locked++; if (test_bit(R5_LOCKED, &dev->flags))
if (test_bit(R5_UPTODATE, &dev->flags)) s.uptodate++; s->locked++;
if (test_bit(R5_Wantcompute, &dev->flags)) s.compute++; if (test_bit(R5_UPTODATE, &dev->flags))
s->uptodate++;
if (test_bit(R5_Wantcompute, &dev->flags))
s->compute++;
if (test_bit(R5_Wantfill, &dev->flags)) if (test_bit(R5_Wantfill, &dev->flags))
s.to_fill++; s->to_fill++;
else if (dev->toread) else if (dev->toread)
s.to_read++; s->to_read++;
if (dev->towrite) { if (dev->towrite) {
s.to_write++; s->to_write++;
if (!test_bit(R5_OVERWRITE, &dev->flags)) if (!test_bit(R5_OVERWRITE, &dev->flags))
s.non_overwrite++; s->non_overwrite++;
} }
if (dev->written) if (dev->written)
s.written++; s->written++;
rdev = rcu_dereference(conf->disks[i].rdev); rdev = rcu_dereference(conf->disks[i].rdev);
if (s.blocked_rdev == NULL && if (s->blocked_rdev == NULL &&
rdev && unlikely(test_bit(Blocked, &rdev->flags))) { rdev && unlikely(test_bit(Blocked, &rdev->flags))) {
s.blocked_rdev = rdev; s->blocked_rdev = rdev;
atomic_inc(&rdev->nr_pending); atomic_inc(&rdev->nr_pending);
} }
clear_bit(R5_Insync, &dev->flags); clear_bit(R5_Insync, &dev->flags);
...@@ -3078,62 +3070,62 @@ static void handle_stripe5(struct stripe_head *sh) ...@@ -3078,62 +3070,62 @@ static void handle_stripe5(struct stripe_head *sh)
if (test_bit(R5_ReadError, &dev->flags)) if (test_bit(R5_ReadError, &dev->flags))
clear_bit(R5_Insync, &dev->flags); clear_bit(R5_Insync, &dev->flags);
if (!test_bit(R5_Insync, &dev->flags)) { if (!test_bit(R5_Insync, &dev->flags)) {
s.failed++; s->failed++;
s.failed_num[0] = i; s->failed_num[0] = i;
} }
} }
spin_unlock_irq(&conf->device_lock); spin_unlock_irq(&conf->device_lock);
rcu_read_unlock(); rcu_read_unlock();
if (unlikely(s.blocked_rdev)) { if (unlikely(s->blocked_rdev)) {
if (s.syncing || s.expanding || s.expanded || if (s->syncing || s->expanding || s->expanded ||
s.to_write || s.written) { s->to_write || s->written) {
set_bit(STRIPE_HANDLE, &sh->state); set_bit(STRIPE_HANDLE, &sh->state);
goto unlock; return;
} }
/* There is nothing for the blocked_rdev to block */ /* There is nothing for the blocked_rdev to block */
rdev_dec_pending(s.blocked_rdev, conf->mddev); rdev_dec_pending(s->blocked_rdev, conf->mddev);
s.blocked_rdev = NULL; s->blocked_rdev = NULL;
} }
if (s.to_fill && !test_bit(STRIPE_BIOFILL_RUN, &sh->state)) { if (s->to_fill && !test_bit(STRIPE_BIOFILL_RUN, &sh->state)) {
set_bit(STRIPE_OP_BIOFILL, &s.ops_request); set_bit(STRIPE_OP_BIOFILL, &s->ops_request);
set_bit(STRIPE_BIOFILL_RUN, &sh->state); set_bit(STRIPE_BIOFILL_RUN, &sh->state);
} }
pr_debug("locked=%d uptodate=%d to_read=%d" pr_debug("locked=%d uptodate=%d to_read=%d"
" to_write=%d failed=%d failed_num=%d\n", " to_write=%d failed=%d failed_num=%d\n",
s.locked, s.uptodate, s.to_read, s.to_write, s->locked, s->uptodate, s->to_read, s->to_write,
s.failed, s.failed_num[0]); s->failed, s->failed_num[0]);
/* check if the array has lost two devices and, if so, some requests might /* check if the array has lost two devices and, if so, some requests might
* need to be failed * need to be failed
*/ */
if (s.failed > 1 && s.to_read+s.to_write+s.written) if (s->failed > 1 && s->to_read+s->to_write+s->written)
handle_failed_stripe(conf, sh, &s, disks, &s.return_bi); handle_failed_stripe(conf, sh, s, disks, &s->return_bi);
if (s.failed > 1 && s.syncing) { if (s->failed > 1 && s->syncing) {
md_done_sync(conf->mddev, STRIPE_SECTORS,0); md_done_sync(conf->mddev, STRIPE_SECTORS,0);
clear_bit(STRIPE_SYNCING, &sh->state); clear_bit(STRIPE_SYNCING, &sh->state);
s.syncing = 0; s->syncing = 0;
} }
/* might be able to return some write requests if the parity block /* might be able to return some write requests if the parity block
* is safe, or on a failed drive * is safe, or on a failed drive
*/ */
dev = &sh->dev[sh->pd_idx]; dev = &sh->dev[sh->pd_idx];
if ( s.written && if (s->written &&
((test_bit(R5_Insync, &dev->flags) && ((test_bit(R5_Insync, &dev->flags) &&
!test_bit(R5_LOCKED, &dev->flags) && !test_bit(R5_LOCKED, &dev->flags) &&
test_bit(R5_UPTODATE, &dev->flags)) || test_bit(R5_UPTODATE, &dev->flags)) ||
(s.failed == 1 && s.failed_num[0] == sh->pd_idx))) (s->failed == 1 && s->failed_num[0] == sh->pd_idx)))
handle_stripe_clean_event(conf, sh, disks, &s.return_bi); handle_stripe_clean_event(conf, sh, disks, &s->return_bi);
/* Now we might consider reading some blocks, either to check/generate /* Now we might consider reading some blocks, either to check/generate
* parity, or to satisfy requests * parity, or to satisfy requests
* or to load a block that is being partially written. * or to load a block that is being partially written.
*/ */
if (s.to_read || s.non_overwrite || if (s->to_read || s->non_overwrite ||
(s.syncing && (s.uptodate + s.compute < disks)) || s.expanding) (s->syncing && (s->uptodate + s->compute < disks)) || s->expanding)
handle_stripe_fill5(sh, &s, disks); handle_stripe_fill5(sh, s, disks);
/* Now we check to see if any write operations have recently /* Now we check to see if any write operations have recently
* completed * completed
...@@ -3158,12 +3150,12 @@ static void handle_stripe5(struct stripe_head *sh) ...@@ -3158,12 +3150,12 @@ static void handle_stripe5(struct stripe_head *sh)
if (prexor) if (prexor)
continue; continue;
if (!test_bit(R5_Insync, &dev->flags) || if (!test_bit(R5_Insync, &dev->flags) ||
(i == sh->pd_idx && s.failed == 0)) (i == sh->pd_idx && s->failed == 0))
set_bit(STRIPE_INSYNC, &sh->state); set_bit(STRIPE_INSYNC, &sh->state);
} }
} }
if (test_and_clear_bit(STRIPE_PREREAD_ACTIVE, &sh->state)) if (test_and_clear_bit(STRIPE_PREREAD_ACTIVE, &sh->state))
s.dec_preread_active = 1; s->dec_preread_active = 1;
} }
/* Now to consider new write requests and what else, if anything /* Now to consider new write requests and what else, if anything
...@@ -3172,8 +3164,8 @@ static void handle_stripe5(struct stripe_head *sh) ...@@ -3172,8 +3164,8 @@ static void handle_stripe5(struct stripe_head *sh)
* 2/ A 'check' operation is in flight, as it may clobber the parity * 2/ A 'check' operation is in flight, as it may clobber the parity
* block. * block.
*/ */
if (s.to_write && !sh->reconstruct_state && !sh->check_state) if (s->to_write && !sh->reconstruct_state && !sh->check_state)
handle_stripe_dirtying5(conf, sh, &s, disks); handle_stripe_dirtying5(conf, sh, s, disks);
/* maybe we need to check and possibly fix the parity for this stripe /* maybe we need to check and possibly fix the parity for this stripe
* Any reads will already have been scheduled, so we just see if enough * Any reads will already have been scheduled, so we just see if enough
...@@ -3181,12 +3173,13 @@ static void handle_stripe5(struct stripe_head *sh) ...@@ -3181,12 +3173,13 @@ static void handle_stripe5(struct stripe_head *sh)
* dependent operations are in flight. * dependent operations are in flight.
*/ */
if (sh->check_state || if (sh->check_state ||
(s.syncing && s.locked == 0 && (s->syncing && s->locked == 0 &&
!test_bit(STRIPE_COMPUTE_RUN, &sh->state) && !test_bit(STRIPE_COMPUTE_RUN, &sh->state) &&
!test_bit(STRIPE_INSYNC, &sh->state))) !test_bit(STRIPE_INSYNC, &sh->state)))
handle_parity_checks5(conf, sh, &s, disks); handle_parity_checks5(conf, sh, s, disks);
if (s.syncing && s.locked == 0 && test_bit(STRIPE_INSYNC, &sh->state)) { if (s->syncing && s->locked == 0
&& test_bit(STRIPE_INSYNC, &sh->state)) {
md_done_sync(conf->mddev, STRIPE_SECTORS,1); md_done_sync(conf->mddev, STRIPE_SECTORS,1);
clear_bit(STRIPE_SYNCING, &sh->state); clear_bit(STRIPE_SYNCING, &sh->state);
} }
...@@ -3194,22 +3187,22 @@ static void handle_stripe5(struct stripe_head *sh) ...@@ -3194,22 +3187,22 @@ static void handle_stripe5(struct stripe_head *sh)
/* If the failed drive is just a ReadError, then we might need to progress /* If the failed drive is just a ReadError, then we might need to progress
* the repair/check process * the repair/check process
*/ */
if (s.failed == 1 && !conf->mddev->ro && if (s->failed == 1 && !conf->mddev->ro &&
test_bit(R5_ReadError, &sh->dev[s.failed_num[0]].flags) test_bit(R5_ReadError, &sh->dev[s->failed_num[0]].flags)
&& !test_bit(R5_LOCKED, &sh->dev[s.failed_num[0]].flags) && !test_bit(R5_LOCKED, &sh->dev[s->failed_num[0]].flags)
&& test_bit(R5_UPTODATE, &sh->dev[s.failed_num[0]].flags) && test_bit(R5_UPTODATE, &sh->dev[s->failed_num[0]].flags)
) { ) {
dev = &sh->dev[s.failed_num[0]]; dev = &sh->dev[s->failed_num[0]];
if (!test_bit(R5_ReWrite, &dev->flags)) { if (!test_bit(R5_ReWrite, &dev->flags)) {
set_bit(R5_Wantwrite, &dev->flags); set_bit(R5_Wantwrite, &dev->flags);
set_bit(R5_ReWrite, &dev->flags); set_bit(R5_ReWrite, &dev->flags);
set_bit(R5_LOCKED, &dev->flags); set_bit(R5_LOCKED, &dev->flags);
s.locked++; s->locked++;
} else { } else {
/* let's read it back */ /* let's read it back */
set_bit(R5_Wantread, &dev->flags); set_bit(R5_Wantread, &dev->flags);
set_bit(R5_LOCKED, &dev->flags); set_bit(R5_LOCKED, &dev->flags);
s.locked++; s->locked++;
} }
} }
...@@ -3227,7 +3220,7 @@ static void handle_stripe5(struct stripe_head *sh) ...@@ -3227,7 +3220,7 @@ static void handle_stripe5(struct stripe_head *sh)
&sh2->state)) &sh2->state))
atomic_inc(&conf->preread_active_stripes); atomic_inc(&conf->preread_active_stripes);
release_stripe(sh2); release_stripe(sh2);
goto unlock; return;
} }
if (sh2) if (sh2)
release_stripe(sh2); release_stripe(sh2);
...@@ -3237,69 +3230,35 @@ static void handle_stripe5(struct stripe_head *sh) ...@@ -3237,69 +3230,35 @@ static void handle_stripe5(struct stripe_head *sh)
for (i = conf->raid_disks; i--; ) { for (i = conf->raid_disks; i--; ) {
set_bit(R5_Wantwrite, &sh->dev[i].flags); set_bit(R5_Wantwrite, &sh->dev[i].flags);
set_bit(R5_LOCKED, &sh->dev[i].flags); set_bit(R5_LOCKED, &sh->dev[i].flags);
s.locked++; s->locked++;
} }
} }
if (s.expanded && test_bit(STRIPE_EXPANDING, &sh->state) && if (s->expanded && test_bit(STRIPE_EXPANDING, &sh->state) &&
!sh->reconstruct_state) { !sh->reconstruct_state) {
/* Need to write out all blocks after computing parity */ /* Need to write out all blocks after computing parity */
sh->disks = conf->raid_disks; sh->disks = conf->raid_disks;
stripe_set_idx(sh->sector, conf, 0, sh); stripe_set_idx(sh->sector, conf, 0, sh);
schedule_reconstruction(sh, &s, 1, 1); schedule_reconstruction(sh, s, 1, 1);
} else if (s.expanded && !sh->reconstruct_state && s.locked == 0) { } else if (s->expanded && !sh->reconstruct_state && s->locked == 0) {
clear_bit(STRIPE_EXPAND_READY, &sh->state); clear_bit(STRIPE_EXPAND_READY, &sh->state);
atomic_dec(&conf->reshape_stripes); atomic_dec(&conf->reshape_stripes);
wake_up(&conf->wait_for_overlap); wake_up(&conf->wait_for_overlap);
md_done_sync(conf->mddev, STRIPE_SECTORS, 1); md_done_sync(conf->mddev, STRIPE_SECTORS, 1);
} }
if (s.expanding && s.locked == 0 && if (s->expanding && s->locked == 0 &&
!test_bit(STRIPE_COMPUTE_RUN, &sh->state)) !test_bit(STRIPE_COMPUTE_RUN, &sh->state))
handle_stripe_expansion(conf, sh, NULL); handle_stripe_expansion(conf, sh, NULL);
unlock:
/* wait for this device to become unblocked */
if (unlikely(s.blocked_rdev))
md_wait_for_blocked_rdev(s.blocked_rdev, conf->mddev);
if (s.ops_request)
raid_run_ops(sh, s.ops_request);
ops_run_io(sh, &s);
if (s.dec_preread_active) {
/* We delay this until after ops_run_io so that if make_request
* is waiting on a flush, it won't continue until the writes
* have actually been submitted.
*/
atomic_dec(&conf->preread_active_stripes);
if (atomic_read(&conf->preread_active_stripes) <
IO_THRESHOLD)
md_wakeup_thread(conf->mddev->thread);
}
return_io(s.return_bi);
} }
static void handle_stripe6(struct stripe_head *sh) static void handle_stripe6(struct stripe_head *sh, struct stripe_head_state *s)
{ {
raid5_conf_t *conf = sh->raid_conf; raid5_conf_t *conf = sh->raid_conf;
int disks = sh->disks; int disks = sh->disks;
int i, pd_idx = sh->pd_idx, qd_idx = sh->qd_idx; int i, pd_idx = sh->pd_idx, qd_idx = sh->qd_idx;
struct stripe_head_state s;
struct r5dev *dev, *pdev, *qdev; struct r5dev *dev, *pdev, *qdev;
pr_debug("handling stripe %llu, state=%#lx cnt=%d, "
"pd_idx=%d, qd_idx=%d\n, check:%d, reconstruct:%d\n",
(unsigned long long)sh->sector, sh->state,
atomic_read(&sh->count), pd_idx, qd_idx,
sh->check_state, sh->reconstruct_state);
memset(&s, 0, sizeof(s));
s.syncing = test_bit(STRIPE_SYNCING, &sh->state);
s.expanding = test_bit(STRIPE_EXPAND_SOURCE, &sh->state);
s.expanded = test_bit(STRIPE_EXPAND_READY, &sh->state);
/* Now to look around and see what can be done */ /* Now to look around and see what can be done */
rcu_read_lock(); rcu_read_lock();
...@@ -3320,28 +3279,30 @@ static void handle_stripe6(struct stripe_head *sh) ...@@ -3320,28 +3279,30 @@ static void handle_stripe6(struct stripe_head *sh)
set_bit(R5_Wantfill, &dev->flags); set_bit(R5_Wantfill, &dev->flags);
/* now count some things */ /* now count some things */
if (test_bit(R5_LOCKED, &dev->flags)) s.locked++; if (test_bit(R5_LOCKED, &dev->flags))
if (test_bit(R5_UPTODATE, &dev->flags)) s.uptodate++; s->locked++;
if (test_bit(R5_UPTODATE, &dev->flags))
s->uptodate++;
if (test_bit(R5_Wantcompute, &dev->flags)) { if (test_bit(R5_Wantcompute, &dev->flags)) {
s.compute++; s->compute++;
BUG_ON(s.compute > 2); BUG_ON(s->compute > 2);
} }
if (test_bit(R5_Wantfill, &dev->flags)) { if (test_bit(R5_Wantfill, &dev->flags)) {
s.to_fill++; s->to_fill++;
} else if (dev->toread) } else if (dev->toread)
s.to_read++; s->to_read++;
if (dev->towrite) { if (dev->towrite) {
s.to_write++; s->to_write++;
if (!test_bit(R5_OVERWRITE, &dev->flags)) if (!test_bit(R5_OVERWRITE, &dev->flags))
s.non_overwrite++; s->non_overwrite++;
} }
if (dev->written) if (dev->written)
s.written++; s->written++;
rdev = rcu_dereference(conf->disks[i].rdev); rdev = rcu_dereference(conf->disks[i].rdev);
if (s.blocked_rdev == NULL && if (s->blocked_rdev == NULL &&
rdev && unlikely(test_bit(Blocked, &rdev->flags))) { rdev && unlikely(test_bit(Blocked, &rdev->flags))) {
s.blocked_rdev = rdev; s->blocked_rdev = rdev;
atomic_inc(&rdev->nr_pending); atomic_inc(&rdev->nr_pending);
} }
clear_bit(R5_Insync, &dev->flags); clear_bit(R5_Insync, &dev->flags);
...@@ -3362,43 +3323,43 @@ static void handle_stripe6(struct stripe_head *sh) ...@@ -3362,43 +3323,43 @@ static void handle_stripe6(struct stripe_head *sh)
if (test_bit(R5_ReadError, &dev->flags)) if (test_bit(R5_ReadError, &dev->flags))
clear_bit(R5_Insync, &dev->flags); clear_bit(R5_Insync, &dev->flags);
if (!test_bit(R5_Insync, &dev->flags)) { if (!test_bit(R5_Insync, &dev->flags)) {
if (s.failed < 2) if (s->failed < 2)
s.failed_num[s.failed] = i; s->failed_num[s->failed] = i;
s.failed++; s->failed++;
} }
} }
spin_unlock_irq(&conf->device_lock); spin_unlock_irq(&conf->device_lock);
rcu_read_unlock(); rcu_read_unlock();
if (unlikely(s.blocked_rdev)) { if (unlikely(s->blocked_rdev)) {
if (s.syncing || s.expanding || s.expanded || if (s->syncing || s->expanding || s->expanded ||
s.to_write || s.written) { s->to_write || s->written) {
set_bit(STRIPE_HANDLE, &sh->state); set_bit(STRIPE_HANDLE, &sh->state);
goto unlock; return;
} }
/* There is nothing for the blocked_rdev to block */ /* There is nothing for the blocked_rdev to block */
rdev_dec_pending(s.blocked_rdev, conf->mddev); rdev_dec_pending(s->blocked_rdev, conf->mddev);
s.blocked_rdev = NULL; s->blocked_rdev = NULL;
} }
if (s.to_fill && !test_bit(STRIPE_BIOFILL_RUN, &sh->state)) { if (s->to_fill && !test_bit(STRIPE_BIOFILL_RUN, &sh->state)) {
set_bit(STRIPE_OP_BIOFILL, &s.ops_request); set_bit(STRIPE_OP_BIOFILL, &s->ops_request);
set_bit(STRIPE_BIOFILL_RUN, &sh->state); set_bit(STRIPE_BIOFILL_RUN, &sh->state);
} }
pr_debug("locked=%d uptodate=%d to_read=%d" pr_debug("locked=%d uptodate=%d to_read=%d"
" to_write=%d failed=%d failed_num=%d,%d\n", " to_write=%d failed=%d failed_num=%d,%d\n",
s.locked, s.uptodate, s.to_read, s.to_write, s.failed, s->locked, s->uptodate, s->to_read, s->to_write, s->failed,
s.failed_num[0], s.failed_num[1]); s->failed_num[0], s->failed_num[1]);
/* check if the array has lost >2 devices and, if so, some requests /* check if the array has lost >2 devices and, if so, some requests
* might need to be failed * might need to be failed
*/ */
if (s.failed > 2 && s.to_read+s.to_write+s.written) if (s->failed > 2 && s->to_read+s->to_write+s->written)
handle_failed_stripe(conf, sh, &s, disks, &s.return_bi); handle_failed_stripe(conf, sh, s, disks, &s->return_bi);
if (s.failed > 2 && s.syncing) { if (s->failed > 2 && s->syncing) {
md_done_sync(conf->mddev, STRIPE_SECTORS,0); md_done_sync(conf->mddev, STRIPE_SECTORS,0);
clear_bit(STRIPE_SYNCING, &sh->state); clear_bit(STRIPE_SYNCING, &sh->state);
s.syncing = 0; s->syncing = 0;
} }
/* /*
...@@ -3406,28 +3367,28 @@ static void handle_stripe6(struct stripe_head *sh) ...@@ -3406,28 +3367,28 @@ static void handle_stripe6(struct stripe_head *sh)
* are safe, or on a failed drive * are safe, or on a failed drive
*/ */
pdev = &sh->dev[pd_idx]; pdev = &sh->dev[pd_idx];
s.p_failed = (s.failed >= 1 && s.failed_num[0] == pd_idx) s->p_failed = (s->failed >= 1 && s->failed_num[0] == pd_idx)
|| (s.failed >= 2 && s.failed_num[1] == pd_idx); || (s->failed >= 2 && s->failed_num[1] == pd_idx);
qdev = &sh->dev[qd_idx]; qdev = &sh->dev[qd_idx];
s.q_failed = (s.failed >= 1 && s.failed_num[0] == qd_idx) s->q_failed = (s->failed >= 1 && s->failed_num[0] == qd_idx)
|| (s.failed >= 2 && s.failed_num[1] == qd_idx); || (s->failed >= 2 && s->failed_num[1] == qd_idx);
if (s.written && if (s->written &&
(s.p_failed || ((test_bit(R5_Insync, &pdev->flags) (s->p_failed || ((test_bit(R5_Insync, &pdev->flags)
&& !test_bit(R5_LOCKED, &pdev->flags) && !test_bit(R5_LOCKED, &pdev->flags)
&& test_bit(R5_UPTODATE, &pdev->flags)))) && && test_bit(R5_UPTODATE, &pdev->flags)))) &&
(s.q_failed || ((test_bit(R5_Insync, &qdev->flags) (s->q_failed || ((test_bit(R5_Insync, &qdev->flags)
&& !test_bit(R5_LOCKED, &qdev->flags) && !test_bit(R5_LOCKED, &qdev->flags)
&& test_bit(R5_UPTODATE, &qdev->flags))))) && test_bit(R5_UPTODATE, &qdev->flags)))))
handle_stripe_clean_event(conf, sh, disks, &s.return_bi); handle_stripe_clean_event(conf, sh, disks, &s->return_bi);
/* Now we might consider reading some blocks, either to check/generate /* Now we might consider reading some blocks, either to check/generate
* parity, or to satisfy requests * parity, or to satisfy requests
* or to load a block that is being partially written. * or to load a block that is being partially written.
*/ */
if (s.to_read || s.non_overwrite || (s.to_write && s.failed) || if (s->to_read || s->non_overwrite || (s->to_write && s->failed) ||
(s.syncing && (s.uptodate + s.compute < disks)) || s.expanding) (s->syncing && (s->uptodate + s->compute < disks)) || s->expanding)
handle_stripe_fill6(sh, &s, disks); handle_stripe_fill6(sh, s, disks);
/* Now we check to see if any write operations have recently /* Now we check to see if any write operations have recently
* completed * completed
...@@ -3450,12 +3411,12 @@ static void handle_stripe6(struct stripe_head *sh) ...@@ -3450,12 +3411,12 @@ static void handle_stripe6(struct stripe_head *sh)
set_bit(R5_Wantwrite, &dev->flags); set_bit(R5_Wantwrite, &dev->flags);
if (!test_bit(R5_Insync, &dev->flags) || if (!test_bit(R5_Insync, &dev->flags) ||
((i == sh->pd_idx || i == qd_idx) && ((i == sh->pd_idx || i == qd_idx) &&
s.failed == 0)) s->failed == 0))
set_bit(STRIPE_INSYNC, &sh->state); set_bit(STRIPE_INSYNC, &sh->state);
} }
} }
if (test_and_clear_bit(STRIPE_PREREAD_ACTIVE, &sh->state)) if (test_and_clear_bit(STRIPE_PREREAD_ACTIVE, &sh->state))
s.dec_preread_active = 1; s->dec_preread_active = 1;
} }
/* Now to consider new write requests and what else, if anything /* Now to consider new write requests and what else, if anything
...@@ -3464,8 +3425,8 @@ static void handle_stripe6(struct stripe_head *sh) ...@@ -3464,8 +3425,8 @@ static void handle_stripe6(struct stripe_head *sh)
* 2/ A 'check' operation is in flight, as it may clobber the parity * 2/ A 'check' operation is in flight, as it may clobber the parity
* block. * block.
*/ */
if (s.to_write && !sh->reconstruct_state && !sh->check_state) if (s->to_write && !sh->reconstruct_state && !sh->check_state)
handle_stripe_dirtying6(conf, sh, &s, disks); handle_stripe_dirtying6(conf, sh, s, disks);
/* maybe we need to check and possibly fix the parity for this stripe /* maybe we need to check and possibly fix the parity for this stripe
* Any reads will already have been scheduled, so we just see if enough * Any reads will already have been scheduled, so we just see if enough
...@@ -3473,12 +3434,13 @@ static void handle_stripe6(struct stripe_head *sh) ...@@ -3473,12 +3434,13 @@ static void handle_stripe6(struct stripe_head *sh)
* dependent operations are in flight. * dependent operations are in flight.
*/ */
if (sh->check_state || if (sh->check_state ||
(s.syncing && s.locked == 0 && (s->syncing && s->locked == 0 &&
!test_bit(STRIPE_COMPUTE_RUN, &sh->state) && !test_bit(STRIPE_COMPUTE_RUN, &sh->state) &&
!test_bit(STRIPE_INSYNC, &sh->state))) !test_bit(STRIPE_INSYNC, &sh->state)))
handle_parity_checks6(conf, sh, &s, disks); handle_parity_checks6(conf, sh, s, disks);
if (s.syncing && s.locked == 0 && test_bit(STRIPE_INSYNC, &sh->state)) { if (s->syncing && s->locked == 0
&& test_bit(STRIPE_INSYNC, &sh->state)) {
md_done_sync(conf->mddev, STRIPE_SECTORS,1); md_done_sync(conf->mddev, STRIPE_SECTORS,1);
clear_bit(STRIPE_SYNCING, &sh->state); clear_bit(STRIPE_SYNCING, &sh->state);
} }
...@@ -3486,9 +3448,9 @@ static void handle_stripe6(struct stripe_head *sh) ...@@ -3486,9 +3448,9 @@ static void handle_stripe6(struct stripe_head *sh)
/* If the failed drives are just a ReadError, then we might need /* If the failed drives are just a ReadError, then we might need
* to progress the repair/check process * to progress the repair/check process
*/ */
if (s.failed <= 2 && !conf->mddev->ro) if (s->failed <= 2 && !conf->mddev->ro)
for (i = 0; i < s.failed; i++) { for (i = 0; i < s->failed; i++) {
dev = &sh->dev[s.failed_num[i]]; dev = &sh->dev[s->failed_num[i]];
if (test_bit(R5_ReadError, &dev->flags) if (test_bit(R5_ReadError, &dev->flags)
&& !test_bit(R5_LOCKED, &dev->flags) && !test_bit(R5_LOCKED, &dev->flags)
&& test_bit(R5_UPTODATE, &dev->flags) && test_bit(R5_UPTODATE, &dev->flags)
...@@ -3497,12 +3459,12 @@ static void handle_stripe6(struct stripe_head *sh) ...@@ -3497,12 +3459,12 @@ static void handle_stripe6(struct stripe_head *sh)
set_bit(R5_Wantwrite, &dev->flags); set_bit(R5_Wantwrite, &dev->flags);
set_bit(R5_ReWrite, &dev->flags); set_bit(R5_ReWrite, &dev->flags);
set_bit(R5_LOCKED, &dev->flags); set_bit(R5_LOCKED, &dev->flags);
s.locked++; s->locked++;
} else { } else {
/* let's read it back */ /* let's read it back */
set_bit(R5_Wantread, &dev->flags); set_bit(R5_Wantread, &dev->flags);
set_bit(R5_LOCKED, &dev->flags); set_bit(R5_LOCKED, &dev->flags);
s.locked++; s->locked++;
} }
} }
} }
...@@ -3514,11 +3476,11 @@ static void handle_stripe6(struct stripe_head *sh) ...@@ -3514,11 +3476,11 @@ static void handle_stripe6(struct stripe_head *sh)
for (i = conf->raid_disks; i--; ) { for (i = conf->raid_disks; i--; ) {
set_bit(R5_Wantwrite, &sh->dev[i].flags); set_bit(R5_Wantwrite, &sh->dev[i].flags);
set_bit(R5_LOCKED, &sh->dev[i].flags); set_bit(R5_LOCKED, &sh->dev[i].flags);
s.locked++; s->locked++;
} }
} }
if (s.expanded && test_bit(STRIPE_EXPANDING, &sh->state) && if (s->expanded && test_bit(STRIPE_EXPANDING, &sh->state) &&
!sh->reconstruct_state) { !sh->reconstruct_state) {
struct stripe_head *sh2 struct stripe_head *sh2
= get_active_stripe(conf, sh->sector, 1, 1, 1); = get_active_stripe(conf, sh->sector, 1, 1, 1);
...@@ -3532,7 +3494,7 @@ static void handle_stripe6(struct stripe_head *sh) ...@@ -3532,7 +3494,7 @@ static void handle_stripe6(struct stripe_head *sh)
&sh2->state)) &sh2->state))
atomic_inc(&conf->preread_active_stripes); atomic_inc(&conf->preread_active_stripes);
release_stripe(sh2); release_stripe(sh2);
goto unlock; return;
} }
if (sh2) if (sh2)
release_stripe(sh2); release_stripe(sh2);
...@@ -3540,19 +3502,54 @@ static void handle_stripe6(struct stripe_head *sh) ...@@ -3540,19 +3502,54 @@ static void handle_stripe6(struct stripe_head *sh)
/* Need to write out all blocks after computing P&Q */ /* Need to write out all blocks after computing P&Q */
sh->disks = conf->raid_disks; sh->disks = conf->raid_disks;
stripe_set_idx(sh->sector, conf, 0, sh); stripe_set_idx(sh->sector, conf, 0, sh);
schedule_reconstruction(sh, &s, 1, 1); schedule_reconstruction(sh, s, 1, 1);
} else if (s.expanded && !sh->reconstruct_state && s.locked == 0) { } else if (s->expanded && !sh->reconstruct_state && s->locked == 0) {
clear_bit(STRIPE_EXPAND_READY, &sh->state); clear_bit(STRIPE_EXPAND_READY, &sh->state);
atomic_dec(&conf->reshape_stripes); atomic_dec(&conf->reshape_stripes);
wake_up(&conf->wait_for_overlap); wake_up(&conf->wait_for_overlap);
md_done_sync(conf->mddev, STRIPE_SECTORS, 1); md_done_sync(conf->mddev, STRIPE_SECTORS, 1);
} }
if (s.expanding && s.locked == 0 && if (s->expanding && s->locked == 0 &&
!test_bit(STRIPE_COMPUTE_RUN, &sh->state)) !test_bit(STRIPE_COMPUTE_RUN, &sh->state))
handle_stripe_expansion(conf, sh, &s); handle_stripe_expansion(conf, sh, s);
}
static void handle_stripe(struct stripe_head *sh)
{
struct stripe_head_state s;
raid5_conf_t *conf = sh->raid_conf;
clear_bit(STRIPE_HANDLE, &sh->state);
if (test_and_set_bit(STRIPE_ACTIVE, &sh->state)) {
/* already being handled, ensure it gets handled
* again when current action finishes */
set_bit(STRIPE_HANDLE, &sh->state);
return;
}
if (test_and_clear_bit(STRIPE_SYNC_REQUESTED, &sh->state)) {
set_bit(STRIPE_SYNCING, &sh->state);
clear_bit(STRIPE_INSYNC, &sh->state);
}
clear_bit(STRIPE_DELAYED, &sh->state);
pr_debug("handling stripe %llu, state=%#lx cnt=%d, "
"pd_idx=%d, qd_idx=%d\n, check:%d, reconstruct:%d\n",
(unsigned long long)sh->sector, sh->state,
atomic_read(&sh->count), sh->pd_idx, sh->qd_idx,
sh->check_state, sh->reconstruct_state);
memset(&s, 0, sizeof(s));
s.syncing = test_bit(STRIPE_SYNCING, &sh->state);
s.expanding = test_bit(STRIPE_EXPAND_SOURCE, &sh->state);
s.expanded = test_bit(STRIPE_EXPAND_READY, &sh->state);
if (conf->level == 6)
handle_stripe6(sh, &s);
else
handle_stripe5(sh, &s);
unlock:
/* wait for this device to become unblocked */ /* wait for this device to become unblocked */
if (unlikely(s.blocked_rdev)) if (unlikely(s.blocked_rdev))
...@@ -3576,28 +3573,7 @@ static void handle_stripe6(struct stripe_head *sh) ...@@ -3576,28 +3573,7 @@ static void handle_stripe6(struct stripe_head *sh)
} }
return_io(s.return_bi); return_io(s.return_bi);
}
static void handle_stripe(struct stripe_head *sh)
{
clear_bit(STRIPE_HANDLE, &sh->state);
if (test_and_set_bit(STRIPE_ACTIVE, &sh->state)) {
/* already being handled, ensure it gets handled
* again when current action finishes */
set_bit(STRIPE_HANDLE, &sh->state);
return;
}
if (test_and_clear_bit(STRIPE_SYNC_REQUESTED, &sh->state)) {
set_bit(STRIPE_SYNCING, &sh->state);
clear_bit(STRIPE_INSYNC, &sh->state);
}
clear_bit(STRIPE_DELAYED, &sh->state);
if (sh->raid_conf->level == 6)
handle_stripe6(sh);
else
handle_stripe5(sh);
clear_bit(STRIPE_ACTIVE, &sh->state); clear_bit(STRIPE_ACTIVE, &sh->state);
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment