Commit 83de75cc authored by Dan Williams's avatar Dan Williams Committed by Neil Brown

md: replace STRIPE_OP_BIOFILL with STRIPE_BIOFILL_RUN

From: Dan Williams <dan.j.williams@intel.com>

Track the state of read operations (copying data from the stripe cache to bio
buffers outside the lock) with a state flag.  Reduce the scope of the
STRIPE_OP_BIOFILL flag to only tracking whether a biofill operation has been
requested via the ops_request field of struct stripe_head_state.

This is another step towards the removal of ops.{pending,ack,complete,count},
i.e. STRIPE_OP_BIOFILL only requests an operation and does not track the state
of the operation.
Signed-off-by: default avatarDan Williams <dan.j.williams@intel.com>
Signed-off-by: default avatarNeil Brown <neilb@suse.de>
parent ecc65c9b
...@@ -523,38 +523,34 @@ static void ops_complete_biofill(void *stripe_head_ref) ...@@ -523,38 +523,34 @@ static void ops_complete_biofill(void *stripe_head_ref)
(unsigned long long)sh->sector); (unsigned long long)sh->sector);
/* clear completed biofills */ /* clear completed biofills */
spin_lock_irq(&conf->device_lock);
for (i = sh->disks; i--; ) { for (i = sh->disks; i--; ) {
struct r5dev *dev = &sh->dev[i]; struct r5dev *dev = &sh->dev[i];
/* acknowledge completion of a biofill operation */ /* acknowledge completion of a biofill operation */
/* and check if we need to reply to a read request, /* and check if we need to reply to a read request,
* new R5_Wantfill requests are held off until * new R5_Wantfill requests are held off until
* !test_bit(STRIPE_OP_BIOFILL, &sh->ops.pending) * !STRIPE_BIOFILL_RUN
*/ */
if (test_and_clear_bit(R5_Wantfill, &dev->flags)) { if (test_and_clear_bit(R5_Wantfill, &dev->flags)) {
struct bio *rbi, *rbi2; struct bio *rbi, *rbi2;
/* The access to dev->read is outside of the
* spin_lock_irq(&conf->device_lock), but is protected
* by the STRIPE_OP_BIOFILL pending bit
*/
BUG_ON(!dev->read); BUG_ON(!dev->read);
rbi = dev->read; rbi = dev->read;
dev->read = NULL; dev->read = NULL;
while (rbi && rbi->bi_sector < while (rbi && rbi->bi_sector <
dev->sector + STRIPE_SECTORS) { dev->sector + STRIPE_SECTORS) {
rbi2 = r5_next_bio(rbi, dev->sector); rbi2 = r5_next_bio(rbi, dev->sector);
spin_lock_irq(&conf->device_lock);
if (--rbi->bi_phys_segments == 0) { if (--rbi->bi_phys_segments == 0) {
rbi->bi_next = return_bi; rbi->bi_next = return_bi;
return_bi = rbi; return_bi = rbi;
} }
spin_unlock_irq(&conf->device_lock);
rbi = rbi2; rbi = rbi2;
} }
} }
} }
set_bit(STRIPE_OP_BIOFILL, &sh->ops.complete); spin_unlock_irq(&conf->device_lock);
clear_bit(STRIPE_BIOFILL_RUN, &sh->state);
return_io(return_bi); return_io(return_bi);
...@@ -880,7 +876,7 @@ static void raid5_run_ops(struct stripe_head *sh, unsigned long pending, ...@@ -880,7 +876,7 @@ static void raid5_run_ops(struct stripe_head *sh, unsigned long pending,
int overlap_clear = 0, i, disks = sh->disks; int overlap_clear = 0, i, disks = sh->disks;
struct dma_async_tx_descriptor *tx = NULL; struct dma_async_tx_descriptor *tx = NULL;
if (test_bit(STRIPE_OP_BIOFILL, &pending)) { if (test_bit(STRIPE_OP_BIOFILL, &ops_request)) {
ops_run_biofill(sh); ops_run_biofill(sh);
overlap_clear++; overlap_clear++;
} }
...@@ -2630,15 +2626,8 @@ static void handle_stripe5(struct stripe_head *sh) ...@@ -2630,15 +2626,8 @@ static void handle_stripe5(struct stripe_head *sh)
s.syncing = test_bit(STRIPE_SYNCING, &sh->state); s.syncing = test_bit(STRIPE_SYNCING, &sh->state);
s.expanding = test_bit(STRIPE_EXPAND_SOURCE, &sh->state); s.expanding = test_bit(STRIPE_EXPAND_SOURCE, &sh->state);
s.expanded = test_bit(STRIPE_EXPAND_READY, &sh->state); s.expanded = test_bit(STRIPE_EXPAND_READY, &sh->state);
/* Now to look around and see what can be done */
/* clean-up completed biofill operations */
if (test_bit(STRIPE_OP_BIOFILL, &sh->ops.complete)) {
clear_bit(STRIPE_OP_BIOFILL, &sh->ops.pending);
clear_bit(STRIPE_OP_BIOFILL, &sh->ops.ack);
clear_bit(STRIPE_OP_BIOFILL, &sh->ops.complete);
}
/* Now to look around and see what can be done */
rcu_read_lock(); rcu_read_lock();
for (i=disks; i--; ) { for (i=disks; i--; ) {
mdk_rdev_t *rdev; mdk_rdev_t *rdev;
...@@ -2652,10 +2641,10 @@ static void handle_stripe5(struct stripe_head *sh) ...@@ -2652,10 +2641,10 @@ static void handle_stripe5(struct stripe_head *sh)
/* maybe we can request a biofill operation /* maybe we can request a biofill operation
* *
* new wantfill requests are only permitted while * new wantfill requests are only permitted while
* STRIPE_OP_BIOFILL is clear * ops_complete_biofill is guaranteed to be inactive
*/ */
if (test_bit(R5_UPTODATE, &dev->flags) && dev->toread && if (test_bit(R5_UPTODATE, &dev->flags) && dev->toread &&
!test_bit(STRIPE_OP_BIOFILL, &sh->ops.pending)) !test_bit(STRIPE_BIOFILL_RUN, &sh->state))
set_bit(R5_Wantfill, &dev->flags); set_bit(R5_Wantfill, &dev->flags);
/* now count some things */ /* now count some things */
...@@ -2699,8 +2688,10 @@ static void handle_stripe5(struct stripe_head *sh) ...@@ -2699,8 +2688,10 @@ static void handle_stripe5(struct stripe_head *sh)
goto unlock; goto unlock;
} }
if (s.to_fill && !test_and_set_bit(STRIPE_OP_BIOFILL, &sh->ops.pending)) if (s.to_fill && !test_bit(STRIPE_BIOFILL_RUN, &sh->state)) {
sh->ops.count++; set_bit(STRIPE_OP_BIOFILL, &s.ops_request);
set_bit(STRIPE_BIOFILL_RUN, &sh->state);
}
pr_debug("locked=%d uptodate=%d to_read=%d" pr_debug("locked=%d uptodate=%d to_read=%d"
" to_write=%d failed=%d failed_num=%d\n", " to_write=%d failed=%d failed_num=%d\n",
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment