struct dma_async_tx_descriptor *tx)
{
int disks = sh->disks;
- struct page **xor_srcs = flex_array_get(percpu->scribble, 0);
+ struct page **srcs = flex_array_get(percpu->scribble, 0);
int count = 0, pd_idx = sh->pd_idx, i;
struct async_submit_ctl submit;
* differently.
*/
if (sh->reconstruct_state == reconstruct_state_prexor_drain_run) {
- /* rmw: xor old data and parity from updated disks */
- for (i = disks; i--;) {
- struct r5dev *dev = &sh->dev[i];
- if (test_bit(R5_Wantdrain, &dev->flags) || i == pd_idx)
- xor_srcs[count++] = dev->page;
- }
+ /*
+ * rmw: xor old data and parity from updated disks
+ * This is calculated earlier by ops_run_prexor5() so just copy
+ * the parity dev page.
+ */
+ srcs[count++] = sh->dev[pd_idx].page;
} else if (sh->reconstruct_state == reconstruct_state_drain_run) {
/* rcw: xor data from all not updated disks */
for (i = disks; i--;) {
struct r5dev *dev = &sh->dev[i];
if (test_bit(R5_UPTODATE, &dev->flags))
- xor_srcs[count++] = dev->page;
+ srcs[count++] = dev->page;
}
} else {
return tx;
+ sizeof(struct page *) * (sh->disks + 2));
if (count == 1)
- tx = async_memcpy(sh->ppl_page, xor_srcs[0], 0, 0, PAGE_SIZE,
+ tx = async_memcpy(sh->ppl_page, srcs[0], 0, 0, PAGE_SIZE,
&submit);
else
- tx = async_xor(sh->ppl_page, xor_srcs, 0, count, PAGE_SIZE,
+ tx = async_xor(sh->ppl_page, srcs, 0, count, PAGE_SIZE,
&submit);
return tx;
async_tx_ack(tx);
}
- if (test_bit(STRIPE_OP_PARTIAL_PARITY, &ops_request))
- tx = ops_run_partial_parity(sh, percpu, tx);
-
if (test_bit(STRIPE_OP_PREXOR, &ops_request)) {
if (level < 6)
tx = ops_run_prexor5(sh, percpu, tx);
tx = ops_run_prexor6(sh, percpu, tx);
}
+ if (test_bit(STRIPE_OP_PARTIAL_PARITY, &ops_request))
+ tx = ops_run_partial_parity(sh, percpu, tx);
+
if (test_bit(STRIPE_OP_BIODRAIN, &ops_request)) {
tx = ops_run_biodrain(sh, tx);
overlap_clear++;