Lines Matching refs:pd_idx

141 	if (idx == sh->pd_idx)  in raid6_idx_to_slot()
274 WARN_ON(test_bit(R5_InJournal, &sh->dev[sh->pd_idx].flags)); in do_release_stripe()
793 while (dd_idx == sh->pd_idx || dd_idx == sh->qd_idx) in stripe_add_to_batch_list()
1467 if (i == sh->qd_idx || i == sh->pd_idx || in set_syndrome_sources()
1673 int count = 0, pd_idx = sh->pd_idx, i; in ops_run_prexor5() local
1677 struct page *xor_dest = xor_srcs[count++] = sh->dev[pd_idx].page; in ops_run_prexor5()
1794 int pd_idx = sh->pd_idx; in ops_complete_reconstruct() local
1811 if (dev->written || i == pd_idx || i == qd_idx) { in ops_complete_reconstruct()
1844 int count, pd_idx = sh->pd_idx, i; in ops_run_reconstruct5() local
1856 if (pd_idx == i) in ops_run_reconstruct5()
1863 set_bit(R5_Discard, &sh->dev[pd_idx].flags); in ops_run_reconstruct5()
1875 xor_dest = xor_srcs[count++] = sh->dev[pd_idx].page; in ops_run_reconstruct5()
1883 xor_dest = sh->dev[pd_idx].page; in ops_run_reconstruct5()
1886 if (i != pd_idx) in ops_run_reconstruct5()
1939 if (sh->pd_idx == i || sh->qd_idx == i) in ops_run_reconstruct6()
1946 set_bit(R5_Discard, &sh->dev[sh->pd_idx].flags); in ops_run_reconstruct6()
1999 int pd_idx = sh->pd_idx; in ops_run_check_p() local
2013 xor_dest = sh->dev[pd_idx].page; in ops_run_check_p()
2016 if (i == pd_idx || i == qd_idx) in ops_run_check_p()
2568 if (sh->qd_idx >= 0 && sh->pd_idx == i) in raid5_end_read_request()
2720 int pd_idx, qd_idx; in raid5_compute_sector() local
2748 pd_idx = qd_idx = -1; in raid5_compute_sector()
2751 pd_idx = data_disks; in raid5_compute_sector()
2756 pd_idx = data_disks - sector_div(stripe2, raid_disks); in raid5_compute_sector()
2757 if (*dd_idx >= pd_idx) in raid5_compute_sector()
2761 pd_idx = sector_div(stripe2, raid_disks); in raid5_compute_sector()
2762 if (*dd_idx >= pd_idx) in raid5_compute_sector()
2766 pd_idx = data_disks - sector_div(stripe2, raid_disks); in raid5_compute_sector()
2767 *dd_idx = (pd_idx + 1 + *dd_idx) % raid_disks; in raid5_compute_sector()
2770 pd_idx = sector_div(stripe2, raid_disks); in raid5_compute_sector()
2771 *dd_idx = (pd_idx + 1 + *dd_idx) % raid_disks; in raid5_compute_sector()
2774 pd_idx = 0; in raid5_compute_sector()
2778 pd_idx = data_disks; in raid5_compute_sector()
2788 pd_idx = raid_disks - 1 - sector_div(stripe2, raid_disks); in raid5_compute_sector()
2789 qd_idx = pd_idx + 1; in raid5_compute_sector()
2790 if (pd_idx == raid_disks-1) { in raid5_compute_sector()
2793 } else if (*dd_idx >= pd_idx) in raid5_compute_sector()
2797 pd_idx = sector_div(stripe2, raid_disks); in raid5_compute_sector()
2798 qd_idx = pd_idx + 1; in raid5_compute_sector()
2799 if (pd_idx == raid_disks-1) { in raid5_compute_sector()
2802 } else if (*dd_idx >= pd_idx) in raid5_compute_sector()
2806 pd_idx = raid_disks - 1 - sector_div(stripe2, raid_disks); in raid5_compute_sector()
2807 qd_idx = (pd_idx + 1) % raid_disks; in raid5_compute_sector()
2808 *dd_idx = (pd_idx + 2 + *dd_idx) % raid_disks; in raid5_compute_sector()
2811 pd_idx = sector_div(stripe2, raid_disks); in raid5_compute_sector()
2812 qd_idx = (pd_idx + 1) % raid_disks; in raid5_compute_sector()
2813 *dd_idx = (pd_idx + 2 + *dd_idx) % raid_disks; in raid5_compute_sector()
2817 pd_idx = 0; in raid5_compute_sector()
2822 pd_idx = data_disks; in raid5_compute_sector()
2830 pd_idx = sector_div(stripe2, raid_disks); in raid5_compute_sector()
2831 qd_idx = pd_idx + 1; in raid5_compute_sector()
2832 if (pd_idx == raid_disks-1) { in raid5_compute_sector()
2835 } else if (*dd_idx >= pd_idx) in raid5_compute_sector()
2846 pd_idx = raid_disks - 1 - sector_div(stripe2, raid_disks); in raid5_compute_sector()
2847 qd_idx = pd_idx + 1; in raid5_compute_sector()
2848 if (pd_idx == raid_disks-1) { in raid5_compute_sector()
2851 } else if (*dd_idx >= pd_idx) in raid5_compute_sector()
2858 pd_idx = raid_disks - 1 - sector_div(stripe2, raid_disks); in raid5_compute_sector()
2859 qd_idx = (pd_idx + raid_disks - 1) % raid_disks; in raid5_compute_sector()
2860 *dd_idx = (pd_idx + 1 + *dd_idx) % raid_disks; in raid5_compute_sector()
2866 pd_idx = data_disks - sector_div(stripe2, raid_disks-1); in raid5_compute_sector()
2867 if (*dd_idx >= pd_idx) in raid5_compute_sector()
2873 pd_idx = sector_div(stripe2, raid_disks-1); in raid5_compute_sector()
2874 if (*dd_idx >= pd_idx) in raid5_compute_sector()
2880 pd_idx = data_disks - sector_div(stripe2, raid_disks-1); in raid5_compute_sector()
2881 *dd_idx = (pd_idx + 1 + *dd_idx) % (raid_disks-1); in raid5_compute_sector()
2886 pd_idx = sector_div(stripe2, raid_disks-1); in raid5_compute_sector()
2887 *dd_idx = (pd_idx + 1 + *dd_idx) % (raid_disks-1); in raid5_compute_sector()
2892 pd_idx = 0; in raid5_compute_sector()
2904 sh->pd_idx = pd_idx; in raid5_compute_sector()
2935 if (i == sh->pd_idx) in raid5_compute_blocknr()
2943 if (i > sh->pd_idx) in raid5_compute_blocknr()
2948 if (i < sh->pd_idx) in raid5_compute_blocknr()
2950 i -= (sh->pd_idx + 1); in raid5_compute_blocknr()
2969 if (sh->pd_idx == raid_disks-1) in raid5_compute_blocknr()
2971 else if (i > sh->pd_idx) in raid5_compute_blocknr()
2976 if (sh->pd_idx == raid_disks-1) in raid5_compute_blocknr()
2980 if (i < sh->pd_idx) in raid5_compute_blocknr()
2982 i -= (sh->pd_idx + 2); in raid5_compute_blocknr()
2992 if (sh->pd_idx == 0) in raid5_compute_blocknr()
2996 if (i < sh->pd_idx) in raid5_compute_blocknr()
2998 i -= (sh->pd_idx + 1); in raid5_compute_blocknr()
3003 if (i > sh->pd_idx) in raid5_compute_blocknr()
3008 if (i < sh->pd_idx) in raid5_compute_blocknr()
3010 i -= (sh->pd_idx + 1); in raid5_compute_blocknr()
3026 if (check != sh->sector || dummy1 != dd_idx || sh2.pd_idx != sh->pd_idx in raid5_compute_blocknr()
3095 int i, pd_idx = sh->pd_idx, qd_idx = sh->qd_idx, disks = sh->disks; in schedule_reconstruction() local
3141 BUG_ON(!(test_bit(R5_UPTODATE, &sh->dev[pd_idx].flags) || in schedule_reconstruction()
3142 test_bit(R5_Wantcompute, &sh->dev[pd_idx].flags))); in schedule_reconstruction()
3149 if (i == pd_idx || i == qd_idx) in schedule_reconstruction()
3176 set_bit(R5_LOCKED, &sh->dev[pd_idx].flags); in schedule_reconstruction()
3177 clear_bit(R5_UPTODATE, &sh->dev[pd_idx].flags); in schedule_reconstruction()
3192 test_bit(R5_Insync, &sh->dev[pd_idx].flags)) in schedule_reconstruction()
3250 if (i != sh->pd_idx && in add_stripe_bio()
3470 if (test_and_clear_bit(R5_Overlap, &sh->dev[sh->pd_idx].flags)) in handle_failed_sync()
3607 if (s->failed_num[i] != sh->pd_idx && in need_this_block()
3647 ((sh->qd_idx >= 0 && sh->pd_idx == disk_idx) || in fetch_block()
3809 test_bit(R5_Discard, &sh->dev[sh->pd_idx].flags)) { in handle_stripe_clean_event()
3811 clear_bit(R5_Discard, &sh->dev[sh->pd_idx].flags); in handle_stripe_clean_event()
3812 clear_bit(R5_UPTODATE, &sh->dev[sh->pd_idx].flags); in handle_stripe_clean_event()
3894 i == sh->pd_idx || i == sh->qd_idx || in handle_stripe_dirtying()
3906 i != sh->pd_idx && i != sh->qd_idx && in handle_stripe_dirtying()
3930 !test_bit(R5_LOCKED, &sh->dev[sh->pd_idx].flags)) { in handle_stripe_dirtying()
3959 i == sh->pd_idx || i == sh->qd_idx || in handle_stripe_dirtying()
3986 i != sh->pd_idx && i != sh->qd_idx && in handle_stripe_dirtying()
4048 clear_bit(R5_UPTODATE, &sh->dev[sh->pd_idx].flags); in handle_parity_checks5()
4057 dev = &sh->dev[sh->pd_idx]; in handle_parity_checks5()
4109 &sh->dev[sh->pd_idx].flags); in handle_parity_checks5()
4110 sh->ops.target = sh->pd_idx; in handle_parity_checks5()
4130 int pd_idx = sh->pd_idx; in handle_parity_checks6() local
4170 clear_bit(R5_UPTODATE, &sh->dev[pd_idx].flags); in handle_parity_checks6()
4209 dev = &sh->dev[pd_idx]; in handle_parity_checks6()
4279 &sh->dev[pd_idx].flags); in handle_parity_checks6()
4280 *target = pd_idx; in handle_parity_checks6()
4314 if (i != sh->pd_idx && i != sh->qd_idx) { in handle_stripe_expansion()
4345 if (j != sh2->pd_idx && in handle_stripe_expansion()
4714 atomic_read(&sh->count), sh->pd_idx, sh->qd_idx, in handle_stripe()
4779 BUG_ON(!test_bit(R5_UPTODATE, &sh->dev[sh->pd_idx].flags) && in handle_stripe()
4780 !test_bit(R5_Discard, &sh->dev[sh->pd_idx].flags)); in handle_stripe()
4787 (i == sh->pd_idx || i == sh->qd_idx || in handle_stripe()
4797 ((i == sh->pd_idx || i == sh->qd_idx) && in handle_stripe()
4810 pdev = &sh->dev[sh->pd_idx]; in handle_stripe()
4811 s.p_failed = (s.failed >= 1 && s.failed_num[0] == sh->pd_idx) in handle_stripe()
4812 || (s.failed >= 2 && s.failed_num[1] == sh->pd_idx); in handle_stripe()
4925 if (test_and_clear_bit(R5_Overlap, &sh->dev[sh->pd_idx].flags)) in handle_stripe()
5523 set_bit(R5_Overlap, &sh->dev[sh->pd_idx].flags); in make_discard_request()
5529 clear_bit(R5_Overlap, &sh->dev[sh->pd_idx].flags); in make_discard_request()
5532 if (d == sh->pd_idx || d == sh->qd_idx) in make_discard_request()
5546 if (d == sh->pd_idx || d == sh->qd_idx) in make_discard_request()
5924 if (j == sh->pd_idx) in reshape_request()