Lines Matching refs:pd_idx
148 if (idx == sh->pd_idx) in raid6_idx_to_slot()
283 WARN_ON(test_bit(R5_InJournal, &sh->dev[sh->pd_idx].flags)); in do_release_stripe()
945 while (dd_idx == sh->pd_idx || dd_idx == sh->qd_idx) in stripe_add_to_batch_list()
1638 if (i == sh->qd_idx || i == sh->pd_idx || in set_syndrome_sources()
1869 int count = 0, pd_idx = sh->pd_idx, i; in ops_run_prexor5() local
1873 unsigned int off_dest = off_srcs[count] = sh->dev[pd_idx].offset; in ops_run_prexor5()
1874 struct page *xor_dest = xor_srcs[count++] = sh->dev[pd_idx].page; in ops_run_prexor5()
2002 int pd_idx = sh->pd_idx; in ops_complete_reconstruct() local
2019 if (dev->written || i == pd_idx || i == qd_idx) { in ops_complete_reconstruct()
2053 int count, pd_idx = sh->pd_idx, i; in ops_run_reconstruct5() local
2066 if (pd_idx == i) in ops_run_reconstruct5()
2073 set_bit(R5_Discard, &sh->dev[pd_idx].flags); in ops_run_reconstruct5()
2086 off_dest = off_srcs[count] = sh->dev[pd_idx].offset; in ops_run_reconstruct5()
2087 xor_dest = xor_srcs[count++] = sh->dev[pd_idx].page; in ops_run_reconstruct5()
2097 xor_dest = sh->dev[pd_idx].page; in ops_run_reconstruct5()
2098 off_dest = sh->dev[pd_idx].offset; in ops_run_reconstruct5()
2101 if (i != pd_idx) { in ops_run_reconstruct5()
2159 if (sh->pd_idx == i || sh->qd_idx == i) in ops_run_reconstruct6()
2166 set_bit(R5_Discard, &sh->dev[sh->pd_idx].flags); in ops_run_reconstruct6()
2221 int pd_idx = sh->pd_idx; in ops_run_check_p() local
2237 xor_dest = sh->dev[pd_idx].page; in ops_run_check_p()
2238 off_dest = sh->dev[pd_idx].offset; in ops_run_check_p()
2242 if (i == pd_idx || i == qd_idx) in ops_run_check_p()
2853 if (sh->qd_idx >= 0 && sh->pd_idx == i) in raid5_end_read_request()
3000 int pd_idx, qd_idx; in raid5_compute_sector() local
3028 pd_idx = qd_idx = -1; in raid5_compute_sector()
3031 pd_idx = data_disks; in raid5_compute_sector()
3036 pd_idx = data_disks - sector_div(stripe2, raid_disks); in raid5_compute_sector()
3037 if (*dd_idx >= pd_idx) in raid5_compute_sector()
3041 pd_idx = sector_div(stripe2, raid_disks); in raid5_compute_sector()
3042 if (*dd_idx >= pd_idx) in raid5_compute_sector()
3046 pd_idx = data_disks - sector_div(stripe2, raid_disks); in raid5_compute_sector()
3047 *dd_idx = (pd_idx + 1 + *dd_idx) % raid_disks; in raid5_compute_sector()
3050 pd_idx = sector_div(stripe2, raid_disks); in raid5_compute_sector()
3051 *dd_idx = (pd_idx + 1 + *dd_idx) % raid_disks; in raid5_compute_sector()
3054 pd_idx = 0; in raid5_compute_sector()
3058 pd_idx = data_disks; in raid5_compute_sector()
3068 pd_idx = raid_disks - 1 - sector_div(stripe2, raid_disks); in raid5_compute_sector()
3069 qd_idx = pd_idx + 1; in raid5_compute_sector()
3070 if (pd_idx == raid_disks-1) { in raid5_compute_sector()
3073 } else if (*dd_idx >= pd_idx) in raid5_compute_sector()
3077 pd_idx = sector_div(stripe2, raid_disks); in raid5_compute_sector()
3078 qd_idx = pd_idx + 1; in raid5_compute_sector()
3079 if (pd_idx == raid_disks-1) { in raid5_compute_sector()
3082 } else if (*dd_idx >= pd_idx) in raid5_compute_sector()
3086 pd_idx = raid_disks - 1 - sector_div(stripe2, raid_disks); in raid5_compute_sector()
3087 qd_idx = (pd_idx + 1) % raid_disks; in raid5_compute_sector()
3088 *dd_idx = (pd_idx + 2 + *dd_idx) % raid_disks; in raid5_compute_sector()
3091 pd_idx = sector_div(stripe2, raid_disks); in raid5_compute_sector()
3092 qd_idx = (pd_idx + 1) % raid_disks; in raid5_compute_sector()
3093 *dd_idx = (pd_idx + 2 + *dd_idx) % raid_disks; in raid5_compute_sector()
3097 pd_idx = 0; in raid5_compute_sector()
3102 pd_idx = data_disks; in raid5_compute_sector()
3110 pd_idx = sector_div(stripe2, raid_disks); in raid5_compute_sector()
3111 qd_idx = pd_idx + 1; in raid5_compute_sector()
3112 if (pd_idx == raid_disks-1) { in raid5_compute_sector()
3115 } else if (*dd_idx >= pd_idx) in raid5_compute_sector()
3126 pd_idx = raid_disks - 1 - sector_div(stripe2, raid_disks); in raid5_compute_sector()
3127 qd_idx = pd_idx + 1; in raid5_compute_sector()
3128 if (pd_idx == raid_disks-1) { in raid5_compute_sector()
3131 } else if (*dd_idx >= pd_idx) in raid5_compute_sector()
3138 pd_idx = raid_disks - 1 - sector_div(stripe2, raid_disks); in raid5_compute_sector()
3139 qd_idx = (pd_idx + raid_disks - 1) % raid_disks; in raid5_compute_sector()
3140 *dd_idx = (pd_idx + 1 + *dd_idx) % raid_disks; in raid5_compute_sector()
3146 pd_idx = data_disks - sector_div(stripe2, raid_disks-1); in raid5_compute_sector()
3147 if (*dd_idx >= pd_idx) in raid5_compute_sector()
3153 pd_idx = sector_div(stripe2, raid_disks-1); in raid5_compute_sector()
3154 if (*dd_idx >= pd_idx) in raid5_compute_sector()
3160 pd_idx = data_disks - sector_div(stripe2, raid_disks-1); in raid5_compute_sector()
3161 *dd_idx = (pd_idx + 1 + *dd_idx) % (raid_disks-1); in raid5_compute_sector()
3166 pd_idx = sector_div(stripe2, raid_disks-1); in raid5_compute_sector()
3167 *dd_idx = (pd_idx + 1 + *dd_idx) % (raid_disks-1); in raid5_compute_sector()
3172 pd_idx = 0; in raid5_compute_sector()
3184 sh->pd_idx = pd_idx; in raid5_compute_sector()
3215 if (i == sh->pd_idx) in raid5_compute_blocknr()
3223 if (i > sh->pd_idx) in raid5_compute_blocknr()
3228 if (i < sh->pd_idx) in raid5_compute_blocknr()
3230 i -= (sh->pd_idx + 1); in raid5_compute_blocknr()
3249 if (sh->pd_idx == raid_disks-1) in raid5_compute_blocknr()
3251 else if (i > sh->pd_idx) in raid5_compute_blocknr()
3256 if (sh->pd_idx == raid_disks-1) in raid5_compute_blocknr()
3260 if (i < sh->pd_idx) in raid5_compute_blocknr()
3262 i -= (sh->pd_idx + 2); in raid5_compute_blocknr()
3272 if (sh->pd_idx == 0) in raid5_compute_blocknr()
3276 if (i < sh->pd_idx) in raid5_compute_blocknr()
3278 i -= (sh->pd_idx + 1); in raid5_compute_blocknr()
3283 if (i > sh->pd_idx) in raid5_compute_blocknr()
3288 if (i < sh->pd_idx) in raid5_compute_blocknr()
3290 i -= (sh->pd_idx + 1); in raid5_compute_blocknr()
3306 if (check != sh->sector || dummy1 != dd_idx || sh2.pd_idx != sh->pd_idx in raid5_compute_blocknr()
3375 int i, pd_idx = sh->pd_idx, qd_idx = sh->qd_idx, disks = sh->disks; in schedule_reconstruction() local
3421 BUG_ON(!(test_bit(R5_UPTODATE, &sh->dev[pd_idx].flags) || in schedule_reconstruction()
3422 test_bit(R5_Wantcompute, &sh->dev[pd_idx].flags))); in schedule_reconstruction()
3429 if (i == pd_idx || i == qd_idx) in schedule_reconstruction()
3456 set_bit(R5_LOCKED, &sh->dev[pd_idx].flags); in schedule_reconstruction()
3457 clear_bit(R5_UPTODATE, &sh->dev[pd_idx].flags); in schedule_reconstruction()
3472 test_bit(R5_Insync, &sh->dev[pd_idx].flags)) in schedule_reconstruction()
3522 if (i != sh->pd_idx && in stripe_bio_overlaps()
3774 if (test_and_clear_bit(R5_Overlap, &sh->dev[sh->pd_idx].flags)) in handle_failed_sync()
3900 s->failed_num[i] == sh->pd_idx || in need_this_block()
3922 if (s->failed_num[i] != sh->pd_idx && in need_this_block()
3962 ((sh->qd_idx >= 0 && sh->pd_idx == disk_idx) || in fetch_block()
4124 test_bit(R5_Discard, &sh->dev[sh->pd_idx].flags)) { in handle_stripe_clean_event()
4126 clear_bit(R5_Discard, &sh->dev[sh->pd_idx].flags); in handle_stripe_clean_event()
4127 clear_bit(R5_UPTODATE, &sh->dev[sh->pd_idx].flags); in handle_stripe_clean_event()
4209 i == sh->pd_idx || i == sh->qd_idx || in handle_stripe_dirtying()
4221 i != sh->pd_idx && i != sh->qd_idx && in handle_stripe_dirtying()
4245 !test_bit(R5_LOCKED, &sh->dev[sh->pd_idx].flags)) { in handle_stripe_dirtying()
4274 i == sh->pd_idx || i == sh->qd_idx || in handle_stripe_dirtying()
4299 i != sh->pd_idx && i != sh->qd_idx && in handle_stripe_dirtying()
4359 clear_bit(R5_UPTODATE, &sh->dev[sh->pd_idx].flags); in handle_parity_checks5()
4368 dev = &sh->dev[sh->pd_idx]; in handle_parity_checks5()
4420 &sh->dev[sh->pd_idx].flags); in handle_parity_checks5()
4421 sh->ops.target = sh->pd_idx; in handle_parity_checks5()
4441 int pd_idx = sh->pd_idx; in handle_parity_checks6() local
4481 clear_bit(R5_UPTODATE, &sh->dev[pd_idx].flags); in handle_parity_checks6()
4520 dev = &sh->dev[pd_idx]; in handle_parity_checks6()
4590 &sh->dev[pd_idx].flags); in handle_parity_checks6()
4591 *target = pd_idx; in handle_parity_checks6()
4625 if (i != sh->pd_idx && i != sh->qd_idx) { in handle_stripe_expansion()
4658 if (j != sh2->pd_idx && in handle_stripe_expansion()
5032 atomic_read(&sh->count), sh->pd_idx, sh->qd_idx, in handle_stripe()
5097 BUG_ON(!test_bit(R5_UPTODATE, &sh->dev[sh->pd_idx].flags) && in handle_stripe()
5098 !test_bit(R5_Discard, &sh->dev[sh->pd_idx].flags)); in handle_stripe()
5105 (i == sh->pd_idx || i == sh->qd_idx || in handle_stripe()
5115 ((i == sh->pd_idx || i == sh->qd_idx) && in handle_stripe()
5128 pdev = &sh->dev[sh->pd_idx]; in handle_stripe()
5129 s.p_failed = (s.failed >= 1 && s.failed_num[0] == sh->pd_idx) in handle_stripe()
5130 || (s.failed >= 2 && s.failed_num[1] == sh->pd_idx); in handle_stripe()
5243 if (test_and_clear_bit(R5_Overlap, &sh->dev[sh->pd_idx].flags)) in handle_stripe()
5825 set_bit(R5_Overlap, &sh->dev[sh->pd_idx].flags); in make_discard_request()
5831 clear_bit(R5_Overlap, &sh->dev[sh->pd_idx].flags); in make_discard_request()
5834 if (d == sh->pd_idx || d == sh->qd_idx) in make_discard_request()
5848 if (d == sh->pd_idx || d == sh->qd_idx) in make_discard_request()
5901 if (dd_idx == sh->pd_idx) in stripe_ahead_of_reshape()
5932 if (dd_idx == sh->pd_idx || dd_idx == sh->qd_idx) in add_all_stripe_bios()
5952 if (dd_idx == sh->pd_idx || dd_idx == sh->qd_idx) in add_all_stripe_bios()
6373 if (j == sh->pd_idx) in reshape_request()