/drivers/md/ |
D | raid5.c | 151 if (idx == sh->pd_idx) in raid6_idx_to_slot() 284 WARN_ON(test_bit(R5_InJournal, &sh->dev[sh->pd_idx].flags)); in do_release_stripe() 801 while (dd_idx == sh->pd_idx || dd_idx == sh->qd_idx) in stripe_add_to_batch_list() 1476 if (i == sh->qd_idx || i == sh->pd_idx || in set_syndrome_sources() 1682 int count = 0, pd_idx = sh->pd_idx, i; in ops_run_prexor5() local 1686 struct page *xor_dest = xor_srcs[count++] = sh->dev[pd_idx].page; in ops_run_prexor5() 1803 int pd_idx = sh->pd_idx; in ops_complete_reconstruct() local 1820 if (dev->written || i == pd_idx || i == qd_idx) { in ops_complete_reconstruct() 1853 int count, pd_idx = sh->pd_idx, i; in ops_run_reconstruct5() local 1865 if (pd_idx == i) in ops_run_reconstruct5() [all …]
|
D | raid5-ppl.c | 162 int count = 0, pd_idx = sh->pd_idx, i; in ops_run_partial_parity() local 179 srcs[count++] = sh->dev[pd_idx].page; in ops_run_partial_parity() 300 if (i != sh->pd_idx && test_bit(R5_Wantwrite, &dev->flags)) { in ppl_log_stripe() 337 e->parity_disk = cpu_to_le32(sh->pd_idx); in ppl_log_stripe() 366 !test_bit(R5_Wantwrite, &sh->dev[sh->pd_idx].flags) || in ppl_write_stripe() 367 !test_bit(R5_Insync, &sh->dev[sh->pd_idx].flags)) { in ppl_write_stripe() 372 log = &ppl_conf->child_logs[sh->pd_idx]; in ppl_write_stripe() 810 BUG_ON(sh.pd_idx != le32_to_cpu(e->parity_disk)); in ppl_recover_entry() 811 parity_rdev = conf->disks[sh.pd_idx].rdev; in ppl_recover_entry()
|
D | raid5-cache.c | 515 set_bit(R5_InJournal, &sh->dev[sh->pd_idx].flags); in r5c_finish_cache_stripe() 520 set_bit(R5_InJournal, &sh->dev[sh->pd_idx].flags); in r5c_finish_cache_stripe() 942 if (i == sh->pd_idx || i == sh->qd_idx) in r5l_log_stripe() 961 sh->sector, sh->dev[sh->pd_idx].log_checksum, in r5l_log_stripe() 963 r5l_append_payload_page(log, sh->dev[sh->pd_idx].page); in r5l_log_stripe() 967 sh->sector, sh->dev[sh->pd_idx].log_checksum, in r5l_log_stripe() 969 r5l_append_payload_page(log, sh->dev[sh->pd_idx].page); in r5l_log_stripe() 1018 if (sh->log_io || !test_bit(R5_Wantwrite, &sh->dev[sh->pd_idx].flags) || in r5l_write_stripe() 1860 r5l_recovery_read_page(log, ctx, sh->dev[sh->pd_idx].page, log_offset); in r5l_recovery_load_parity() 1861 sh->dev[sh->pd_idx].log_checksum = in r5l_recovery_load_parity() [all …]
|
D | raid5.h | 206 short pd_idx; /* parity disk index */ member
|
/drivers/infiniband/hw/i40iw/ |
D | i40iw_hmc.h | 70 #define I40IW_INVALIDATE_PF_HMC_PD(hw, sd_idx, pd_idx) \ argument 74 ((pd_idx) << I40E_PFHMC_PDINV_PMPDIDX_SHIFT))) 83 #define I40IW_INVALIDATE_VF_HMC_PD(hw, sd_idx, pd_idx, hmc_fn_id) \ argument 86 (pd_idx << I40E_PFHMC_PDINV_PMPDIDX_SHIFT)))
|
D | i40iw_hmc.c | 92 u32 *pd_idx, in i40iw_find_pd_index_limit() argument 100 *(pd_idx) = (u32)(fpm_adr / I40IW_HMC_PAGED_BP_SIZE); in i40iw_find_pd_index_limit() 300 u32 pd_idx = 0, pd_lmt = 0; in i40iw_sc_create_hmc_obj() local 329 info->start_idx, info->count, &pd_idx, &pd_lmt); in i40iw_sc_create_hmc_obj() 343 pd_idx1 = max(pd_idx, (j * I40IW_HMC_MAX_BP_COUNT)); in i40iw_sc_create_hmc_obj() 377 pd_idx1 = max(pd_idx, in i40iw_sc_create_hmc_obj() 453 u32 pd_idx, pd_lmt, rel_pd_idx; in i40iw_sc_del_hmc_obj() local 482 info->start_idx, info->count, &pd_idx, &pd_lmt); in i40iw_sc_del_hmc_obj() 484 for (j = pd_idx; j < pd_lmt; j++) { in i40iw_sc_del_hmc_obj()
|
D | i40iw_pble.c | 122 idx->pd_idx = (u32)(pble_rsrc->next_fpm_addr) / I40IW_HMC_PAGED_BP_SIZE; in get_sd_pd_idx() 123 idx->rel_pd_idx = (idx->pd_idx % I40IW_HMC_PD_CNT_IN_SD); in get_sd_pd_idx() 263 u32 pd_idx = info->idx.pd_idx; in add_bp_pages() local 291 status = i40iw_add_pd_table_entry(dev->hw, hmc_info, pd_idx++, &mem); in add_bp_pages()
|
D | i40iw_pble.h | 78 u32 pd_idx; member
|
/drivers/net/ethernet/intel/i40e/ |
D | i40e_lan_hmc.c | 307 u32 pd_idx = 0, pd_lmt = 0; in i40e_create_lan_hmc_object() local 354 info->start_idx, info->count, &pd_idx, in i40e_create_lan_hmc_object() 383 pd_idx1 = max(pd_idx, (j * I40E_HMC_MAX_BP_COUNT)); in i40e_create_lan_hmc_object() 431 pd_idx1 = max(pd_idx, in i40e_create_lan_hmc_object() 550 u32 pd_idx, pd_lmt, rel_pd_idx; in i40e_delete_lan_hmc_object() local 597 info->start_idx, info->count, &pd_idx, in i40e_delete_lan_hmc_object() 600 for (j = pd_idx; j < pd_lmt; j++) { in i40e_delete_lan_hmc_object() 1007 u32 pd_idx, pd_lmt, rel_pd_idx; in i40e_hmc_get_object_va() local 1047 &pd_idx, &pd_lmt); in i40e_hmc_get_object_va() 1048 rel_pd_idx = pd_idx % I40E_HMC_PD_CNT_IN_SD; in i40e_hmc_get_object_va()
|
D | i40e_hmc.h | 162 #define I40E_INVALIDATE_PF_HMC_PD(hw, sd_idx, pd_idx) \ argument 165 ((pd_idx) << I40E_PFHMC_PDINV_PMPDIDX_SHIFT)))
|
/drivers/net/ethernet/intel/i40evf/ |
D | i40e_hmc.h | 162 #define I40E_INVALIDATE_PF_HMC_PD(hw, sd_idx, pd_idx) \ argument 165 ((pd_idx) << I40E_PFHMC_PDINV_PMPDIDX_SHIFT)))
|
/drivers/infiniband/hw/ocrdma/ |
D | ocrdma_verbs.c | 388 u16 pd_idx = 0; in ocrdma_get_pd_num() local 395 pd_idx = _ocrdma_pd_mgr_get_bitmap(dev, true); in ocrdma_get_pd_num() 396 pd->id = dev->pd_mgr->pd_dpp_start + pd_idx; in ocrdma_get_pd_num() 397 pd->dpp_page = dev->pd_mgr->dpp_page_index + pd_idx; in ocrdma_get_pd_num() 400 pd_idx = _ocrdma_pd_mgr_get_bitmap(dev, false); in ocrdma_get_pd_num() 401 pd->id = dev->pd_mgr->pd_norm_start + pd_idx; in ocrdma_get_pd_num() 408 pd_idx = _ocrdma_pd_mgr_get_bitmap(dev, false); in ocrdma_get_pd_num() 409 pd->id = dev->pd_mgr->pd_norm_start + pd_idx; in ocrdma_get_pd_num()
|