Home
last modified time | relevance | path

Searched refs:pd_idx (Results 1 – 12 of 12) sorted by relevance

/drivers/md/
Draid5.c151 if (idx == sh->pd_idx) in raid6_idx_to_slot()
284 WARN_ON(test_bit(R5_InJournal, &sh->dev[sh->pd_idx].flags)); in do_release_stripe()
801 while (dd_idx == sh->pd_idx || dd_idx == sh->qd_idx) in stripe_add_to_batch_list()
1476 if (i == sh->qd_idx || i == sh->pd_idx || in set_syndrome_sources()
1682 int count = 0, pd_idx = sh->pd_idx, i; in ops_run_prexor5() local
1686 struct page *xor_dest = xor_srcs[count++] = sh->dev[pd_idx].page; in ops_run_prexor5()
1803 int pd_idx = sh->pd_idx; in ops_complete_reconstruct() local
1820 if (dev->written || i == pd_idx || i == qd_idx) { in ops_complete_reconstruct()
1853 int count, pd_idx = sh->pd_idx, i; in ops_run_reconstruct5() local
1865 if (pd_idx == i) in ops_run_reconstruct5()
[all …]
Draid5-ppl.c162 int count = 0, pd_idx = sh->pd_idx, i; in ops_run_partial_parity() local
179 srcs[count++] = sh->dev[pd_idx].page; in ops_run_partial_parity()
300 if (i != sh->pd_idx && test_bit(R5_Wantwrite, &dev->flags)) { in ppl_log_stripe()
337 e->parity_disk = cpu_to_le32(sh->pd_idx); in ppl_log_stripe()
366 !test_bit(R5_Wantwrite, &sh->dev[sh->pd_idx].flags) || in ppl_write_stripe()
367 !test_bit(R5_Insync, &sh->dev[sh->pd_idx].flags)) { in ppl_write_stripe()
372 log = &ppl_conf->child_logs[sh->pd_idx]; in ppl_write_stripe()
810 BUG_ON(sh.pd_idx != le32_to_cpu(e->parity_disk)); in ppl_recover_entry()
811 parity_rdev = conf->disks[sh.pd_idx].rdev; in ppl_recover_entry()
Draid5-cache.c515 set_bit(R5_InJournal, &sh->dev[sh->pd_idx].flags); in r5c_finish_cache_stripe()
520 set_bit(R5_InJournal, &sh->dev[sh->pd_idx].flags); in r5c_finish_cache_stripe()
942 if (i == sh->pd_idx || i == sh->qd_idx) in r5l_log_stripe()
961 sh->sector, sh->dev[sh->pd_idx].log_checksum, in r5l_log_stripe()
963 r5l_append_payload_page(log, sh->dev[sh->pd_idx].page); in r5l_log_stripe()
967 sh->sector, sh->dev[sh->pd_idx].log_checksum, in r5l_log_stripe()
969 r5l_append_payload_page(log, sh->dev[sh->pd_idx].page); in r5l_log_stripe()
1018 if (sh->log_io || !test_bit(R5_Wantwrite, &sh->dev[sh->pd_idx].flags) || in r5l_write_stripe()
1860 r5l_recovery_read_page(log, ctx, sh->dev[sh->pd_idx].page, log_offset); in r5l_recovery_load_parity()
1861 sh->dev[sh->pd_idx].log_checksum = in r5l_recovery_load_parity()
[all …]
Draid5.h206 short pd_idx; /* parity disk index */ member
/drivers/infiniband/hw/i40iw/
Di40iw_hmc.h70 #define I40IW_INVALIDATE_PF_HMC_PD(hw, sd_idx, pd_idx) \ argument
74 ((pd_idx) << I40E_PFHMC_PDINV_PMPDIDX_SHIFT)))
83 #define I40IW_INVALIDATE_VF_HMC_PD(hw, sd_idx, pd_idx, hmc_fn_id) \ argument
86 (pd_idx << I40E_PFHMC_PDINV_PMPDIDX_SHIFT)))
Di40iw_hmc.c92 u32 *pd_idx, in i40iw_find_pd_index_limit() argument
100 *(pd_idx) = (u32)(fpm_adr / I40IW_HMC_PAGED_BP_SIZE); in i40iw_find_pd_index_limit()
300 u32 pd_idx = 0, pd_lmt = 0; in i40iw_sc_create_hmc_obj() local
329 info->start_idx, info->count, &pd_idx, &pd_lmt); in i40iw_sc_create_hmc_obj()
343 pd_idx1 = max(pd_idx, (j * I40IW_HMC_MAX_BP_COUNT)); in i40iw_sc_create_hmc_obj()
377 pd_idx1 = max(pd_idx, in i40iw_sc_create_hmc_obj()
453 u32 pd_idx, pd_lmt, rel_pd_idx; in i40iw_sc_del_hmc_obj() local
482 info->start_idx, info->count, &pd_idx, &pd_lmt); in i40iw_sc_del_hmc_obj()
484 for (j = pd_idx; j < pd_lmt; j++) { in i40iw_sc_del_hmc_obj()
Di40iw_pble.c122 idx->pd_idx = (u32)(pble_rsrc->next_fpm_addr) / I40IW_HMC_PAGED_BP_SIZE; in get_sd_pd_idx()
123 idx->rel_pd_idx = (idx->pd_idx % I40IW_HMC_PD_CNT_IN_SD); in get_sd_pd_idx()
263 u32 pd_idx = info->idx.pd_idx; in add_bp_pages() local
291 status = i40iw_add_pd_table_entry(dev->hw, hmc_info, pd_idx++, &mem); in add_bp_pages()
Di40iw_pble.h78 u32 pd_idx; member
/drivers/net/ethernet/intel/i40e/
Di40e_lan_hmc.c307 u32 pd_idx = 0, pd_lmt = 0; in i40e_create_lan_hmc_object() local
354 info->start_idx, info->count, &pd_idx, in i40e_create_lan_hmc_object()
383 pd_idx1 = max(pd_idx, (j * I40E_HMC_MAX_BP_COUNT)); in i40e_create_lan_hmc_object()
431 pd_idx1 = max(pd_idx, in i40e_create_lan_hmc_object()
550 u32 pd_idx, pd_lmt, rel_pd_idx; in i40e_delete_lan_hmc_object() local
597 info->start_idx, info->count, &pd_idx, in i40e_delete_lan_hmc_object()
600 for (j = pd_idx; j < pd_lmt; j++) { in i40e_delete_lan_hmc_object()
1007 u32 pd_idx, pd_lmt, rel_pd_idx; in i40e_hmc_get_object_va() local
1047 &pd_idx, &pd_lmt); in i40e_hmc_get_object_va()
1048 rel_pd_idx = pd_idx % I40E_HMC_PD_CNT_IN_SD; in i40e_hmc_get_object_va()
Di40e_hmc.h162 #define I40E_INVALIDATE_PF_HMC_PD(hw, sd_idx, pd_idx) \ argument
165 ((pd_idx) << I40E_PFHMC_PDINV_PMPDIDX_SHIFT)))
/drivers/net/ethernet/intel/i40evf/
Di40e_hmc.h162 #define I40E_INVALIDATE_PF_HMC_PD(hw, sd_idx, pd_idx) \ argument
165 ((pd_idx) << I40E_PFHMC_PDINV_PMPDIDX_SHIFT)))
/drivers/infiniband/hw/ocrdma/
Docrdma_verbs.c388 u16 pd_idx = 0; in ocrdma_get_pd_num() local
395 pd_idx = _ocrdma_pd_mgr_get_bitmap(dev, true); in ocrdma_get_pd_num()
396 pd->id = dev->pd_mgr->pd_dpp_start + pd_idx; in ocrdma_get_pd_num()
397 pd->dpp_page = dev->pd_mgr->dpp_page_index + pd_idx; in ocrdma_get_pd_num()
400 pd_idx = _ocrdma_pd_mgr_get_bitmap(dev, false); in ocrdma_get_pd_num()
401 pd->id = dev->pd_mgr->pd_norm_start + pd_idx; in ocrdma_get_pd_num()
408 pd_idx = _ocrdma_pd_mgr_get_bitmap(dev, false); in ocrdma_get_pd_num()
409 pd->id = dev->pd_mgr->pd_norm_start + pd_idx; in ocrdma_get_pd_num()