/drivers/infiniband/hw/bnxt_re/ |
D | qplib_res.c | 66 dma_free_coherent(&pdev->dev, pbl->pg_size, in __free_pbl() 82 pbl->pg_size = 0; in __free_pbl() 87 u32 nmaps, u32 pg_size) in __alloc_pbl() argument 105 pbl->pg_size = pg_size; in __alloc_pbl() 110 pbl->pg_size, in __alloc_pbl() 164 u32 pg_size, enum bnxt_qplib_hwq_type hwq_type) in bnxt_qplib_alloc_init_hwq() argument 176 aux_pages = (slots * aux_size) / pg_size; in bnxt_qplib_alloc_init_hwq() 177 if ((slots * aux_size) % pg_size) in bnxt_qplib_alloc_init_hwq() 187 pages = (slots * size) / pg_size + aux_pages; in bnxt_qplib_alloc_init_hwq() 188 if ((slots * size) % pg_size) in bnxt_qplib_alloc_init_hwq() [all …]
|
D | qplib_fp.c | 568 (pbl->pg_size == ROCE_PG_SIZE_4K ? in bnxt_qplib_create_srq() 570 pbl->pg_size == ROCE_PG_SIZE_8K ? in bnxt_qplib_create_srq() 572 pbl->pg_size == ROCE_PG_SIZE_64K ? in bnxt_qplib_create_srq() 574 pbl->pg_size == ROCE_PG_SIZE_2M ? in bnxt_qplib_create_srq() 576 pbl->pg_size == ROCE_PG_SIZE_8M ? in bnxt_qplib_create_srq() 578 pbl->pg_size == ROCE_PG_SIZE_1G ? in bnxt_qplib_create_srq() 760 (pbl->pg_size == ROCE_PG_SIZE_4K ? in bnxt_qplib_create_qp1() 762 pbl->pg_size == ROCE_PG_SIZE_8K ? in bnxt_qplib_create_qp1() 764 pbl->pg_size == ROCE_PG_SIZE_64K ? in bnxt_qplib_create_qp1() 766 pbl->pg_size == ROCE_PG_SIZE_2M ? in bnxt_qplib_create_qp1() [all …]
|
D | qplib_rcfw.c | 461 return (pbl->pg_size == ROCE_PG_SIZE_4K ? in __get_pbl_pg_idx() 463 pbl->pg_size == ROCE_PG_SIZE_8K ? in __get_pbl_pg_idx() 465 pbl->pg_size == ROCE_PG_SIZE_64K ? in __get_pbl_pg_idx() 467 pbl->pg_size == ROCE_PG_SIZE_2M ? in __get_pbl_pg_idx() 469 pbl->pg_size == ROCE_PG_SIZE_8M ? in __get_pbl_pg_idx() 471 pbl->pg_size == ROCE_PG_SIZE_1G ? in __get_pbl_pg_idx()
|
D | qplib_res.h | 83 u32 pg_size; member 241 u32 elements_per_page, u32 aux, u32 pg_size,
|
D | qplib_sp.c | 664 u32 pg_size; in bnxt_qplib_reg_mr() local 710 pg_size = PAGE_SIZE; in bnxt_qplib_reg_mr() 715 pg_size = buf_pg_size ? buf_pg_size : PAGE_SIZE; in bnxt_qplib_reg_mr() 717 ((ilog2(pg_size) << in bnxt_qplib_reg_mr()
|
D | ib_verbs.c | 3644 resp.pg_size = PAGE_SIZE; in bnxt_re_alloc_ucontext()
|
/drivers/net/ethernet/huawei/hinic/ |
D | hinic_hw_eqs.c | 26 #define GET_EQ_NUM_PAGES(eq, pg_size) \ argument 27 (ALIGN((eq)->q_len * (eq)->elem_size, pg_size) / (pg_size)) 29 #define GET_EQ_NUM_ELEMS_IN_PG(eq, pg_size) ((pg_size) / (eq)->elem_size) argument
|
/drivers/crypto/chelsio/chtls/ |
D | chtls_io.c | 1076 int pg_size = PAGE_SIZE; in chtls_sendmsg() local 1081 pg_size = page_size(page); in chtls_sendmsg() 1082 if (off < pg_size && in chtls_sendmsg() 1092 if (page && off == pg_size) { in chtls_sendmsg() 1095 pg_size = PAGE_SIZE; in chtls_sendmsg() 1108 pg_size <<= order; in chtls_sendmsg() 1112 pg_size = PAGE_SIZE; in chtls_sendmsg() 1119 if (copy > pg_size - off) in chtls_sendmsg() 1120 copy = pg_size - off; in chtls_sendmsg() 1141 if (off + copy < pg_size) { in chtls_sendmsg()
|
/drivers/net/wireless/realtek/rtw88/ |
D | fw.c | 592 u32 pg_size; in rtw_download_drv_rsvd_page() local 596 pg_size = rtwdev->chip->page_size; in rtw_download_drv_rsvd_page() 597 pg_num = size / pg_size + ((size & (pg_size - 1)) ? 1 : 0); in rtw_download_drv_rsvd_page()
|
/drivers/infiniband/hw/i40iw/ |
D | i40iw_verbs.c | 1340 static bool i40iw_check_mem_contiguous(u64 *arr, u32 npages, u32 pg_size) in i40iw_check_mem_contiguous() argument 1345 if ((*arr + (pg_size * pg_idx)) != arr[pg_idx]) in i40iw_check_mem_contiguous() 1356 static bool i40iw_check_mr_contiguous(struct i40iw_pble_alloc *palloc, u32 pg_size) in i40iw_check_mr_contiguous() argument 1367 ret = i40iw_check_mem_contiguous(arr, palloc->total_cnt, pg_size); in i40iw_check_mr_contiguous() 1375 if ((*start_addr + (i * pg_size * PBLE_PER_PAGE)) != *arr) in i40iw_check_mr_contiguous() 1377 ret = i40iw_check_mem_contiguous(arr, leaf->cnt, pg_size); in i40iw_check_mr_contiguous() 1443 u32 pg_size; in i40iw_handle_q_mem() local 1449 pg_size = iwmr->page_size; in i40iw_handle_q_mem() 1469 ret = i40iw_check_mem_contiguous(arr, req->sq_pages, pg_size); in i40iw_handle_q_mem() 1471 ret = i40iw_check_mem_contiguous(&arr[req->sq_pages], req->rq_pages, pg_size); in i40iw_handle_q_mem() [all …]
|
/drivers/crypto/caam/ |
D | ctrl.c | 586 int pg_size; in caam_probe() local 658 pg_size = (comp_params & CTPR_MS_PG_SZ_MASK) >> CTPR_MS_PG_SZ_SHIFT; in caam_probe() 659 if (pg_size == 0) in caam_probe()
|
/drivers/iommu/ |
D | iommu.c | 646 unsigned long pg_size; in iommu_group_create_direct_mappings() local 654 pg_size = 1UL << __ffs(domain->pgsize_bitmap); in iommu_group_create_direct_mappings() 666 start = ALIGN(entry->start, pg_size); in iommu_group_create_direct_mappings() 667 end = ALIGN(entry->start + entry->length, pg_size); in iommu_group_create_direct_mappings() 673 for (addr = start; addr < end; addr += pg_size) { in iommu_group_create_direct_mappings() 680 ret = iommu_map(domain, addr, addr, pg_size, entry->prot); in iommu_group_create_direct_mappings()
|
D | amd_iommu.c | 496 unsigned long pte_mask, pg_size, cnt; in first_pte_l7() local 499 pg_size = PTE_PAGE_SIZE(*pte); in first_pte_l7() 500 cnt = PAGE_SIZE_PTE_COUNT(pg_size); in first_pte_l7() 505 *page_size = pg_size; in first_pte_l7()
|
/drivers/target/sbp/ |
D | sbp_target.c | 1244 int tcode, sg_miter_flags, max_payload, pg_size, speed, node_id, in sbp_rw_data() local 1263 pg_size = CMDBLK_ORB_PG_SIZE(be32_to_cpu(req->orb.misc)); in sbp_rw_data() 1264 if (pg_size) { in sbp_rw_data() 1266 pg_size = 0x100 << pg_size; in sbp_rw_data()
|
/drivers/net/ethernet/chelsio/cxgb3/ |
D | t3_hw.c | 2586 unsigned int pg_size) in pm_num_pages() argument 2588 unsigned int n = mem_size / pg_size; in pm_num_pages()
|
/drivers/net/ethernet/broadcom/bnxt/ |
D | bnxt.c | 6502 u8 pg_size = 0; in bnxt_hwrm_set_pg_attr() local 6505 pg_size = 1 << 4; in bnxt_hwrm_set_pg_attr() 6507 pg_size = 2 << 4; in bnxt_hwrm_set_pg_attr() 6509 *pg_attr = pg_size; in bnxt_hwrm_set_pg_attr()
|