Home
last modified time | relevance | path

Searched refs:page_size (Results 1 – 25 of 185) sorted by relevance

12345678

/drivers/misc/habanalabs/common/mmu/
Dmmu.c16 return hl_mem_area_inside_range(virt_addr, prop->dmmu.page_size, in hl_is_dram_va()
145 int hl_mmu_unmap_page(struct hl_ctx *ctx, u64 virt_addr, u32 page_size, in hl_mmu_unmap_page() argument
163 else if ((page_size % prop->pmmu_huge.page_size) == 0) in hl_mmu_unmap_page()
173 if ((page_size % mmu_prop->page_size) == 0) { in hl_mmu_unmap_page()
174 real_page_size = mmu_prop->page_size; in hl_mmu_unmap_page()
182 if (is_dram_addr && ((page_size % prop->dram_page_size) == 0)) { in hl_mmu_unmap_page()
187 page_size, mmu_prop->page_size >> 10); in hl_mmu_unmap_page()
193 npages = page_size / real_page_size; in hl_mmu_unmap_page()
234 u32 page_size, bool flush_pte) in hl_mmu_map_page() argument
252 else if ((page_size % prop->pmmu_huge.page_size) == 0) in hl_mmu_map_page()
[all …]
/drivers/pci/endpoint/
Dpci-epc-mem.c26 unsigned int page_shift = ilog2(mem->window.page_size); in pci_epc_mem_get_order()
54 size_t page_size; in pci_epc_multi_mem_init() local
70 page_size = windows[i].page_size; in pci_epc_multi_mem_init()
71 if (page_size < PAGE_SIZE) in pci_epc_multi_mem_init()
72 page_size = PAGE_SIZE; in pci_epc_multi_mem_init()
73 page_shift = ilog2(page_size); in pci_epc_multi_mem_init()
94 mem->window.page_size = page_size; in pci_epc_multi_mem_init()
119 size_t size, size_t page_size) in pci_epc_mem_init() argument
125 mem_window.page_size = page_size; in pci_epc_mem_init()
182 align_size = ALIGN(size, mem->window.page_size); in pci_epc_mem_alloc_addr()
[all …]
/drivers/misc/
Dvmw_balloon.c247 enum vmballoon_page_size_type page_size; member
576 unsigned int vmballoon_page_order(enum vmballoon_page_size_type page_size) in vmballoon_page_order() argument
578 return page_size == VMW_BALLOON_2M_PAGE ? VMW_BALLOON_2M_ORDER : 0; in vmballoon_page_order()
588 vmballoon_page_in_frames(enum vmballoon_page_size_type page_size) in vmballoon_page_in_frames() argument
590 return 1 << vmballoon_page_order(page_size); in vmballoon_page_in_frames()
600 enum vmballoon_page_size_type page_size) in vmballoon_mark_page_offline() argument
604 for (i = 0; i < vmballoon_page_in_frames(page_size); i++) in vmballoon_mark_page_offline()
615 enum vmballoon_page_size_type page_size) in vmballoon_mark_page_online() argument
619 for (i = 0; i < vmballoon_page_in_frames(page_size); i++) in vmballoon_mark_page_online()
679 if (ctl->page_size == VMW_BALLOON_2M_PAGE) in vmballoon_alloc_page_list()
[all …]
/drivers/mtd/spi-nor/
Dxilinx.c33 offset = addr % nor->page_size; in s3an_convert_addr()
34 page = addr / nor->page_size; in s3an_convert_addr()
35 page <<= (nor->page_size > 512) ? 10 : 9; in s3an_convert_addr()
67 nor->page_size = (nor->page_size == 264) ? 256 : 512; in xilinx_nor_setup()
68 nor->mtd.writebufsize = nor->page_size; in xilinx_nor_setup()
69 nor->params->size = 8 * nor->page_size * nor->info->n_sectors; in xilinx_nor_setup()
71 nor->mtd.erasesize = 8 * nor->page_size; in xilinx_nor_setup()
Dcore.h270 u32 page_size; member
330 u16 page_size; member
404 .page_size = 256, \
419 .page_size = 256, \
425 .page_size = (_page_size), \
438 .page_size = _page_size, \
/drivers/net/ethernet/qlogic/qed/
Dqed_chain.c22 params->page_size); in qed_chain_init()
24 params->page_size, in qed_chain_init()
33 chain->page_size = params->page_size; in qed_chain_init()
88 dma_free_coherent(dev, chain->page_size, virt, phys); in qed_chain_free_next_ptr()
101 dma_free_coherent(&cdev->pdev->dev, chain->page_size, in qed_chain_free_single()
119 dma_free_coherent(dev, chain->page_size, entry->virt_addr, in qed_chain_free_pbl()
164 chain_size = ELEMS_PER_PAGE(params->elem_size, params->page_size); in qed_chain_alloc_sanity_check()
207 virt = dma_alloc_coherent(dev, chain->page_size, &phys, in qed_chain_alloc_next_ptr()
238 virt = dma_alloc_coherent(&cdev->pdev->dev, chain->page_size, in qed_chain_alloc_single()
290 virt = dma_alloc_coherent(dev, chain->page_size, &phys, in qed_chain_alloc_pbl()
[all …]
/drivers/misc/habanalabs/common/
Dmemory.c60 u32 num_curr_pgs, page_size; in alloc_device_memory() local
65 page_size = hdev->asic_prop.dram_page_size; in alloc_device_memory()
66 num_pgs = DIV_ROUND_UP_ULL(args->alloc.mem_size, page_size); in alloc_device_memory()
67 total_size = num_pgs * page_size; in alloc_device_memory()
95 phys_pg_pack->page_size = page_size; in alloc_device_memory()
108 phys_pg_pack->pages[i] = paddr + i * page_size; in alloc_device_memory()
113 page_size); in alloc_device_memory()
153 page_size); in alloc_device_memory()
302 phys_pg_pack->page_size); in free_phys_pg_pack()
309 phys_pg_pack->page_size); in free_phys_pg_pack()
[all …]
/drivers/infiniband/hw/mlx5/
Dmem.c43 void mlx5_ib_populate_pas(struct ib_umem *umem, size_t page_size, __be64 *pas, in mlx5_ib_populate_pas() argument
48 rdma_umem_for_each_dma_block (umem, &biter, page_size) { in mlx5_ib_populate_pas()
66 unsigned long page_size; in __mlx5_umem_find_best_quantized_pgoff() local
69 page_size = ib_umem_find_best_pgoff(umem, pgsz_bitmap, pgoff_bitmask); in __mlx5_umem_find_best_quantized_pgoff()
70 if (!page_size) in __mlx5_umem_find_best_quantized_pgoff()
80 page_offset = ib_umem_dma_offset(umem, page_size); in __mlx5_umem_find_best_quantized_pgoff()
81 while (page_offset & ~(u64)(page_offset_mask * (page_size / scale))) { in __mlx5_umem_find_best_quantized_pgoff()
82 page_size /= 2; in __mlx5_umem_find_best_quantized_pgoff()
83 page_offset = ib_umem_dma_offset(umem, page_size); in __mlx5_umem_find_best_quantized_pgoff()
90 if (!(pgsz_bitmap & page_size)) in __mlx5_umem_find_best_quantized_pgoff()
[all …]
/drivers/gpu/drm/i915/gem/selftests/
Dhuge_pages.c37 unsigned int page_size = page_sizes[i]; in get_largest_page_size() local
39 if (HAS_PAGE_SIZES(i915, page_size) && rem >= page_size) in get_largest_page_size()
40 return page_size; in get_largest_page_size()
89 unsigned int page_size = BIT(bit); in get_huge_pages() local
90 int order = get_order(page_size); in get_huge_pages()
100 sg_set_page(sg, page, page_size, 0); in get_huge_pages()
101 sg_page_sizes |= page_size; in get_huge_pages()
104 rem -= page_size; in get_huge_pages()
111 } while ((rem - ((page_size-1) & page_mask)) >= page_size); in get_huge_pages()
113 page_mask &= (page_size-1); in get_huge_pages()
[all …]
/drivers/iommu/amd/
Dio_pgtable.c50 static u64 *first_pte_l7(u64 *pte, unsigned long *page_size, in first_pte_l7() argument
61 if (page_size) in first_pte_l7()
62 *page_size = pg_size; in first_pte_l7()
223 unsigned long page_size, in alloc_pte() argument
231 BUG_ON(!is_power_of_2(page_size)); in alloc_pte()
245 address = PAGE_SIZE_ALIGN(address, page_size); in alloc_pte()
246 end_lvl = PAGE_SIZE_LEVEL(page_size); in alloc_pte()
318 unsigned long *page_size) in fetch_pte() argument
323 *page_size = 0; in fetch_pte()
330 *page_size = PTE_LEVEL_PAGE_SIZE(level); in fetch_pte()
[all …]
/drivers/mtd/devices/
Dmtd_dataflash.c91 unsigned int page_size; /* of bytes per page */ member
168 unsigned blocksize = priv->page_size << 3; in dataflash_erase()
175 div_u64_rem(instr->len, priv->page_size, &rem); in dataflash_erase()
178 div_u64_rem(instr->addr, priv->page_size, &rem); in dataflash_erase()
197 pageaddr = div_u64(instr->addr, priv->page_size); in dataflash_erase()
227 instr->addr += priv->page_size; in dataflash_erase()
228 instr->len -= priv->page_size; in dataflash_erase()
257 addr = (((unsigned)from / priv->page_size) << priv->page_offset) in dataflash_read()
258 + ((unsigned)from % priv->page_size); in dataflash_read()
329 pageaddr = ((unsigned)to / priv->page_size); in dataflash_write()
[all …]
Dspear_smi.c202 u32 page_size; member
680 u32 page_offset, page_size; in spear_mtd_write() local
695 page_offset = (u32)to % flash->page_size; in spear_mtd_write()
698 if (page_offset + len <= flash->page_size) { in spear_mtd_write()
706 page_size = flash->page_size - page_offset; in spear_mtd_write()
709 page_size); in spear_mtd_write()
713 *retlen += page_size; in spear_mtd_write()
716 for (i = page_size; i < len; i += page_size) { in spear_mtd_write()
717 page_size = len - i; in spear_mtd_write()
718 if (page_size > flash->page_size) in spear_mtd_write()
[all …]
Dmchp48l640.c29 unsigned int page_size; member
217 size_t page_sz = flash->caps->page_size; in mchp48l640_write()
275 size_t page_sz = flash->caps->page_size; in mchp48l640_read()
295 .page_size = 32,
331 flash->mtd.writesize = flash->caps->page_size; in mchp48l640_probe()
/drivers/misc/cxl/
Dcxllib.c203 unsigned long *page_size) in get_vma_info() argument
215 *page_size = vma_kernel_pagesize(vma); in get_vma_info()
227 unsigned long page_size; in cxllib_handle_fault() local
238 rc = get_vma_info(mm, addr, &vma_start, &vma_end, &page_size); in cxllib_handle_fault()
242 for (dar = (addr & ~(page_size - 1)); dar < (addr + size); in cxllib_handle_fault()
243 dar += page_size) { in cxllib_handle_fault()
259 &page_size); in cxllib_handle_fault()
/drivers/staging/rtl8723bs/hal/
Dhal_sdio.c62 u32 page_size; in rtw_hal_set_sdio_tx_max_length() local
65 rtw_hal_get_def_var(padapter, HAL_DEF_TX_PAGE_SIZE, &page_size); in rtw_hal_set_sdio_tx_max_length()
67 lenHQ = ((numHQ + numPubQ) >> 1) * page_size; in rtw_hal_set_sdio_tx_max_length()
68 lenNQ = ((numNQ + numPubQ) >> 1) * page_size; in rtw_hal_set_sdio_tx_max_length()
69 lenLQ = ((numLQ + numPubQ) >> 1) * page_size; in rtw_hal_set_sdio_tx_max_length()
/drivers/block/
Dps3vram.c63 unsigned int page_size; member
318 if (ps3vram_upload(dev, CACHE_OFFSET + entry * cache->page_size, in ps3vram_cache_evict()
320 cache->page_size / DMA_PAGE_SIZE) < 0) { in ps3vram_cache_evict()
323 entry * cache->page_size, cache->tags[entry].address, in ps3vram_cache_evict()
324 cache->page_size); in ps3vram_cache_evict()
337 CACHE_OFFSET + entry * cache->page_size, in ps3vram_cache_load()
339 cache->page_size / DMA_PAGE_SIZE) < 0) { in ps3vram_cache_load()
342 address, entry * cache->page_size, cache->page_size); in ps3vram_cache_load()
373 offset = (unsigned int) (address & (cache->page_size - 1)); in ps3vram_cache_match()
403 priv->cache.page_size = CACHE_PAGE_SIZE; in ps3vram_cache_init()
[all …]
/drivers/gpu/drm/i915/gem/
Di915_gem_region.c35 resource_size_t page_size, in i915_gem_object_create_region() argument
54 if (page_size) in i915_gem_object_create_region()
55 default_page_size = page_size; in i915_gem_object_create_region()
72 err = mem->ops->init_object(mem, obj, size, page_size, flags); in i915_gem_object_create_region()
/drivers/memstick/core/
Dms_block.c435 msb->page_size); in h_msb_read_page()
443 msb->current_sg_offset += msb->page_size; in h_msb_read_page()
454 msb->current_sg_offset += msb->page_size; in h_msb_read_page()
561 msb->page_size) < msb->page_size) in h_msb_write_block()
571 msb->current_sg_offset += msb->page_size; in h_msb_write_block()
867 size_t len = msb->page_size; in msb_read_page()
987 NULL, &sg, page * msb->page_size); in msb_verify_block()
1005 BUG_ON(sg->length < msb->page_size); in msb_write_block()
1182 p->attr.page_size = be16_to_cpu(p->attr.page_size); in msb_fix_boot_page_endianness()
1278 page = data_offset / msb->page_size; in msb_read_bad_block_table()
[all …]
Dmspro_block.c74 __be16 page_size; member
144 unsigned short page_size; member
341 be16_to_cpu(x_sys->page_size)); in mspro_block_attr_show_sysinfo()
608 / msb->page_size)) { in h_mspro_block_transfer_data()
632 t_offset += msb->current_page * msb->page_size; in h_mspro_block_transfer_data()
637 msb->page_size, offset_in_page(t_offset)); in h_mspro_block_transfer_data()
676 .data_count = cpu_to_be16((uint16_t)(length / msb->page_size)), in h_mspro_block_setup_cmd()
682 do_div(offset, msb->page_size); in h_mspro_block_setup_cmd()
758 / msb->page_size; in mspro_block_complete_req()
763 t_len *= msb->page_size; in mspro_block_complete_req()
[all …]
Dms_block.h79 unsigned short page_size; member
112 unsigned short page_size; member
170 unsigned short page_size; member
/drivers/net/ethernet/huawei/hinic/
Dhinic_hw_eqs.c62 #define PAGE_IN_4K(page_size) ((page_size) >> 12) argument
63 #define EQ_SET_HW_PAGE_SIZE_VAL(eq) (ilog2(PAGE_IN_4K((eq)->page_size)))
650 eq->page_size, in alloc_eq_pages()
680 dma_free_coherent(&pdev->dev, eq->page_size, in alloc_eq_pages()
702 dma_free_coherent(&pdev->dev, eq->page_size, in free_eq_pages()
723 enum hinic_eq_type type, int q_id, u32 q_len, u32 page_size, in init_eq() argument
733 eq->page_size = page_size; in init_eq()
751 eq->num_pages = GET_EQ_NUM_PAGES(eq, page_size); in init_eq()
752 eq->num_elem_in_pg = GET_EQ_NUM_ELEMS_IN_PG(eq, page_size); in init_eq()
861 int num_aeqs, u32 q_len, u32 page_size, in hinic_aeqs_init() argument
[all …]
/drivers/misc/eeprom/
Dat24.c88 u16 page_size; member
379 next_page = roundup(offset + 1, at24->page_size); in at24_adjust_write_count()
592 u32 byte_len, page_size, flags, addrw; in at24_probe() local
611 err = device_property_read_u32(dev, "pagesize", &page_size); in at24_probe()
618 page_size = 1; in at24_probe()
649 page_size = 1; in at24_probe()
651 if (!page_size) { in at24_probe()
656 if (!is_power_of_2(page_size)) in at24_probe()
689 at24->page_size = page_size; in at24_probe()
704 page_size, at24_io_limit); in at24_probe()
/drivers/infiniband/hw/hns/
Dhns_roce_alloc.c70 u32 trunk_size, page_size, alloced_size; in hns_roce_buf_alloc() local
86 page_size = 1 << buf->page_shift; in hns_roce_buf_alloc()
93 buf->trunk_shift = order_base_2(ALIGN(page_size, PAGE_SIZE)); in hns_roce_buf_alloc()
127 buf->npages = DIV_ROUND_UP(alloced_size, page_size); in hns_roce_buf_alloc()
/drivers/mtd/nand/raw/
Dsh_flctl.c234 if (flctl->page_size) { in set_addr()
346 if (flctl->page_size) in wait_recfifo_ready()
546 if (flctl->page_size) in set_cmd_regs()
628 page_sectors = flctl->page_size ? 4 : 1; in execmd_read_page_sector()
674 int page_sectors = flctl->page_size ? 4 : 1; in execmd_read_oob()
698 page_sectors = flctl->page_size ? 4 : 1; in execmd_write_page_sector()
724 page_sectors = flctl->page_size ? 4 : 1; in execmd_write_oob()
761 if (flctl->page_size) in flctl_cmdfunc()
782 if (flctl->page_size) { in flctl_cmdfunc()
797 if (flctl->page_size) in flctl_cmdfunc()
[all …]
/drivers/gpu/drm/i915/gt/
Dgen8_ppgtt.c478 unsigned int page_size; in gen8_ppgtt_insert_huge() local
488 page_size = I915_GTT_PAGE_SIZE_2M; in gen8_ppgtt_insert_huge()
496 page_size = I915_GTT_PAGE_SIZE; in gen8_ppgtt_insert_huge()
509 GEM_BUG_ON(sg_dma_len(iter->sg) < page_size); in gen8_ppgtt_insert_huge()
512 start += page_size; in gen8_ppgtt_insert_huge()
513 iter->dma += page_size; in gen8_ppgtt_insert_huge()
514 rem -= page_size; in gen8_ppgtt_insert_huge()
533 if (unlikely(!IS_ALIGNED(iter->dma, page_size))) in gen8_ppgtt_insert_huge()
536 } while (rem >= page_size && index < I915_PDES); in gen8_ppgtt_insert_huge()
554 page_size = I915_GTT_PAGE_SIZE_64K; in gen8_ppgtt_insert_huge()
[all …]

12345678