/drivers/misc/habanalabs/common/mmu/ |
D | mmu.c | 16 return hl_mem_area_inside_range(virt_addr, prop->dmmu.page_size, in hl_is_dram_va() 145 int hl_mmu_unmap_page(struct hl_ctx *ctx, u64 virt_addr, u32 page_size, in hl_mmu_unmap_page() argument 163 else if ((page_size % prop->pmmu_huge.page_size) == 0) in hl_mmu_unmap_page() 173 if ((page_size % mmu_prop->page_size) == 0) { in hl_mmu_unmap_page() 174 real_page_size = mmu_prop->page_size; in hl_mmu_unmap_page() 182 if (is_dram_addr && ((page_size % prop->dram_page_size) == 0)) { in hl_mmu_unmap_page() 187 page_size, mmu_prop->page_size >> 10); in hl_mmu_unmap_page() 193 npages = page_size / real_page_size; in hl_mmu_unmap_page() 234 u32 page_size, bool flush_pte) in hl_mmu_map_page() argument 252 else if ((page_size % prop->pmmu_huge.page_size) == 0) in hl_mmu_map_page() [all …]
|
/drivers/pci/endpoint/ |
D | pci-epc-mem.c | 26 unsigned int page_shift = ilog2(mem->window.page_size); in pci_epc_mem_get_order() 54 size_t page_size; in pci_epc_multi_mem_init() local 70 page_size = windows[i].page_size; in pci_epc_multi_mem_init() 71 if (page_size < PAGE_SIZE) in pci_epc_multi_mem_init() 72 page_size = PAGE_SIZE; in pci_epc_multi_mem_init() 73 page_shift = ilog2(page_size); in pci_epc_multi_mem_init() 94 mem->window.page_size = page_size; in pci_epc_multi_mem_init() 119 size_t size, size_t page_size) in pci_epc_mem_init() argument 125 mem_window.page_size = page_size; in pci_epc_mem_init() 182 align_size = ALIGN(size, mem->window.page_size); in pci_epc_mem_alloc_addr() [all …]
|
/drivers/misc/ |
D | vmw_balloon.c | 247 enum vmballoon_page_size_type page_size; member 576 unsigned int vmballoon_page_order(enum vmballoon_page_size_type page_size) in vmballoon_page_order() argument 578 return page_size == VMW_BALLOON_2M_PAGE ? VMW_BALLOON_2M_ORDER : 0; in vmballoon_page_order() 588 vmballoon_page_in_frames(enum vmballoon_page_size_type page_size) in vmballoon_page_in_frames() argument 590 return 1 << vmballoon_page_order(page_size); in vmballoon_page_in_frames() 600 enum vmballoon_page_size_type page_size) in vmballoon_mark_page_offline() argument 604 for (i = 0; i < vmballoon_page_in_frames(page_size); i++) in vmballoon_mark_page_offline() 615 enum vmballoon_page_size_type page_size) in vmballoon_mark_page_online() argument 619 for (i = 0; i < vmballoon_page_in_frames(page_size); i++) in vmballoon_mark_page_online() 679 if (ctl->page_size == VMW_BALLOON_2M_PAGE) in vmballoon_alloc_page_list() [all …]
|
/drivers/mtd/spi-nor/ |
D | xilinx.c | 33 offset = addr % nor->page_size; in s3an_convert_addr() 34 page = addr / nor->page_size; in s3an_convert_addr() 35 page <<= (nor->page_size > 512) ? 10 : 9; in s3an_convert_addr() 67 nor->page_size = (nor->page_size == 264) ? 256 : 512; in xilinx_nor_setup() 68 nor->mtd.writebufsize = nor->page_size; in xilinx_nor_setup() 69 nor->params->size = 8 * nor->page_size * nor->info->n_sectors; in xilinx_nor_setup() 71 nor->mtd.erasesize = 8 * nor->page_size; in xilinx_nor_setup()
|
D | core.h | 270 u32 page_size; member 330 u16 page_size; member 404 .page_size = 256, \ 419 .page_size = 256, \ 425 .page_size = (_page_size), \ 438 .page_size = _page_size, \
|
/drivers/net/ethernet/qlogic/qed/ |
D | qed_chain.c | 22 params->page_size); in qed_chain_init() 24 params->page_size, in qed_chain_init() 33 chain->page_size = params->page_size; in qed_chain_init() 88 dma_free_coherent(dev, chain->page_size, virt, phys); in qed_chain_free_next_ptr() 101 dma_free_coherent(&cdev->pdev->dev, chain->page_size, in qed_chain_free_single() 119 dma_free_coherent(dev, chain->page_size, entry->virt_addr, in qed_chain_free_pbl() 164 chain_size = ELEMS_PER_PAGE(params->elem_size, params->page_size); in qed_chain_alloc_sanity_check() 207 virt = dma_alloc_coherent(dev, chain->page_size, &phys, in qed_chain_alloc_next_ptr() 238 virt = dma_alloc_coherent(&cdev->pdev->dev, chain->page_size, in qed_chain_alloc_single() 290 virt = dma_alloc_coherent(dev, chain->page_size, &phys, in qed_chain_alloc_pbl() [all …]
|
/drivers/misc/habanalabs/common/ |
D | memory.c | 60 u32 num_curr_pgs, page_size; in alloc_device_memory() local 65 page_size = hdev->asic_prop.dram_page_size; in alloc_device_memory() 66 num_pgs = DIV_ROUND_UP_ULL(args->alloc.mem_size, page_size); in alloc_device_memory() 67 total_size = num_pgs * page_size; in alloc_device_memory() 95 phys_pg_pack->page_size = page_size; in alloc_device_memory() 108 phys_pg_pack->pages[i] = paddr + i * page_size; in alloc_device_memory() 113 page_size); in alloc_device_memory() 153 page_size); in alloc_device_memory() 302 phys_pg_pack->page_size); in free_phys_pg_pack() 309 phys_pg_pack->page_size); in free_phys_pg_pack() [all …]
|
/drivers/infiniband/hw/mlx5/ |
D | mem.c | 43 void mlx5_ib_populate_pas(struct ib_umem *umem, size_t page_size, __be64 *pas, in mlx5_ib_populate_pas() argument 48 rdma_umem_for_each_dma_block (umem, &biter, page_size) { in mlx5_ib_populate_pas() 66 unsigned long page_size; in __mlx5_umem_find_best_quantized_pgoff() local 69 page_size = ib_umem_find_best_pgoff(umem, pgsz_bitmap, pgoff_bitmask); in __mlx5_umem_find_best_quantized_pgoff() 70 if (!page_size) in __mlx5_umem_find_best_quantized_pgoff() 80 page_offset = ib_umem_dma_offset(umem, page_size); in __mlx5_umem_find_best_quantized_pgoff() 81 while (page_offset & ~(u64)(page_offset_mask * (page_size / scale))) { in __mlx5_umem_find_best_quantized_pgoff() 82 page_size /= 2; in __mlx5_umem_find_best_quantized_pgoff() 83 page_offset = ib_umem_dma_offset(umem, page_size); in __mlx5_umem_find_best_quantized_pgoff() 90 if (!(pgsz_bitmap & page_size)) in __mlx5_umem_find_best_quantized_pgoff() [all …]
|
/drivers/gpu/drm/i915/gem/selftests/ |
D | huge_pages.c | 37 unsigned int page_size = page_sizes[i]; in get_largest_page_size() local 39 if (HAS_PAGE_SIZES(i915, page_size) && rem >= page_size) in get_largest_page_size() 40 return page_size; in get_largest_page_size() 89 unsigned int page_size = BIT(bit); in get_huge_pages() local 90 int order = get_order(page_size); in get_huge_pages() 100 sg_set_page(sg, page, page_size, 0); in get_huge_pages() 101 sg_page_sizes |= page_size; in get_huge_pages() 104 rem -= page_size; in get_huge_pages() 111 } while ((rem - ((page_size-1) & page_mask)) >= page_size); in get_huge_pages() 113 page_mask &= (page_size-1); in get_huge_pages() [all …]
|
/drivers/iommu/amd/ |
D | io_pgtable.c | 50 static u64 *first_pte_l7(u64 *pte, unsigned long *page_size, in first_pte_l7() argument 61 if (page_size) in first_pte_l7() 62 *page_size = pg_size; in first_pte_l7() 223 unsigned long page_size, in alloc_pte() argument 231 BUG_ON(!is_power_of_2(page_size)); in alloc_pte() 245 address = PAGE_SIZE_ALIGN(address, page_size); in alloc_pte() 246 end_lvl = PAGE_SIZE_LEVEL(page_size); in alloc_pte() 318 unsigned long *page_size) in fetch_pte() argument 323 *page_size = 0; in fetch_pte() 330 *page_size = PTE_LEVEL_PAGE_SIZE(level); in fetch_pte() [all …]
|
/drivers/mtd/devices/ |
D | mtd_dataflash.c | 91 unsigned int page_size; /* of bytes per page */ member 168 unsigned blocksize = priv->page_size << 3; in dataflash_erase() 175 div_u64_rem(instr->len, priv->page_size, &rem); in dataflash_erase() 178 div_u64_rem(instr->addr, priv->page_size, &rem); in dataflash_erase() 197 pageaddr = div_u64(instr->addr, priv->page_size); in dataflash_erase() 227 instr->addr += priv->page_size; in dataflash_erase() 228 instr->len -= priv->page_size; in dataflash_erase() 257 addr = (((unsigned)from / priv->page_size) << priv->page_offset) in dataflash_read() 258 + ((unsigned)from % priv->page_size); in dataflash_read() 329 pageaddr = ((unsigned)to / priv->page_size); in dataflash_write() [all …]
|
D | spear_smi.c | 202 u32 page_size; member 680 u32 page_offset, page_size; in spear_mtd_write() local 695 page_offset = (u32)to % flash->page_size; in spear_mtd_write() 698 if (page_offset + len <= flash->page_size) { in spear_mtd_write() 706 page_size = flash->page_size - page_offset; in spear_mtd_write() 709 page_size); in spear_mtd_write() 713 *retlen += page_size; in spear_mtd_write() 716 for (i = page_size; i < len; i += page_size) { in spear_mtd_write() 717 page_size = len - i; in spear_mtd_write() 718 if (page_size > flash->page_size) in spear_mtd_write() [all …]
|
D | mchp48l640.c | 29 unsigned int page_size; member 217 size_t page_sz = flash->caps->page_size; in mchp48l640_write() 275 size_t page_sz = flash->caps->page_size; in mchp48l640_read() 295 .page_size = 32, 331 flash->mtd.writesize = flash->caps->page_size; in mchp48l640_probe()
|
/drivers/misc/cxl/ |
D | cxllib.c | 203 unsigned long *page_size) in get_vma_info() argument 215 *page_size = vma_kernel_pagesize(vma); in get_vma_info() 227 unsigned long page_size; in cxllib_handle_fault() local 238 rc = get_vma_info(mm, addr, &vma_start, &vma_end, &page_size); in cxllib_handle_fault() 242 for (dar = (addr & ~(page_size - 1)); dar < (addr + size); in cxllib_handle_fault() 243 dar += page_size) { in cxllib_handle_fault() 259 &page_size); in cxllib_handle_fault()
|
/drivers/staging/rtl8723bs/hal/ |
D | hal_sdio.c | 62 u32 page_size; in rtw_hal_set_sdio_tx_max_length() local 65 rtw_hal_get_def_var(padapter, HAL_DEF_TX_PAGE_SIZE, &page_size); in rtw_hal_set_sdio_tx_max_length() 67 lenHQ = ((numHQ + numPubQ) >> 1) * page_size; in rtw_hal_set_sdio_tx_max_length() 68 lenNQ = ((numNQ + numPubQ) >> 1) * page_size; in rtw_hal_set_sdio_tx_max_length() 69 lenLQ = ((numLQ + numPubQ) >> 1) * page_size; in rtw_hal_set_sdio_tx_max_length()
|
/drivers/block/ |
D | ps3vram.c | 63 unsigned int page_size; member 318 if (ps3vram_upload(dev, CACHE_OFFSET + entry * cache->page_size, in ps3vram_cache_evict() 320 cache->page_size / DMA_PAGE_SIZE) < 0) { in ps3vram_cache_evict() 323 entry * cache->page_size, cache->tags[entry].address, in ps3vram_cache_evict() 324 cache->page_size); in ps3vram_cache_evict() 337 CACHE_OFFSET + entry * cache->page_size, in ps3vram_cache_load() 339 cache->page_size / DMA_PAGE_SIZE) < 0) { in ps3vram_cache_load() 342 address, entry * cache->page_size, cache->page_size); in ps3vram_cache_load() 373 offset = (unsigned int) (address & (cache->page_size - 1)); in ps3vram_cache_match() 403 priv->cache.page_size = CACHE_PAGE_SIZE; in ps3vram_cache_init() [all …]
|
/drivers/gpu/drm/i915/gem/ |
D | i915_gem_region.c | 35 resource_size_t page_size, in i915_gem_object_create_region() argument 54 if (page_size) in i915_gem_object_create_region() 55 default_page_size = page_size; in i915_gem_object_create_region() 72 err = mem->ops->init_object(mem, obj, size, page_size, flags); in i915_gem_object_create_region()
|
/drivers/memstick/core/ |
D | ms_block.c | 435 msb->page_size); in h_msb_read_page() 443 msb->current_sg_offset += msb->page_size; in h_msb_read_page() 454 msb->current_sg_offset += msb->page_size; in h_msb_read_page() 561 msb->page_size) < msb->page_size) in h_msb_write_block() 571 msb->current_sg_offset += msb->page_size; in h_msb_write_block() 867 size_t len = msb->page_size; in msb_read_page() 987 NULL, &sg, page * msb->page_size); in msb_verify_block() 1005 BUG_ON(sg->length < msb->page_size); in msb_write_block() 1182 p->attr.page_size = be16_to_cpu(p->attr.page_size); in msb_fix_boot_page_endianness() 1278 page = data_offset / msb->page_size; in msb_read_bad_block_table() [all …]
|
D | mspro_block.c | 74 __be16 page_size; member 144 unsigned short page_size; member 341 be16_to_cpu(x_sys->page_size)); in mspro_block_attr_show_sysinfo() 608 / msb->page_size)) { in h_mspro_block_transfer_data() 632 t_offset += msb->current_page * msb->page_size; in h_mspro_block_transfer_data() 637 msb->page_size, offset_in_page(t_offset)); in h_mspro_block_transfer_data() 676 .data_count = cpu_to_be16((uint16_t)(length / msb->page_size)), in h_mspro_block_setup_cmd() 682 do_div(offset, msb->page_size); in h_mspro_block_setup_cmd() 758 / msb->page_size; in mspro_block_complete_req() 763 t_len *= msb->page_size; in mspro_block_complete_req() [all …]
|
D | ms_block.h | 79 unsigned short page_size; member 112 unsigned short page_size; member 170 unsigned short page_size; member
|
/drivers/net/ethernet/huawei/hinic/ |
D | hinic_hw_eqs.c | 62 #define PAGE_IN_4K(page_size) ((page_size) >> 12) argument 63 #define EQ_SET_HW_PAGE_SIZE_VAL(eq) (ilog2(PAGE_IN_4K((eq)->page_size))) 650 eq->page_size, in alloc_eq_pages() 680 dma_free_coherent(&pdev->dev, eq->page_size, in alloc_eq_pages() 702 dma_free_coherent(&pdev->dev, eq->page_size, in free_eq_pages() 723 enum hinic_eq_type type, int q_id, u32 q_len, u32 page_size, in init_eq() argument 733 eq->page_size = page_size; in init_eq() 751 eq->num_pages = GET_EQ_NUM_PAGES(eq, page_size); in init_eq() 752 eq->num_elem_in_pg = GET_EQ_NUM_ELEMS_IN_PG(eq, page_size); in init_eq() 861 int num_aeqs, u32 q_len, u32 page_size, in hinic_aeqs_init() argument [all …]
|
/drivers/misc/eeprom/ |
D | at24.c | 88 u16 page_size; member 379 next_page = roundup(offset + 1, at24->page_size); in at24_adjust_write_count() 592 u32 byte_len, page_size, flags, addrw; in at24_probe() local 611 err = device_property_read_u32(dev, "pagesize", &page_size); in at24_probe() 618 page_size = 1; in at24_probe() 649 page_size = 1; in at24_probe() 651 if (!page_size) { in at24_probe() 656 if (!is_power_of_2(page_size)) in at24_probe() 689 at24->page_size = page_size; in at24_probe() 704 page_size, at24_io_limit); in at24_probe()
|
/drivers/infiniband/hw/hns/ |
D | hns_roce_alloc.c | 70 u32 trunk_size, page_size, alloced_size; in hns_roce_buf_alloc() local 86 page_size = 1 << buf->page_shift; in hns_roce_buf_alloc() 93 buf->trunk_shift = order_base_2(ALIGN(page_size, PAGE_SIZE)); in hns_roce_buf_alloc() 127 buf->npages = DIV_ROUND_UP(alloced_size, page_size); in hns_roce_buf_alloc()
|
/drivers/mtd/nand/raw/ |
D | sh_flctl.c | 234 if (flctl->page_size) { in set_addr() 346 if (flctl->page_size) in wait_recfifo_ready() 546 if (flctl->page_size) in set_cmd_regs() 628 page_sectors = flctl->page_size ? 4 : 1; in execmd_read_page_sector() 674 int page_sectors = flctl->page_size ? 4 : 1; in execmd_read_oob() 698 page_sectors = flctl->page_size ? 4 : 1; in execmd_write_page_sector() 724 page_sectors = flctl->page_size ? 4 : 1; in execmd_write_oob() 761 if (flctl->page_size) in flctl_cmdfunc() 782 if (flctl->page_size) { in flctl_cmdfunc() 797 if (flctl->page_size) in flctl_cmdfunc() [all …]
|
/drivers/gpu/drm/i915/gt/ |
D | gen8_ppgtt.c | 478 unsigned int page_size; in gen8_ppgtt_insert_huge() local 488 page_size = I915_GTT_PAGE_SIZE_2M; in gen8_ppgtt_insert_huge() 496 page_size = I915_GTT_PAGE_SIZE; in gen8_ppgtt_insert_huge() 509 GEM_BUG_ON(sg_dma_len(iter->sg) < page_size); in gen8_ppgtt_insert_huge() 512 start += page_size; in gen8_ppgtt_insert_huge() 513 iter->dma += page_size; in gen8_ppgtt_insert_huge() 514 rem -= page_size; in gen8_ppgtt_insert_huge() 533 if (unlikely(!IS_ALIGNED(iter->dma, page_size))) in gen8_ppgtt_insert_huge() 536 } while (rem >= page_size && index < I915_PDES); in gen8_ppgtt_insert_huge() 554 page_size = I915_GTT_PAGE_SIZE_64K; in gen8_ppgtt_insert_huge() [all …]
|