/kernel/linux/linux-5.10/arch/microblaze/mm/ |
D | pgtable.c | 60 p = addr & PAGE_MASK; in __ioremap() 117 return (void __iomem *) (v + ((unsigned long)addr & ~PAGE_MASK)); in __ioremap() 130 vfree((void *) (PAGE_MASK & (unsigned long) addr)); in iounmap() 202 pgd = pgd_offset(mm, addr & PAGE_MASK); in get_pteptr() 204 p4d = p4d_offset(pgd, addr & PAGE_MASK); in get_pteptr() 205 pud = pud_offset(p4d, addr & PAGE_MASK); in get_pteptr() 206 pmd = pmd_offset(pud, addr & PAGE_MASK); in get_pteptr() 208 pte = pte_offset_kernel(pmd, addr & PAGE_MASK); in get_pteptr() 238 pa = (pte_val(*pte) & PAGE_MASK) | (addr & ~PAGE_MASK); in iopa()
|
/kernel/linux/linux-5.10/arch/m68k/mm/ |
D | cache.c | 50 return (mmusr & PAGE_MASK) | (vaddr & ~PAGE_MASK); in virt_to_phys_slow() 68 return (*descaddr & PAGE_MASK) | (vaddr & ~PAGE_MASK); in virt_to_phys_slow() 88 address &= PAGE_MASK; in flush_icache_user_range()
|
D | memory.c | 111 pushcl040(paddr & PAGE_MASK); in cache_clear() 117 paddr &= PAGE_MASK; in cache_clear() 165 paddr &= PAGE_MASK; in cache_push()
|
/kernel/linux/linux-5.10/arch/m68k/kernel/ |
D | sys_m68k.c | 63 _paddr = (_mmusr & MMU_R_040) ? (_mmusr & PAGE_MASK) : 0; \ 104 paddr += addr & ~(PAGE_MASK | 15); in cache_flush_040() 107 unsigned long tmp = PAGE_SIZE - (addr & ~PAGE_MASK); in cache_flush_040() 125 i = (PAGE_SIZE - (paddr & ~PAGE_MASK)) >> 4; in cache_flush_040() 180 len += (addr & ~PAGE_MASK) + (PAGE_SIZE - 1); in cache_flush_040() 268 unsigned long tmp = PAGE_SIZE - (addr & ~PAGE_MASK); in cache_flush_060() 286 i = (PAGE_SIZE - (paddr & ~PAGE_MASK)) >> 4; in cache_flush_060() 319 addr &= PAGE_MASK; in cache_flush_060() 341 len += (addr & ~PAGE_MASK) + (PAGE_SIZE - 1); in cache_flush_060() 342 addr &= PAGE_MASK; /* Workaround for bug in some in cache_flush_060()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/subdev/devinit/ |
D | fbmem.h | 63 u8 __iomem *p = io_mapping_map_atomic_wc(fb, off & PAGE_MASK); in fbmem_peek() 64 u32 val = ioread32(p + (off & ~PAGE_MASK)); in fbmem_peek() 72 u8 __iomem *p = io_mapping_map_atomic_wc(fb, off & PAGE_MASK); in fbmem_poke() 73 iowrite32(val, p + (off & ~PAGE_MASK)); in fbmem_poke()
|
/kernel/linux/linux-5.10/arch/sh/mm/ |
D | tlbflush_32.c | 25 page &= PAGE_MASK; in local_flush_tlb_page() 60 start &= PAGE_MASK; in local_flush_tlb_range() 62 end &= PAGE_MASK; in local_flush_tlb_range() 93 start &= PAGE_MASK; in local_flush_tlb_kernel_range() 95 end &= PAGE_MASK; in local_flush_tlb_kernel_range()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx5/ |
D | doorbell.c | 56 if (page->user_virt == (virt & PAGE_MASK)) in mlx5_ib_db_map_user() 65 page->user_virt = (virt & PAGE_MASK); in mlx5_ib_db_map_user() 67 page->umem = ib_umem_get(context->ibucontext.device, virt & PAGE_MASK, in mlx5_ib_db_map_user() 78 db->dma = sg_dma_address(page->umem->sg_head.sgl) + (virt & ~PAGE_MASK); in mlx5_ib_db_map_user()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx4/ |
D | doorbell.c | 56 if (page->user_virt == (virt & PAGE_MASK)) in mlx4_ib_db_map_user() 65 page->user_virt = (virt & PAGE_MASK); in mlx4_ib_db_map_user() 67 page->umem = ib_umem_get(context->ibucontext.device, virt & PAGE_MASK, in mlx4_ib_db_map_user() 78 db->dma = sg_dma_address(page->umem->sg_head.sgl) + (virt & ~PAGE_MASK); in mlx4_ib_db_map_user()
|
/kernel/linux/linux-5.10/arch/microblaze/pci/ |
D | indirect_pci.c | 148 resource_size_t base = cfg_addr & PAGE_MASK; in setup_indirect_pci() 152 hose->cfg_addr = mbase + (cfg_addr & ~PAGE_MASK); in setup_indirect_pci() 153 if ((cfg_data & PAGE_MASK) != base) in setup_indirect_pci() 154 mbase = ioremap(cfg_data & PAGE_MASK, PAGE_SIZE); in setup_indirect_pci() 155 hose->cfg_data = mbase + (cfg_data & ~PAGE_MASK); in setup_indirect_pci()
|
/kernel/linux/linux-5.10/arch/ia64/mm/ |
D | ioremap.c | 69 page_base = phys_addr & PAGE_MASK; in ioremap() 77 offset = phys_addr & ~PAGE_MASK; in ioremap() 78 phys_addr &= PAGE_MASK; in ioremap() 121 vunmap((void *) ((unsigned long) addr & PAGE_MASK)); in iounmap()
|
/kernel/linux/linux-5.10/arch/mips/mm/ |
D | tlb-r3k.c | 89 start &= PAGE_MASK; in local_flush_tlb_range() 91 end &= PAGE_MASK; in local_flush_tlb_range() 125 start &= PAGE_MASK; in local_flush_tlb_kernel_range() 127 end &= PAGE_MASK; in local_flush_tlb_kernel_range() 162 page &= PAGE_MASK; in local_flush_tlb_page() 203 address &= PAGE_MASK; in __update_tlb()
|
D | ioremap.c | 91 offset = phys_addr & ~PAGE_MASK; in ioremap_prot() 92 phys_addr &= PAGE_MASK; in ioremap_prot() 117 vunmap((void *)((unsigned long)addr & PAGE_MASK)); in iounmap()
|
/kernel/linux/linux-5.10/arch/mips/lib/ |
D | r3k_dump_tlb.c | 48 if ((entryhi & PAGE_MASK) != KSEG0 && in dump_tlb() 58 entryhi & PAGE_MASK, in dump_tlb() 60 entrylo0 & PAGE_MASK, in dump_tlb()
|
/kernel/linux/linux-5.10/arch/sparc/mm/ |
D | iommu.c | 164 start &= PAGE_MASK; in iommu_flush_iotlb() 188 unsigned long off = paddr & ~PAGE_MASK; in __sbus_iommu_map_page() 208 for (p = vaddr & PAGE_MASK; p < vaddr + len; p += PAGE_SIZE) in __sbus_iommu_map_page() 283 unsigned int busa = dma_addr & PAGE_MASK; in sbus_iommu_unmap_page() 284 unsigned long off = dma_addr & ~PAGE_MASK; in sbus_iommu_unmap_page() 334 BUG_ON((va & ~PAGE_MASK) != 0); in sbus_iommu_alloc() 335 BUG_ON((addr & ~PAGE_MASK) != 0); in sbus_iommu_alloc() 336 BUG_ON((len & ~PAGE_MASK) != 0); in sbus_iommu_alloc() 406 BUG_ON((busa & ~PAGE_MASK) != 0); in sbus_iommu_free() 407 BUG_ON((len & ~PAGE_MASK) != 0); in sbus_iommu_free()
|
/kernel/linux/linux-5.10/arch/powerpc/mm/ |
D | ioremap_64.c | 24 paligned = addr & PAGE_MASK; in __ioremap_caller() 25 offset = addr & ~PAGE_MASK; in __ioremap_caller() 57 addr = (void *)((unsigned long __force)PCI_FIX_ADDR(token) & PAGE_MASK); in iounmap()
|
D | ioremap_32.c | 30 p = addr & PAGE_MASK; in __ioremap_caller() 31 offset = addr & ~PAGE_MASK; in __ioremap_caller() 91 vunmap((void *)(PAGE_MASK & (unsigned long)addr)); in iounmap()
|
/kernel/linux/linux-5.10/arch/powerpc/sysdev/ |
D | indirect_pci.c | 163 resource_size_t base = cfg_addr & PAGE_MASK; in setup_indirect_pci() 167 hose->cfg_addr = mbase + (cfg_addr & ~PAGE_MASK); in setup_indirect_pci() 168 if ((cfg_data & PAGE_MASK) != base) in setup_indirect_pci() 169 mbase = ioremap(cfg_data & PAGE_MASK, PAGE_SIZE); in setup_indirect_pci() 170 hose->cfg_data = mbase + (cfg_data & ~PAGE_MASK); in setup_indirect_pci()
|
/kernel/linux/linux-5.10/arch/xtensa/mm/ |
D | ioremap.c | 17 unsigned long offset = paddr & ~PAGE_MASK; in xtensa_ioremap() 23 paddr &= PAGE_MASK; in xtensa_ioremap() 61 void *addr = (void *)(PAGE_MASK & (unsigned long)io_addr); in xtensa_iounmap()
|
/kernel/linux/linux-5.10/arch/parisc/mm/ |
D | ioremap.c | 71 offset = phys_addr & ~PAGE_MASK; in ioremap() 72 phys_addr &= PAGE_MASK; in ioremap() 95 unsigned long addr = (unsigned long)io_addr & PAGE_MASK; in iounmap()
|
/kernel/linux/linux-5.10/arch/openrisc/mm/ |
D | ioremap.c | 52 offset = addr & ~PAGE_MASK; in ioremap() 53 p = addr & PAGE_MASK; in ioremap() 104 return vfree((void *)(PAGE_MASK & (unsigned long)addr)); in iounmap()
|
/kernel/linux/linux-5.10/arch/arm64/mm/ |
D | ioremap.c | 25 unsigned long offset = phys_addr & ~PAGE_MASK; in __ioremap_caller() 34 phys_addr &= PAGE_MASK; in __ioremap_caller() 74 unsigned long addr = (unsigned long)io_addr & PAGE_MASK; in iounmap()
|
/kernel/linux/linux-5.10/arch/powerpc/kernel/ |
D | isa-bridge.c | 44 WARN_ON_ONCE(ISA_IO_BASE & ~PAGE_MASK); in remap_isa_base() 45 WARN_ON_ONCE(pa & ~PAGE_MASK); in remap_isa_base() 46 WARN_ON_ONCE(size & ~PAGE_MASK); in remap_isa_base() 253 if ((cbase & ~PAGE_MASK) || (pbase & ~PAGE_MASK)) { in isa_bridge_init_non_pci()
|
/kernel/linux/linux-5.10/arch/powerpc/include/asm/book3s/32/ |
D | hash.h | 41 #define _PMD_PRESENT_MASK (PAGE_MASK) 42 #define _PMD_BAD (~PAGE_MASK)
|
/kernel/linux/linux-5.10/net/ceph/ |
D | pagevec.c | 66 int po = off & ~PAGE_MASK; in ceph_copy_user_to_page_vector() 92 size_t po = off & ~PAGE_MASK; in ceph_copy_to_page_vector() 115 size_t po = off & ~PAGE_MASK; in ceph_copy_from_page_vector() 141 off &= ~PAGE_MASK; in ceph_zero_page_vector_range()
|
/kernel/linux/linux-5.10/arch/powerpc/mm/book3s32/ |
D | tlb.c | 40 ptephys = __pa(ptep) & PAGE_MASK; in flush_hash_entry() 82 start &= PAGE_MASK; in flush_range() 92 end = (end - 1) | ~PAGE_MASK; in flush_range()
|