/arch/avr32/mm/ |
D | dma-coherent.c | 38 static struct page *__dma_alloc(struct device *dev, size_t size, in __dma_alloc() 41 struct page *page, *free, *end; in __dma_alloc() local 54 page = alloc_pages(gfp, order); in __dma_alloc() 55 if (!page) in __dma_alloc() 57 split_page(page, order); in __dma_alloc() 68 invalidate_dcache_region(phys_to_virt(page_to_phys(page)), size); in __dma_alloc() 70 *handle = page_to_bus(page); in __dma_alloc() 71 free = page + (size >> PAGE_SHIFT); in __dma_alloc() 72 end = page + (1 << order); in __dma_alloc() 82 return page; in __dma_alloc() [all …]
|
/arch/mips/include/asm/ |
D | cacheflush.h | 38 extern void (*flush_cache_page)(struct vm_area_struct *vma, unsigned long page, unsigned long pfn); 39 extern void __flush_dcache_page(struct page *page); 41 static inline void flush_dcache_page(struct page *page) in flush_dcache_page() argument 44 __flush_dcache_page(page); in flush_dcache_page() 52 extern void __flush_anon_page(struct page *, unsigned long); 54 struct page *page, unsigned long vmaddr) in flush_anon_page() argument 56 if (cpu_has_dc_aliases && PageAnon(page)) in flush_anon_page() 57 __flush_anon_page(page, vmaddr); in flush_anon_page() 61 struct page *page) in flush_icache_page() argument 85 struct page *page, unsigned long vaddr, void *dst, const void *src, [all …]
|
D | highmem.h | 47 extern void * kmap_high(struct page *page); 48 extern void kunmap_high(struct page *page); 50 extern void *__kmap(struct page *page); 51 extern void __kunmap(struct page *page); 52 extern void *__kmap_atomic(struct page *page, enum km_type type); 55 extern struct page *__kmap_atomic_to_page(void *ptr);
|
/arch/sparc/include/asm/ |
D | cacheflush_64.h | 18 #define flush_cache_page(vma, page, pfn) \ argument 30 extern void flush_dcache_page_impl(struct page *page); 32 extern void smp_flush_dcache_page_impl(struct page *page, int cpu); 33 extern void flush_dcache_page_all(struct mm_struct *mm, struct page *page); 35 #define smp_flush_dcache_page_impl(page,cpu) flush_dcache_page_impl(page) argument 36 #define flush_dcache_page_all(mm,page) flush_dcache_page_impl(page) argument 40 extern void flush_dcache_page(struct page *page); 45 extern void flush_ptrace_access(struct vm_area_struct *, struct page *, 49 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 51 flush_cache_page(vma, vaddr, page_to_pfn(page)); \ [all …]
|
D | highmem.h | 54 extern void *kmap_high(struct page *page); 55 extern void kunmap_high(struct page *page); 57 static inline void *kmap(struct page *page) in kmap() argument 60 if (!PageHighMem(page)) in kmap() 61 return page_address(page); in kmap() 62 return kmap_high(page); in kmap() 65 static inline void kunmap(struct page *page) in kunmap() argument 68 if (!PageHighMem(page)) in kunmap() 70 kunmap_high(page); in kunmap() 73 extern void *kmap_atomic(struct page *page, enum km_type type); [all …]
|
/arch/s390/mm/ |
D | pgtable.c | 58 struct page *page = alloc_pages(GFP_KERNEL, ALLOC_ORDER); in crst_table_alloc() local 60 if (!page) in crst_table_alloc() 62 page->index = 0; in crst_table_alloc() 64 struct page *shadow = alloc_pages(GFP_KERNEL, ALLOC_ORDER); in crst_table_alloc() 66 __free_pages(page, ALLOC_ORDER); in crst_table_alloc() 69 page->index = page_to_phys(shadow); in crst_table_alloc() 72 list_add(&page->lru, &mm->context.crst_list); in crst_table_alloc() 74 return (unsigned long *) page_to_phys(page); in crst_table_alloc() 80 struct page *page = virt_to_page(table); in crst_table_free() local 83 list_del(&page->lru); in crst_table_free() [all …]
|
/arch/powerpc/include/asm/ |
D | highmem.h | 62 extern void *kmap_high(struct page *page); 63 extern void kunmap_high(struct page *page); 65 static inline void *kmap(struct page *page) in kmap() argument 68 if (!PageHighMem(page)) in kmap() 69 return page_address(page); in kmap() 70 return kmap_high(page); in kmap() 73 static inline void kunmap(struct page *page) in kunmap() argument 76 if (!PageHighMem(page)) in kunmap() 78 kunmap_high(page); in kunmap() 87 static inline void *kmap_atomic_prot(struct page *page, enum km_type type, pgprot_t prot) in kmap_atomic_prot() argument [all …]
|
D | cacheflush.h | 24 #define flush_icache_page(vma, page) do { } while (0) argument 28 extern void flush_dcache_page(struct page *page); 40 struct page *page, unsigned long addr, 43 extern void flush_dcache_icache_page(struct page *page); 58 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 61 flush_icache_user_range(vma, page, vaddr, len); \ 63 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument 70 void kernel_map_pages(struct page *page, int numpages, int enable);
|
/arch/x86/include/asm/ |
D | highmem.h | 57 extern void *kmap_high(struct page *page); 58 extern void kunmap_high(struct page *page); 60 void *kmap(struct page *page); 61 void kunmap(struct page *page); 62 void *kmap_atomic_prot(struct page *page, enum km_type type, pgprot_t prot); 63 void *kmap_atomic(struct page *page, enum km_type type); 66 struct page *kmap_atomic_to_page(void *ptr); 69 #define kmap_atomic_pte(page, type) kmap_atomic(page, type) argument
|
D | cacheflush.h | 13 #define flush_dcache_page(page) do { } while (0) argument 22 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 24 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument 94 int set_pages_uc(struct page *page, int numpages); 95 int set_pages_wb(struct page *page, int numpages); 96 int set_pages_x(struct page *page, int numpages); 97 int set_pages_nx(struct page *page, int numpages); 98 int set_pages_ro(struct page *page, int numpages); 99 int set_pages_rw(struct page *page, int numpages);
|
/arch/xtensa/mm/ |
D | cache.c | 70 void flush_dcache_page(struct page *page) in flush_dcache_page() argument 72 struct address_space *mapping = page_mapping(page); in flush_dcache_page() 81 if (!test_bit(PG_arch_1, &page->flags)) in flush_dcache_page() 82 set_bit(PG_arch_1, &page->flags); in flush_dcache_page() 87 unsigned long phys = page_to_phys(page); in flush_dcache_page() 88 unsigned long temp = page->index << PAGE_SHIFT; in flush_dcache_page() 102 __flush_invalidate_dcache_page((long)page_address(page)); in flush_dcache_page() 153 struct page *page; in update_mmu_cache() local 158 page = pfn_to_page(pfn); in update_mmu_cache() 167 if (!PageReserved(page) && test_bit(PG_arch_1, &page->flags)) { in update_mmu_cache() [all …]
|
/arch/frv/mm/ |
D | pgalloc.c | 33 struct page *page; in pte_alloc_one() local 36 page = alloc_pages(GFP_KERNEL|__GFP_HIGHMEM|__GFP_REPEAT, 0); in pte_alloc_one() 38 page = alloc_pages(GFP_KERNEL|__GFP_REPEAT, 0); in pte_alloc_one() 40 if (page) { in pte_alloc_one() 41 clear_highpage(page); in pte_alloc_one() 42 pgtable_page_ctor(page); in pte_alloc_one() 43 flush_dcache_page(page); in pte_alloc_one() 45 return page; in pte_alloc_one() 83 struct page *pgd_list; 87 struct page *page = virt_to_page(pgd); in pgd_list_add() local [all …]
|
D | highmem.c | 14 void *kmap(struct page *page) in kmap() argument 17 if (!PageHighMem(page)) in kmap() 18 return page_address(page); in kmap() 19 return kmap_high(page); in kmap() 24 void kunmap(struct page *page) in kunmap() argument 28 if (!PageHighMem(page)) in kunmap() 30 kunmap_high(page); in kunmap() 35 struct page *kmap_atomic_to_page(void *ptr) in kmap_atomic_to_page()
|
/arch/parisc/include/asm/ |
D | cacheflush.h | 45 extern void flush_dcache_page(struct page *page); 52 #define flush_icache_page(vma,page) do { \ argument 53 flush_kernel_dcache_page(page); \ 54 flush_kernel_icache_page(page_address(page)); \ 62 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 64 flush_cache_page(vma, vaddr, page_to_pfn(page)); \ 69 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument 71 flush_cache_page(vma, vaddr, page_to_pfn(page)); \ 81 flush_anon_page(struct vm_area_struct *vma, struct page *page, unsigned long vmaddr) in flush_anon_page() argument 83 if (PageAnon(page)) in flush_anon_page() [all …]
|
/arch/mips/kernel/ |
D | smtc-proc.c | 38 static int proc_read_smtc(char *page, char **start, off_t off, in proc_read_smtc() argument 46 len = sprintf(page, "SMTC Status Word: 0x%08x\n", smtc_status); in proc_read_smtc() 48 page += len; in proc_read_smtc() 49 len = sprintf(page, "Config7: 0x%08x\n", read_c0_config7()); in proc_read_smtc() 51 page += len; in proc_read_smtc() 52 len = sprintf(page, "EBASE: 0x%08lx\n", ebase); in proc_read_smtc() 54 page += len; in proc_read_smtc() 55 len = sprintf(page, "Counter Interrupts taken per CPU (TC)\n"); in proc_read_smtc() 57 page += len; in proc_read_smtc() 59 len = sprintf(page, "%d: %ld\n", i, smtc_cpu_stats[i].timerints); in proc_read_smtc() [all …]
|
/arch/sh/mm/ |
D | pg-sh4.c | 34 static inline void *kmap_coherent(struct page *page, unsigned long addr) in kmap_coherent() argument 44 pte = mk_pte(page, PAGE_KERNEL); in kmap_coherent() 57 static inline void kunmap_coherent(struct page *page) in kunmap_coherent() argument 69 void clear_user_page(void *to, unsigned long address, struct page *page) in clear_user_page() argument 71 __set_bit(PG_mapped, &page->flags); in clear_user_page() 78 void copy_to_user_page(struct vm_area_struct *vma, struct page *page, in copy_to_user_page() argument 84 __set_bit(PG_mapped, &page->flags); in copy_to_user_page() 86 vto = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); in copy_to_user_page() 91 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_to_user_page() 94 void copy_from_user_page(struct vm_area_struct *vma, struct page *page, in copy_from_user_page() argument [all …]
|
/arch/mips/nxp/pnx8550/common/ |
D | proc.c | 29 static int pnx8550_timers_read(char* page, char** start, off_t offset, int count, int* eof, void* d… in pnx8550_timers_read() argument 35 len += sprintf(&page[len], "Timer: count, compare, tc, status\n"); in pnx8550_timers_read() 36 len += sprintf(&page[len], " 1: %11i, %8i, %1i, %s\n", in pnx8550_timers_read() 39 len += sprintf(&page[len], " 2: %11i, %8i, %1i, %s\n", in pnx8550_timers_read() 42 len += sprintf(&page[len], " 3: %11i, %8i, %1i, %s\n", in pnx8550_timers_read() 50 static int pnx8550_registers_read(char* page, char** start, off_t offset, int count, int* eof, void… in pnx8550_registers_read() argument 55 len += sprintf(&page[len], "config1: %#10.8x\n", read_c0_config1()); in pnx8550_registers_read() 56 len += sprintf(&page[len], "config2: %#10.8x\n", read_c0_config2()); in pnx8550_registers_read() 57 len += sprintf(&page[len], "config3: %#10.8x\n", read_c0_config3()); in pnx8550_registers_read() 58 len += sprintf(&page[len], "configPR: %#10.8x\n", read_c0_config7()); in pnx8550_registers_read() [all …]
|
/arch/m68k/include/asm/ |
D | motorola_pgalloc.h | 32 struct page *page = alloc_pages(GFP_KERNEL|__GFP_REPEAT|__GFP_ZERO, 0); in pte_alloc_one() local 35 if(!page) in pte_alloc_one() 38 pte = kmap(page); in pte_alloc_one() 45 pgtable_page_ctor(page); in pte_alloc_one() 46 return page; in pte_alloc_one() 49 static inline void pte_free(struct mm_struct *mm, pgtable_t page) in pte_free() argument 51 pgtable_page_dtor(page); in pte_free() 52 cache_page(kmap(page)); in pte_free() 53 kunmap(page); in pte_free() 54 __free_page(page); in pte_free() [all …]
|
D | sun3_pgalloc.h | 29 static inline void pte_free(struct mm_struct *mm, pgtable_t page) in pte_free() argument 31 pgtable_page_dtor(page); in pte_free() 32 __free_page(page); in pte_free() 44 unsigned long page = __get_free_page(GFP_KERNEL|__GFP_REPEAT); in pte_alloc_one_kernel() local 46 if (!page) in pte_alloc_one_kernel() 49 memset((void *)page, 0, PAGE_SIZE); in pte_alloc_one_kernel() 50 return (pte_t *) (page); in pte_alloc_one_kernel() 56 struct page *page = alloc_pages(GFP_KERNEL|__GFP_REPEAT, 0); in pte_alloc_one() local 58 if (page == NULL) in pte_alloc_one() 61 clear_highpage(page); in pte_alloc_one() [all …]
|
/arch/arm/mm/ |
D | flush.c | 92 void flush_ptrace_access(struct vm_area_struct *vma, struct page *page, in flush_ptrace_access() argument 105 flush_pfn_alias(page_to_pfn(page), uaddr); in flush_ptrace_access() 121 void __flush_dcache_page(struct address_space *mapping, struct page *page) in __flush_dcache_page() argument 128 __cpuc_flush_dcache_page(page_address(page)); in __flush_dcache_page() 136 flush_pfn_alias(page_to_pfn(page), in __flush_dcache_page() 137 page->index << PAGE_CACHE_SHIFT); in __flush_dcache_page() 140 static void __flush_dcache_aliases(struct address_space *mapping, struct page *page) in __flush_dcache_aliases() argument 153 pgoff = page->index << (PAGE_CACHE_SHIFT - PAGE_SHIFT); in __flush_dcache_aliases() 167 flush_cache_page(mpnt, mpnt->vm_start + offset, page_to_pfn(page)); in __flush_dcache_aliases() 190 void flush_dcache_page(struct page *page) in flush_dcache_page() argument [all …]
|
/arch/mips/mm/ |
D | cache.c | 30 void (*flush_cache_page)(struct vm_area_struct *vma, unsigned long page, 75 void __flush_dcache_page(struct page *page) in __flush_dcache_page() argument 77 struct address_space *mapping = page_mapping(page); in __flush_dcache_page() 80 if (PageHighMem(page)) in __flush_dcache_page() 83 SetPageDcacheDirty(page); in __flush_dcache_page() 92 addr = (unsigned long) page_address(page); in __flush_dcache_page() 98 void __flush_anon_page(struct page *page, unsigned long vmaddr) in __flush_anon_page() argument 100 unsigned long addr = (unsigned long) page_address(page); in __flush_anon_page() 103 if (page_mapped(page) && !Page_dcache_dirty(page)) { in __flush_anon_page() 106 kaddr = kmap_coherent(page, vmaddr); in __flush_anon_page() [all …]
|
D | highmem.c | 5 void *__kmap(struct page *page) in __kmap() argument 10 if (!PageHighMem(page)) in __kmap() 11 return page_address(page); in __kmap() 12 addr = kmap_high(page); in __kmap() 18 void __kunmap(struct page *page) in __kunmap() argument 22 if (!PageHighMem(page)) in __kunmap() 24 kunmap_high(page); in __kunmap() 36 void *__kmap_atomic(struct page *page, enum km_type type) in __kmap_atomic() argument 43 if (!PageHighMem(page)) in __kmap_atomic() 44 return page_address(page); in __kmap_atomic() [all …]
|
/arch/x86/mm/ |
D | highmem_32.c | 4 void *kmap(struct page *page) in kmap() argument 7 if (!PageHighMem(page)) in kmap() 8 return page_address(page); in kmap() 9 return kmap_high(page); in kmap() 12 void kunmap(struct page *page) in kunmap() argument 16 if (!PageHighMem(page)) in kunmap() 18 kunmap_high(page); in kunmap() 72 void *kmap_atomic_prot(struct page *page, enum km_type type, pgprot_t prot) in kmap_atomic_prot() argument 80 if (!PageHighMem(page)) in kmap_atomic_prot() 81 return page_address(page); in kmap_atomic_prot() [all …]
|
D | gup.c | 71 unsigned long end, int write, struct page **pages, int *nr) in gup_pte_range() 83 struct page *page; in gup_pte_range() local 90 page = pte_page(pte); in gup_pte_range() 91 get_page(page); in gup_pte_range() 92 pages[*nr] = page; in gup_pte_range() 101 static inline void get_head_page_multiple(struct page *page, int nr) in get_head_page_multiple() argument 103 VM_BUG_ON(page != compound_head(page)); in get_head_page_multiple() 104 VM_BUG_ON(page_count(page) == 0); in get_head_page_multiple() 105 atomic_add(nr, &page->_count); in get_head_page_multiple() 109 unsigned long end, int write, struct page **pages, int *nr) in gup_huge_pmd() [all …]
|
/arch/sh/include/asm/ |
D | cacheflush.h | 19 #define flush_dcache_page(page) do { } while (0) argument 47 static inline void flush_kernel_dcache_page(struct page *page) in flush_kernel_dcache_page() argument 49 flush_dcache_page(page); in flush_kernel_dcache_page() 54 struct page *page, unsigned long vaddr, void *dst, const void *src, 58 struct page *page, unsigned long vaddr, void *dst, const void *src, 61 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 63 flush_cache_page(vma, vaddr, page_to_pfn(page));\ 65 flush_icache_user_range(vma, page, vaddr, len); \ 68 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument 70 flush_cache_page(vma, vaddr, page_to_pfn(page));\
|