Home
last modified time | relevance | path

Searched refs:page (Results 1 – 25 of 477) sorted by relevance

12345678910>>...20

/arch/avr32/mm/
Ddma-coherent.c38 static struct page *__dma_alloc(struct device *dev, size_t size, in __dma_alloc()
41 struct page *page, *free, *end; in __dma_alloc() local
54 page = alloc_pages(gfp, order); in __dma_alloc()
55 if (!page) in __dma_alloc()
57 split_page(page, order); in __dma_alloc()
68 invalidate_dcache_region(phys_to_virt(page_to_phys(page)), size); in __dma_alloc()
70 *handle = page_to_bus(page); in __dma_alloc()
71 free = page + (size >> PAGE_SHIFT); in __dma_alloc()
72 end = page + (1 << order); in __dma_alloc()
82 return page; in __dma_alloc()
[all …]
/arch/mips/include/asm/
Dcacheflush.h38 extern void (*flush_cache_page)(struct vm_area_struct *vma, unsigned long page, unsigned long pfn);
39 extern void __flush_dcache_page(struct page *page);
41 static inline void flush_dcache_page(struct page *page) in flush_dcache_page() argument
44 __flush_dcache_page(page); in flush_dcache_page()
52 extern void __flush_anon_page(struct page *, unsigned long);
54 struct page *page, unsigned long vmaddr) in flush_anon_page() argument
56 if (cpu_has_dc_aliases && PageAnon(page)) in flush_anon_page()
57 __flush_anon_page(page, vmaddr); in flush_anon_page()
61 struct page *page) in flush_icache_page() argument
85 struct page *page, unsigned long vaddr, void *dst, const void *src,
[all …]
Dhighmem.h47 extern void * kmap_high(struct page *page);
48 extern void kunmap_high(struct page *page);
50 extern void *__kmap(struct page *page);
51 extern void __kunmap(struct page *page);
52 extern void *__kmap_atomic(struct page *page, enum km_type type);
55 extern struct page *__kmap_atomic_to_page(void *ptr);
/arch/sparc/include/asm/
Dcacheflush_64.h18 #define flush_cache_page(vma, page, pfn) \ argument
30 extern void flush_dcache_page_impl(struct page *page);
32 extern void smp_flush_dcache_page_impl(struct page *page, int cpu);
33 extern void flush_dcache_page_all(struct mm_struct *mm, struct page *page);
35 #define smp_flush_dcache_page_impl(page,cpu) flush_dcache_page_impl(page) argument
36 #define flush_dcache_page_all(mm,page) flush_dcache_page_impl(page) argument
40 extern void flush_dcache_page(struct page *page);
45 extern void flush_ptrace_access(struct vm_area_struct *, struct page *,
49 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument
51 flush_cache_page(vma, vaddr, page_to_pfn(page)); \
[all …]
Dhighmem.h54 extern void *kmap_high(struct page *page);
55 extern void kunmap_high(struct page *page);
57 static inline void *kmap(struct page *page) in kmap() argument
60 if (!PageHighMem(page)) in kmap()
61 return page_address(page); in kmap()
62 return kmap_high(page); in kmap()
65 static inline void kunmap(struct page *page) in kunmap() argument
68 if (!PageHighMem(page)) in kunmap()
70 kunmap_high(page); in kunmap()
73 extern void *kmap_atomic(struct page *page, enum km_type type);
[all …]
/arch/s390/mm/
Dpgtable.c58 struct page *page = alloc_pages(GFP_KERNEL, ALLOC_ORDER); in crst_table_alloc() local
60 if (!page) in crst_table_alloc()
62 page->index = 0; in crst_table_alloc()
64 struct page *shadow = alloc_pages(GFP_KERNEL, ALLOC_ORDER); in crst_table_alloc()
66 __free_pages(page, ALLOC_ORDER); in crst_table_alloc()
69 page->index = page_to_phys(shadow); in crst_table_alloc()
72 list_add(&page->lru, &mm->context.crst_list); in crst_table_alloc()
74 return (unsigned long *) page_to_phys(page); in crst_table_alloc()
80 struct page *page = virt_to_page(table); in crst_table_free() local
83 list_del(&page->lru); in crst_table_free()
[all …]
/arch/powerpc/include/asm/
Dhighmem.h62 extern void *kmap_high(struct page *page);
63 extern void kunmap_high(struct page *page);
65 static inline void *kmap(struct page *page) in kmap() argument
68 if (!PageHighMem(page)) in kmap()
69 return page_address(page); in kmap()
70 return kmap_high(page); in kmap()
73 static inline void kunmap(struct page *page) in kunmap() argument
76 if (!PageHighMem(page)) in kunmap()
78 kunmap_high(page); in kunmap()
87 static inline void *kmap_atomic_prot(struct page *page, enum km_type type, pgprot_t prot) in kmap_atomic_prot() argument
[all …]
Dcacheflush.h24 #define flush_icache_page(vma, page) do { } while (0) argument
28 extern void flush_dcache_page(struct page *page);
40 struct page *page, unsigned long addr,
43 extern void flush_dcache_icache_page(struct page *page);
58 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument
61 flush_icache_user_range(vma, page, vaddr, len); \
63 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
70 void kernel_map_pages(struct page *page, int numpages, int enable);
/arch/x86/include/asm/
Dhighmem.h57 extern void *kmap_high(struct page *page);
58 extern void kunmap_high(struct page *page);
60 void *kmap(struct page *page);
61 void kunmap(struct page *page);
62 void *kmap_atomic_prot(struct page *page, enum km_type type, pgprot_t prot);
63 void *kmap_atomic(struct page *page, enum km_type type);
66 struct page *kmap_atomic_to_page(void *ptr);
69 #define kmap_atomic_pte(page, type) kmap_atomic(page, type) argument
Dcacheflush.h13 #define flush_dcache_page(page) do { } while (0) argument
22 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument
24 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
94 int set_pages_uc(struct page *page, int numpages);
95 int set_pages_wb(struct page *page, int numpages);
96 int set_pages_x(struct page *page, int numpages);
97 int set_pages_nx(struct page *page, int numpages);
98 int set_pages_ro(struct page *page, int numpages);
99 int set_pages_rw(struct page *page, int numpages);
/arch/xtensa/mm/
Dcache.c70 void flush_dcache_page(struct page *page) in flush_dcache_page() argument
72 struct address_space *mapping = page_mapping(page); in flush_dcache_page()
81 if (!test_bit(PG_arch_1, &page->flags)) in flush_dcache_page()
82 set_bit(PG_arch_1, &page->flags); in flush_dcache_page()
87 unsigned long phys = page_to_phys(page); in flush_dcache_page()
88 unsigned long temp = page->index << PAGE_SHIFT; in flush_dcache_page()
102 __flush_invalidate_dcache_page((long)page_address(page)); in flush_dcache_page()
153 struct page *page; in update_mmu_cache() local
158 page = pfn_to_page(pfn); in update_mmu_cache()
167 if (!PageReserved(page) && test_bit(PG_arch_1, &page->flags)) { in update_mmu_cache()
[all …]
/arch/frv/mm/
Dpgalloc.c33 struct page *page; in pte_alloc_one() local
36 page = alloc_pages(GFP_KERNEL|__GFP_HIGHMEM|__GFP_REPEAT, 0); in pte_alloc_one()
38 page = alloc_pages(GFP_KERNEL|__GFP_REPEAT, 0); in pte_alloc_one()
40 if (page) { in pte_alloc_one()
41 clear_highpage(page); in pte_alloc_one()
42 pgtable_page_ctor(page); in pte_alloc_one()
43 flush_dcache_page(page); in pte_alloc_one()
45 return page; in pte_alloc_one()
83 struct page *pgd_list;
87 struct page *page = virt_to_page(pgd); in pgd_list_add() local
[all …]
Dhighmem.c14 void *kmap(struct page *page) in kmap() argument
17 if (!PageHighMem(page)) in kmap()
18 return page_address(page); in kmap()
19 return kmap_high(page); in kmap()
24 void kunmap(struct page *page) in kunmap() argument
28 if (!PageHighMem(page)) in kunmap()
30 kunmap_high(page); in kunmap()
35 struct page *kmap_atomic_to_page(void *ptr) in kmap_atomic_to_page()
/arch/parisc/include/asm/
Dcacheflush.h45 extern void flush_dcache_page(struct page *page);
52 #define flush_icache_page(vma,page) do { \ argument
53 flush_kernel_dcache_page(page); \
54 flush_kernel_icache_page(page_address(page)); \
62 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument
64 flush_cache_page(vma, vaddr, page_to_pfn(page)); \
69 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
71 flush_cache_page(vma, vaddr, page_to_pfn(page)); \
81 flush_anon_page(struct vm_area_struct *vma, struct page *page, unsigned long vmaddr) in flush_anon_page() argument
83 if (PageAnon(page)) in flush_anon_page()
[all …]
/arch/mips/kernel/
Dsmtc-proc.c38 static int proc_read_smtc(char *page, char **start, off_t off, in proc_read_smtc() argument
46 len = sprintf(page, "SMTC Status Word: 0x%08x\n", smtc_status); in proc_read_smtc()
48 page += len; in proc_read_smtc()
49 len = sprintf(page, "Config7: 0x%08x\n", read_c0_config7()); in proc_read_smtc()
51 page += len; in proc_read_smtc()
52 len = sprintf(page, "EBASE: 0x%08lx\n", ebase); in proc_read_smtc()
54 page += len; in proc_read_smtc()
55 len = sprintf(page, "Counter Interrupts taken per CPU (TC)\n"); in proc_read_smtc()
57 page += len; in proc_read_smtc()
59 len = sprintf(page, "%d: %ld\n", i, smtc_cpu_stats[i].timerints); in proc_read_smtc()
[all …]
/arch/sh/mm/
Dpg-sh4.c34 static inline void *kmap_coherent(struct page *page, unsigned long addr) in kmap_coherent() argument
44 pte = mk_pte(page, PAGE_KERNEL); in kmap_coherent()
57 static inline void kunmap_coherent(struct page *page) in kunmap_coherent() argument
69 void clear_user_page(void *to, unsigned long address, struct page *page) in clear_user_page() argument
71 __set_bit(PG_mapped, &page->flags); in clear_user_page()
78 void copy_to_user_page(struct vm_area_struct *vma, struct page *page, in copy_to_user_page() argument
84 __set_bit(PG_mapped, &page->flags); in copy_to_user_page()
86 vto = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); in copy_to_user_page()
91 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_to_user_page()
94 void copy_from_user_page(struct vm_area_struct *vma, struct page *page, in copy_from_user_page() argument
[all …]
/arch/mips/nxp/pnx8550/common/
Dproc.c29 static int pnx8550_timers_read(char* page, char** start, off_t offset, int count, int* eof, void* d… in pnx8550_timers_read() argument
35 len += sprintf(&page[len], "Timer: count, compare, tc, status\n"); in pnx8550_timers_read()
36 len += sprintf(&page[len], " 1: %11i, %8i, %1i, %s\n", in pnx8550_timers_read()
39 len += sprintf(&page[len], " 2: %11i, %8i, %1i, %s\n", in pnx8550_timers_read()
42 len += sprintf(&page[len], " 3: %11i, %8i, %1i, %s\n", in pnx8550_timers_read()
50 static int pnx8550_registers_read(char* page, char** start, off_t offset, int count, int* eof, void… in pnx8550_registers_read() argument
55 len += sprintf(&page[len], "config1: %#10.8x\n", read_c0_config1()); in pnx8550_registers_read()
56 len += sprintf(&page[len], "config2: %#10.8x\n", read_c0_config2()); in pnx8550_registers_read()
57 len += sprintf(&page[len], "config3: %#10.8x\n", read_c0_config3()); in pnx8550_registers_read()
58 len += sprintf(&page[len], "configPR: %#10.8x\n", read_c0_config7()); in pnx8550_registers_read()
[all …]
/arch/m68k/include/asm/
Dmotorola_pgalloc.h32 struct page *page = alloc_pages(GFP_KERNEL|__GFP_REPEAT|__GFP_ZERO, 0); in pte_alloc_one() local
35 if(!page) in pte_alloc_one()
38 pte = kmap(page); in pte_alloc_one()
45 pgtable_page_ctor(page); in pte_alloc_one()
46 return page; in pte_alloc_one()
49 static inline void pte_free(struct mm_struct *mm, pgtable_t page) in pte_free() argument
51 pgtable_page_dtor(page); in pte_free()
52 cache_page(kmap(page)); in pte_free()
53 kunmap(page); in pte_free()
54 __free_page(page); in pte_free()
[all …]
Dsun3_pgalloc.h29 static inline void pte_free(struct mm_struct *mm, pgtable_t page) in pte_free() argument
31 pgtable_page_dtor(page); in pte_free()
32 __free_page(page); in pte_free()
44 unsigned long page = __get_free_page(GFP_KERNEL|__GFP_REPEAT); in pte_alloc_one_kernel() local
46 if (!page) in pte_alloc_one_kernel()
49 memset((void *)page, 0, PAGE_SIZE); in pte_alloc_one_kernel()
50 return (pte_t *) (page); in pte_alloc_one_kernel()
56 struct page *page = alloc_pages(GFP_KERNEL|__GFP_REPEAT, 0); in pte_alloc_one() local
58 if (page == NULL) in pte_alloc_one()
61 clear_highpage(page); in pte_alloc_one()
[all …]
/arch/arm/mm/
Dflush.c92 void flush_ptrace_access(struct vm_area_struct *vma, struct page *page, in flush_ptrace_access() argument
105 flush_pfn_alias(page_to_pfn(page), uaddr); in flush_ptrace_access()
121 void __flush_dcache_page(struct address_space *mapping, struct page *page) in __flush_dcache_page() argument
128 __cpuc_flush_dcache_page(page_address(page)); in __flush_dcache_page()
136 flush_pfn_alias(page_to_pfn(page), in __flush_dcache_page()
137 page->index << PAGE_CACHE_SHIFT); in __flush_dcache_page()
140 static void __flush_dcache_aliases(struct address_space *mapping, struct page *page) in __flush_dcache_aliases() argument
153 pgoff = page->index << (PAGE_CACHE_SHIFT - PAGE_SHIFT); in __flush_dcache_aliases()
167 flush_cache_page(mpnt, mpnt->vm_start + offset, page_to_pfn(page)); in __flush_dcache_aliases()
190 void flush_dcache_page(struct page *page) in flush_dcache_page() argument
[all …]
/arch/mips/mm/
Dcache.c30 void (*flush_cache_page)(struct vm_area_struct *vma, unsigned long page,
75 void __flush_dcache_page(struct page *page) in __flush_dcache_page() argument
77 struct address_space *mapping = page_mapping(page); in __flush_dcache_page()
80 if (PageHighMem(page)) in __flush_dcache_page()
83 SetPageDcacheDirty(page); in __flush_dcache_page()
92 addr = (unsigned long) page_address(page); in __flush_dcache_page()
98 void __flush_anon_page(struct page *page, unsigned long vmaddr) in __flush_anon_page() argument
100 unsigned long addr = (unsigned long) page_address(page); in __flush_anon_page()
103 if (page_mapped(page) && !Page_dcache_dirty(page)) { in __flush_anon_page()
106 kaddr = kmap_coherent(page, vmaddr); in __flush_anon_page()
[all …]
Dhighmem.c5 void *__kmap(struct page *page) in __kmap() argument
10 if (!PageHighMem(page)) in __kmap()
11 return page_address(page); in __kmap()
12 addr = kmap_high(page); in __kmap()
18 void __kunmap(struct page *page) in __kunmap() argument
22 if (!PageHighMem(page)) in __kunmap()
24 kunmap_high(page); in __kunmap()
36 void *__kmap_atomic(struct page *page, enum km_type type) in __kmap_atomic() argument
43 if (!PageHighMem(page)) in __kmap_atomic()
44 return page_address(page); in __kmap_atomic()
[all …]
/arch/x86/mm/
Dhighmem_32.c4 void *kmap(struct page *page) in kmap() argument
7 if (!PageHighMem(page)) in kmap()
8 return page_address(page); in kmap()
9 return kmap_high(page); in kmap()
12 void kunmap(struct page *page) in kunmap() argument
16 if (!PageHighMem(page)) in kunmap()
18 kunmap_high(page); in kunmap()
72 void *kmap_atomic_prot(struct page *page, enum km_type type, pgprot_t prot) in kmap_atomic_prot() argument
80 if (!PageHighMem(page)) in kmap_atomic_prot()
81 return page_address(page); in kmap_atomic_prot()
[all …]
Dgup.c71 unsigned long end, int write, struct page **pages, int *nr) in gup_pte_range()
83 struct page *page; in gup_pte_range() local
90 page = pte_page(pte); in gup_pte_range()
91 get_page(page); in gup_pte_range()
92 pages[*nr] = page; in gup_pte_range()
101 static inline void get_head_page_multiple(struct page *page, int nr) in get_head_page_multiple() argument
103 VM_BUG_ON(page != compound_head(page)); in get_head_page_multiple()
104 VM_BUG_ON(page_count(page) == 0); in get_head_page_multiple()
105 atomic_add(nr, &page->_count); in get_head_page_multiple()
109 unsigned long end, int write, struct page **pages, int *nr) in gup_huge_pmd()
[all …]
/arch/sh/include/asm/
Dcacheflush.h19 #define flush_dcache_page(page) do { } while (0) argument
47 static inline void flush_kernel_dcache_page(struct page *page) in flush_kernel_dcache_page() argument
49 flush_dcache_page(page); in flush_kernel_dcache_page()
54 struct page *page, unsigned long vaddr, void *dst, const void *src,
58 struct page *page, unsigned long vaddr, void *dst, const void *src,
61 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument
63 flush_cache_page(vma, vaddr, page_to_pfn(page));\
65 flush_icache_user_range(vma, page, vaddr, len); \
68 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
70 flush_cache_page(vma, vaddr, page_to_pfn(page));\

12345678910>>...20