1 #ifndef _ALPHA_TLBFLUSH_H
2 #define _ALPHA_TLBFLUSH_H
3
4 #include <linux/mm.h>
5 #include <asm/compiler.h>
6 #include <asm/pgalloc.h>
7
8 #ifndef __EXTERN_INLINE
9 #define __EXTERN_INLINE extern inline
10 #define __MMU_EXTERN_INLINE
11 #endif
12
13 extern void __load_new_mm_context(struct mm_struct *);
14
15
16 /* Use a few helper functions to hide the ugly broken ASN
17 numbers on early Alphas (ev4 and ev45). */
18
19 __EXTERN_INLINE void
ev4_flush_tlb_current(struct mm_struct * mm)20 ev4_flush_tlb_current(struct mm_struct *mm)
21 {
22 __load_new_mm_context(mm);
23 tbiap();
24 }
25
26 __EXTERN_INLINE void
ev5_flush_tlb_current(struct mm_struct * mm)27 ev5_flush_tlb_current(struct mm_struct *mm)
28 {
29 __load_new_mm_context(mm);
30 }
31
32 /* Flush just one page in the current TLB set. We need to be very
33 careful about the icache here, there is no way to invalidate a
34 specific icache page. */
35
36 __EXTERN_INLINE void
ev4_flush_tlb_current_page(struct mm_struct * mm,struct vm_area_struct * vma,unsigned long addr)37 ev4_flush_tlb_current_page(struct mm_struct * mm,
38 struct vm_area_struct *vma,
39 unsigned long addr)
40 {
41 int tbi_flag = 2;
42 if (vma->vm_flags & VM_EXEC) {
43 __load_new_mm_context(mm);
44 tbi_flag = 3;
45 }
46 tbi(tbi_flag, addr);
47 }
48
49 __EXTERN_INLINE void
ev5_flush_tlb_current_page(struct mm_struct * mm,struct vm_area_struct * vma,unsigned long addr)50 ev5_flush_tlb_current_page(struct mm_struct * mm,
51 struct vm_area_struct *vma,
52 unsigned long addr)
53 {
54 if (vma->vm_flags & VM_EXEC)
55 __load_new_mm_context(mm);
56 else
57 tbi(2, addr);
58 }
59
60
61 #ifdef CONFIG_ALPHA_GENERIC
62 # define flush_tlb_current alpha_mv.mv_flush_tlb_current
63 # define flush_tlb_current_page alpha_mv.mv_flush_tlb_current_page
64 #else
65 # ifdef CONFIG_ALPHA_EV4
66 # define flush_tlb_current ev4_flush_tlb_current
67 # define flush_tlb_current_page ev4_flush_tlb_current_page
68 # else
69 # define flush_tlb_current ev5_flush_tlb_current
70 # define flush_tlb_current_page ev5_flush_tlb_current_page
71 # endif
72 #endif
73
74 #ifdef __MMU_EXTERN_INLINE
75 #undef __EXTERN_INLINE
76 #undef __MMU_EXTERN_INLINE
77 #endif
78
79 /* Flush current user mapping. */
80 static inline void
flush_tlb(void)81 flush_tlb(void)
82 {
83 flush_tlb_current(current->active_mm);
84 }
85
86 /* Flush someone else's user mapping. */
87 static inline void
flush_tlb_other(struct mm_struct * mm)88 flush_tlb_other(struct mm_struct *mm)
89 {
90 unsigned long *mmc = &mm->context[smp_processor_id()];
91 /* Check it's not zero first to avoid cacheline ping pong
92 when possible. */
93 if (*mmc) *mmc = 0;
94 }
95
96 #ifndef CONFIG_SMP
97 /* Flush everything (kernel mapping may also have changed
98 due to vmalloc/vfree). */
flush_tlb_all(void)99 static inline void flush_tlb_all(void)
100 {
101 tbia();
102 }
103
104 /* Flush a specified user mapping. */
105 static inline void
flush_tlb_mm(struct mm_struct * mm)106 flush_tlb_mm(struct mm_struct *mm)
107 {
108 if (mm == current->active_mm)
109 flush_tlb_current(mm);
110 else
111 flush_tlb_other(mm);
112 }
113
114 /* Page-granular tlb flush. */
115 static inline void
flush_tlb_page(struct vm_area_struct * vma,unsigned long addr)116 flush_tlb_page(struct vm_area_struct *vma, unsigned long addr)
117 {
118 struct mm_struct *mm = vma->vm_mm;
119
120 if (mm == current->active_mm)
121 flush_tlb_current_page(mm, vma, addr);
122 else
123 flush_tlb_other(mm);
124 }
125
126 /* Flush a specified range of user mapping. On the Alpha we flush
127 the whole user tlb. */
128 static inline void
flush_tlb_range(struct vm_area_struct * vma,unsigned long start,unsigned long end)129 flush_tlb_range(struct vm_area_struct *vma, unsigned long start,
130 unsigned long end)
131 {
132 flush_tlb_mm(vma->vm_mm);
133 }
134
135 #else /* CONFIG_SMP */
136
137 extern void flush_tlb_all(void);
138 extern void flush_tlb_mm(struct mm_struct *);
139 extern void flush_tlb_page(struct vm_area_struct *, unsigned long);
140 extern void flush_tlb_range(struct vm_area_struct *, unsigned long,
141 unsigned long);
142
143 #endif /* CONFIG_SMP */
144
flush_tlb_kernel_range(unsigned long start,unsigned long end)145 static inline void flush_tlb_kernel_range(unsigned long start,
146 unsigned long end)
147 {
148 flush_tlb_all();
149 }
150
151 #endif /* _ALPHA_TLBFLUSH_H */
152