1 #ifndef _ALPHA_TLBFLUSH_H
2 #define _ALPHA_TLBFLUSH_H
5 #include <asm/compiler.h>
6 #include <asm/pgalloc.h>
8 #ifndef __EXTERN_INLINE
9 #define __EXTERN_INLINE extern inline
10 #define __MMU_EXTERN_INLINE
13 extern void __load_new_mm_context(struct mm_struct
*);
16 /* Use a few helper functions to hide the ugly broken ASN
17 numbers on early Alphas (ev4 and ev45). */
20 ev4_flush_tlb_current(struct mm_struct
*mm
)
22 __load_new_mm_context(mm
);
27 ev5_flush_tlb_current(struct mm_struct
*mm
)
29 __load_new_mm_context(mm
);
32 /* Flush just one page in the current TLB set. We need to be very
33 careful about the icache here, there is no way to invalidate a
34 specific icache page. */
37 ev4_flush_tlb_current_page(struct mm_struct
* mm
,
38 struct vm_area_struct
*vma
,
42 if (vma
->vm_flags
& VM_EXEC
) {
43 __load_new_mm_context(mm
);
50 ev5_flush_tlb_current_page(struct mm_struct
* mm
,
51 struct vm_area_struct
*vma
,
54 if (vma
->vm_flags
& VM_EXEC
)
55 __load_new_mm_context(mm
);
61 #ifdef CONFIG_ALPHA_GENERIC
62 # define flush_tlb_current alpha_mv.mv_flush_tlb_current
63 # define flush_tlb_current_page alpha_mv.mv_flush_tlb_current_page
65 # ifdef CONFIG_ALPHA_EV4
66 # define flush_tlb_current ev4_flush_tlb_current
67 # define flush_tlb_current_page ev4_flush_tlb_current_page
69 # define flush_tlb_current ev5_flush_tlb_current
70 # define flush_tlb_current_page ev5_flush_tlb_current_page
74 #ifdef __MMU_EXTERN_INLINE
75 #undef __EXTERN_INLINE
76 #undef __MMU_EXTERN_INLINE
79 /* Flush current user mapping. */
83 flush_tlb_current(current
->active_mm
);
86 /* Flush someone else's user mapping. */
88 flush_tlb_other(struct mm_struct
*mm
)
90 unsigned long *mmc
= &mm
->context
[smp_processor_id()];
91 /* Check it's not zero first to avoid cacheline ping pong
97 /* Flush everything (kernel mapping may also have changed
98 due to vmalloc/vfree). */
99 static inline void flush_tlb_all(void)
104 /* Flush a specified user mapping. */
106 flush_tlb_mm(struct mm_struct
*mm
)
108 if (mm
== current
->active_mm
)
109 flush_tlb_current(mm
);
114 /* Page-granular tlb flush. */
116 flush_tlb_page(struct vm_area_struct
*vma
, unsigned long addr
)
118 struct mm_struct
*mm
= vma
->vm_mm
;
120 if (mm
== current
->active_mm
)
121 flush_tlb_current_page(mm
, vma
, addr
);
126 /* Flush a specified range of user mapping. On the Alpha we flush
127 the whole user tlb. */
129 flush_tlb_range(struct vm_area_struct
*vma
, unsigned long start
,
132 flush_tlb_mm(vma
->vm_mm
);
135 #else /* CONFIG_SMP */
137 extern void flush_tlb_all(void);
138 extern void flush_tlb_mm(struct mm_struct
*);
139 extern void flush_tlb_page(struct vm_area_struct
*, unsigned long);
140 extern void flush_tlb_range(struct vm_area_struct
*, unsigned long,
143 #endif /* CONFIG_SMP */
145 static inline void flush_tlb_kernel_range(unsigned long start
,
151 #endif /* _ALPHA_TLBFLUSH_H */