2 * Flush the host cpu caches.
4 * This work is licensed under the terms of the GNU GPL, version 2 or later.
5 * See the COPYING file in the top-level directory.
8 #include "qemu/osdep.h"
9 #include "qemu/cacheflush.h"
10 #include "qemu/cacheinfo.h"
11 #include "qemu/bitops.h"
14 #if defined(__i386__) || defined(__x86_64__) || defined(__s390__)
16 /* Caches are coherent and do not require flushing; symbol inline. */
18 #elif defined(__aarch64__)
21 /* Apple does not expose CTR_EL0, so we must use system interfaces. */
22 extern void sys_icache_invalidate(void *start
, size_t len
);
23 extern void sys_dcache_flush(void *start
, size_t len
);
24 void flush_idcache_range(uintptr_t rx
, uintptr_t rw
, size_t len
)
26 sys_dcache_flush((void *)rw
, len
);
27 sys_icache_invalidate((void *)rx
, len
);
32 * TODO: unify this with cacheinfo.c.
33 * We want to save the whole contents of CTR_EL0, so that we
34 * have more than the linesize, but also IDC and DIC.
36 static uint64_t save_ctr_el0
;
37 static void __attribute__((constructor
)) init_ctr_el0(void)
39 asm volatile("mrs\t%0, ctr_el0" : "=r"(save_ctr_el0
));
43 * This is a copy of gcc's __aarch64_sync_cache_range, modified
44 * to fit this three-operand interface.
46 void flush_idcache_range(uintptr_t rx
, uintptr_t rw
, size_t len
)
48 const unsigned CTR_IDC
= 1u << 28;
49 const unsigned CTR_DIC
= 1u << 29;
50 const uint64_t ctr_el0
= save_ctr_el0
;
51 const uintptr_t icache_lsize
= 4 << extract64(ctr_el0
, 0, 4);
52 const uintptr_t dcache_lsize
= 4 << extract64(ctr_el0
, 16, 4);
56 * If CTR_EL0.IDC is enabled, Data cache clean to the Point of Unification
57 * is not required for instruction to data coherence.
59 if (!(ctr_el0
& CTR_IDC
)) {
61 * Loop over the address range, clearing one cache line at once.
62 * Data cache must be flushed to unification first to make sure
63 * the instruction cache fetches the updated data.
65 for (p
= rw
& -dcache_lsize
; p
< rw
+ len
; p
+= dcache_lsize
) {
66 asm volatile("dc\tcvau, %0" : : "r" (p
) : "memory");
68 asm volatile("dsb\tish" : : : "memory");
72 * If CTR_EL0.DIC is enabled, Instruction cache cleaning to the Point
73 * of Unification is not required for instruction to data coherence.
75 if (!(ctr_el0
& CTR_DIC
)) {
76 for (p
= rx
& -icache_lsize
; p
< rx
+ len
; p
+= icache_lsize
) {
77 asm volatile("ic\tivau, %0" : : "r"(p
) : "memory");
79 asm volatile ("dsb\tish" : : : "memory");
82 asm volatile("isb" : : : "memory");
84 #endif /* CONFIG_DARWIN */
86 #elif defined(__mips__)
89 #include <machine/sysarch.h>
91 #include <sys/cachectl.h>
94 void flush_idcache_range(uintptr_t rx
, uintptr_t rw
, size_t len
)
97 cacheflush((void *)rw
, len
, DCACHE
);
99 cacheflush((void *)rx
, len
, ICACHE
);
102 #elif defined(__powerpc__)
104 void flush_idcache_range(uintptr_t rx
, uintptr_t rw
, size_t len
)
107 size_t dsize
= qemu_dcache_linesize
;
108 size_t isize
= qemu_icache_linesize
;
110 b
= rw
& ~(dsize
- 1);
111 e
= (rw
+ len
+ dsize
- 1) & ~(dsize
- 1);
112 for (p
= b
; p
< e
; p
+= dsize
) {
113 asm volatile ("dcbst 0,%0" : : "r"(p
) : "memory");
115 asm volatile ("sync" : : : "memory");
117 b
= rx
& ~(isize
- 1);
118 e
= (rx
+ len
+ isize
- 1) & ~(isize
- 1);
119 for (p
= b
; p
< e
; p
+= isize
) {
120 asm volatile ("icbi 0,%0" : : "r"(p
) : "memory");
122 asm volatile ("sync" : : : "memory");
123 asm volatile ("isync" : : : "memory");
126 #elif defined(__sparc__)
128 void flush_idcache_range(uintptr_t rx
, uintptr_t rw
, size_t len
)
130 /* No additional data flush to the RW virtual address required. */
131 uintptr_t p
, end
= (rx
+ len
+ 7) & -8;
132 for (p
= rx
& -8; p
< end
; p
+= 8) {
133 __asm__
__volatile__("flush\t%0" : : "r" (p
));
139 void flush_idcache_range(uintptr_t rx
, uintptr_t rw
, size_t len
)
142 __builtin___clear_cache((char *)rw
, (char *)rw
+ len
);
144 __builtin___clear_cache((char *)rx
, (char *)rx
+ len
);