2 * PowerPC emulation special registers manipulation helpers for qemu.
4 * Copyright (c) 2003-2007 Jocelyn Mayer
6 * This library is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU Lesser General Public
8 * License as published by the Free Software Foundation; either
9 * version 2 of the License, or (at your option) any later version.
11 * This library is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14 * Lesser General Public License for more details.
16 * You should have received a copy of the GNU Lesser General Public
17 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
23 /* Swap temporary saved registers with GPRs */
24 static inline void hreg_swap_gpr_tgpr(CPUPPCState
*env
)
29 env
->gpr
[0] = env
->tgpr
[0];
32 env
->gpr
[1] = env
->tgpr
[1];
35 env
->gpr
[2] = env
->tgpr
[2];
38 env
->gpr
[3] = env
->tgpr
[3];
42 static inline void hreg_compute_mem_idx(CPUPPCState
*env
)
44 /* This is our encoding for server processors. The architecture
45 * specifies that there is no such thing as userspace with
46 * translation off, however it appears that MacOS does it and
47 * some 32-bit CPUs support it. Weird...
49 * 0 = Guest User space virtual mode
50 * 1 = Guest Kernel space virtual mode
51 * 2 = Guest User space real mode
52 * 3 = Guest Kernel space real mode
53 * 4 = HV User space virtual mode
54 * 5 = HV Kernel space virtual mode
55 * 6 = HV User space real mode
56 * 7 = HV Kernel space real mode
58 * For BookE, we need 8 MMU modes as follow:
60 * 0 = AS 0 HV User space
61 * 1 = AS 0 HV Kernel space
62 * 2 = AS 1 HV User space
63 * 3 = AS 1 HV Kernel space
64 * 4 = AS 0 Guest User space
65 * 5 = AS 0 Guest Kernel space
66 * 6 = AS 1 Guest User space
67 * 7 = AS 1 Guest Kernel space
69 if (env
->mmu_model
& POWERPC_MMU_BOOKE
) {
70 env
->immu_idx
= env
->dmmu_idx
= msr_pr
? 0 : 1;
71 env
->immu_idx
+= msr_is
? 2 : 0;
72 env
->dmmu_idx
+= msr_ds
? 2 : 0;
73 env
->immu_idx
+= msr_gs
? 4 : 0;
74 env
->dmmu_idx
+= msr_gs
? 4 : 0;
76 env
->immu_idx
= env
->dmmu_idx
= msr_pr
? 0 : 1;
77 env
->immu_idx
+= msr_ir
? 0 : 2;
78 env
->dmmu_idx
+= msr_dr
? 0 : 2;
79 env
->immu_idx
+= msr_hv
? 4 : 0;
80 env
->dmmu_idx
+= msr_hv
? 4 : 0;
84 static inline void hreg_compute_hflags(CPUPPCState
*env
)
86 target_ulong hflags_mask
;
88 /* We 'forget' FE0 & FE1: we'll never generate imprecise exceptions */
89 hflags_mask
= (1 << MSR_VR
) | (1 << MSR_AP
) | (1 << MSR_SA
) |
90 (1 << MSR_PR
) | (1 << MSR_FP
) | (1 << MSR_SE
) | (1 << MSR_BE
) |
91 (1 << MSR_LE
) | (1 << MSR_VSX
) | (1 << MSR_IR
) | (1 << MSR_DR
);
92 hflags_mask
|= (1ULL << MSR_CM
) | (1ULL << MSR_SF
) | MSR_HVB
;
93 hreg_compute_mem_idx(env
);
94 env
->hflags
= env
->msr
& hflags_mask
;
95 /* Merge with hflags coming from other registers */
96 env
->hflags
|= env
->hflags_nmsr
;
99 static inline int hreg_store_msr(CPUPPCState
*env
, target_ulong value
,
103 #if !defined(CONFIG_USER_ONLY)
104 CPUState
*cs
= CPU(ppc_env_get_cpu(env
));
108 value
&= env
->msr_mask
;
109 #if !defined(CONFIG_USER_ONLY)
110 /* Neither mtmsr nor guest state can alter HV */
111 if (!alter_hv
|| !(env
->msr
& MSR_HVB
)) {
113 value
|= env
->msr
& MSR_HVB
;
115 if (((value
>> MSR_IR
) & 1) != msr_ir
||
116 ((value
>> MSR_DR
) & 1) != msr_dr
) {
117 cs
->interrupt_request
|= CPU_INTERRUPT_EXITTB
;
119 if ((env
->mmu_model
& POWERPC_MMU_BOOKE
) &&
120 ((value
>> MSR_GS
) & 1) != msr_gs
) {
121 cs
->interrupt_request
|= CPU_INTERRUPT_EXITTB
;
123 if (unlikely((env
->flags
& POWERPC_FLAG_TGPR
) &&
124 ((value
^ env
->msr
) & (1 << MSR_TGPR
)))) {
125 /* Swap temporary saved registers with GPRs */
126 hreg_swap_gpr_tgpr(env
);
128 if (unlikely((value
>> MSR_EP
) & 1) != msr_ep
) {
129 /* Change the exception prefix on PowerPC 601 */
130 env
->excp_prefix
= ((value
>> MSR_EP
) & 1) * 0xFFF00000;
132 /* If PR=1 then EE, IR and DR must be 1
134 * Note: We only enforce this on 64-bit processors. It appears that
135 * 32-bit implementations supports PR=1 and EE/DR/IR=0 and MacOS
138 if ((env
->insns_flags
& PPC_64B
) && ((value
>> MSR_PR
) & 1)) {
139 value
|= (1 << MSR_EE
) | (1 << MSR_DR
) | (1 << MSR_IR
);
143 hreg_compute_hflags(env
);
144 #if !defined(CONFIG_USER_ONLY)
145 if (unlikely(msr_pow
== 1)) {
146 if (!env
->pending_interrupts
&& (*env
->check_pow
)(env
)) {
156 #if !defined(CONFIG_USER_ONLY)
157 static inline void check_tlb_flush(CPUPPCState
*env
, bool global
)
159 CPUState
*cs
= CPU(ppc_env_get_cpu(env
));
160 if (env
->tlb_need_flush
& TLB_NEED_LOCAL_FLUSH
) {
162 env
->tlb_need_flush
&= ~TLB_NEED_LOCAL_FLUSH
;
165 /* Propagate TLB invalidations to other CPUs when the guest uses broadcast
166 * TLB invalidation instructions.
168 if (global
&& (env
->tlb_need_flush
& TLB_NEED_GLOBAL_FLUSH
)) {
170 CPU_FOREACH(other_cs
) {
171 if (other_cs
!= cs
) {
172 PowerPCCPU
*cpu
= POWERPC_CPU(other_cs
);
173 CPUPPCState
*other_env
= &cpu
->env
;
175 other_env
->tlb_need_flush
&= ~TLB_NEED_LOCAL_FLUSH
;
176 tlb_flush(other_cs
, 1);
179 env
->tlb_need_flush
&= ~TLB_NEED_GLOBAL_FLUSH
;
183 static inline void check_tlb_flush(CPUPPCState
*env
, bool global
) { }
186 #endif /* HELPER_REGS_H */