2 * Helpers for emulation of CP0-related MIPS instructions.
4 * Copyright (C) 2004-2005 Jocelyn Mayer
5 * Copyright (C) 2020 Wave Computing, Inc.
6 * Copyright (C) 2020 Aleksandar Markovic <amarkovic@wavecomp.com>
8 * This library is free software; you can redistribute it and/or
9 * modify it under the terms of the GNU Lesser General Public
10 * License as published by the Free Software Foundation; either
11 * version 2 of the License, or (at your option) any later version.
13 * This library is distributed in the hope that it will be useful,
14 * but WITHOUT ANY WARRANTY; without even the implied warranty of
15 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
16 * Lesser General Public License for more details.
18 * You should have received a copy of the GNU Lesser General Public
19 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
23 #include "qemu/osdep.h"
24 #include "qemu/main-loop.h"
27 #include "qemu/host-utils.h"
28 #include "exec/helper-proto.h"
29 #include "exec/exec-all.h"
30 #include "exec/cpu_ldst.h"
31 #include "exec/memop.h"
32 #include "sysemu/kvm.h"
35 #ifndef CONFIG_USER_ONLY
37 static bool mips_vpe_is_wfi(MIPSCPU
*c
)
39 CPUState
*cpu
= CPU(c
);
40 CPUMIPSState
*env
= &c
->env
;
43 * If the VPE is halted but otherwise active, it means it's waiting for
46 return cpu
->halted
&& mips_vpe_active(env
);
49 static bool mips_vp_is_wfi(MIPSCPU
*c
)
51 CPUState
*cpu
= CPU(c
);
52 CPUMIPSState
*env
= &c
->env
;
54 return cpu
->halted
&& mips_vp_active(env
);
57 static inline void mips_vpe_wake(MIPSCPU
*c
)
60 * Don't set ->halted = 0 directly, let it be done via cpu_has_work
61 * because there might be other conditions that state that c should
64 qemu_mutex_lock_iothread();
65 cpu_interrupt(CPU(c
), CPU_INTERRUPT_WAKE
);
66 qemu_mutex_unlock_iothread();
69 static inline void mips_vpe_sleep(MIPSCPU
*cpu
)
71 CPUState
*cs
= CPU(cpu
);
74 * The VPE was shut off, really go to bed.
75 * Reset any old _WAKE requests.
78 cpu_reset_interrupt(cs
, CPU_INTERRUPT_WAKE
);
81 static inline void mips_tc_wake(MIPSCPU
*cpu
, int tc
)
83 CPUMIPSState
*c
= &cpu
->env
;
85 /* FIXME: TC reschedule. */
86 if (mips_vpe_active(c
) && !mips_vpe_is_wfi(cpu
)) {
91 static inline void mips_tc_sleep(MIPSCPU
*cpu
, int tc
)
93 CPUMIPSState
*c
= &cpu
->env
;
95 /* FIXME: TC reschedule. */
96 if (!mips_vpe_active(c
)) {
103 * @env: CPU from which mapping is performed.
104 * @tc: Should point to an int with the value of the global TC index.
106 * This function will transform @tc into a local index within the
107 * returned #CPUMIPSState.
111 * FIXME: This code assumes that all VPEs have the same number of TCs,
112 * which depends on runtime setup. Can probably be fixed by
113 * walking the list of CPUMIPSStates.
115 static CPUMIPSState
*mips_cpu_map_tc(CPUMIPSState
*env
, int *tc
)
123 if (!(env
->CP0_VPEConf0
& (1 << CP0VPEC0_MVP
))) {
124 /* Not allowed to address other CPUs. */
125 *tc
= env
->current_tc
;
130 vpe_idx
= tc_idx
/ cs
->nr_threads
;
131 *tc
= tc_idx
% cs
->nr_threads
;
132 other_cs
= qemu_get_cpu(vpe_idx
);
133 if (other_cs
== NULL
) {
136 cpu
= MIPS_CPU(other_cs
);
141 * The per VPE CP0_Status register shares some fields with the per TC
142 * CP0_TCStatus registers. These fields are wired to the same registers,
143 * so changes to either of them should be reflected on both registers.
145 * Also, EntryHi shares the bottom 8 bit ASID with TCStauts.
147 * These helper call synchronizes the regs for a given cpu.
151 * Called for updates to CP0_Status. Defined in "cpu.h" for gdbstub.c.
152 * static inline void sync_c0_status(CPUMIPSState *env, CPUMIPSState *cpu,
156 /* Called for updates to CP0_TCStatus. */
157 static void sync_c0_tcstatus(CPUMIPSState
*cpu
, int tc
,
161 uint32_t tcu
, tmx
, tasid
, tksu
;
162 uint32_t mask
= ((1U << CP0St_CU3
)
169 tcu
= (v
>> CP0TCSt_TCU0
) & 0xf;
170 tmx
= (v
>> CP0TCSt_TMX
) & 0x1;
171 tasid
= v
& cpu
->CP0_EntryHi_ASID_mask
;
172 tksu
= (v
>> CP0TCSt_TKSU
) & 0x3;
174 status
= tcu
<< CP0St_CU0
;
175 status
|= tmx
<< CP0St_MX
;
176 status
|= tksu
<< CP0St_KSU
;
178 cpu
->CP0_Status
&= ~mask
;
179 cpu
->CP0_Status
|= status
;
181 /* Sync the TASID with EntryHi. */
182 cpu
->CP0_EntryHi
&= ~cpu
->CP0_EntryHi_ASID_mask
;
183 cpu
->CP0_EntryHi
|= tasid
;
188 /* Called for updates to CP0_EntryHi. */
189 static void sync_c0_entryhi(CPUMIPSState
*cpu
, int tc
)
192 uint32_t asid
, v
= cpu
->CP0_EntryHi
;
194 asid
= v
& cpu
->CP0_EntryHi_ASID_mask
;
196 if (tc
== cpu
->current_tc
) {
197 tcst
= &cpu
->active_tc
.CP0_TCStatus
;
199 tcst
= &cpu
->tcs
[tc
].CP0_TCStatus
;
202 *tcst
&= ~cpu
->CP0_EntryHi_ASID_mask
;
207 target_ulong
helper_mfc0_mvpcontrol(CPUMIPSState
*env
)
209 return env
->mvp
->CP0_MVPControl
;
212 target_ulong
helper_mfc0_mvpconf0(CPUMIPSState
*env
)
214 return env
->mvp
->CP0_MVPConf0
;
217 target_ulong
helper_mfc0_mvpconf1(CPUMIPSState
*env
)
219 return env
->mvp
->CP0_MVPConf1
;
222 target_ulong
helper_mfc0_random(CPUMIPSState
*env
)
224 return (int32_t)cpu_mips_get_random(env
);
227 target_ulong
helper_mfc0_tcstatus(CPUMIPSState
*env
)
229 return env
->active_tc
.CP0_TCStatus
;
232 target_ulong
helper_mftc0_tcstatus(CPUMIPSState
*env
)
234 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
235 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
237 if (other_tc
== other
->current_tc
) {
238 return other
->active_tc
.CP0_TCStatus
;
240 return other
->tcs
[other_tc
].CP0_TCStatus
;
244 target_ulong
helper_mfc0_tcbind(CPUMIPSState
*env
)
246 return env
->active_tc
.CP0_TCBind
;
249 target_ulong
helper_mftc0_tcbind(CPUMIPSState
*env
)
251 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
252 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
254 if (other_tc
== other
->current_tc
) {
255 return other
->active_tc
.CP0_TCBind
;
257 return other
->tcs
[other_tc
].CP0_TCBind
;
261 target_ulong
helper_mfc0_tcrestart(CPUMIPSState
*env
)
263 return env
->active_tc
.PC
;
266 target_ulong
helper_mftc0_tcrestart(CPUMIPSState
*env
)
268 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
269 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
271 if (other_tc
== other
->current_tc
) {
272 return other
->active_tc
.PC
;
274 return other
->tcs
[other_tc
].PC
;
278 target_ulong
helper_mfc0_tchalt(CPUMIPSState
*env
)
280 return env
->active_tc
.CP0_TCHalt
;
283 target_ulong
helper_mftc0_tchalt(CPUMIPSState
*env
)
285 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
286 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
288 if (other_tc
== other
->current_tc
) {
289 return other
->active_tc
.CP0_TCHalt
;
291 return other
->tcs
[other_tc
].CP0_TCHalt
;
295 target_ulong
helper_mfc0_tccontext(CPUMIPSState
*env
)
297 return env
->active_tc
.CP0_TCContext
;
300 target_ulong
helper_mftc0_tccontext(CPUMIPSState
*env
)
302 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
303 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
305 if (other_tc
== other
->current_tc
) {
306 return other
->active_tc
.CP0_TCContext
;
308 return other
->tcs
[other_tc
].CP0_TCContext
;
312 target_ulong
helper_mfc0_tcschedule(CPUMIPSState
*env
)
314 return env
->active_tc
.CP0_TCSchedule
;
317 target_ulong
helper_mftc0_tcschedule(CPUMIPSState
*env
)
319 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
320 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
322 if (other_tc
== other
->current_tc
) {
323 return other
->active_tc
.CP0_TCSchedule
;
325 return other
->tcs
[other_tc
].CP0_TCSchedule
;
329 target_ulong
helper_mfc0_tcschefback(CPUMIPSState
*env
)
331 return env
->active_tc
.CP0_TCScheFBack
;
334 target_ulong
helper_mftc0_tcschefback(CPUMIPSState
*env
)
336 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
337 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
339 if (other_tc
== other
->current_tc
) {
340 return other
->active_tc
.CP0_TCScheFBack
;
342 return other
->tcs
[other_tc
].CP0_TCScheFBack
;
346 target_ulong
helper_mfc0_count(CPUMIPSState
*env
)
348 return (int32_t)cpu_mips_get_count(env
);
351 target_ulong
helper_mfc0_saar(CPUMIPSState
*env
)
353 if ((env
->CP0_SAARI
& 0x3f) < 2) {
354 return (int32_t) env
->CP0_SAAR
[env
->CP0_SAARI
& 0x3f];
359 target_ulong
helper_mfhc0_saar(CPUMIPSState
*env
)
361 if ((env
->CP0_SAARI
& 0x3f) < 2) {
362 return env
->CP0_SAAR
[env
->CP0_SAARI
& 0x3f] >> 32;
367 target_ulong
helper_mftc0_entryhi(CPUMIPSState
*env
)
369 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
370 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
372 return other
->CP0_EntryHi
;
375 target_ulong
helper_mftc0_cause(CPUMIPSState
*env
)
377 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
379 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
381 if (other_tc
== other
->current_tc
) {
382 tccause
= other
->CP0_Cause
;
384 tccause
= other
->CP0_Cause
;
390 target_ulong
helper_mftc0_status(CPUMIPSState
*env
)
392 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
393 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
395 return other
->CP0_Status
;
398 target_ulong
helper_mfc0_lladdr(CPUMIPSState
*env
)
400 return (int32_t)(env
->CP0_LLAddr
>> env
->CP0_LLAddr_shift
);
403 target_ulong
helper_mfc0_maar(CPUMIPSState
*env
)
405 return (int32_t) env
->CP0_MAAR
[env
->CP0_MAARI
];
408 target_ulong
helper_mfhc0_maar(CPUMIPSState
*env
)
410 return env
->CP0_MAAR
[env
->CP0_MAARI
] >> 32;
413 target_ulong
helper_mfc0_watchlo(CPUMIPSState
*env
, uint32_t sel
)
415 return (int32_t)env
->CP0_WatchLo
[sel
];
418 target_ulong
helper_mfc0_watchhi(CPUMIPSState
*env
, uint32_t sel
)
420 return (int32_t) env
->CP0_WatchHi
[sel
];
423 target_ulong
helper_mfhc0_watchhi(CPUMIPSState
*env
, uint32_t sel
)
425 return env
->CP0_WatchHi
[sel
] >> 32;
428 target_ulong
helper_mfc0_debug(CPUMIPSState
*env
)
430 target_ulong t0
= env
->CP0_Debug
;
431 if (env
->hflags
& MIPS_HFLAG_DM
) {
438 target_ulong
helper_mftc0_debug(CPUMIPSState
*env
)
440 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
442 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
444 if (other_tc
== other
->current_tc
) {
445 tcstatus
= other
->active_tc
.CP0_Debug_tcstatus
;
447 tcstatus
= other
->tcs
[other_tc
].CP0_Debug_tcstatus
;
450 /* XXX: Might be wrong, check with EJTAG spec. */
451 return (other
->CP0_Debug
& ~((1 << CP0DB_SSt
) | (1 << CP0DB_Halt
))) |
452 (tcstatus
& ((1 << CP0DB_SSt
) | (1 << CP0DB_Halt
)));
455 #if defined(TARGET_MIPS64)
456 target_ulong
helper_dmfc0_tcrestart(CPUMIPSState
*env
)
458 return env
->active_tc
.PC
;
461 target_ulong
helper_dmfc0_tchalt(CPUMIPSState
*env
)
463 return env
->active_tc
.CP0_TCHalt
;
466 target_ulong
helper_dmfc0_tccontext(CPUMIPSState
*env
)
468 return env
->active_tc
.CP0_TCContext
;
471 target_ulong
helper_dmfc0_tcschedule(CPUMIPSState
*env
)
473 return env
->active_tc
.CP0_TCSchedule
;
476 target_ulong
helper_dmfc0_tcschefback(CPUMIPSState
*env
)
478 return env
->active_tc
.CP0_TCScheFBack
;
481 target_ulong
helper_dmfc0_lladdr(CPUMIPSState
*env
)
483 return env
->CP0_LLAddr
>> env
->CP0_LLAddr_shift
;
486 target_ulong
helper_dmfc0_maar(CPUMIPSState
*env
)
488 return env
->CP0_MAAR
[env
->CP0_MAARI
];
491 target_ulong
helper_dmfc0_watchlo(CPUMIPSState
*env
, uint32_t sel
)
493 return env
->CP0_WatchLo
[sel
];
496 target_ulong
helper_dmfc0_watchhi(CPUMIPSState
*env
, uint32_t sel
)
498 return env
->CP0_WatchHi
[sel
];
501 target_ulong
helper_dmfc0_saar(CPUMIPSState
*env
)
503 if ((env
->CP0_SAARI
& 0x3f) < 2) {
504 return env
->CP0_SAAR
[env
->CP0_SAARI
& 0x3f];
508 #endif /* TARGET_MIPS64 */
510 void helper_mtc0_index(CPUMIPSState
*env
, target_ulong arg1
)
512 uint32_t index_p
= env
->CP0_Index
& 0x80000000;
513 uint32_t tlb_index
= arg1
& 0x7fffffff;
514 if (tlb_index
< env
->tlb
->nb_tlb
) {
515 if (env
->insn_flags
& ISA_MIPS32R6
) {
516 index_p
|= arg1
& 0x80000000;
518 env
->CP0_Index
= index_p
| tlb_index
;
522 void helper_mtc0_mvpcontrol(CPUMIPSState
*env
, target_ulong arg1
)
527 if (env
->CP0_VPEConf0
& (1 << CP0VPEC0_MVP
)) {
528 mask
|= (1 << CP0MVPCo_CPA
) | (1 << CP0MVPCo_VPC
) |
531 if (env
->mvp
->CP0_MVPControl
& (1 << CP0MVPCo_VPC
)) {
532 mask
|= (1 << CP0MVPCo_STLB
);
534 newval
= (env
->mvp
->CP0_MVPControl
& ~mask
) | (arg1
& mask
);
536 /* TODO: Enable/disable shared TLB, enable/disable VPEs. */
538 env
->mvp
->CP0_MVPControl
= newval
;
541 void helper_mtc0_vpecontrol(CPUMIPSState
*env
, target_ulong arg1
)
546 mask
= (1 << CP0VPECo_YSI
) | (1 << CP0VPECo_GSI
) |
547 (1 << CP0VPECo_TE
) | (0xff << CP0VPECo_TargTC
);
548 newval
= (env
->CP0_VPEControl
& ~mask
) | (arg1
& mask
);
551 * Yield scheduler intercept not implemented.
552 * Gating storage scheduler intercept not implemented.
555 /* TODO: Enable/disable TCs. */
557 env
->CP0_VPEControl
= newval
;
560 void helper_mttc0_vpecontrol(CPUMIPSState
*env
, target_ulong arg1
)
562 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
563 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
567 mask
= (1 << CP0VPECo_YSI
) | (1 << CP0VPECo_GSI
) |
568 (1 << CP0VPECo_TE
) | (0xff << CP0VPECo_TargTC
);
569 newval
= (other
->CP0_VPEControl
& ~mask
) | (arg1
& mask
);
571 /* TODO: Enable/disable TCs. */
573 other
->CP0_VPEControl
= newval
;
576 target_ulong
helper_mftc0_vpecontrol(CPUMIPSState
*env
)
578 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
579 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
580 /* FIXME: Mask away return zero on read bits. */
581 return other
->CP0_VPEControl
;
584 target_ulong
helper_mftc0_vpeconf0(CPUMIPSState
*env
)
586 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
587 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
589 return other
->CP0_VPEConf0
;
592 void helper_mtc0_vpeconf0(CPUMIPSState
*env
, target_ulong arg1
)
597 if (env
->CP0_VPEConf0
& (1 << CP0VPEC0_MVP
)) {
598 if (env
->CP0_VPEConf0
& (1 << CP0VPEC0_VPA
)) {
599 mask
|= (0xff << CP0VPEC0_XTC
);
601 mask
|= (1 << CP0VPEC0_MVP
) | (1 << CP0VPEC0_VPA
);
603 newval
= (env
->CP0_VPEConf0
& ~mask
) | (arg1
& mask
);
605 /* TODO: TC exclusive handling due to ERL/EXL. */
607 env
->CP0_VPEConf0
= newval
;
610 void helper_mttc0_vpeconf0(CPUMIPSState
*env
, target_ulong arg1
)
612 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
613 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
617 mask
|= (1 << CP0VPEC0_MVP
) | (1 << CP0VPEC0_VPA
);
618 newval
= (other
->CP0_VPEConf0
& ~mask
) | (arg1
& mask
);
620 /* TODO: TC exclusive handling due to ERL/EXL. */
621 other
->CP0_VPEConf0
= newval
;
624 void helper_mtc0_vpeconf1(CPUMIPSState
*env
, target_ulong arg1
)
629 if (env
->mvp
->CP0_MVPControl
& (1 << CP0MVPCo_VPC
))
630 mask
|= (0xff << CP0VPEC1_NCX
) | (0xff << CP0VPEC1_NCP2
) |
631 (0xff << CP0VPEC1_NCP1
);
632 newval
= (env
->CP0_VPEConf1
& ~mask
) | (arg1
& mask
);
634 /* UDI not implemented. */
635 /* CP2 not implemented. */
637 /* TODO: Handle FPU (CP1) binding. */
639 env
->CP0_VPEConf1
= newval
;
642 void helper_mtc0_yqmask(CPUMIPSState
*env
, target_ulong arg1
)
644 /* Yield qualifier inputs not implemented. */
645 env
->CP0_YQMask
= 0x00000000;
648 void helper_mtc0_vpeopt(CPUMIPSState
*env
, target_ulong arg1
)
650 env
->CP0_VPEOpt
= arg1
& 0x0000ffff;
653 #define MTC0_ENTRYLO_MASK(env) ((env->PAMask >> 6) & 0x3FFFFFFF)
655 void helper_mtc0_entrylo0(CPUMIPSState
*env
, target_ulong arg1
)
657 /* 1k pages not implemented */
658 target_ulong rxi
= arg1
& (env
->CP0_PageGrain
& (3u << CP0PG_XIE
));
659 env
->CP0_EntryLo0
= (arg1
& MTC0_ENTRYLO_MASK(env
))
660 | (rxi
<< (CP0EnLo_XI
- 30));
663 #if defined(TARGET_MIPS64)
664 #define DMTC0_ENTRYLO_MASK(env) (env->PAMask >> 6)
666 void helper_dmtc0_entrylo0(CPUMIPSState
*env
, uint64_t arg1
)
668 uint64_t rxi
= arg1
& ((env
->CP0_PageGrain
& (3ull << CP0PG_XIE
)) << 32);
669 env
->CP0_EntryLo0
= (arg1
& DMTC0_ENTRYLO_MASK(env
)) | rxi
;
673 void helper_mtc0_tcstatus(CPUMIPSState
*env
, target_ulong arg1
)
675 uint32_t mask
= env
->CP0_TCStatus_rw_bitmask
;
678 newval
= (env
->active_tc
.CP0_TCStatus
& ~mask
) | (arg1
& mask
);
680 env
->active_tc
.CP0_TCStatus
= newval
;
681 sync_c0_tcstatus(env
, env
->current_tc
, newval
);
684 void helper_mttc0_tcstatus(CPUMIPSState
*env
, target_ulong arg1
)
686 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
687 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
689 if (other_tc
== other
->current_tc
) {
690 other
->active_tc
.CP0_TCStatus
= arg1
;
692 other
->tcs
[other_tc
].CP0_TCStatus
= arg1
;
694 sync_c0_tcstatus(other
, other_tc
, arg1
);
697 void helper_mtc0_tcbind(CPUMIPSState
*env
, target_ulong arg1
)
699 uint32_t mask
= (1 << CP0TCBd_TBE
);
702 if (env
->mvp
->CP0_MVPControl
& (1 << CP0MVPCo_VPC
)) {
703 mask
|= (1 << CP0TCBd_CurVPE
);
705 newval
= (env
->active_tc
.CP0_TCBind
& ~mask
) | (arg1
& mask
);
706 env
->active_tc
.CP0_TCBind
= newval
;
709 void helper_mttc0_tcbind(CPUMIPSState
*env
, target_ulong arg1
)
711 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
712 uint32_t mask
= (1 << CP0TCBd_TBE
);
714 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
716 if (other
->mvp
->CP0_MVPControl
& (1 << CP0MVPCo_VPC
)) {
717 mask
|= (1 << CP0TCBd_CurVPE
);
719 if (other_tc
== other
->current_tc
) {
720 newval
= (other
->active_tc
.CP0_TCBind
& ~mask
) | (arg1
& mask
);
721 other
->active_tc
.CP0_TCBind
= newval
;
723 newval
= (other
->tcs
[other_tc
].CP0_TCBind
& ~mask
) | (arg1
& mask
);
724 other
->tcs
[other_tc
].CP0_TCBind
= newval
;
728 void helper_mtc0_tcrestart(CPUMIPSState
*env
, target_ulong arg1
)
730 env
->active_tc
.PC
= arg1
;
731 env
->active_tc
.CP0_TCStatus
&= ~(1 << CP0TCSt_TDS
);
734 /* MIPS16 not implemented. */
737 void helper_mttc0_tcrestart(CPUMIPSState
*env
, target_ulong arg1
)
739 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
740 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
742 if (other_tc
== other
->current_tc
) {
743 other
->active_tc
.PC
= arg1
;
744 other
->active_tc
.CP0_TCStatus
&= ~(1 << CP0TCSt_TDS
);
745 other
->CP0_LLAddr
= 0;
747 /* MIPS16 not implemented. */
749 other
->tcs
[other_tc
].PC
= arg1
;
750 other
->tcs
[other_tc
].CP0_TCStatus
&= ~(1 << CP0TCSt_TDS
);
751 other
->CP0_LLAddr
= 0;
753 /* MIPS16 not implemented. */
757 void helper_mtc0_tchalt(CPUMIPSState
*env
, target_ulong arg1
)
759 MIPSCPU
*cpu
= env_archcpu(env
);
761 env
->active_tc
.CP0_TCHalt
= arg1
& 0x1;
763 /* TODO: Halt TC / Restart (if allocated+active) TC. */
764 if (env
->active_tc
.CP0_TCHalt
& 1) {
765 mips_tc_sleep(cpu
, env
->current_tc
);
767 mips_tc_wake(cpu
, env
->current_tc
);
771 void helper_mttc0_tchalt(CPUMIPSState
*env
, target_ulong arg1
)
773 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
774 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
775 MIPSCPU
*other_cpu
= env_archcpu(other
);
777 /* TODO: Halt TC / Restart (if allocated+active) TC. */
779 if (other_tc
== other
->current_tc
) {
780 other
->active_tc
.CP0_TCHalt
= arg1
;
782 other
->tcs
[other_tc
].CP0_TCHalt
= arg1
;
786 mips_tc_sleep(other_cpu
, other_tc
);
788 mips_tc_wake(other_cpu
, other_tc
);
792 void helper_mtc0_tccontext(CPUMIPSState
*env
, target_ulong arg1
)
794 env
->active_tc
.CP0_TCContext
= arg1
;
797 void helper_mttc0_tccontext(CPUMIPSState
*env
, target_ulong arg1
)
799 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
800 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
802 if (other_tc
== other
->current_tc
) {
803 other
->active_tc
.CP0_TCContext
= arg1
;
805 other
->tcs
[other_tc
].CP0_TCContext
= arg1
;
809 void helper_mtc0_tcschedule(CPUMIPSState
*env
, target_ulong arg1
)
811 env
->active_tc
.CP0_TCSchedule
= arg1
;
814 void helper_mttc0_tcschedule(CPUMIPSState
*env
, target_ulong arg1
)
816 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
817 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
819 if (other_tc
== other
->current_tc
) {
820 other
->active_tc
.CP0_TCSchedule
= arg1
;
822 other
->tcs
[other_tc
].CP0_TCSchedule
= arg1
;
826 void helper_mtc0_tcschefback(CPUMIPSState
*env
, target_ulong arg1
)
828 env
->active_tc
.CP0_TCScheFBack
= arg1
;
831 void helper_mttc0_tcschefback(CPUMIPSState
*env
, target_ulong arg1
)
833 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
834 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
836 if (other_tc
== other
->current_tc
) {
837 other
->active_tc
.CP0_TCScheFBack
= arg1
;
839 other
->tcs
[other_tc
].CP0_TCScheFBack
= arg1
;
843 void helper_mtc0_entrylo1(CPUMIPSState
*env
, target_ulong arg1
)
845 /* 1k pages not implemented */
846 target_ulong rxi
= arg1
& (env
->CP0_PageGrain
& (3u << CP0PG_XIE
));
847 env
->CP0_EntryLo1
= (arg1
& MTC0_ENTRYLO_MASK(env
))
848 | (rxi
<< (CP0EnLo_XI
- 30));
851 #if defined(TARGET_MIPS64)
852 void helper_dmtc0_entrylo1(CPUMIPSState
*env
, uint64_t arg1
)
854 uint64_t rxi
= arg1
& ((env
->CP0_PageGrain
& (3ull << CP0PG_XIE
)) << 32);
855 env
->CP0_EntryLo1
= (arg1
& DMTC0_ENTRYLO_MASK(env
)) | rxi
;
859 void helper_mtc0_context(CPUMIPSState
*env
, target_ulong arg1
)
861 env
->CP0_Context
= (env
->CP0_Context
& 0x007FFFFF) | (arg1
& ~0x007FFFFF);
864 void helper_mtc0_memorymapid(CPUMIPSState
*env
, target_ulong arg1
)
867 old
= env
->CP0_MemoryMapID
;
868 env
->CP0_MemoryMapID
= (int32_t) arg1
;
869 /* If the MemoryMapID changes, flush qemu's TLB. */
870 if (old
!= env
->CP0_MemoryMapID
) {
871 cpu_mips_tlb_flush(env
);
875 void update_pagemask(CPUMIPSState
*env
, target_ulong arg1
, int32_t *pagemask
)
877 uint64_t mask
= arg1
>> (TARGET_PAGE_BITS
+ 1);
878 if (!(env
->insn_flags
& ISA_MIPS32R6
) || (arg1
== ~0) ||
879 (mask
== 0x0000 || mask
== 0x0003 || mask
== 0x000F ||
880 mask
== 0x003F || mask
== 0x00FF || mask
== 0x03FF ||
881 mask
== 0x0FFF || mask
== 0x3FFF || mask
== 0xFFFF)) {
882 env
->CP0_PageMask
= arg1
& (0x1FFFFFFF & (TARGET_PAGE_MASK
<< 1));
886 void helper_mtc0_pagemask(CPUMIPSState
*env
, target_ulong arg1
)
888 update_pagemask(env
, arg1
, &env
->CP0_PageMask
);
891 void helper_mtc0_pagegrain(CPUMIPSState
*env
, target_ulong arg1
)
893 /* SmartMIPS not implemented */
894 /* 1k pages not implemented */
895 env
->CP0_PageGrain
= (arg1
& env
->CP0_PageGrain_rw_bitmask
) |
896 (env
->CP0_PageGrain
& ~env
->CP0_PageGrain_rw_bitmask
);
901 void helper_mtc0_segctl0(CPUMIPSState
*env
, target_ulong arg1
)
903 CPUState
*cs
= env_cpu(env
);
905 env
->CP0_SegCtl0
= arg1
& CP0SC0_MASK
;
909 void helper_mtc0_segctl1(CPUMIPSState
*env
, target_ulong arg1
)
911 CPUState
*cs
= env_cpu(env
);
913 env
->CP0_SegCtl1
= arg1
& CP0SC1_MASK
;
917 void helper_mtc0_segctl2(CPUMIPSState
*env
, target_ulong arg1
)
919 CPUState
*cs
= env_cpu(env
);
921 env
->CP0_SegCtl2
= arg1
& CP0SC2_MASK
;
925 void helper_mtc0_pwfield(CPUMIPSState
*env
, target_ulong arg1
)
927 #if defined(TARGET_MIPS64)
928 uint64_t mask
= 0x3F3FFFFFFFULL
;
929 uint32_t old_ptei
= (env
->CP0_PWField
>> CP0PF_PTEI
) & 0x3FULL
;
930 uint32_t new_ptei
= (arg1
>> CP0PF_PTEI
) & 0x3FULL
;
932 if ((env
->insn_flags
& ISA_MIPS32R6
)) {
933 if (((arg1
>> CP0PF_BDI
) & 0x3FULL
) < 12) {
934 mask
&= ~(0x3FULL
<< CP0PF_BDI
);
936 if (((arg1
>> CP0PF_GDI
) & 0x3FULL
) < 12) {
937 mask
&= ~(0x3FULL
<< CP0PF_GDI
);
939 if (((arg1
>> CP0PF_UDI
) & 0x3FULL
) < 12) {
940 mask
&= ~(0x3FULL
<< CP0PF_UDI
);
942 if (((arg1
>> CP0PF_MDI
) & 0x3FULL
) < 12) {
943 mask
&= ~(0x3FULL
<< CP0PF_MDI
);
945 if (((arg1
>> CP0PF_PTI
) & 0x3FULL
) < 12) {
946 mask
&= ~(0x3FULL
<< CP0PF_PTI
);
949 env
->CP0_PWField
= arg1
& mask
;
951 if ((new_ptei
>= 32) ||
952 ((env
->insn_flags
& ISA_MIPS32R6
) &&
953 (new_ptei
== 0 || new_ptei
== 1))) {
954 env
->CP0_PWField
= (env
->CP0_PWField
& ~0x3FULL
) |
955 (old_ptei
<< CP0PF_PTEI
);
958 uint32_t mask
= 0x3FFFFFFF;
959 uint32_t old_ptew
= (env
->CP0_PWField
>> CP0PF_PTEW
) & 0x3F;
960 uint32_t new_ptew
= (arg1
>> CP0PF_PTEW
) & 0x3F;
962 if ((env
->insn_flags
& ISA_MIPS32R6
)) {
963 if (((arg1
>> CP0PF_GDW
) & 0x3F) < 12) {
964 mask
&= ~(0x3F << CP0PF_GDW
);
966 if (((arg1
>> CP0PF_UDW
) & 0x3F) < 12) {
967 mask
&= ~(0x3F << CP0PF_UDW
);
969 if (((arg1
>> CP0PF_MDW
) & 0x3F) < 12) {
970 mask
&= ~(0x3F << CP0PF_MDW
);
972 if (((arg1
>> CP0PF_PTW
) & 0x3F) < 12) {
973 mask
&= ~(0x3F << CP0PF_PTW
);
976 env
->CP0_PWField
= arg1
& mask
;
978 if ((new_ptew
>= 32) ||
979 ((env
->insn_flags
& ISA_MIPS32R6
) &&
980 (new_ptew
== 0 || new_ptew
== 1))) {
981 env
->CP0_PWField
= (env
->CP0_PWField
& ~0x3F) |
982 (old_ptew
<< CP0PF_PTEW
);
987 void helper_mtc0_pwsize(CPUMIPSState
*env
, target_ulong arg1
)
989 #if defined(TARGET_MIPS64)
990 env
->CP0_PWSize
= arg1
& 0x3F7FFFFFFFULL
;
992 env
->CP0_PWSize
= arg1
& 0x3FFFFFFF;
996 void helper_mtc0_wired(CPUMIPSState
*env
, target_ulong arg1
)
998 if (env
->insn_flags
& ISA_MIPS32R6
) {
999 if (arg1
< env
->tlb
->nb_tlb
) {
1000 env
->CP0_Wired
= arg1
;
1003 env
->CP0_Wired
= arg1
% env
->tlb
->nb_tlb
;
1007 void helper_mtc0_pwctl(CPUMIPSState
*env
, target_ulong arg1
)
1009 #if defined(TARGET_MIPS64)
1010 /* PWEn = 0. Hardware page table walking is not implemented. */
1011 env
->CP0_PWCtl
= (env
->CP0_PWCtl
& 0x000000C0) | (arg1
& 0x5C00003F);
1013 env
->CP0_PWCtl
= (arg1
& 0x800000FF);
1017 void helper_mtc0_srsconf0(CPUMIPSState
*env
, target_ulong arg1
)
1019 env
->CP0_SRSConf0
|= arg1
& env
->CP0_SRSConf0_rw_bitmask
;
1022 void helper_mtc0_srsconf1(CPUMIPSState
*env
, target_ulong arg1
)
1024 env
->CP0_SRSConf1
|= arg1
& env
->CP0_SRSConf1_rw_bitmask
;
1027 void helper_mtc0_srsconf2(CPUMIPSState
*env
, target_ulong arg1
)
1029 env
->CP0_SRSConf2
|= arg1
& env
->CP0_SRSConf2_rw_bitmask
;
1032 void helper_mtc0_srsconf3(CPUMIPSState
*env
, target_ulong arg1
)
1034 env
->CP0_SRSConf3
|= arg1
& env
->CP0_SRSConf3_rw_bitmask
;
1037 void helper_mtc0_srsconf4(CPUMIPSState
*env
, target_ulong arg1
)
1039 env
->CP0_SRSConf4
|= arg1
& env
->CP0_SRSConf4_rw_bitmask
;
1042 void helper_mtc0_hwrena(CPUMIPSState
*env
, target_ulong arg1
)
1044 uint32_t mask
= 0x0000000F;
1046 if ((env
->CP0_Config1
& (1 << CP0C1_PC
)) &&
1047 (env
->insn_flags
& ISA_MIPS32R6
)) {
1050 if (env
->insn_flags
& ISA_MIPS32R6
) {
1053 if (env
->CP0_Config3
& (1 << CP0C3_ULRI
)) {
1056 if (arg1
& (1 << 29)) {
1057 env
->hflags
|= MIPS_HFLAG_HWRENA_ULR
;
1059 env
->hflags
&= ~MIPS_HFLAG_HWRENA_ULR
;
1063 env
->CP0_HWREna
= arg1
& mask
;
1066 void helper_mtc0_count(CPUMIPSState
*env
, target_ulong arg1
)
1068 cpu_mips_store_count(env
, arg1
);
1071 void helper_mtc0_saari(CPUMIPSState
*env
, target_ulong arg1
)
1073 uint32_t target
= arg1
& 0x3f;
1075 env
->CP0_SAARI
= target
;
1079 void helper_mtc0_saar(CPUMIPSState
*env
, target_ulong arg1
)
1081 uint32_t target
= env
->CP0_SAARI
& 0x3f;
1083 env
->CP0_SAAR
[target
] = arg1
& 0x00000ffffffff03fULL
;
1087 itc_reconfigure(env
->itu
);
1094 void helper_mthc0_saar(CPUMIPSState
*env
, target_ulong arg1
)
1096 uint32_t target
= env
->CP0_SAARI
& 0x3f;
1098 env
->CP0_SAAR
[target
] =
1099 (((uint64_t) arg1
<< 32) & 0x00000fff00000000ULL
) |
1100 (env
->CP0_SAAR
[target
] & 0x00000000ffffffffULL
);
1104 itc_reconfigure(env
->itu
);
1111 void helper_mtc0_entryhi(CPUMIPSState
*env
, target_ulong arg1
)
1113 target_ulong old
, val
, mask
;
1114 mask
= (TARGET_PAGE_MASK
<< 1) | env
->CP0_EntryHi_ASID_mask
;
1115 if (((env
->CP0_Config4
>> CP0C4_IE
) & 0x3) >= 2) {
1116 mask
|= 1 << CP0EnHi_EHINV
;
1119 /* 1k pages not implemented */
1120 #if defined(TARGET_MIPS64)
1121 if (env
->insn_flags
& ISA_MIPS32R6
) {
1122 int entryhi_r
= extract64(arg1
, 62, 2);
1123 int config0_at
= extract32(env
->CP0_Config0
, 13, 2);
1124 bool no_supervisor
= (env
->CP0_Status_rw_bitmask
& 0x8) == 0;
1125 if ((entryhi_r
== 2) ||
1126 (entryhi_r
== 1 && (no_supervisor
|| config0_at
== 1))) {
1127 /* skip EntryHi.R field if new value is reserved */
1128 mask
&= ~(0x3ull
<< 62);
1131 mask
&= env
->SEGMask
;
1133 old
= env
->CP0_EntryHi
;
1134 val
= (arg1
& mask
) | (old
& ~mask
);
1135 env
->CP0_EntryHi
= val
;
1136 if (env
->CP0_Config3
& (1 << CP0C3_MT
)) {
1137 sync_c0_entryhi(env
, env
->current_tc
);
1139 /* If the ASID changes, flush qemu's TLB. */
1140 if ((old
& env
->CP0_EntryHi_ASID_mask
) !=
1141 (val
& env
->CP0_EntryHi_ASID_mask
)) {
1142 tlb_flush(env_cpu(env
));
1146 void helper_mttc0_entryhi(CPUMIPSState
*env
, target_ulong arg1
)
1148 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
1149 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
1151 other
->CP0_EntryHi
= arg1
;
1152 sync_c0_entryhi(other
, other_tc
);
1155 void helper_mtc0_compare(CPUMIPSState
*env
, target_ulong arg1
)
1157 cpu_mips_store_compare(env
, arg1
);
1160 void helper_mtc0_status(CPUMIPSState
*env
, target_ulong arg1
)
1164 old
= env
->CP0_Status
;
1165 cpu_mips_store_status(env
, arg1
);
1166 val
= env
->CP0_Status
;
1168 if (qemu_loglevel_mask(CPU_LOG_EXEC
)) {
1169 qemu_log("Status %08x (%08x) => %08x (%08x) Cause %08x",
1170 old
, old
& env
->CP0_Cause
& CP0Ca_IP_mask
,
1171 val
, val
& env
->CP0_Cause
& CP0Ca_IP_mask
,
1173 switch (cpu_mmu_index(env
, false)) {
1175 qemu_log(", ERL\n");
1187 cpu_abort(env_cpu(env
), "Invalid MMU mode!\n");
1193 void helper_mttc0_status(CPUMIPSState
*env
, target_ulong arg1
)
1195 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
1196 uint32_t mask
= env
->CP0_Status_rw_bitmask
& ~0xf1000018;
1197 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
1199 other
->CP0_Status
= (other
->CP0_Status
& ~mask
) | (arg1
& mask
);
1200 sync_c0_status(env
, other
, other_tc
);
1203 void helper_mtc0_intctl(CPUMIPSState
*env
, target_ulong arg1
)
1205 env
->CP0_IntCtl
= (env
->CP0_IntCtl
& ~0x000003e0) | (arg1
& 0x000003e0);
1208 void helper_mtc0_srsctl(CPUMIPSState
*env
, target_ulong arg1
)
1210 uint32_t mask
= (0xf << CP0SRSCtl_ESS
) | (0xf << CP0SRSCtl_PSS
);
1211 env
->CP0_SRSCtl
= (env
->CP0_SRSCtl
& ~mask
) | (arg1
& mask
);
1214 void helper_mtc0_cause(CPUMIPSState
*env
, target_ulong arg1
)
1216 cpu_mips_store_cause(env
, arg1
);
1219 void helper_mttc0_cause(CPUMIPSState
*env
, target_ulong arg1
)
1221 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
1222 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
1224 cpu_mips_store_cause(other
, arg1
);
1227 target_ulong
helper_mftc0_epc(CPUMIPSState
*env
)
1229 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
1230 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
1232 return other
->CP0_EPC
;
1235 target_ulong
helper_mftc0_ebase(CPUMIPSState
*env
)
1237 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
1238 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
1240 return other
->CP0_EBase
;
1243 void helper_mtc0_ebase(CPUMIPSState
*env
, target_ulong arg1
)
1245 target_ulong mask
= 0x3FFFF000 | env
->CP0_EBaseWG_rw_bitmask
;
1246 if (arg1
& env
->CP0_EBaseWG_rw_bitmask
) {
1247 mask
|= ~0x3FFFFFFF;
1249 env
->CP0_EBase
= (env
->CP0_EBase
& ~mask
) | (arg1
& mask
);
1252 void helper_mttc0_ebase(CPUMIPSState
*env
, target_ulong arg1
)
1254 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
1255 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
1256 target_ulong mask
= 0x3FFFF000 | env
->CP0_EBaseWG_rw_bitmask
;
1257 if (arg1
& env
->CP0_EBaseWG_rw_bitmask
) {
1258 mask
|= ~0x3FFFFFFF;
1260 other
->CP0_EBase
= (other
->CP0_EBase
& ~mask
) | (arg1
& mask
);
1263 target_ulong
helper_mftc0_configx(CPUMIPSState
*env
, target_ulong idx
)
1265 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
1266 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
1269 case 0: return other
->CP0_Config0
;
1270 case 1: return other
->CP0_Config1
;
1271 case 2: return other
->CP0_Config2
;
1272 case 3: return other
->CP0_Config3
;
1273 /* 4 and 5 are reserved. */
1274 case 6: return other
->CP0_Config6
;
1275 case 7: return other
->CP0_Config7
;
1282 void helper_mtc0_config0(CPUMIPSState
*env
, target_ulong arg1
)
1284 env
->CP0_Config0
= (env
->CP0_Config0
& 0x81FFFFF8) | (arg1
& 0x00000007);
1287 void helper_mtc0_config2(CPUMIPSState
*env
, target_ulong arg1
)
1289 /* tertiary/secondary caches not implemented */
1290 env
->CP0_Config2
= (env
->CP0_Config2
& 0x8FFF0FFF);
1293 void helper_mtc0_config3(CPUMIPSState
*env
, target_ulong arg1
)
1295 if (env
->insn_flags
& ASE_MICROMIPS
) {
1296 env
->CP0_Config3
= (env
->CP0_Config3
& ~(1 << CP0C3_ISA_ON_EXC
)) |
1297 (arg1
& (1 << CP0C3_ISA_ON_EXC
));
1301 void helper_mtc0_config4(CPUMIPSState
*env
, target_ulong arg1
)
1303 env
->CP0_Config4
= (env
->CP0_Config4
& (~env
->CP0_Config4_rw_bitmask
)) |
1304 (arg1
& env
->CP0_Config4_rw_bitmask
);
1307 void helper_mtc0_config5(CPUMIPSState
*env
, target_ulong arg1
)
1309 env
->CP0_Config5
= (env
->CP0_Config5
& (~env
->CP0_Config5_rw_bitmask
)) |
1310 (arg1
& env
->CP0_Config5_rw_bitmask
);
1311 env
->CP0_EntryHi_ASID_mask
= (env
->CP0_Config5
& (1 << CP0C5_MI
)) ?
1312 0x0 : (env
->CP0_Config4
& (1 << CP0C4_AE
)) ? 0x3ff : 0xff;
1313 compute_hflags(env
);
1316 void helper_mtc0_lladdr(CPUMIPSState
*env
, target_ulong arg1
)
1318 target_long mask
= env
->CP0_LLAddr_rw_bitmask
;
1319 arg1
= arg1
<< env
->CP0_LLAddr_shift
;
1320 env
->CP0_LLAddr
= (env
->CP0_LLAddr
& ~mask
) | (arg1
& mask
);
1323 #define MTC0_MAAR_MASK(env) \
1324 ((0x1ULL << 63) | ((env->PAMask >> 4) & ~0xFFFull) | 0x3)
1326 void helper_mtc0_maar(CPUMIPSState
*env
, target_ulong arg1
)
1328 env
->CP0_MAAR
[env
->CP0_MAARI
] = arg1
& MTC0_MAAR_MASK(env
);
1331 void helper_mthc0_maar(CPUMIPSState
*env
, target_ulong arg1
)
1333 env
->CP0_MAAR
[env
->CP0_MAARI
] =
1334 (((uint64_t) arg1
<< 32) & MTC0_MAAR_MASK(env
)) |
1335 (env
->CP0_MAAR
[env
->CP0_MAARI
] & 0x00000000ffffffffULL
);
1338 void helper_mtc0_maari(CPUMIPSState
*env
, target_ulong arg1
)
1340 int index
= arg1
& 0x3f;
1341 if (index
== 0x3f) {
1343 * Software may write all ones to INDEX to determine the
1344 * maximum value supported.
1346 env
->CP0_MAARI
= MIPS_MAAR_MAX
- 1;
1347 } else if (index
< MIPS_MAAR_MAX
) {
1348 env
->CP0_MAARI
= index
;
1351 * Other than the all ones, if the value written is not supported,
1352 * then INDEX is unchanged from its previous value.
1356 void helper_mtc0_watchlo(CPUMIPSState
*env
, target_ulong arg1
, uint32_t sel
)
1359 * Watch exceptions for instructions, data loads, data stores
1362 env
->CP0_WatchLo
[sel
] = (arg1
& ~0x7);
1365 void helper_mtc0_watchhi(CPUMIPSState
*env
, target_ulong arg1
, uint32_t sel
)
1367 uint64_t mask
= 0x40000FF8 | (env
->CP0_EntryHi_ASID_mask
<< CP0WH_ASID
);
1368 if ((env
->CP0_Config5
>> CP0C5_MI
) & 1) {
1369 mask
|= 0xFFFFFFFF00000000ULL
; /* MMID */
1371 env
->CP0_WatchHi
[sel
] = arg1
& mask
;
1372 env
->CP0_WatchHi
[sel
] &= ~(env
->CP0_WatchHi
[sel
] & arg1
& 0x7);
1375 void helper_mthc0_watchhi(CPUMIPSState
*env
, target_ulong arg1
, uint32_t sel
)
1377 env
->CP0_WatchHi
[sel
] = ((uint64_t) (arg1
) << 32) |
1378 (env
->CP0_WatchHi
[sel
] & 0x00000000ffffffffULL
);
1381 void helper_mtc0_xcontext(CPUMIPSState
*env
, target_ulong arg1
)
1383 target_ulong mask
= (1ULL << (env
->SEGBITS
- 7)) - 1;
1384 env
->CP0_XContext
= (env
->CP0_XContext
& mask
) | (arg1
& ~mask
);
1387 void helper_mtc0_framemask(CPUMIPSState
*env
, target_ulong arg1
)
1389 env
->CP0_Framemask
= arg1
; /* XXX */
1392 void helper_mtc0_debug(CPUMIPSState
*env
, target_ulong arg1
)
1394 env
->CP0_Debug
= (env
->CP0_Debug
& 0x8C03FC1F) | (arg1
& 0x13300120);
1395 if (arg1
& (1 << CP0DB_DM
)) {
1396 env
->hflags
|= MIPS_HFLAG_DM
;
1398 env
->hflags
&= ~MIPS_HFLAG_DM
;
1402 void helper_mttc0_debug(CPUMIPSState
*env
, target_ulong arg1
)
1404 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
1405 uint32_t val
= arg1
& ((1 << CP0DB_SSt
) | (1 << CP0DB_Halt
));
1406 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
1408 /* XXX: Might be wrong, check with EJTAG spec. */
1409 if (other_tc
== other
->current_tc
) {
1410 other
->active_tc
.CP0_Debug_tcstatus
= val
;
1412 other
->tcs
[other_tc
].CP0_Debug_tcstatus
= val
;
1414 other
->CP0_Debug
= (other
->CP0_Debug
&
1415 ((1 << CP0DB_SSt
) | (1 << CP0DB_Halt
))) |
1416 (arg1
& ~((1 << CP0DB_SSt
) | (1 << CP0DB_Halt
)));
1419 void helper_mtc0_performance0(CPUMIPSState
*env
, target_ulong arg1
)
1421 env
->CP0_Performance0
= arg1
& 0x000007ff;
1424 void helper_mtc0_errctl(CPUMIPSState
*env
, target_ulong arg1
)
1426 int32_t wst
= arg1
& (1 << CP0EC_WST
);
1427 int32_t spr
= arg1
& (1 << CP0EC_SPR
);
1428 int32_t itc
= env
->itc_tag
? (arg1
& (1 << CP0EC_ITC
)) : 0;
1430 env
->CP0_ErrCtl
= wst
| spr
| itc
;
1432 if (itc
&& !wst
&& !spr
) {
1433 env
->hflags
|= MIPS_HFLAG_ITC_CACHE
;
1435 env
->hflags
&= ~MIPS_HFLAG_ITC_CACHE
;
1439 void helper_mtc0_taglo(CPUMIPSState
*env
, target_ulong arg1
)
1441 if (env
->hflags
& MIPS_HFLAG_ITC_CACHE
) {
1443 * If CACHE instruction is configured for ITC tags then make all
1444 * CP0.TagLo bits writable. The actual write to ITC Configuration
1445 * Tag will take care of the read-only bits.
1447 env
->CP0_TagLo
= arg1
;
1449 env
->CP0_TagLo
= arg1
& 0xFFFFFCF6;
1453 void helper_mtc0_datalo(CPUMIPSState
*env
, target_ulong arg1
)
1455 env
->CP0_DataLo
= arg1
; /* XXX */
1458 void helper_mtc0_taghi(CPUMIPSState
*env
, target_ulong arg1
)
1460 env
->CP0_TagHi
= arg1
; /* XXX */
1463 void helper_mtc0_datahi(CPUMIPSState
*env
, target_ulong arg1
)
1465 env
->CP0_DataHi
= arg1
; /* XXX */
1468 /* MIPS MT functions */
1469 target_ulong
helper_mftgpr(CPUMIPSState
*env
, uint32_t sel
)
1471 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
1472 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
1474 if (other_tc
== other
->current_tc
) {
1475 return other
->active_tc
.gpr
[sel
];
1477 return other
->tcs
[other_tc
].gpr
[sel
];
1481 target_ulong
helper_mftlo(CPUMIPSState
*env
, uint32_t sel
)
1483 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
1484 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
1486 if (other_tc
== other
->current_tc
) {
1487 return other
->active_tc
.LO
[sel
];
1489 return other
->tcs
[other_tc
].LO
[sel
];
1493 target_ulong
helper_mfthi(CPUMIPSState
*env
, uint32_t sel
)
1495 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
1496 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
1498 if (other_tc
== other
->current_tc
) {
1499 return other
->active_tc
.HI
[sel
];
1501 return other
->tcs
[other_tc
].HI
[sel
];
1505 target_ulong
helper_mftacx(CPUMIPSState
*env
, uint32_t sel
)
1507 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
1508 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
1510 if (other_tc
== other
->current_tc
) {
1511 return other
->active_tc
.ACX
[sel
];
1513 return other
->tcs
[other_tc
].ACX
[sel
];
1517 target_ulong
helper_mftdsp(CPUMIPSState
*env
)
1519 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
1520 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
1522 if (other_tc
== other
->current_tc
) {
1523 return other
->active_tc
.DSPControl
;
1525 return other
->tcs
[other_tc
].DSPControl
;
1529 void helper_mttgpr(CPUMIPSState
*env
, target_ulong arg1
, uint32_t sel
)
1531 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
1532 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
1534 if (other_tc
== other
->current_tc
) {
1535 other
->active_tc
.gpr
[sel
] = arg1
;
1537 other
->tcs
[other_tc
].gpr
[sel
] = arg1
;
1541 void helper_mttlo(CPUMIPSState
*env
, target_ulong arg1
, uint32_t sel
)
1543 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
1544 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
1546 if (other_tc
== other
->current_tc
) {
1547 other
->active_tc
.LO
[sel
] = arg1
;
1549 other
->tcs
[other_tc
].LO
[sel
] = arg1
;
1553 void helper_mtthi(CPUMIPSState
*env
, target_ulong arg1
, uint32_t sel
)
1555 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
1556 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
1558 if (other_tc
== other
->current_tc
) {
1559 other
->active_tc
.HI
[sel
] = arg1
;
1561 other
->tcs
[other_tc
].HI
[sel
] = arg1
;
1565 void helper_mttacx(CPUMIPSState
*env
, target_ulong arg1
, uint32_t sel
)
1567 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
1568 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
1570 if (other_tc
== other
->current_tc
) {
1571 other
->active_tc
.ACX
[sel
] = arg1
;
1573 other
->tcs
[other_tc
].ACX
[sel
] = arg1
;
1577 void helper_mttdsp(CPUMIPSState
*env
, target_ulong arg1
)
1579 int other_tc
= env
->CP0_VPEControl
& (0xff << CP0VPECo_TargTC
);
1580 CPUMIPSState
*other
= mips_cpu_map_tc(env
, &other_tc
);
1582 if (other_tc
== other
->current_tc
) {
1583 other
->active_tc
.DSPControl
= arg1
;
1585 other
->tcs
[other_tc
].DSPControl
= arg1
;
1589 /* MIPS MT functions */
1590 target_ulong
helper_dmt(void)
1596 target_ulong
helper_emt(void)
1602 target_ulong
helper_dvpe(CPUMIPSState
*env
)
1604 CPUState
*other_cs
= first_cpu
;
1605 target_ulong prev
= env
->mvp
->CP0_MVPControl
;
1607 CPU_FOREACH(other_cs
) {
1608 MIPSCPU
*other_cpu
= MIPS_CPU(other_cs
);
1609 /* Turn off all VPEs except the one executing the dvpe. */
1610 if (&other_cpu
->env
!= env
) {
1611 other_cpu
->env
.mvp
->CP0_MVPControl
&= ~(1 << CP0MVPCo_EVP
);
1612 mips_vpe_sleep(other_cpu
);
1618 target_ulong
helper_evpe(CPUMIPSState
*env
)
1620 CPUState
*other_cs
= first_cpu
;
1621 target_ulong prev
= env
->mvp
->CP0_MVPControl
;
1623 CPU_FOREACH(other_cs
) {
1624 MIPSCPU
*other_cpu
= MIPS_CPU(other_cs
);
1626 if (&other_cpu
->env
!= env
1627 /* If the VPE is WFI, don't disturb its sleep. */
1628 && !mips_vpe_is_wfi(other_cpu
)) {
1629 /* Enable the VPE. */
1630 other_cpu
->env
.mvp
->CP0_MVPControl
|= (1 << CP0MVPCo_EVP
);
1631 mips_vpe_wake(other_cpu
); /* And wake it up. */
1636 #endif /* !CONFIG_USER_ONLY */
1638 /* R6 Multi-threading */
1639 #ifndef CONFIG_USER_ONLY
1640 target_ulong
helper_dvp(CPUMIPSState
*env
)
1642 CPUState
*other_cs
= first_cpu
;
1643 target_ulong prev
= env
->CP0_VPControl
;
1645 if (!((env
->CP0_VPControl
>> CP0VPCtl_DIS
) & 1)) {
1646 CPU_FOREACH(other_cs
) {
1647 MIPSCPU
*other_cpu
= MIPS_CPU(other_cs
);
1648 /* Turn off all VPs except the one executing the dvp. */
1649 if (&other_cpu
->env
!= env
) {
1650 mips_vpe_sleep(other_cpu
);
1653 env
->CP0_VPControl
|= (1 << CP0VPCtl_DIS
);
1658 target_ulong
helper_evp(CPUMIPSState
*env
)
1660 CPUState
*other_cs
= first_cpu
;
1661 target_ulong prev
= env
->CP0_VPControl
;
1663 if ((env
->CP0_VPControl
>> CP0VPCtl_DIS
) & 1) {
1664 CPU_FOREACH(other_cs
) {
1665 MIPSCPU
*other_cpu
= MIPS_CPU(other_cs
);
1666 if ((&other_cpu
->env
!= env
) && !mips_vp_is_wfi(other_cpu
)) {
1668 * If the VP is WFI, don't disturb its sleep.
1669 * Otherwise, wake it up.
1671 mips_vpe_wake(other_cpu
);
1674 env
->CP0_VPControl
&= ~(1 << CP0VPCtl_DIS
);
1678 #endif /* !CONFIG_USER_ONLY */