1 #if !defined (__MIPS_CPU_H__)
8 #define CPUArchState struct CPUMIPSState
10 #include "qemu-common.h"
11 #include "mips-defs.h"
12 #include "exec/cpu-defs.h"
13 #include "fpu/softfloat.h"
17 typedef struct r4k_tlb_t r4k_tlb_t
;
37 #if !defined(CONFIG_USER_ONLY)
38 typedef struct CPUMIPSTLBContext CPUMIPSTLBContext
;
39 struct CPUMIPSTLBContext
{
42 int (*map_address
) (struct CPUMIPSState
*env
, hwaddr
*physical
, int *prot
, target_ulong address
, int rw
, int access_type
);
43 void (*helper_tlbwi
)(struct CPUMIPSState
*env
);
44 void (*helper_tlbwr
)(struct CPUMIPSState
*env
);
45 void (*helper_tlbp
)(struct CPUMIPSState
*env
);
46 void (*helper_tlbr
)(struct CPUMIPSState
*env
);
47 void (*helper_tlbinv
)(struct CPUMIPSState
*env
);
48 void (*helper_tlbinvf
)(struct CPUMIPSState
*env
);
51 r4k_tlb_t tlb
[MIPS_TLB_MAX
];
58 #define MSA_WRLEN (128)
60 enum CPUMIPSMSADataFormat
{
67 typedef union wr_t wr_t
;
69 int8_t b
[MSA_WRLEN
/8];
70 int16_t h
[MSA_WRLEN
/16];
71 int32_t w
[MSA_WRLEN
/32];
72 int64_t d
[MSA_WRLEN
/64];
75 typedef union fpr_t fpr_t
;
77 float64 fd
; /* ieee double precision */
78 float32 fs
[2];/* ieee single precision */
79 uint64_t d
; /* binary double fixed-point */
80 uint32_t w
[2]; /* binary single fixed-point */
81 /* FPU/MSA register mapping is not tested on big-endian hosts. */
82 wr_t wr
; /* vector data */
84 /* define FP_ENDIAN_IDX to access the same location
85 * in the fpr_t union regardless of the host endianness
87 #if defined(HOST_WORDS_BIGENDIAN)
88 # define FP_ENDIAN_IDX 1
90 # define FP_ENDIAN_IDX 0
93 typedef struct CPUMIPSFPUContext CPUMIPSFPUContext
;
94 struct CPUMIPSFPUContext
{
95 /* Floating point registers */
97 float_status fp_status
;
98 /* fpu implementation/revision register (fir) */
102 #define FCR0_HAS2008 23
114 #define FCR31_ABS2008 19
115 #define FCR31_NAN2008 18
116 #define SET_FP_COND(num,env) do { ((env).fcr31) |= ((num) ? (1 << ((num) + 24)) : (1 << 23)); } while(0)
117 #define CLEAR_FP_COND(num,env) do { ((env).fcr31) &= ~((num) ? (1 << ((num) + 24)) : (1 << 23)); } while(0)
118 #define GET_FP_COND(env) ((((env).fcr31 >> 24) & 0xfe) | (((env).fcr31 >> 23) & 0x1))
119 #define GET_FP_CAUSE(reg) (((reg) >> 12) & 0x3f)
120 #define GET_FP_ENABLE(reg) (((reg) >> 7) & 0x1f)
121 #define GET_FP_FLAGS(reg) (((reg) >> 2) & 0x1f)
122 #define SET_FP_CAUSE(reg,v) do { (reg) = ((reg) & ~(0x3f << 12)) | ((v & 0x3f) << 12); } while(0)
123 #define SET_FP_ENABLE(reg,v) do { (reg) = ((reg) & ~(0x1f << 7)) | ((v & 0x1f) << 7); } while(0)
124 #define SET_FP_FLAGS(reg,v) do { (reg) = ((reg) & ~(0x1f << 2)) | ((v & 0x1f) << 2); } while(0)
125 #define UPDATE_FP_FLAGS(reg,v) do { (reg) |= ((v & 0x1f) << 2); } while(0)
127 #define FP_UNDERFLOW 2
128 #define FP_OVERFLOW 4
130 #define FP_INVALID 16
131 #define FP_UNIMPLEMENTED 32
134 #define NB_MMU_MODES 3
135 #define TARGET_INSN_START_EXTRA_WORDS 2
137 typedef struct CPUMIPSMVPContext CPUMIPSMVPContext
;
138 struct CPUMIPSMVPContext
{
139 int32_t CP0_MVPControl
;
140 #define CP0MVPCo_CPA 3
141 #define CP0MVPCo_STLB 2
142 #define CP0MVPCo_VPC 1
143 #define CP0MVPCo_EVP 0
144 int32_t CP0_MVPConf0
;
145 #define CP0MVPC0_M 31
146 #define CP0MVPC0_TLBS 29
147 #define CP0MVPC0_GS 28
148 #define CP0MVPC0_PCP 27
149 #define CP0MVPC0_PTLBE 16
150 #define CP0MVPC0_TCA 15
151 #define CP0MVPC0_PVPE 10
152 #define CP0MVPC0_PTC 0
153 int32_t CP0_MVPConf1
;
154 #define CP0MVPC1_CIM 31
155 #define CP0MVPC1_CIF 30
156 #define CP0MVPC1_PCX 20
157 #define CP0MVPC1_PCP2 10
158 #define CP0MVPC1_PCP1 0
161 typedef struct mips_def_t mips_def_t
;
163 #define MIPS_SHADOW_SET_MAX 16
164 #define MIPS_TC_MAX 5
165 #define MIPS_FPU_MAX 1
166 #define MIPS_DSP_ACC 4
167 #define MIPS_KSCRATCH_NUM 6
168 #define MIPS_MAAR_MAX 16 /* Must be an even number. */
170 typedef struct TCState TCState
;
172 target_ulong gpr
[32];
174 target_ulong HI
[MIPS_DSP_ACC
];
175 target_ulong LO
[MIPS_DSP_ACC
];
176 target_ulong ACX
[MIPS_DSP_ACC
];
177 target_ulong DSPControl
;
178 int32_t CP0_TCStatus
;
179 #define CP0TCSt_TCU3 31
180 #define CP0TCSt_TCU2 30
181 #define CP0TCSt_TCU1 29
182 #define CP0TCSt_TCU0 28
183 #define CP0TCSt_TMX 27
184 #define CP0TCSt_RNST 23
185 #define CP0TCSt_TDS 21
186 #define CP0TCSt_DT 20
187 #define CP0TCSt_DA 15
189 #define CP0TCSt_TKSU 11
190 #define CP0TCSt_IXMT 10
191 #define CP0TCSt_TASID 0
193 #define CP0TCBd_CurTC 21
194 #define CP0TCBd_TBE 17
195 #define CP0TCBd_CurVPE 0
196 target_ulong CP0_TCHalt
;
197 target_ulong CP0_TCContext
;
198 target_ulong CP0_TCSchedule
;
199 target_ulong CP0_TCScheFBack
;
200 int32_t CP0_Debug_tcstatus
;
201 target_ulong CP0_UserLocal
;
206 #define MSACSR_FS_MASK (1 << MSACSR_FS)
208 #define MSACSR_NX_MASK (1 << MSACSR_NX)
210 #define MSACSR_CEF_MASK (0xffff << MSACSR_CEF)
212 #define MSACSR_RM_MASK (0x3 << MSACSR_RM)
213 #define MSACSR_MASK (MSACSR_RM_MASK | MSACSR_CEF_MASK | MSACSR_NX_MASK | \
216 float_status msa_fp_status
;
219 typedef struct CPUMIPSState CPUMIPSState
;
220 struct CPUMIPSState
{
222 CPUMIPSFPUContext active_fpu
;
225 uint32_t current_fpu
;
229 #if defined(TARGET_MIPS64)
230 # define PABITS_BASE 36
232 # define PABITS_BASE 32
234 target_ulong SEGMask
;
236 #define PAMASK_BASE ((1ULL << PABITS_BASE) - 1)
239 #define MSAIR_ProcID 8
243 /* CP0_MVP* are per MVP registers. */
244 int32_t CP0_VPControl
;
245 #define CP0VPCtl_DIS 0
247 int32_t CP0_VPEControl
;
248 #define CP0VPECo_YSI 21
249 #define CP0VPECo_GSI 20
250 #define CP0VPECo_EXCPT 16
251 #define CP0VPECo_TE 15
252 #define CP0VPECo_TargTC 0
253 int32_t CP0_VPEConf0
;
254 #define CP0VPEC0_M 31
255 #define CP0VPEC0_XTC 21
256 #define CP0VPEC0_TCS 19
257 #define CP0VPEC0_SCS 18
258 #define CP0VPEC0_DSC 17
259 #define CP0VPEC0_ICS 16
260 #define CP0VPEC0_MVP 1
261 #define CP0VPEC0_VPA 0
262 int32_t CP0_VPEConf1
;
263 #define CP0VPEC1_NCX 20
264 #define CP0VPEC1_NCP2 10
265 #define CP0VPEC1_NCP1 0
266 target_ulong CP0_YQMask
;
267 target_ulong CP0_VPESchedule
;
268 target_ulong CP0_VPEScheFBack
;
270 #define CP0VPEOpt_IWX7 15
271 #define CP0VPEOpt_IWX6 14
272 #define CP0VPEOpt_IWX5 13
273 #define CP0VPEOpt_IWX4 12
274 #define CP0VPEOpt_IWX3 11
275 #define CP0VPEOpt_IWX2 10
276 #define CP0VPEOpt_IWX1 9
277 #define CP0VPEOpt_IWX0 8
278 #define CP0VPEOpt_DWX7 7
279 #define CP0VPEOpt_DWX6 6
280 #define CP0VPEOpt_DWX5 5
281 #define CP0VPEOpt_DWX4 4
282 #define CP0VPEOpt_DWX3 3
283 #define CP0VPEOpt_DWX2 2
284 #define CP0VPEOpt_DWX1 1
285 #define CP0VPEOpt_DWX0 0
286 uint64_t CP0_EntryLo0
;
287 uint64_t CP0_EntryLo1
;
288 #if defined(TARGET_MIPS64)
289 # define CP0EnLo_RI 63
290 # define CP0EnLo_XI 62
292 # define CP0EnLo_RI 31
293 # define CP0EnLo_XI 30
295 int32_t CP0_GlobalNumber
;
297 target_ulong CP0_Context
;
298 target_ulong CP0_KScratch
[MIPS_KSCRATCH_NUM
];
299 int32_t CP0_PageMask
;
300 int32_t CP0_PageGrain_rw_bitmask
;
301 int32_t CP0_PageGrain
;
304 #define CP0PG_ELPA 29
307 int32_t CP0_SRSConf0_rw_bitmask
;
308 int32_t CP0_SRSConf0
;
309 #define CP0SRSC0_M 31
310 #define CP0SRSC0_SRS3 20
311 #define CP0SRSC0_SRS2 10
312 #define CP0SRSC0_SRS1 0
313 int32_t CP0_SRSConf1_rw_bitmask
;
314 int32_t CP0_SRSConf1
;
315 #define CP0SRSC1_M 31
316 #define CP0SRSC1_SRS6 20
317 #define CP0SRSC1_SRS5 10
318 #define CP0SRSC1_SRS4 0
319 int32_t CP0_SRSConf2_rw_bitmask
;
320 int32_t CP0_SRSConf2
;
321 #define CP0SRSC2_M 31
322 #define CP0SRSC2_SRS9 20
323 #define CP0SRSC2_SRS8 10
324 #define CP0SRSC2_SRS7 0
325 int32_t CP0_SRSConf3_rw_bitmask
;
326 int32_t CP0_SRSConf3
;
327 #define CP0SRSC3_M 31
328 #define CP0SRSC3_SRS12 20
329 #define CP0SRSC3_SRS11 10
330 #define CP0SRSC3_SRS10 0
331 int32_t CP0_SRSConf4_rw_bitmask
;
332 int32_t CP0_SRSConf4
;
333 #define CP0SRSC4_SRS15 20
334 #define CP0SRSC4_SRS14 10
335 #define CP0SRSC4_SRS13 0
337 target_ulong CP0_BadVAddr
;
338 uint32_t CP0_BadInstr
;
339 uint32_t CP0_BadInstrP
;
341 target_ulong CP0_EntryHi
;
342 #define CP0EnHi_EHINV 10
367 #define CP0IntCtl_IPTI 29
368 #define CP0IntCtl_IPPCI 26
369 #define CP0IntCtl_VS 5
371 #define CP0SRSCtl_HSS 26
372 #define CP0SRSCtl_EICSS 18
373 #define CP0SRSCtl_ESS 12
374 #define CP0SRSCtl_PSS 6
375 #define CP0SRSCtl_CSS 0
377 #define CP0SRSMap_SSV7 28
378 #define CP0SRSMap_SSV6 24
379 #define CP0SRSMap_SSV5 20
380 #define CP0SRSMap_SSV4 16
381 #define CP0SRSMap_SSV3 12
382 #define CP0SRSMap_SSV2 8
383 #define CP0SRSMap_SSV1 4
384 #define CP0SRSMap_SSV0 0
394 #define CP0Ca_IP_mask 0x0000FF00
396 target_ulong CP0_EPC
;
399 target_ulong CP0_CMGCRBase
;
442 #define CP0C3_CMGCR 29
443 #define CP0C3_MSAP 28
446 #define CP0C3_IPLW 21
447 #define CP0C3_MMAR 18
449 #define CP0C3_ISA_ON_EXC 16
451 #define CP0C3_ULRI 13
453 #define CP0C3_DSP2P 11
454 #define CP0C3_DSPP 10
464 int32_t CP0_Config4_rw_bitmask
;
467 #define CP0C4_KScrExist 16
468 #define CP0C4_MMUExtDef 14
469 #define CP0C4_FTLBPageSize 8
470 #define CP0C4_FTLBWays 4
471 #define CP0C4_FTLBSets 0
472 #define CP0C4_MMUSizeExt 0
474 int32_t CP0_Config5_rw_bitmask
;
479 #define CP0C5_MSAEn 27
489 #define CP0C5_NFExists 0
492 uint64_t CP0_MAAR
[MIPS_MAAR_MAX
];
494 /* XXX: Maybe make LLAddr per-TC? */
497 target_ulong llnewval
;
499 uint64_t CP0_LLAddr_rw_bitmask
;
500 int CP0_LLAddr_shift
;
501 target_ulong CP0_WatchLo
[8];
502 int32_t CP0_WatchHi
[8];
503 target_ulong CP0_XContext
;
504 int32_t CP0_Framemask
;
508 #define CP0DB_LSNM 28
509 #define CP0DB_Doze 27
510 #define CP0DB_Halt 26
512 #define CP0DB_IBEP 24
513 #define CP0DB_DBEP 21
514 #define CP0DB_IEXI 20
524 target_ulong CP0_DEPC
;
525 int32_t CP0_Performance0
;
534 target_ulong CP0_ErrorEPC
;
536 /* We waste some space so we can handle shadow registers like TCs. */
537 TCState tcs
[MIPS_SHADOW_SET_MAX
];
538 CPUMIPSFPUContext fpus
[MIPS_FPU_MAX
];
541 #define EXCP_TLB_NOMATCH 0x1
542 #define EXCP_INST_NOTAVAIL 0x2 /* No valid instruction word for BadInstr */
543 uint32_t hflags
; /* CPU State */
544 /* TMASK defines different execution modes */
545 #define MIPS_HFLAG_TMASK 0xF5807FF
546 #define MIPS_HFLAG_MODE 0x00007 /* execution modes */
547 /* The KSU flags must be the lowest bits in hflags. The flag order
548 must be the same as defined for CP0 Status. This allows to use
549 the bits as the value of mmu_idx. */
550 #define MIPS_HFLAG_KSU 0x00003 /* kernel/supervisor/user mode mask */
551 #define MIPS_HFLAG_UM 0x00002 /* user mode flag */
552 #define MIPS_HFLAG_SM 0x00001 /* supervisor mode flag */
553 #define MIPS_HFLAG_KM 0x00000 /* kernel mode flag */
554 #define MIPS_HFLAG_DM 0x00004 /* Debug mode */
555 #define MIPS_HFLAG_64 0x00008 /* 64-bit instructions enabled */
556 #define MIPS_HFLAG_CP0 0x00010 /* CP0 enabled */
557 #define MIPS_HFLAG_FPU 0x00020 /* FPU enabled */
558 #define MIPS_HFLAG_F64 0x00040 /* 64-bit FPU enabled */
559 /* True if the MIPS IV COP1X instructions can be used. This also
560 controls the non-COP1X instructions RECIP.S, RECIP.D, RSQRT.S
562 #define MIPS_HFLAG_COP1X 0x00080 /* COP1X instructions enabled */
563 #define MIPS_HFLAG_RE 0x00100 /* Reversed endianness */
564 #define MIPS_HFLAG_AWRAP 0x00200 /* 32-bit compatibility address wrapping */
565 #define MIPS_HFLAG_M16 0x00400 /* MIPS16 mode flag */
566 #define MIPS_HFLAG_M16_SHIFT 10
567 /* If translation is interrupted between the branch instruction and
568 * the delay slot, record what type of branch it is so that we can
569 * resume translation properly. It might be possible to reduce
570 * this from three bits to two. */
571 #define MIPS_HFLAG_BMASK_BASE 0x803800
572 #define MIPS_HFLAG_B 0x00800 /* Unconditional branch */
573 #define MIPS_HFLAG_BC 0x01000 /* Conditional branch */
574 #define MIPS_HFLAG_BL 0x01800 /* Likely branch */
575 #define MIPS_HFLAG_BR 0x02000 /* branch to register (can't link TB) */
576 /* Extra flags about the current pending branch. */
577 #define MIPS_HFLAG_BMASK_EXT 0x7C000
578 #define MIPS_HFLAG_B16 0x04000 /* branch instruction was 16 bits */
579 #define MIPS_HFLAG_BDS16 0x08000 /* branch requires 16-bit delay slot */
580 #define MIPS_HFLAG_BDS32 0x10000 /* branch requires 32-bit delay slot */
581 #define MIPS_HFLAG_BDS_STRICT 0x20000 /* Strict delay slot size */
582 #define MIPS_HFLAG_BX 0x40000 /* branch exchanges execution mode */
583 #define MIPS_HFLAG_BMASK (MIPS_HFLAG_BMASK_BASE | MIPS_HFLAG_BMASK_EXT)
584 /* MIPS DSP resources access. */
585 #define MIPS_HFLAG_DSP 0x080000 /* Enable access to MIPS DSP resources. */
586 #define MIPS_HFLAG_DSPR2 0x100000 /* Enable access to MIPS DSPR2 resources. */
587 /* Extra flag about HWREna register. */
588 #define MIPS_HFLAG_HWRENA_ULR 0x200000 /* ULR bit from HWREna is set. */
589 #define MIPS_HFLAG_SBRI 0x400000 /* R6 SDBBP causes RI excpt. in user mode */
590 #define MIPS_HFLAG_FBNSLOT 0x800000 /* Forbidden slot */
591 #define MIPS_HFLAG_MSA 0x1000000
592 #define MIPS_HFLAG_FRE 0x2000000 /* FRE enabled */
593 #define MIPS_HFLAG_ELPA 0x4000000
594 #define MIPS_HFLAG_ITC_CACHE 0x8000000 /* CACHE instr. operates on ITC tag */
595 target_ulong btarget
; /* Jump / branch target */
596 target_ulong bcond
; /* Branch condition (if needed) */
598 int SYNCI_Step
; /* Address step size for SYNCI */
599 int CCRes
; /* Cycle count resolution/divisor */
600 uint32_t CP0_Status_rw_bitmask
; /* Read/write bits in CP0_Status */
601 uint32_t CP0_TCStatus_rw_bitmask
; /* Read/write bits in CP0_TCStatus */
602 int insn_flags
; /* Supported instruction set */
606 /* Fields from here on are preserved across CPU reset. */
607 CPUMIPSMVPContext
*mvp
;
608 #if !defined(CONFIG_USER_ONLY)
609 CPUMIPSTLBContext
*tlb
;
612 const mips_def_t
*cpu_model
;
614 QEMUTimer
*timer
; /* Internal timer */
615 MemoryRegion
*itc_tag
; /* ITC Configuration Tags */
620 #if !defined(CONFIG_USER_ONLY)
621 int no_mmu_map_address (CPUMIPSState
*env
, hwaddr
*physical
, int *prot
,
622 target_ulong address
, int rw
, int access_type
);
623 int fixed_mmu_map_address (CPUMIPSState
*env
, hwaddr
*physical
, int *prot
,
624 target_ulong address
, int rw
, int access_type
);
625 int r4k_map_address (CPUMIPSState
*env
, hwaddr
*physical
, int *prot
,
626 target_ulong address
, int rw
, int access_type
);
627 void r4k_helper_tlbwi(CPUMIPSState
*env
);
628 void r4k_helper_tlbwr(CPUMIPSState
*env
);
629 void r4k_helper_tlbp(CPUMIPSState
*env
);
630 void r4k_helper_tlbr(CPUMIPSState
*env
);
631 void r4k_helper_tlbinv(CPUMIPSState
*env
);
632 void r4k_helper_tlbinvf(CPUMIPSState
*env
);
634 void mips_cpu_unassigned_access(CPUState
*cpu
, hwaddr addr
,
635 bool is_write
, bool is_exec
, int unused
,
639 void mips_cpu_list (FILE *f
, fprintf_function cpu_fprintf
);
641 #define cpu_exec cpu_mips_exec
642 #define cpu_signal_handler cpu_mips_signal_handler
643 #define cpu_list mips_cpu_list
645 extern void cpu_wrdsp(uint32_t rs
, uint32_t mask_num
, CPUMIPSState
*env
);
646 extern uint32_t cpu_rddsp(uint32_t mask_num
, CPUMIPSState
*env
);
648 /* MMU modes definitions. We carefully match the indices with our
650 #define MMU_MODE0_SUFFIX _kernel
651 #define MMU_MODE1_SUFFIX _super
652 #define MMU_MODE2_SUFFIX _user
653 #define MMU_USER_IDX 2
654 static inline int cpu_mmu_index (CPUMIPSState
*env
, bool ifetch
)
656 return env
->hflags
& MIPS_HFLAG_KSU
;
659 static inline bool cpu_mips_hw_interrupts_enabled(CPUMIPSState
*env
)
661 return (env
->CP0_Status
& (1 << CP0St_IE
)) &&
662 !(env
->CP0_Status
& (1 << CP0St_EXL
)) &&
663 !(env
->CP0_Status
& (1 << CP0St_ERL
)) &&
664 !(env
->hflags
& MIPS_HFLAG_DM
) &&
665 /* Note that the TCStatus IXMT field is initialized to zero,
666 and only MT capable cores can set it to one. So we don't
667 need to check for MT capabilities here. */
668 !(env
->active_tc
.CP0_TCStatus
& (1 << CP0TCSt_IXMT
));
671 /* Check if there is pending and not masked out interrupt */
672 static inline bool cpu_mips_hw_interrupts_pending(CPUMIPSState
*env
)
678 pending
= env
->CP0_Cause
& CP0Ca_IP_mask
;
679 status
= env
->CP0_Status
& CP0Ca_IP_mask
;
681 if (env
->CP0_Config3
& (1 << CP0C3_VEIC
)) {
682 /* A MIPS configured with a vectorizing external interrupt controller
683 will feed a vector into the Cause pending lines. The core treats
684 the status lines as a vector level, not as indiviual masks. */
685 r
= pending
> status
;
687 /* A MIPS configured with compatibility or VInt (Vectored Interrupts)
688 treats the pending lines as individual interrupt lines, the status
689 lines are individual masks. */
690 r
= (pending
& status
) != 0;
695 #include "exec/cpu-all.h"
697 /* Memory access type :
698 * may be needed for precise access rights control and precise exceptions.
701 /* 1 bit to define user level / supervisor access */
704 /* 1 bit to indicate direction */
706 /* Type of instruction that generated the access */
707 ACCESS_CODE
= 0x10, /* Code fetch access */
708 ACCESS_INT
= 0x20, /* Integer load/store access */
709 ACCESS_FLOAT
= 0x30, /* floating point load/store access */
723 EXCP_EXT_INTERRUPT
, /* 8 */
739 EXCP_DWATCH
, /* 24 */
754 EXCP_LAST
= EXCP_TLBRI
,
756 /* Dummy exception for conditional stores. */
757 #define EXCP_SC 0x100
760 * This is an interrnally generated WAKE request line.
761 * It is driven by the CPU itself. Raised when the MT
762 * block wants to wake a VPE from an inactive state and
763 * cleared when VPE goes from active to inactive.
765 #define CPU_INTERRUPT_WAKE CPU_INTERRUPT_TGT_INT_0
767 int cpu_mips_exec(CPUState
*cpu
);
768 void mips_tcg_init(void);
769 MIPSCPU
*cpu_mips_init(const char *cpu_model
);
770 int cpu_mips_signal_handler(int host_signum
, void *pinfo
, void *puc
);
772 #define cpu_init(cpu_model) CPU(cpu_mips_init(cpu_model))
773 bool cpu_supports_cps_smp(const char *cpu_model
);
775 /* TODO QOM'ify CPU reset and remove */
776 void cpu_state_reset(CPUMIPSState
*s
);
779 uint32_t cpu_mips_get_random (CPUMIPSState
*env
);
780 uint32_t cpu_mips_get_count (CPUMIPSState
*env
);
781 void cpu_mips_store_count (CPUMIPSState
*env
, uint32_t value
);
782 void cpu_mips_store_compare (CPUMIPSState
*env
, uint32_t value
);
783 void cpu_mips_start_count(CPUMIPSState
*env
);
784 void cpu_mips_stop_count(CPUMIPSState
*env
);
787 void cpu_mips_soft_irq(CPUMIPSState
*env
, int irq
, int level
);
790 int mips_cpu_handle_mmu_fault(CPUState
*cpu
, vaddr address
, int rw
,
792 #if !defined(CONFIG_USER_ONLY)
793 void r4k_invalidate_tlb (CPUMIPSState
*env
, int idx
, int use_extra
);
794 hwaddr
cpu_mips_translate_address (CPUMIPSState
*env
, target_ulong address
,
797 target_ulong
exception_resume_pc (CPUMIPSState
*env
);
800 extern unsigned int ieee_rm
[];
801 int ieee_ex_to_mips(int xcpt
);
803 static inline void restore_rounding_mode(CPUMIPSState
*env
)
805 set_float_rounding_mode(ieee_rm
[env
->active_fpu
.fcr31
& 3],
806 &env
->active_fpu
.fp_status
);
809 static inline void restore_flush_mode(CPUMIPSState
*env
)
811 set_flush_to_zero((env
->active_fpu
.fcr31
& (1 << 24)) != 0,
812 &env
->active_fpu
.fp_status
);
815 static inline void restore_fp_status(CPUMIPSState
*env
)
817 restore_rounding_mode(env
);
818 restore_flush_mode(env
);
821 static inline void restore_msa_fp_status(CPUMIPSState
*env
)
823 float_status
*status
= &env
->active_tc
.msa_fp_status
;
824 int rounding_mode
= (env
->active_tc
.msacsr
& MSACSR_RM_MASK
) >> MSACSR_RM
;
825 bool flush_to_zero
= (env
->active_tc
.msacsr
& MSACSR_FS_MASK
) != 0;
827 set_float_rounding_mode(ieee_rm
[rounding_mode
], status
);
828 set_flush_to_zero(flush_to_zero
, status
);
829 set_flush_inputs_to_zero(flush_to_zero
, status
);
832 static inline void restore_pamask(CPUMIPSState
*env
)
834 if (env
->hflags
& MIPS_HFLAG_ELPA
) {
835 env
->PAMask
= (1ULL << env
->PABITS
) - 1;
837 env
->PAMask
= PAMASK_BASE
;
841 static inline void cpu_get_tb_cpu_state(CPUMIPSState
*env
, target_ulong
*pc
,
842 target_ulong
*cs_base
, int *flags
)
844 *pc
= env
->active_tc
.PC
;
846 *flags
= env
->hflags
& (MIPS_HFLAG_TMASK
| MIPS_HFLAG_BMASK
|
847 MIPS_HFLAG_HWRENA_ULR
);
850 static inline int mips_vpe_active(CPUMIPSState
*env
)
854 /* Check that the VPE is enabled. */
855 if (!(env
->mvp
->CP0_MVPControl
& (1 << CP0MVPCo_EVP
))) {
858 /* Check that the VPE is activated. */
859 if (!(env
->CP0_VPEConf0
& (1 << CP0VPEC0_VPA
))) {
863 /* Now verify that there are active thread contexts in the VPE.
865 This assumes the CPU model will internally reschedule threads
866 if the active one goes to sleep. If there are no threads available
867 the active one will be in a sleeping state, and we can turn off
869 if (!(env
->active_tc
.CP0_TCStatus
& (1 << CP0TCSt_A
))) {
870 /* TC is not activated. */
873 if (env
->active_tc
.CP0_TCHalt
& 1) {
874 /* TC is in halt state. */
881 static inline int mips_vp_active(CPUMIPSState
*env
)
883 CPUState
*other_cs
= first_cpu
;
885 /* Check if the VP disabled other VPs (which means the VP is enabled) */
886 if ((env
->CP0_VPControl
>> CP0VPCtl_DIS
) & 1) {
890 /* Check if the virtual processor is disabled due to a DVP */
891 CPU_FOREACH(other_cs
) {
892 MIPSCPU
*other_cpu
= MIPS_CPU(other_cs
);
893 if ((&other_cpu
->env
!= env
) &&
894 ((other_cpu
->env
.CP0_VPControl
>> CP0VPCtl_DIS
) & 1)) {
901 #include "exec/exec-all.h"
903 static inline void compute_hflags(CPUMIPSState
*env
)
905 env
->hflags
&= ~(MIPS_HFLAG_COP1X
| MIPS_HFLAG_64
| MIPS_HFLAG_CP0
|
906 MIPS_HFLAG_F64
| MIPS_HFLAG_FPU
| MIPS_HFLAG_KSU
|
907 MIPS_HFLAG_AWRAP
| MIPS_HFLAG_DSP
| MIPS_HFLAG_DSPR2
|
908 MIPS_HFLAG_SBRI
| MIPS_HFLAG_MSA
| MIPS_HFLAG_FRE
|
910 if (!(env
->CP0_Status
& (1 << CP0St_EXL
)) &&
911 !(env
->CP0_Status
& (1 << CP0St_ERL
)) &&
912 !(env
->hflags
& MIPS_HFLAG_DM
)) {
913 env
->hflags
|= (env
->CP0_Status
>> CP0St_KSU
) & MIPS_HFLAG_KSU
;
915 #if defined(TARGET_MIPS64)
916 if ((env
->insn_flags
& ISA_MIPS3
) &&
917 (((env
->hflags
& MIPS_HFLAG_KSU
) != MIPS_HFLAG_UM
) ||
918 (env
->CP0_Status
& (1 << CP0St_PX
)) ||
919 (env
->CP0_Status
& (1 << CP0St_UX
)))) {
920 env
->hflags
|= MIPS_HFLAG_64
;
923 if (!(env
->insn_flags
& ISA_MIPS3
)) {
924 env
->hflags
|= MIPS_HFLAG_AWRAP
;
925 } else if (((env
->hflags
& MIPS_HFLAG_KSU
) == MIPS_HFLAG_UM
) &&
926 !(env
->CP0_Status
& (1 << CP0St_UX
))) {
927 env
->hflags
|= MIPS_HFLAG_AWRAP
;
928 } else if (env
->insn_flags
& ISA_MIPS64R6
) {
929 /* Address wrapping for Supervisor and Kernel is specified in R6 */
930 if ((((env
->hflags
& MIPS_HFLAG_KSU
) == MIPS_HFLAG_SM
) &&
931 !(env
->CP0_Status
& (1 << CP0St_SX
))) ||
932 (((env
->hflags
& MIPS_HFLAG_KSU
) == MIPS_HFLAG_KM
) &&
933 !(env
->CP0_Status
& (1 << CP0St_KX
)))) {
934 env
->hflags
|= MIPS_HFLAG_AWRAP
;
938 if (((env
->CP0_Status
& (1 << CP0St_CU0
)) &&
939 !(env
->insn_flags
& ISA_MIPS32R6
)) ||
940 !(env
->hflags
& MIPS_HFLAG_KSU
)) {
941 env
->hflags
|= MIPS_HFLAG_CP0
;
943 if (env
->CP0_Status
& (1 << CP0St_CU1
)) {
944 env
->hflags
|= MIPS_HFLAG_FPU
;
946 if (env
->CP0_Status
& (1 << CP0St_FR
)) {
947 env
->hflags
|= MIPS_HFLAG_F64
;
949 if (((env
->hflags
& MIPS_HFLAG_KSU
) != MIPS_HFLAG_KM
) &&
950 (env
->CP0_Config5
& (1 << CP0C5_SBRI
))) {
951 env
->hflags
|= MIPS_HFLAG_SBRI
;
953 if (env
->insn_flags
& ASE_DSPR2
) {
954 /* Enables access MIPS DSP resources, now our cpu is DSP ASER2,
955 so enable to access DSPR2 resources. */
956 if (env
->CP0_Status
& (1 << CP0St_MX
)) {
957 env
->hflags
|= MIPS_HFLAG_DSP
| MIPS_HFLAG_DSPR2
;
960 } else if (env
->insn_flags
& ASE_DSP
) {
961 /* Enables access MIPS DSP resources, now our cpu is DSP ASE,
962 so enable to access DSP resources. */
963 if (env
->CP0_Status
& (1 << CP0St_MX
)) {
964 env
->hflags
|= MIPS_HFLAG_DSP
;
968 if (env
->insn_flags
& ISA_MIPS32R2
) {
969 if (env
->active_fpu
.fcr0
& (1 << FCR0_F64
)) {
970 env
->hflags
|= MIPS_HFLAG_COP1X
;
972 } else if (env
->insn_flags
& ISA_MIPS32
) {
973 if (env
->hflags
& MIPS_HFLAG_64
) {
974 env
->hflags
|= MIPS_HFLAG_COP1X
;
976 } else if (env
->insn_flags
& ISA_MIPS4
) {
977 /* All supported MIPS IV CPUs use the XX (CU3) to enable
978 and disable the MIPS IV extensions to the MIPS III ISA.
979 Some other MIPS IV CPUs ignore the bit, so the check here
980 would be too restrictive for them. */
981 if (env
->CP0_Status
& (1U << CP0St_CU3
)) {
982 env
->hflags
|= MIPS_HFLAG_COP1X
;
985 if (env
->insn_flags
& ASE_MSA
) {
986 if (env
->CP0_Config5
& (1 << CP0C5_MSAEn
)) {
987 env
->hflags
|= MIPS_HFLAG_MSA
;
990 if (env
->active_fpu
.fcr0
& (1 << FCR0_FREP
)) {
991 if (env
->CP0_Config5
& (1 << CP0C5_FRE
)) {
992 env
->hflags
|= MIPS_HFLAG_FRE
;
995 if (env
->CP0_Config3
& (1 << CP0C3_LPA
)) {
996 if (env
->CP0_PageGrain
& (1 << CP0PG_ELPA
)) {
997 env
->hflags
|= MIPS_HFLAG_ELPA
;
1002 #ifndef CONFIG_USER_ONLY
1003 static inline void cpu_mips_tlb_flush(CPUMIPSState
*env
, int flush_global
)
1005 MIPSCPU
*cpu
= mips_env_get_cpu(env
);
1007 /* Flush qemu's TLB and discard all shadowed entries. */
1008 tlb_flush(CPU(cpu
), flush_global
);
1009 env
->tlb
->tlb_in_use
= env
->tlb
->nb_tlb
;
1012 /* Called for updates to CP0_Status. */
1013 static inline void sync_c0_status(CPUMIPSState
*env
, CPUMIPSState
*cpu
, int tc
)
1015 int32_t tcstatus
, *tcst
;
1016 uint32_t v
= cpu
->CP0_Status
;
1017 uint32_t cu
, mx
, asid
, ksu
;
1018 uint32_t mask
= ((1 << CP0TCSt_TCU3
)
1019 | (1 << CP0TCSt_TCU2
)
1020 | (1 << CP0TCSt_TCU1
)
1021 | (1 << CP0TCSt_TCU0
)
1022 | (1 << CP0TCSt_TMX
)
1023 | (3 << CP0TCSt_TKSU
)
1024 | (0xff << CP0TCSt_TASID
));
1026 cu
= (v
>> CP0St_CU0
) & 0xf;
1027 mx
= (v
>> CP0St_MX
) & 0x1;
1028 ksu
= (v
>> CP0St_KSU
) & 0x3;
1029 asid
= env
->CP0_EntryHi
& 0xff;
1031 tcstatus
= cu
<< CP0TCSt_TCU0
;
1032 tcstatus
|= mx
<< CP0TCSt_TMX
;
1033 tcstatus
|= ksu
<< CP0TCSt_TKSU
;
1036 if (tc
== cpu
->current_tc
) {
1037 tcst
= &cpu
->active_tc
.CP0_TCStatus
;
1039 tcst
= &cpu
->tcs
[tc
].CP0_TCStatus
;
1044 compute_hflags(cpu
);
1047 static inline void cpu_mips_store_status(CPUMIPSState
*env
, target_ulong val
)
1049 uint32_t mask
= env
->CP0_Status_rw_bitmask
;
1050 target_ulong old
= env
->CP0_Status
;
1052 if (env
->insn_flags
& ISA_MIPS32R6
) {
1053 bool has_supervisor
= extract32(mask
, CP0St_KSU
, 2) == 0x3;
1054 #if defined(TARGET_MIPS64)
1055 uint32_t ksux
= (1 << CP0St_KX
) & val
;
1056 ksux
|= (ksux
>> 1) & val
; /* KX = 0 forces SX to be 0 */
1057 ksux
|= (ksux
>> 1) & val
; /* SX = 0 forces UX to be 0 */
1058 val
= (val
& ~(7 << CP0St_UX
)) | ksux
;
1060 if (has_supervisor
&& extract32(val
, CP0St_KSU
, 2) == 0x3) {
1061 mask
&= ~(3 << CP0St_KSU
);
1063 mask
&= ~(((1 << CP0St_SR
) | (1 << CP0St_NMI
)) & val
);
1066 env
->CP0_Status
= (old
& ~mask
) | (val
& mask
);
1067 #if defined(TARGET_MIPS64)
1068 if ((env
->CP0_Status
^ old
) & (old
& (7 << CP0St_UX
))) {
1069 /* Access to at least one of the 64-bit segments has been disabled */
1070 cpu_mips_tlb_flush(env
, 1);
1073 if (env
->CP0_Config3
& (1 << CP0C3_MT
)) {
1074 sync_c0_status(env
, env
, env
->current_tc
);
1076 compute_hflags(env
);
1080 static inline void cpu_mips_store_cause(CPUMIPSState
*env
, target_ulong val
)
1082 uint32_t mask
= 0x00C00300;
1083 uint32_t old
= env
->CP0_Cause
;
1086 if (env
->insn_flags
& ISA_MIPS32R2
) {
1087 mask
|= 1 << CP0Ca_DC
;
1089 if (env
->insn_flags
& ISA_MIPS32R6
) {
1090 mask
&= ~((1 << CP0Ca_WP
) & val
);
1093 env
->CP0_Cause
= (env
->CP0_Cause
& ~mask
) | (val
& mask
);
1095 if ((old
^ env
->CP0_Cause
) & (1 << CP0Ca_DC
)) {
1096 if (env
->CP0_Cause
& (1 << CP0Ca_DC
)) {
1097 cpu_mips_stop_count(env
);
1099 cpu_mips_start_count(env
);
1103 /* Set/reset software interrupts */
1104 for (i
= 0 ; i
< 2 ; i
++) {
1105 if ((old
^ env
->CP0_Cause
) & (1 << (CP0Ca_IP
+ i
))) {
1106 cpu_mips_soft_irq(env
, i
, env
->CP0_Cause
& (1 << (CP0Ca_IP
+ i
)));
1112 static inline void QEMU_NORETURN
do_raise_exception_err(CPUMIPSState
*env
,
1117 CPUState
*cs
= CPU(mips_env_get_cpu(env
));
1119 if (exception
< EXCP_SC
) {
1120 qemu_log_mask(CPU_LOG_INT
, "%s: %d %d\n",
1121 __func__
, exception
, error_code
);
1123 cs
->exception_index
= exception
;
1124 env
->error_code
= error_code
;
1126 cpu_loop_exit_restore(cs
, pc
);
1129 static inline void QEMU_NORETURN
do_raise_exception(CPUMIPSState
*env
,
1133 do_raise_exception_err(env
, exception
, 0, pc
);
1136 #endif /* !defined (__MIPS_CPU_H__) */