crypto: skip testing of unsupported cipher algorithms
[qemu/ar7.git] / target-mips / cpu.h
blob1e2b070cc3c5eb3c40215e806d577fa20b8569e4
1 #if !defined (__MIPS_CPU_H__)
2 #define __MIPS_CPU_H__
4 //#define DEBUG_OP
6 #define ALIGNED_ONLY
8 #define CPUArchState struct CPUMIPSState
10 #include "qemu-common.h"
11 #include "mips-defs.h"
12 #include "exec/cpu-defs.h"
13 #include "fpu/softfloat.h"
15 struct CPUMIPSState;
17 typedef struct r4k_tlb_t r4k_tlb_t;
18 struct r4k_tlb_t {
19 target_ulong VPN;
20 uint32_t PageMask;
21 uint8_t ASID;
22 unsigned int G:1;
23 unsigned int C0:3;
24 unsigned int C1:3;
25 unsigned int V0:1;
26 unsigned int V1:1;
27 unsigned int D0:1;
28 unsigned int D1:1;
29 unsigned int XI0:1;
30 unsigned int XI1:1;
31 unsigned int RI0:1;
32 unsigned int RI1:1;
33 unsigned int EHINV:1;
34 uint64_t PFN[2];
37 #if !defined(CONFIG_USER_ONLY)
38 typedef struct CPUMIPSTLBContext CPUMIPSTLBContext;
39 struct CPUMIPSTLBContext {
40 uint32_t nb_tlb;
41 uint32_t tlb_in_use;
42 int (*map_address) (struct CPUMIPSState *env, hwaddr *physical, int *prot, target_ulong address, int rw, int access_type);
43 void (*helper_tlbwi)(struct CPUMIPSState *env);
44 void (*helper_tlbwr)(struct CPUMIPSState *env);
45 void (*helper_tlbp)(struct CPUMIPSState *env);
46 void (*helper_tlbr)(struct CPUMIPSState *env);
47 void (*helper_tlbinv)(struct CPUMIPSState *env);
48 void (*helper_tlbinvf)(struct CPUMIPSState *env);
49 union {
50 struct {
51 r4k_tlb_t tlb[MIPS_TLB_MAX];
52 } r4k;
53 } mmu;
55 #endif
57 /* MSA Context */
58 #define MSA_WRLEN (128)
60 enum CPUMIPSMSADataFormat {
61 DF_BYTE = 0,
62 DF_HALF,
63 DF_WORD,
64 DF_DOUBLE
67 typedef union wr_t wr_t;
68 union wr_t {
69 int8_t b[MSA_WRLEN/8];
70 int16_t h[MSA_WRLEN/16];
71 int32_t w[MSA_WRLEN/32];
72 int64_t d[MSA_WRLEN/64];
75 typedef union fpr_t fpr_t;
76 union fpr_t {
77 float64 fd; /* ieee double precision */
78 float32 fs[2];/* ieee single precision */
79 uint64_t d; /* binary double fixed-point */
80 uint32_t w[2]; /* binary single fixed-point */
81 /* FPU/MSA register mapping is not tested on big-endian hosts. */
82 wr_t wr; /* vector data */
84 /* define FP_ENDIAN_IDX to access the same location
85 * in the fpr_t union regardless of the host endianness
87 #if defined(HOST_WORDS_BIGENDIAN)
88 # define FP_ENDIAN_IDX 1
89 #else
90 # define FP_ENDIAN_IDX 0
91 #endif
93 typedef struct CPUMIPSFPUContext CPUMIPSFPUContext;
94 struct CPUMIPSFPUContext {
95 /* Floating point registers */
96 fpr_t fpr[32];
97 float_status fp_status;
98 /* fpu implementation/revision register (fir) */
99 uint32_t fcr0;
100 #define FCR0_FREP 29
101 #define FCR0_UFRP 28
102 #define FCR0_F64 22
103 #define FCR0_L 21
104 #define FCR0_W 20
105 #define FCR0_3D 19
106 #define FCR0_PS 18
107 #define FCR0_D 17
108 #define FCR0_S 16
109 #define FCR0_PRID 8
110 #define FCR0_REV 0
111 /* fcsr */
112 uint32_t fcr31;
113 #define SET_FP_COND(num,env) do { ((env).fcr31) |= ((num) ? (1 << ((num) + 24)) : (1 << 23)); } while(0)
114 #define CLEAR_FP_COND(num,env) do { ((env).fcr31) &= ~((num) ? (1 << ((num) + 24)) : (1 << 23)); } while(0)
115 #define GET_FP_COND(env) ((((env).fcr31 >> 24) & 0xfe) | (((env).fcr31 >> 23) & 0x1))
116 #define GET_FP_CAUSE(reg) (((reg) >> 12) & 0x3f)
117 #define GET_FP_ENABLE(reg) (((reg) >> 7) & 0x1f)
118 #define GET_FP_FLAGS(reg) (((reg) >> 2) & 0x1f)
119 #define SET_FP_CAUSE(reg,v) do { (reg) = ((reg) & ~(0x3f << 12)) | ((v & 0x3f) << 12); } while(0)
120 #define SET_FP_ENABLE(reg,v) do { (reg) = ((reg) & ~(0x1f << 7)) | ((v & 0x1f) << 7); } while(0)
121 #define SET_FP_FLAGS(reg,v) do { (reg) = ((reg) & ~(0x1f << 2)) | ((v & 0x1f) << 2); } while(0)
122 #define UPDATE_FP_FLAGS(reg,v) do { (reg) |= ((v & 0x1f) << 2); } while(0)
123 #define FP_INEXACT 1
124 #define FP_UNDERFLOW 2
125 #define FP_OVERFLOW 4
126 #define FP_DIV0 8
127 #define FP_INVALID 16
128 #define FP_UNIMPLEMENTED 32
131 #define NB_MMU_MODES 3
132 #define TARGET_INSN_START_EXTRA_WORDS 2
134 typedef struct CPUMIPSMVPContext CPUMIPSMVPContext;
135 struct CPUMIPSMVPContext {
136 int32_t CP0_MVPControl;
137 #define CP0MVPCo_CPA 3
138 #define CP0MVPCo_STLB 2
139 #define CP0MVPCo_VPC 1
140 #define CP0MVPCo_EVP 0
141 int32_t CP0_MVPConf0;
142 #define CP0MVPC0_M 31
143 #define CP0MVPC0_TLBS 29
144 #define CP0MVPC0_GS 28
145 #define CP0MVPC0_PCP 27
146 #define CP0MVPC0_PTLBE 16
147 #define CP0MVPC0_TCA 15
148 #define CP0MVPC0_PVPE 10
149 #define CP0MVPC0_PTC 0
150 int32_t CP0_MVPConf1;
151 #define CP0MVPC1_CIM 31
152 #define CP0MVPC1_CIF 30
153 #define CP0MVPC1_PCX 20
154 #define CP0MVPC1_PCP2 10
155 #define CP0MVPC1_PCP1 0
158 typedef struct mips_def_t mips_def_t;
160 #define MIPS_SHADOW_SET_MAX 16
161 #define MIPS_TC_MAX 5
162 #define MIPS_FPU_MAX 1
163 #define MIPS_DSP_ACC 4
164 #define MIPS_KSCRATCH_NUM 6
166 typedef struct TCState TCState;
167 struct TCState {
168 target_ulong gpr[32];
169 target_ulong PC;
170 target_ulong HI[MIPS_DSP_ACC];
171 target_ulong LO[MIPS_DSP_ACC];
172 target_ulong ACX[MIPS_DSP_ACC];
173 target_ulong DSPControl;
174 int32_t CP0_TCStatus;
175 #define CP0TCSt_TCU3 31
176 #define CP0TCSt_TCU2 30
177 #define CP0TCSt_TCU1 29
178 #define CP0TCSt_TCU0 28
179 #define CP0TCSt_TMX 27
180 #define CP0TCSt_RNST 23
181 #define CP0TCSt_TDS 21
182 #define CP0TCSt_DT 20
183 #define CP0TCSt_DA 15
184 #define CP0TCSt_A 13
185 #define CP0TCSt_TKSU 11
186 #define CP0TCSt_IXMT 10
187 #define CP0TCSt_TASID 0
188 int32_t CP0_TCBind;
189 #define CP0TCBd_CurTC 21
190 #define CP0TCBd_TBE 17
191 #define CP0TCBd_CurVPE 0
192 target_ulong CP0_TCHalt;
193 target_ulong CP0_TCContext;
194 target_ulong CP0_TCSchedule;
195 target_ulong CP0_TCScheFBack;
196 int32_t CP0_Debug_tcstatus;
197 target_ulong CP0_UserLocal;
199 int32_t msacsr;
201 #define MSACSR_FS 24
202 #define MSACSR_FS_MASK (1 << MSACSR_FS)
203 #define MSACSR_NX 18
204 #define MSACSR_NX_MASK (1 << MSACSR_NX)
205 #define MSACSR_CEF 2
206 #define MSACSR_CEF_MASK (0xffff << MSACSR_CEF)
207 #define MSACSR_RM 0
208 #define MSACSR_RM_MASK (0x3 << MSACSR_RM)
209 #define MSACSR_MASK (MSACSR_RM_MASK | MSACSR_CEF_MASK | MSACSR_NX_MASK | \
210 MSACSR_FS_MASK)
212 float_status msa_fp_status;
215 typedef struct CPUMIPSState CPUMIPSState;
216 struct CPUMIPSState {
217 TCState active_tc;
218 CPUMIPSFPUContext active_fpu;
220 uint32_t current_tc;
221 uint32_t current_fpu;
223 uint32_t SEGBITS;
224 uint32_t PABITS;
225 #if defined(TARGET_MIPS64)
226 # define PABITS_BASE 36
227 #else
228 # define PABITS_BASE 32
229 #endif
230 target_ulong SEGMask;
231 uint64_t PAMask;
232 #define PAMASK_BASE ((1ULL << PABITS_BASE) - 1)
234 int32_t msair;
235 #define MSAIR_ProcID 8
236 #define MSAIR_Rev 0
238 int32_t CP0_Index;
239 /* CP0_MVP* are per MVP registers. */
240 int32_t CP0_VPControl;
241 #define CP0VPCtl_DIS 0
242 int32_t CP0_Random;
243 int32_t CP0_VPEControl;
244 #define CP0VPECo_YSI 21
245 #define CP0VPECo_GSI 20
246 #define CP0VPECo_EXCPT 16
247 #define CP0VPECo_TE 15
248 #define CP0VPECo_TargTC 0
249 int32_t CP0_VPEConf0;
250 #define CP0VPEC0_M 31
251 #define CP0VPEC0_XTC 21
252 #define CP0VPEC0_TCS 19
253 #define CP0VPEC0_SCS 18
254 #define CP0VPEC0_DSC 17
255 #define CP0VPEC0_ICS 16
256 #define CP0VPEC0_MVP 1
257 #define CP0VPEC0_VPA 0
258 int32_t CP0_VPEConf1;
259 #define CP0VPEC1_NCX 20
260 #define CP0VPEC1_NCP2 10
261 #define CP0VPEC1_NCP1 0
262 target_ulong CP0_YQMask;
263 target_ulong CP0_VPESchedule;
264 target_ulong CP0_VPEScheFBack;
265 int32_t CP0_VPEOpt;
266 #define CP0VPEOpt_IWX7 15
267 #define CP0VPEOpt_IWX6 14
268 #define CP0VPEOpt_IWX5 13
269 #define CP0VPEOpt_IWX4 12
270 #define CP0VPEOpt_IWX3 11
271 #define CP0VPEOpt_IWX2 10
272 #define CP0VPEOpt_IWX1 9
273 #define CP0VPEOpt_IWX0 8
274 #define CP0VPEOpt_DWX7 7
275 #define CP0VPEOpt_DWX6 6
276 #define CP0VPEOpt_DWX5 5
277 #define CP0VPEOpt_DWX4 4
278 #define CP0VPEOpt_DWX3 3
279 #define CP0VPEOpt_DWX2 2
280 #define CP0VPEOpt_DWX1 1
281 #define CP0VPEOpt_DWX0 0
282 uint64_t CP0_EntryLo0;
283 uint64_t CP0_EntryLo1;
284 #if defined(TARGET_MIPS64)
285 # define CP0EnLo_RI 63
286 # define CP0EnLo_XI 62
287 #else
288 # define CP0EnLo_RI 31
289 # define CP0EnLo_XI 30
290 #endif
291 int32_t CP0_GlobalNumber;
292 #define CP0GN_VPId 0
293 target_ulong CP0_Context;
294 target_ulong CP0_KScratch[MIPS_KSCRATCH_NUM];
295 int32_t CP0_PageMask;
296 int32_t CP0_PageGrain_rw_bitmask;
297 int32_t CP0_PageGrain;
298 #define CP0PG_RIE 31
299 #define CP0PG_XIE 30
300 #define CP0PG_ELPA 29
301 #define CP0PG_IEC 27
302 int32_t CP0_Wired;
303 int32_t CP0_SRSConf0_rw_bitmask;
304 int32_t CP0_SRSConf0;
305 #define CP0SRSC0_M 31
306 #define CP0SRSC0_SRS3 20
307 #define CP0SRSC0_SRS2 10
308 #define CP0SRSC0_SRS1 0
309 int32_t CP0_SRSConf1_rw_bitmask;
310 int32_t CP0_SRSConf1;
311 #define CP0SRSC1_M 31
312 #define CP0SRSC1_SRS6 20
313 #define CP0SRSC1_SRS5 10
314 #define CP0SRSC1_SRS4 0
315 int32_t CP0_SRSConf2_rw_bitmask;
316 int32_t CP0_SRSConf2;
317 #define CP0SRSC2_M 31
318 #define CP0SRSC2_SRS9 20
319 #define CP0SRSC2_SRS8 10
320 #define CP0SRSC2_SRS7 0
321 int32_t CP0_SRSConf3_rw_bitmask;
322 int32_t CP0_SRSConf3;
323 #define CP0SRSC3_M 31
324 #define CP0SRSC3_SRS12 20
325 #define CP0SRSC3_SRS11 10
326 #define CP0SRSC3_SRS10 0
327 int32_t CP0_SRSConf4_rw_bitmask;
328 int32_t CP0_SRSConf4;
329 #define CP0SRSC4_SRS15 20
330 #define CP0SRSC4_SRS14 10
331 #define CP0SRSC4_SRS13 0
332 int32_t CP0_HWREna;
333 target_ulong CP0_BadVAddr;
334 uint32_t CP0_BadInstr;
335 uint32_t CP0_BadInstrP;
336 int32_t CP0_Count;
337 target_ulong CP0_EntryHi;
338 #define CP0EnHi_EHINV 10
339 int32_t CP0_Compare;
340 int32_t CP0_Status;
341 #define CP0St_CU3 31
342 #define CP0St_CU2 30
343 #define CP0St_CU1 29
344 #define CP0St_CU0 28
345 #define CP0St_RP 27
346 #define CP0St_FR 26
347 #define CP0St_RE 25
348 #define CP0St_MX 24
349 #define CP0St_PX 23
350 #define CP0St_BEV 22
351 #define CP0St_TS 21
352 #define CP0St_SR 20
353 #define CP0St_NMI 19
354 #define CP0St_IM 8
355 #define CP0St_KX 7
356 #define CP0St_SX 6
357 #define CP0St_UX 5
358 #define CP0St_KSU 3
359 #define CP0St_ERL 2
360 #define CP0St_EXL 1
361 #define CP0St_IE 0
362 int32_t CP0_IntCtl;
363 #define CP0IntCtl_IPTI 29
364 #define CP0IntCtl_IPPCI 26
365 #define CP0IntCtl_VS 5
366 int32_t CP0_SRSCtl;
367 #define CP0SRSCtl_HSS 26
368 #define CP0SRSCtl_EICSS 18
369 #define CP0SRSCtl_ESS 12
370 #define CP0SRSCtl_PSS 6
371 #define CP0SRSCtl_CSS 0
372 int32_t CP0_SRSMap;
373 #define CP0SRSMap_SSV7 28
374 #define CP0SRSMap_SSV6 24
375 #define CP0SRSMap_SSV5 20
376 #define CP0SRSMap_SSV4 16
377 #define CP0SRSMap_SSV3 12
378 #define CP0SRSMap_SSV2 8
379 #define CP0SRSMap_SSV1 4
380 #define CP0SRSMap_SSV0 0
381 int32_t CP0_Cause;
382 #define CP0Ca_BD 31
383 #define CP0Ca_TI 30
384 #define CP0Ca_CE 28
385 #define CP0Ca_DC 27
386 #define CP0Ca_PCI 26
387 #define CP0Ca_IV 23
388 #define CP0Ca_WP 22
389 #define CP0Ca_IP 8
390 #define CP0Ca_IP_mask 0x0000FF00
391 #define CP0Ca_EC 2
392 target_ulong CP0_EPC;
393 int32_t CP0_PRid;
394 int32_t CP0_EBase;
395 int32_t CP0_Config0;
396 #define CP0C0_M 31
397 #define CP0C0_K23 28
398 #define CP0C0_KU 25
399 #define CP0C0_MDU 20
400 #define CP0C0_MM 18
401 #define CP0C0_BM 16
402 #define CP0C0_BE 15
403 #define CP0C0_AT 13
404 #define CP0C0_AR 10
405 #define CP0C0_MT 7
406 #define CP0C0_VI 3
407 #define CP0C0_K0 0
408 int32_t CP0_Config1;
409 #define CP0C1_M 31
410 #define CP0C1_MMU 25
411 #define CP0C1_IS 22
412 #define CP0C1_IL 19
413 #define CP0C1_IA 16
414 #define CP0C1_DS 13
415 #define CP0C1_DL 10
416 #define CP0C1_DA 7
417 #define CP0C1_C2 6
418 #define CP0C1_MD 5
419 #define CP0C1_PC 4
420 #define CP0C1_WR 3
421 #define CP0C1_CA 2
422 #define CP0C1_EP 1
423 #define CP0C1_FP 0
424 int32_t CP0_Config2;
425 #define CP0C2_M 31
426 #define CP0C2_TU 28
427 #define CP0C2_TS 24
428 #define CP0C2_TL 20
429 #define CP0C2_TA 16
430 #define CP0C2_SU 12
431 #define CP0C2_SS 8
432 #define CP0C2_SL 4
433 #define CP0C2_SA 0
434 int32_t CP0_Config3;
435 #define CP0C3_M 31
436 #define CP0C3_BPG 30
437 #define CP0C3_CMCGR 29
438 #define CP0C3_MSAP 28
439 #define CP0C3_BP 27
440 #define CP0C3_BI 26
441 #define CP0C3_IPLW 21
442 #define CP0C3_MMAR 18
443 #define CP0C3_MCU 17
444 #define CP0C3_ISA_ON_EXC 16
445 #define CP0C3_ISA 14
446 #define CP0C3_ULRI 13
447 #define CP0C3_RXI 12
448 #define CP0C3_DSP2P 11
449 #define CP0C3_DSPP 10
450 #define CP0C3_LPA 7
451 #define CP0C3_VEIC 6
452 #define CP0C3_VInt 5
453 #define CP0C3_SP 4
454 #define CP0C3_CDMM 3
455 #define CP0C3_MT 2
456 #define CP0C3_SM 1
457 #define CP0C3_TL 0
458 int32_t CP0_Config4;
459 int32_t CP0_Config4_rw_bitmask;
460 #define CP0C4_M 31
461 #define CP0C4_IE 29
462 #define CP0C4_KScrExist 16
463 #define CP0C4_MMUExtDef 14
464 #define CP0C4_FTLBPageSize 8
465 #define CP0C4_FTLBWays 4
466 #define CP0C4_FTLBSets 0
467 #define CP0C4_MMUSizeExt 0
468 int32_t CP0_Config5;
469 int32_t CP0_Config5_rw_bitmask;
470 #define CP0C5_M 31
471 #define CP0C5_K 30
472 #define CP0C5_CV 29
473 #define CP0C5_EVA 28
474 #define CP0C5_MSAEn 27
475 #define CP0C5_XNP 13
476 #define CP0C5_UFE 9
477 #define CP0C5_FRE 8
478 #define CP0C5_VP 7
479 #define CP0C5_SBRI 6
480 #define CP0C5_MVH 5
481 #define CP0C5_LLB 4
482 #define CP0C5_UFR 2
483 #define CP0C5_NFExists 0
484 int32_t CP0_Config6;
485 int32_t CP0_Config7;
486 /* XXX: Maybe make LLAddr per-TC? */
487 uint64_t lladdr;
488 target_ulong llval;
489 target_ulong llnewval;
490 target_ulong llreg;
491 uint64_t CP0_LLAddr_rw_bitmask;
492 int CP0_LLAddr_shift;
493 target_ulong CP0_WatchLo[8];
494 int32_t CP0_WatchHi[8];
495 target_ulong CP0_XContext;
496 int32_t CP0_Framemask;
497 int32_t CP0_Debug;
498 #define CP0DB_DBD 31
499 #define CP0DB_DM 30
500 #define CP0DB_LSNM 28
501 #define CP0DB_Doze 27
502 #define CP0DB_Halt 26
503 #define CP0DB_CNT 25
504 #define CP0DB_IBEP 24
505 #define CP0DB_DBEP 21
506 #define CP0DB_IEXI 20
507 #define CP0DB_VER 15
508 #define CP0DB_DEC 10
509 #define CP0DB_SSt 8
510 #define CP0DB_DINT 5
511 #define CP0DB_DIB 4
512 #define CP0DB_DDBS 3
513 #define CP0DB_DDBL 2
514 #define CP0DB_DBp 1
515 #define CP0DB_DSS 0
516 target_ulong CP0_DEPC;
517 int32_t CP0_Performance0;
518 uint64_t CP0_TagLo;
519 int32_t CP0_DataLo;
520 int32_t CP0_TagHi;
521 int32_t CP0_DataHi;
522 target_ulong CP0_ErrorEPC;
523 int32_t CP0_DESAVE;
524 /* We waste some space so we can handle shadow registers like TCs. */
525 TCState tcs[MIPS_SHADOW_SET_MAX];
526 CPUMIPSFPUContext fpus[MIPS_FPU_MAX];
527 /* QEMU */
528 int error_code;
529 #define EXCP_TLB_NOMATCH 0x1
530 #define EXCP_INST_NOTAVAIL 0x2 /* No valid instruction word for BadInstr */
531 uint32_t hflags; /* CPU State */
532 /* TMASK defines different execution modes */
533 #define MIPS_HFLAG_TMASK 0x75807FF
534 #define MIPS_HFLAG_MODE 0x00007 /* execution modes */
535 /* The KSU flags must be the lowest bits in hflags. The flag order
536 must be the same as defined for CP0 Status. This allows to use
537 the bits as the value of mmu_idx. */
538 #define MIPS_HFLAG_KSU 0x00003 /* kernel/supervisor/user mode mask */
539 #define MIPS_HFLAG_UM 0x00002 /* user mode flag */
540 #define MIPS_HFLAG_SM 0x00001 /* supervisor mode flag */
541 #define MIPS_HFLAG_KM 0x00000 /* kernel mode flag */
542 #define MIPS_HFLAG_DM 0x00004 /* Debug mode */
543 #define MIPS_HFLAG_64 0x00008 /* 64-bit instructions enabled */
544 #define MIPS_HFLAG_CP0 0x00010 /* CP0 enabled */
545 #define MIPS_HFLAG_FPU 0x00020 /* FPU enabled */
546 #define MIPS_HFLAG_F64 0x00040 /* 64-bit FPU enabled */
547 /* True if the MIPS IV COP1X instructions can be used. This also
548 controls the non-COP1X instructions RECIP.S, RECIP.D, RSQRT.S
549 and RSQRT.D. */
550 #define MIPS_HFLAG_COP1X 0x00080 /* COP1X instructions enabled */
551 #define MIPS_HFLAG_RE 0x00100 /* Reversed endianness */
552 #define MIPS_HFLAG_AWRAP 0x00200 /* 32-bit compatibility address wrapping */
553 #define MIPS_HFLAG_M16 0x00400 /* MIPS16 mode flag */
554 #define MIPS_HFLAG_M16_SHIFT 10
555 /* If translation is interrupted between the branch instruction and
556 * the delay slot, record what type of branch it is so that we can
557 * resume translation properly. It might be possible to reduce
558 * this from three bits to two. */
559 #define MIPS_HFLAG_BMASK_BASE 0x803800
560 #define MIPS_HFLAG_B 0x00800 /* Unconditional branch */
561 #define MIPS_HFLAG_BC 0x01000 /* Conditional branch */
562 #define MIPS_HFLAG_BL 0x01800 /* Likely branch */
563 #define MIPS_HFLAG_BR 0x02000 /* branch to register (can't link TB) */
564 /* Extra flags about the current pending branch. */
565 #define MIPS_HFLAG_BMASK_EXT 0x7C000
566 #define MIPS_HFLAG_B16 0x04000 /* branch instruction was 16 bits */
567 #define MIPS_HFLAG_BDS16 0x08000 /* branch requires 16-bit delay slot */
568 #define MIPS_HFLAG_BDS32 0x10000 /* branch requires 32-bit delay slot */
569 #define MIPS_HFLAG_BDS_STRICT 0x20000 /* Strict delay slot size */
570 #define MIPS_HFLAG_BX 0x40000 /* branch exchanges execution mode */
571 #define MIPS_HFLAG_BMASK (MIPS_HFLAG_BMASK_BASE | MIPS_HFLAG_BMASK_EXT)
572 /* MIPS DSP resources access. */
573 #define MIPS_HFLAG_DSP 0x080000 /* Enable access to MIPS DSP resources. */
574 #define MIPS_HFLAG_DSPR2 0x100000 /* Enable access to MIPS DSPR2 resources. */
575 /* Extra flag about HWREna register. */
576 #define MIPS_HFLAG_HWRENA_ULR 0x200000 /* ULR bit from HWREna is set. */
577 #define MIPS_HFLAG_SBRI 0x400000 /* R6 SDBBP causes RI excpt. in user mode */
578 #define MIPS_HFLAG_FBNSLOT 0x800000 /* Forbidden slot */
579 #define MIPS_HFLAG_MSA 0x1000000
580 #define MIPS_HFLAG_FRE 0x2000000 /* FRE enabled */
581 #define MIPS_HFLAG_ELPA 0x4000000
582 target_ulong btarget; /* Jump / branch target */
583 target_ulong bcond; /* Branch condition (if needed) */
585 int SYNCI_Step; /* Address step size for SYNCI */
586 int CCRes; /* Cycle count resolution/divisor */
587 uint32_t CP0_Status_rw_bitmask; /* Read/write bits in CP0_Status */
588 uint32_t CP0_TCStatus_rw_bitmask; /* Read/write bits in CP0_TCStatus */
589 int insn_flags; /* Supported instruction set */
591 CPU_COMMON
593 /* Fields from here on are preserved across CPU reset. */
594 CPUMIPSMVPContext *mvp;
595 #if !defined(CONFIG_USER_ONLY)
596 CPUMIPSTLBContext *tlb;
597 #endif
599 const mips_def_t *cpu_model;
600 void *irq[8];
601 QEMUTimer *timer; /* Internal timer */
604 #include "cpu-qom.h"
606 #if !defined(CONFIG_USER_ONLY)
607 int no_mmu_map_address (CPUMIPSState *env, hwaddr *physical, int *prot,
608 target_ulong address, int rw, int access_type);
609 int fixed_mmu_map_address (CPUMIPSState *env, hwaddr *physical, int *prot,
610 target_ulong address, int rw, int access_type);
611 int r4k_map_address (CPUMIPSState *env, hwaddr *physical, int *prot,
612 target_ulong address, int rw, int access_type);
613 void r4k_helper_tlbwi(CPUMIPSState *env);
614 void r4k_helper_tlbwr(CPUMIPSState *env);
615 void r4k_helper_tlbp(CPUMIPSState *env);
616 void r4k_helper_tlbr(CPUMIPSState *env);
617 void r4k_helper_tlbinv(CPUMIPSState *env);
618 void r4k_helper_tlbinvf(CPUMIPSState *env);
620 void mips_cpu_unassigned_access(CPUState *cpu, hwaddr addr,
621 bool is_write, bool is_exec, int unused,
622 unsigned size);
623 #endif
625 void mips_cpu_list (FILE *f, fprintf_function cpu_fprintf);
627 #define cpu_exec cpu_mips_exec
628 #define cpu_signal_handler cpu_mips_signal_handler
629 #define cpu_list mips_cpu_list
631 extern void cpu_wrdsp(uint32_t rs, uint32_t mask_num, CPUMIPSState *env);
632 extern uint32_t cpu_rddsp(uint32_t mask_num, CPUMIPSState *env);
634 /* MMU modes definitions. We carefully match the indices with our
635 hflags layout. */
636 #define MMU_MODE0_SUFFIX _kernel
637 #define MMU_MODE1_SUFFIX _super
638 #define MMU_MODE2_SUFFIX _user
639 #define MMU_USER_IDX 2
640 static inline int cpu_mmu_index (CPUMIPSState *env, bool ifetch)
642 return env->hflags & MIPS_HFLAG_KSU;
645 static inline bool cpu_mips_hw_interrupts_enabled(CPUMIPSState *env)
647 return (env->CP0_Status & (1 << CP0St_IE)) &&
648 !(env->CP0_Status & (1 << CP0St_EXL)) &&
649 !(env->CP0_Status & (1 << CP0St_ERL)) &&
650 !(env->hflags & MIPS_HFLAG_DM) &&
651 /* Note that the TCStatus IXMT field is initialized to zero,
652 and only MT capable cores can set it to one. So we don't
653 need to check for MT capabilities here. */
654 !(env->active_tc.CP0_TCStatus & (1 << CP0TCSt_IXMT));
657 /* Check if there is pending and not masked out interrupt */
658 static inline bool cpu_mips_hw_interrupts_pending(CPUMIPSState *env)
660 int32_t pending;
661 int32_t status;
662 bool r;
664 pending = env->CP0_Cause & CP0Ca_IP_mask;
665 status = env->CP0_Status & CP0Ca_IP_mask;
667 if (env->CP0_Config3 & (1 << CP0C3_VEIC)) {
668 /* A MIPS configured with a vectorizing external interrupt controller
669 will feed a vector into the Cause pending lines. The core treats
670 the status lines as a vector level, not as indiviual masks. */
671 r = pending > status;
672 } else {
673 /* A MIPS configured with compatibility or VInt (Vectored Interrupts)
674 treats the pending lines as individual interrupt lines, the status
675 lines are individual masks. */
676 r = (pending & status) != 0;
678 return r;
681 #include "exec/cpu-all.h"
683 /* Memory access type :
684 * may be needed for precise access rights control and precise exceptions.
686 enum {
687 /* 1 bit to define user level / supervisor access */
688 ACCESS_USER = 0x00,
689 ACCESS_SUPER = 0x01,
690 /* 1 bit to indicate direction */
691 ACCESS_STORE = 0x02,
692 /* Type of instruction that generated the access */
693 ACCESS_CODE = 0x10, /* Code fetch access */
694 ACCESS_INT = 0x20, /* Integer load/store access */
695 ACCESS_FLOAT = 0x30, /* floating point load/store access */
698 /* Exceptions */
699 enum {
700 EXCP_NONE = -1,
701 EXCP_RESET = 0,
702 EXCP_SRESET,
703 EXCP_DSS,
704 EXCP_DINT,
705 EXCP_DDBL,
706 EXCP_DDBS,
707 EXCP_NMI,
708 EXCP_MCHECK,
709 EXCP_EXT_INTERRUPT, /* 8 */
710 EXCP_DFWATCH,
711 EXCP_DIB,
712 EXCP_IWATCH,
713 EXCP_AdEL,
714 EXCP_AdES,
715 EXCP_TLBF,
716 EXCP_IBE,
717 EXCP_DBp, /* 16 */
718 EXCP_SYSCALL,
719 EXCP_BREAK,
720 EXCP_CpU,
721 EXCP_RI,
722 EXCP_OVERFLOW,
723 EXCP_TRAP,
724 EXCP_FPE,
725 EXCP_DWATCH, /* 24 */
726 EXCP_LTLBL,
727 EXCP_TLBL,
728 EXCP_TLBS,
729 EXCP_DBE,
730 EXCP_THREAD,
731 EXCP_MDMX,
732 EXCP_C2E,
733 EXCP_CACHE, /* 32 */
734 EXCP_DSPDIS,
735 EXCP_MSADIS,
736 EXCP_MSAFPE,
737 EXCP_TLBXI,
738 EXCP_TLBRI,
740 EXCP_LAST = EXCP_TLBRI,
742 /* Dummy exception for conditional stores. */
743 #define EXCP_SC 0x100
746 * This is an interrnally generated WAKE request line.
747 * It is driven by the CPU itself. Raised when the MT
748 * block wants to wake a VPE from an inactive state and
749 * cleared when VPE goes from active to inactive.
751 #define CPU_INTERRUPT_WAKE CPU_INTERRUPT_TGT_INT_0
753 int cpu_mips_exec(CPUState *cpu);
754 void mips_tcg_init(void);
755 MIPSCPU *cpu_mips_init(const char *cpu_model);
756 int cpu_mips_signal_handler(int host_signum, void *pinfo, void *puc);
758 #define cpu_init(cpu_model) CPU(cpu_mips_init(cpu_model))
760 /* TODO QOM'ify CPU reset and remove */
761 void cpu_state_reset(CPUMIPSState *s);
763 /* mips_timer.c */
764 uint32_t cpu_mips_get_random (CPUMIPSState *env);
765 uint32_t cpu_mips_get_count (CPUMIPSState *env);
766 void cpu_mips_store_count (CPUMIPSState *env, uint32_t value);
767 void cpu_mips_store_compare (CPUMIPSState *env, uint32_t value);
768 void cpu_mips_start_count(CPUMIPSState *env);
769 void cpu_mips_stop_count(CPUMIPSState *env);
771 /* mips_int.c */
772 void cpu_mips_soft_irq(CPUMIPSState *env, int irq, int level);
774 /* helper.c */
775 int mips_cpu_handle_mmu_fault(CPUState *cpu, vaddr address, int rw,
776 int mmu_idx);
777 #if !defined(CONFIG_USER_ONLY)
778 void r4k_invalidate_tlb (CPUMIPSState *env, int idx, int use_extra);
779 hwaddr cpu_mips_translate_address (CPUMIPSState *env, target_ulong address,
780 int rw);
781 #endif
782 target_ulong exception_resume_pc (CPUMIPSState *env);
784 /* op_helper.c */
785 extern unsigned int ieee_rm[];
786 int ieee_ex_to_mips(int xcpt);
788 static inline void restore_rounding_mode(CPUMIPSState *env)
790 set_float_rounding_mode(ieee_rm[env->active_fpu.fcr31 & 3],
791 &env->active_fpu.fp_status);
794 static inline void restore_flush_mode(CPUMIPSState *env)
796 set_flush_to_zero((env->active_fpu.fcr31 & (1 << 24)) != 0,
797 &env->active_fpu.fp_status);
800 static inline void restore_fp_status(CPUMIPSState *env)
802 restore_rounding_mode(env);
803 restore_flush_mode(env);
806 static inline void restore_msa_fp_status(CPUMIPSState *env)
808 float_status *status = &env->active_tc.msa_fp_status;
809 int rounding_mode = (env->active_tc.msacsr & MSACSR_RM_MASK) >> MSACSR_RM;
810 bool flush_to_zero = (env->active_tc.msacsr & MSACSR_FS_MASK) != 0;
812 set_float_rounding_mode(ieee_rm[rounding_mode], status);
813 set_flush_to_zero(flush_to_zero, status);
814 set_flush_inputs_to_zero(flush_to_zero, status);
817 static inline void restore_pamask(CPUMIPSState *env)
819 if (env->hflags & MIPS_HFLAG_ELPA) {
820 env->PAMask = (1ULL << env->PABITS) - 1;
821 } else {
822 env->PAMask = PAMASK_BASE;
826 static inline void cpu_get_tb_cpu_state(CPUMIPSState *env, target_ulong *pc,
827 target_ulong *cs_base, int *flags)
829 *pc = env->active_tc.PC;
830 *cs_base = 0;
831 *flags = env->hflags & (MIPS_HFLAG_TMASK | MIPS_HFLAG_BMASK |
832 MIPS_HFLAG_HWRENA_ULR);
835 static inline int mips_vpe_active(CPUMIPSState *env)
837 int active = 1;
839 /* Check that the VPE is enabled. */
840 if (!(env->mvp->CP0_MVPControl & (1 << CP0MVPCo_EVP))) {
841 active = 0;
843 /* Check that the VPE is activated. */
844 if (!(env->CP0_VPEConf0 & (1 << CP0VPEC0_VPA))) {
845 active = 0;
848 /* Now verify that there are active thread contexts in the VPE.
850 This assumes the CPU model will internally reschedule threads
851 if the active one goes to sleep. If there are no threads available
852 the active one will be in a sleeping state, and we can turn off
853 the entire VPE. */
854 if (!(env->active_tc.CP0_TCStatus & (1 << CP0TCSt_A))) {
855 /* TC is not activated. */
856 active = 0;
858 if (env->active_tc.CP0_TCHalt & 1) {
859 /* TC is in halt state. */
860 active = 0;
863 return active;
866 static inline int mips_vp_active(CPUMIPSState *env)
868 CPUState *other_cs = first_cpu;
870 /* Check if the VP disabled other VPs (which means the VP is enabled) */
871 if ((env->CP0_VPControl >> CP0VPCtl_DIS) & 1) {
872 return 1;
875 /* Check if the virtual processor is disabled due to a DVP */
876 CPU_FOREACH(other_cs) {
877 MIPSCPU *other_cpu = MIPS_CPU(other_cs);
878 if ((&other_cpu->env != env) &&
879 ((other_cpu->env.CP0_VPControl >> CP0VPCtl_DIS) & 1)) {
880 return 0;
883 return 1;
886 #include "exec/exec-all.h"
888 static inline void compute_hflags(CPUMIPSState *env)
890 env->hflags &= ~(MIPS_HFLAG_COP1X | MIPS_HFLAG_64 | MIPS_HFLAG_CP0 |
891 MIPS_HFLAG_F64 | MIPS_HFLAG_FPU | MIPS_HFLAG_KSU |
892 MIPS_HFLAG_AWRAP | MIPS_HFLAG_DSP | MIPS_HFLAG_DSPR2 |
893 MIPS_HFLAG_SBRI | MIPS_HFLAG_MSA | MIPS_HFLAG_FRE |
894 MIPS_HFLAG_ELPA);
895 if (!(env->CP0_Status & (1 << CP0St_EXL)) &&
896 !(env->CP0_Status & (1 << CP0St_ERL)) &&
897 !(env->hflags & MIPS_HFLAG_DM)) {
898 env->hflags |= (env->CP0_Status >> CP0St_KSU) & MIPS_HFLAG_KSU;
900 #if defined(TARGET_MIPS64)
901 if ((env->insn_flags & ISA_MIPS3) &&
902 (((env->hflags & MIPS_HFLAG_KSU) != MIPS_HFLAG_UM) ||
903 (env->CP0_Status & (1 << CP0St_PX)) ||
904 (env->CP0_Status & (1 << CP0St_UX)))) {
905 env->hflags |= MIPS_HFLAG_64;
908 if (!(env->insn_flags & ISA_MIPS3)) {
909 env->hflags |= MIPS_HFLAG_AWRAP;
910 } else if (((env->hflags & MIPS_HFLAG_KSU) == MIPS_HFLAG_UM) &&
911 !(env->CP0_Status & (1 << CP0St_UX))) {
912 env->hflags |= MIPS_HFLAG_AWRAP;
913 } else if (env->insn_flags & ISA_MIPS64R6) {
914 /* Address wrapping for Supervisor and Kernel is specified in R6 */
915 if ((((env->hflags & MIPS_HFLAG_KSU) == MIPS_HFLAG_SM) &&
916 !(env->CP0_Status & (1 << CP0St_SX))) ||
917 (((env->hflags & MIPS_HFLAG_KSU) == MIPS_HFLAG_KM) &&
918 !(env->CP0_Status & (1 << CP0St_KX)))) {
919 env->hflags |= MIPS_HFLAG_AWRAP;
922 #endif
923 if (((env->CP0_Status & (1 << CP0St_CU0)) &&
924 !(env->insn_flags & ISA_MIPS32R6)) ||
925 !(env->hflags & MIPS_HFLAG_KSU)) {
926 env->hflags |= MIPS_HFLAG_CP0;
928 if (env->CP0_Status & (1 << CP0St_CU1)) {
929 env->hflags |= MIPS_HFLAG_FPU;
931 if (env->CP0_Status & (1 << CP0St_FR)) {
932 env->hflags |= MIPS_HFLAG_F64;
934 if (((env->hflags & MIPS_HFLAG_KSU) != MIPS_HFLAG_KM) &&
935 (env->CP0_Config5 & (1 << CP0C5_SBRI))) {
936 env->hflags |= MIPS_HFLAG_SBRI;
938 if (env->insn_flags & ASE_DSPR2) {
939 /* Enables access MIPS DSP resources, now our cpu is DSP ASER2,
940 so enable to access DSPR2 resources. */
941 if (env->CP0_Status & (1 << CP0St_MX)) {
942 env->hflags |= MIPS_HFLAG_DSP | MIPS_HFLAG_DSPR2;
945 } else if (env->insn_flags & ASE_DSP) {
946 /* Enables access MIPS DSP resources, now our cpu is DSP ASE,
947 so enable to access DSP resources. */
948 if (env->CP0_Status & (1 << CP0St_MX)) {
949 env->hflags |= MIPS_HFLAG_DSP;
953 if (env->insn_flags & ISA_MIPS32R2) {
954 if (env->active_fpu.fcr0 & (1 << FCR0_F64)) {
955 env->hflags |= MIPS_HFLAG_COP1X;
957 } else if (env->insn_flags & ISA_MIPS32) {
958 if (env->hflags & MIPS_HFLAG_64) {
959 env->hflags |= MIPS_HFLAG_COP1X;
961 } else if (env->insn_flags & ISA_MIPS4) {
962 /* All supported MIPS IV CPUs use the XX (CU3) to enable
963 and disable the MIPS IV extensions to the MIPS III ISA.
964 Some other MIPS IV CPUs ignore the bit, so the check here
965 would be too restrictive for them. */
966 if (env->CP0_Status & (1U << CP0St_CU3)) {
967 env->hflags |= MIPS_HFLAG_COP1X;
970 if (env->insn_flags & ASE_MSA) {
971 if (env->CP0_Config5 & (1 << CP0C5_MSAEn)) {
972 env->hflags |= MIPS_HFLAG_MSA;
975 if (env->active_fpu.fcr0 & (1 << FCR0_FREP)) {
976 if (env->CP0_Config5 & (1 << CP0C5_FRE)) {
977 env->hflags |= MIPS_HFLAG_FRE;
980 if (env->CP0_Config3 & (1 << CP0C3_LPA)) {
981 if (env->CP0_PageGrain & (1 << CP0PG_ELPA)) {
982 env->hflags |= MIPS_HFLAG_ELPA;
987 #ifndef CONFIG_USER_ONLY
988 static inline void cpu_mips_tlb_flush(CPUMIPSState *env, int flush_global)
990 MIPSCPU *cpu = mips_env_get_cpu(env);
992 /* Flush qemu's TLB and discard all shadowed entries. */
993 tlb_flush(CPU(cpu), flush_global);
994 env->tlb->tlb_in_use = env->tlb->nb_tlb;
997 /* Called for updates to CP0_Status. */
998 static inline void sync_c0_status(CPUMIPSState *env, CPUMIPSState *cpu, int tc)
1000 int32_t tcstatus, *tcst;
1001 uint32_t v = cpu->CP0_Status;
1002 uint32_t cu, mx, asid, ksu;
1003 uint32_t mask = ((1 << CP0TCSt_TCU3)
1004 | (1 << CP0TCSt_TCU2)
1005 | (1 << CP0TCSt_TCU1)
1006 | (1 << CP0TCSt_TCU0)
1007 | (1 << CP0TCSt_TMX)
1008 | (3 << CP0TCSt_TKSU)
1009 | (0xff << CP0TCSt_TASID));
1011 cu = (v >> CP0St_CU0) & 0xf;
1012 mx = (v >> CP0St_MX) & 0x1;
1013 ksu = (v >> CP0St_KSU) & 0x3;
1014 asid = env->CP0_EntryHi & 0xff;
1016 tcstatus = cu << CP0TCSt_TCU0;
1017 tcstatus |= mx << CP0TCSt_TMX;
1018 tcstatus |= ksu << CP0TCSt_TKSU;
1019 tcstatus |= asid;
1021 if (tc == cpu->current_tc) {
1022 tcst = &cpu->active_tc.CP0_TCStatus;
1023 } else {
1024 tcst = &cpu->tcs[tc].CP0_TCStatus;
1027 *tcst &= ~mask;
1028 *tcst |= tcstatus;
1029 compute_hflags(cpu);
1032 static inline void cpu_mips_store_status(CPUMIPSState *env, target_ulong val)
1034 uint32_t mask = env->CP0_Status_rw_bitmask;
1035 target_ulong old = env->CP0_Status;
1037 if (env->insn_flags & ISA_MIPS32R6) {
1038 bool has_supervisor = extract32(mask, CP0St_KSU, 2) == 0x3;
1039 #if defined(TARGET_MIPS64)
1040 uint32_t ksux = (1 << CP0St_KX) & val;
1041 ksux |= (ksux >> 1) & val; /* KX = 0 forces SX to be 0 */
1042 ksux |= (ksux >> 1) & val; /* SX = 0 forces UX to be 0 */
1043 val = (val & ~(7 << CP0St_UX)) | ksux;
1044 #endif
1045 if (has_supervisor && extract32(val, CP0St_KSU, 2) == 0x3) {
1046 mask &= ~(3 << CP0St_KSU);
1048 mask &= ~(((1 << CP0St_SR) | (1 << CP0St_NMI)) & val);
1051 env->CP0_Status = (old & ~mask) | (val & mask);
1052 #if defined(TARGET_MIPS64)
1053 if ((env->CP0_Status ^ old) & (old & (7 << CP0St_UX))) {
1054 /* Access to at least one of the 64-bit segments has been disabled */
1055 cpu_mips_tlb_flush(env, 1);
1057 #endif
1058 if (env->CP0_Config3 & (1 << CP0C3_MT)) {
1059 sync_c0_status(env, env, env->current_tc);
1060 } else {
1061 compute_hflags(env);
1065 static inline void cpu_mips_store_cause(CPUMIPSState *env, target_ulong val)
1067 uint32_t mask = 0x00C00300;
1068 uint32_t old = env->CP0_Cause;
1069 int i;
1071 if (env->insn_flags & ISA_MIPS32R2) {
1072 mask |= 1 << CP0Ca_DC;
1074 if (env->insn_flags & ISA_MIPS32R6) {
1075 mask &= ~((1 << CP0Ca_WP) & val);
1078 env->CP0_Cause = (env->CP0_Cause & ~mask) | (val & mask);
1080 if ((old ^ env->CP0_Cause) & (1 << CP0Ca_DC)) {
1081 if (env->CP0_Cause & (1 << CP0Ca_DC)) {
1082 cpu_mips_stop_count(env);
1083 } else {
1084 cpu_mips_start_count(env);
1088 /* Set/reset software interrupts */
1089 for (i = 0 ; i < 2 ; i++) {
1090 if ((old ^ env->CP0_Cause) & (1 << (CP0Ca_IP + i))) {
1091 cpu_mips_soft_irq(env, i, env->CP0_Cause & (1 << (CP0Ca_IP + i)));
1095 #endif
1097 static inline void QEMU_NORETURN do_raise_exception_err(CPUMIPSState *env,
1098 uint32_t exception,
1099 int error_code,
1100 uintptr_t pc)
1102 CPUState *cs = CPU(mips_env_get_cpu(env));
1104 if (exception < EXCP_SC) {
1105 qemu_log_mask(CPU_LOG_INT, "%s: %d %d\n",
1106 __func__, exception, error_code);
1108 cs->exception_index = exception;
1109 env->error_code = error_code;
1111 cpu_loop_exit_restore(cs, pc);
1114 static inline void QEMU_NORETURN do_raise_exception(CPUMIPSState *env,
1115 uint32_t exception,
1116 uintptr_t pc)
1118 do_raise_exception_err(env, exception, 0, pc);
1121 #endif /* !defined (__MIPS_CPU_H__) */