1 #include "qemu/osdep.h"
3 #include "qemu/error-report.h"
4 #include "sysemu/kvm.h"
7 #include "migration/cpu.h"
9 static bool vfp_needed(void *opaque
)
13 return (arm_feature(&cpu
->env
, ARM_FEATURE_AARCH64
)
14 ? cpu_isar_feature(aa64_fp_simd
, cpu
)
15 : cpu_isar_feature(aa32_vfp_simd
, cpu
));
18 static int get_fpscr(QEMUFile
*f
, void *opaque
, size_t size
,
19 const VMStateField
*field
)
22 CPUARMState
*env
= &cpu
->env
;
23 uint32_t val
= qemu_get_be32(f
);
25 vfp_set_fpscr(env
, val
);
29 static int put_fpscr(QEMUFile
*f
, void *opaque
, size_t size
,
30 const VMStateField
*field
, JSONWriter
*vmdesc
)
33 CPUARMState
*env
= &cpu
->env
;
35 qemu_put_be32(f
, vfp_get_fpscr(env
));
39 static const VMStateInfo vmstate_fpscr
= {
45 static const VMStateDescription vmstate_vfp
= {
48 .minimum_version_id
= 3,
50 .fields
= (VMStateField
[]) {
51 /* For compatibility, store Qn out of Zn here. */
52 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[0].d
, ARMCPU
, 0, 2),
53 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[1].d
, ARMCPU
, 0, 2),
54 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[2].d
, ARMCPU
, 0, 2),
55 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[3].d
, ARMCPU
, 0, 2),
56 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[4].d
, ARMCPU
, 0, 2),
57 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[5].d
, ARMCPU
, 0, 2),
58 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[6].d
, ARMCPU
, 0, 2),
59 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[7].d
, ARMCPU
, 0, 2),
60 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[8].d
, ARMCPU
, 0, 2),
61 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[9].d
, ARMCPU
, 0, 2),
62 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[10].d
, ARMCPU
, 0, 2),
63 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[11].d
, ARMCPU
, 0, 2),
64 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[12].d
, ARMCPU
, 0, 2),
65 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[13].d
, ARMCPU
, 0, 2),
66 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[14].d
, ARMCPU
, 0, 2),
67 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[15].d
, ARMCPU
, 0, 2),
68 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[16].d
, ARMCPU
, 0, 2),
69 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[17].d
, ARMCPU
, 0, 2),
70 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[18].d
, ARMCPU
, 0, 2),
71 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[19].d
, ARMCPU
, 0, 2),
72 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[20].d
, ARMCPU
, 0, 2),
73 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[21].d
, ARMCPU
, 0, 2),
74 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[22].d
, ARMCPU
, 0, 2),
75 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[23].d
, ARMCPU
, 0, 2),
76 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[24].d
, ARMCPU
, 0, 2),
77 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[25].d
, ARMCPU
, 0, 2),
78 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[26].d
, ARMCPU
, 0, 2),
79 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[27].d
, ARMCPU
, 0, 2),
80 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[28].d
, ARMCPU
, 0, 2),
81 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[29].d
, ARMCPU
, 0, 2),
82 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[30].d
, ARMCPU
, 0, 2),
83 VMSTATE_UINT64_SUB_ARRAY(env
.vfp
.zregs
[31].d
, ARMCPU
, 0, 2),
85 /* The xregs array is a little awkward because element 1 (FPSCR)
86 * requires a specific accessor, so we have to split it up in
89 VMSTATE_UINT32(env
.vfp
.xregs
[0], ARMCPU
),
90 VMSTATE_UINT32_SUB_ARRAY(env
.vfp
.xregs
, ARMCPU
, 2, 14),
94 .size
= sizeof(uint32_t),
95 .info
= &vmstate_fpscr
,
103 static bool iwmmxt_needed(void *opaque
)
105 ARMCPU
*cpu
= opaque
;
106 CPUARMState
*env
= &cpu
->env
;
108 return arm_feature(env
, ARM_FEATURE_IWMMXT
);
111 static const VMStateDescription vmstate_iwmmxt
= {
112 .name
= "cpu/iwmmxt",
114 .minimum_version_id
= 1,
115 .needed
= iwmmxt_needed
,
116 .fields
= (VMStateField
[]) {
117 VMSTATE_UINT64_ARRAY(env
.iwmmxt
.regs
, ARMCPU
, 16),
118 VMSTATE_UINT32_ARRAY(env
.iwmmxt
.cregs
, ARMCPU
, 16),
119 VMSTATE_END_OF_LIST()
123 #ifdef TARGET_AARCH64
124 /* The expression ARM_MAX_VQ - 2 is 0 for pure AArch32 build,
125 * and ARMPredicateReg is actively empty. This triggers errors
126 * in the expansion of the VMSTATE macros.
129 static bool sve_needed(void *opaque
)
131 ARMCPU
*cpu
= opaque
;
133 return cpu_isar_feature(aa64_sve
, cpu
);
136 /* The first two words of each Zreg is stored in VFP state. */
137 static const VMStateDescription vmstate_zreg_hi_reg
= {
138 .name
= "cpu/sve/zreg_hi",
140 .minimum_version_id
= 1,
141 .fields
= (VMStateField
[]) {
142 VMSTATE_UINT64_SUB_ARRAY(d
, ARMVectorReg
, 2, ARM_MAX_VQ
- 2),
143 VMSTATE_END_OF_LIST()
147 static const VMStateDescription vmstate_preg_reg
= {
148 .name
= "cpu/sve/preg",
150 .minimum_version_id
= 1,
151 .fields
= (VMStateField
[]) {
152 VMSTATE_UINT64_ARRAY(p
, ARMPredicateReg
, 2 * ARM_MAX_VQ
/ 8),
153 VMSTATE_END_OF_LIST()
157 static const VMStateDescription vmstate_sve
= {
160 .minimum_version_id
= 1,
161 .needed
= sve_needed
,
162 .fields
= (VMStateField
[]) {
163 VMSTATE_STRUCT_ARRAY(env
.vfp
.zregs
, ARMCPU
, 32, 0,
164 vmstate_zreg_hi_reg
, ARMVectorReg
),
165 VMSTATE_STRUCT_ARRAY(env
.vfp
.pregs
, ARMCPU
, 17, 0,
166 vmstate_preg_reg
, ARMPredicateReg
),
167 VMSTATE_END_OF_LIST()
172 static bool serror_needed(void *opaque
)
174 ARMCPU
*cpu
= opaque
;
175 CPUARMState
*env
= &cpu
->env
;
177 return env
->serror
.pending
!= 0;
180 static const VMStateDescription vmstate_serror
= {
181 .name
= "cpu/serror",
183 .minimum_version_id
= 1,
184 .needed
= serror_needed
,
185 .fields
= (VMStateField
[]) {
186 VMSTATE_UINT8(env
.serror
.pending
, ARMCPU
),
187 VMSTATE_UINT8(env
.serror
.has_esr
, ARMCPU
),
188 VMSTATE_UINT64(env
.serror
.esr
, ARMCPU
),
189 VMSTATE_END_OF_LIST()
193 static bool irq_line_state_needed(void *opaque
)
198 static const VMStateDescription vmstate_irq_line_state
= {
199 .name
= "cpu/irq-line-state",
201 .minimum_version_id
= 1,
202 .needed
= irq_line_state_needed
,
203 .fields
= (VMStateField
[]) {
204 VMSTATE_UINT32(env
.irq_line_state
, ARMCPU
),
205 VMSTATE_END_OF_LIST()
209 static bool m_needed(void *opaque
)
211 ARMCPU
*cpu
= opaque
;
212 CPUARMState
*env
= &cpu
->env
;
214 return arm_feature(env
, ARM_FEATURE_M
);
217 static const VMStateDescription vmstate_m_faultmask_primask
= {
218 .name
= "cpu/m/faultmask-primask",
220 .minimum_version_id
= 1,
222 .fields
= (VMStateField
[]) {
223 VMSTATE_UINT32(env
.v7m
.faultmask
[M_REG_NS
], ARMCPU
),
224 VMSTATE_UINT32(env
.v7m
.primask
[M_REG_NS
], ARMCPU
),
225 VMSTATE_END_OF_LIST()
229 /* CSSELR is in a subsection because we didn't implement it previously.
230 * Migration from an old implementation will leave it at zero, which
231 * is OK since the only CPUs in the old implementation make the
233 * Since there was no version of QEMU which implemented the CSSELR for
234 * just non-secure, we transfer both banks here rather than putting
235 * the secure banked version in the m-security subsection.
237 static bool csselr_vmstate_validate(void *opaque
, int version_id
)
239 ARMCPU
*cpu
= opaque
;
241 return cpu
->env
.v7m
.csselr
[M_REG_NS
] <= R_V7M_CSSELR_INDEX_MASK
242 && cpu
->env
.v7m
.csselr
[M_REG_S
] <= R_V7M_CSSELR_INDEX_MASK
;
245 static bool m_csselr_needed(void *opaque
)
247 ARMCPU
*cpu
= opaque
;
249 return !arm_v7m_csselr_razwi(cpu
);
252 static const VMStateDescription vmstate_m_csselr
= {
253 .name
= "cpu/m/csselr",
255 .minimum_version_id
= 1,
256 .needed
= m_csselr_needed
,
257 .fields
= (VMStateField
[]) {
258 VMSTATE_UINT32_ARRAY(env
.v7m
.csselr
, ARMCPU
, M_REG_NUM_BANKS
),
259 VMSTATE_VALIDATE("CSSELR is valid", csselr_vmstate_validate
),
260 VMSTATE_END_OF_LIST()
264 static const VMStateDescription vmstate_m_scr
= {
267 .minimum_version_id
= 1,
269 .fields
= (VMStateField
[]) {
270 VMSTATE_UINT32(env
.v7m
.scr
[M_REG_NS
], ARMCPU
),
271 VMSTATE_END_OF_LIST()
275 static const VMStateDescription vmstate_m_other_sp
= {
276 .name
= "cpu/m/other-sp",
278 .minimum_version_id
= 1,
280 .fields
= (VMStateField
[]) {
281 VMSTATE_UINT32(env
.v7m
.other_sp
, ARMCPU
),
282 VMSTATE_END_OF_LIST()
286 static bool m_v8m_needed(void *opaque
)
288 ARMCPU
*cpu
= opaque
;
289 CPUARMState
*env
= &cpu
->env
;
291 return arm_feature(env
, ARM_FEATURE_M
) && arm_feature(env
, ARM_FEATURE_V8
);
294 static const VMStateDescription vmstate_m_v8m
= {
297 .minimum_version_id
= 1,
298 .needed
= m_v8m_needed
,
299 .fields
= (VMStateField
[]) {
300 VMSTATE_UINT32_ARRAY(env
.v7m
.msplim
, ARMCPU
, M_REG_NUM_BANKS
),
301 VMSTATE_UINT32_ARRAY(env
.v7m
.psplim
, ARMCPU
, M_REG_NUM_BANKS
),
302 VMSTATE_END_OF_LIST()
306 static const VMStateDescription vmstate_m_fp
= {
309 .minimum_version_id
= 1,
310 .needed
= vfp_needed
,
311 .fields
= (VMStateField
[]) {
312 VMSTATE_UINT32_ARRAY(env
.v7m
.fpcar
, ARMCPU
, M_REG_NUM_BANKS
),
313 VMSTATE_UINT32_ARRAY(env
.v7m
.fpccr
, ARMCPU
, M_REG_NUM_BANKS
),
314 VMSTATE_UINT32_ARRAY(env
.v7m
.fpdscr
, ARMCPU
, M_REG_NUM_BANKS
),
315 VMSTATE_UINT32_ARRAY(env
.v7m
.cpacr
, ARMCPU
, M_REG_NUM_BANKS
),
316 VMSTATE_UINT32(env
.v7m
.nsacr
, ARMCPU
),
317 VMSTATE_END_OF_LIST()
321 static bool mve_needed(void *opaque
)
323 ARMCPU
*cpu
= opaque
;
325 return cpu_isar_feature(aa32_mve
, cpu
);
328 static const VMStateDescription vmstate_m_mve
= {
331 .minimum_version_id
= 1,
332 .needed
= mve_needed
,
333 .fields
= (VMStateField
[]) {
334 VMSTATE_UINT32(env
.v7m
.vpr
, ARMCPU
),
335 VMSTATE_UINT32(env
.v7m
.ltpsize
, ARMCPU
),
336 VMSTATE_END_OF_LIST()
340 static const VMStateDescription vmstate_m
= {
343 .minimum_version_id
= 4,
345 .fields
= (VMStateField
[]) {
346 VMSTATE_UINT32(env
.v7m
.vecbase
[M_REG_NS
], ARMCPU
),
347 VMSTATE_UINT32(env
.v7m
.basepri
[M_REG_NS
], ARMCPU
),
348 VMSTATE_UINT32(env
.v7m
.control
[M_REG_NS
], ARMCPU
),
349 VMSTATE_UINT32(env
.v7m
.ccr
[M_REG_NS
], ARMCPU
),
350 VMSTATE_UINT32(env
.v7m
.cfsr
[M_REG_NS
], ARMCPU
),
351 VMSTATE_UINT32(env
.v7m
.hfsr
, ARMCPU
),
352 VMSTATE_UINT32(env
.v7m
.dfsr
, ARMCPU
),
353 VMSTATE_UINT32(env
.v7m
.mmfar
[M_REG_NS
], ARMCPU
),
354 VMSTATE_UINT32(env
.v7m
.bfar
, ARMCPU
),
355 VMSTATE_UINT32(env
.v7m
.mpu_ctrl
[M_REG_NS
], ARMCPU
),
356 VMSTATE_INT32(env
.v7m
.exception
, ARMCPU
),
357 VMSTATE_END_OF_LIST()
359 .subsections
= (const VMStateDescription
*[]) {
360 &vmstate_m_faultmask_primask
,
371 static bool thumb2ee_needed(void *opaque
)
373 ARMCPU
*cpu
= opaque
;
374 CPUARMState
*env
= &cpu
->env
;
376 return arm_feature(env
, ARM_FEATURE_THUMB2EE
);
379 static const VMStateDescription vmstate_thumb2ee
= {
380 .name
= "cpu/thumb2ee",
382 .minimum_version_id
= 1,
383 .needed
= thumb2ee_needed
,
384 .fields
= (VMStateField
[]) {
385 VMSTATE_UINT32(env
.teecr
, ARMCPU
),
386 VMSTATE_UINT32(env
.teehbr
, ARMCPU
),
387 VMSTATE_END_OF_LIST()
391 static bool pmsav7_needed(void *opaque
)
393 ARMCPU
*cpu
= opaque
;
394 CPUARMState
*env
= &cpu
->env
;
396 return arm_feature(env
, ARM_FEATURE_PMSA
) &&
397 arm_feature(env
, ARM_FEATURE_V7
) &&
398 !arm_feature(env
, ARM_FEATURE_V8
);
401 static bool pmsav7_rgnr_vmstate_validate(void *opaque
, int version_id
)
403 ARMCPU
*cpu
= opaque
;
405 return cpu
->env
.pmsav7
.rnr
[M_REG_NS
] < cpu
->pmsav7_dregion
;
408 static const VMStateDescription vmstate_pmsav7
= {
409 .name
= "cpu/pmsav7",
411 .minimum_version_id
= 1,
412 .needed
= pmsav7_needed
,
413 .fields
= (VMStateField
[]) {
414 VMSTATE_VARRAY_UINT32(env
.pmsav7
.drbar
, ARMCPU
, pmsav7_dregion
, 0,
415 vmstate_info_uint32
, uint32_t),
416 VMSTATE_VARRAY_UINT32(env
.pmsav7
.drsr
, ARMCPU
, pmsav7_dregion
, 0,
417 vmstate_info_uint32
, uint32_t),
418 VMSTATE_VARRAY_UINT32(env
.pmsav7
.dracr
, ARMCPU
, pmsav7_dregion
, 0,
419 vmstate_info_uint32
, uint32_t),
420 VMSTATE_VALIDATE("rgnr is valid", pmsav7_rgnr_vmstate_validate
),
421 VMSTATE_END_OF_LIST()
425 static bool pmsav7_rnr_needed(void *opaque
)
427 ARMCPU
*cpu
= opaque
;
428 CPUARMState
*env
= &cpu
->env
;
430 /* For R profile cores pmsav7.rnr is migrated via the cpreg
431 * "RGNR" definition in helper.h. For M profile we have to
432 * migrate it separately.
434 return arm_feature(env
, ARM_FEATURE_M
);
437 static const VMStateDescription vmstate_pmsav7_rnr
= {
438 .name
= "cpu/pmsav7-rnr",
440 .minimum_version_id
= 1,
441 .needed
= pmsav7_rnr_needed
,
442 .fields
= (VMStateField
[]) {
443 VMSTATE_UINT32(env
.pmsav7
.rnr
[M_REG_NS
], ARMCPU
),
444 VMSTATE_END_OF_LIST()
448 static bool pmsav8_needed(void *opaque
)
450 ARMCPU
*cpu
= opaque
;
451 CPUARMState
*env
= &cpu
->env
;
453 return arm_feature(env
, ARM_FEATURE_PMSA
) &&
454 arm_feature(env
, ARM_FEATURE_V8
);
457 static const VMStateDescription vmstate_pmsav8
= {
458 .name
= "cpu/pmsav8",
460 .minimum_version_id
= 1,
461 .needed
= pmsav8_needed
,
462 .fields
= (VMStateField
[]) {
463 VMSTATE_VARRAY_UINT32(env
.pmsav8
.rbar
[M_REG_NS
], ARMCPU
, pmsav7_dregion
,
464 0, vmstate_info_uint32
, uint32_t),
465 VMSTATE_VARRAY_UINT32(env
.pmsav8
.rlar
[M_REG_NS
], ARMCPU
, pmsav7_dregion
,
466 0, vmstate_info_uint32
, uint32_t),
467 VMSTATE_UINT32(env
.pmsav8
.mair0
[M_REG_NS
], ARMCPU
),
468 VMSTATE_UINT32(env
.pmsav8
.mair1
[M_REG_NS
], ARMCPU
),
469 VMSTATE_END_OF_LIST()
473 static bool s_rnr_vmstate_validate(void *opaque
, int version_id
)
475 ARMCPU
*cpu
= opaque
;
477 return cpu
->env
.pmsav7
.rnr
[M_REG_S
] < cpu
->pmsav7_dregion
;
480 static bool sau_rnr_vmstate_validate(void *opaque
, int version_id
)
482 ARMCPU
*cpu
= opaque
;
484 return cpu
->env
.sau
.rnr
< cpu
->sau_sregion
;
487 static bool m_security_needed(void *opaque
)
489 ARMCPU
*cpu
= opaque
;
490 CPUARMState
*env
= &cpu
->env
;
492 return arm_feature(env
, ARM_FEATURE_M_SECURITY
);
495 static const VMStateDescription vmstate_m_security
= {
496 .name
= "cpu/m-security",
498 .minimum_version_id
= 1,
499 .needed
= m_security_needed
,
500 .fields
= (VMStateField
[]) {
501 VMSTATE_UINT32(env
.v7m
.secure
, ARMCPU
),
502 VMSTATE_UINT32(env
.v7m
.other_ss_msp
, ARMCPU
),
503 VMSTATE_UINT32(env
.v7m
.other_ss_psp
, ARMCPU
),
504 VMSTATE_UINT32(env
.v7m
.basepri
[M_REG_S
], ARMCPU
),
505 VMSTATE_UINT32(env
.v7m
.primask
[M_REG_S
], ARMCPU
),
506 VMSTATE_UINT32(env
.v7m
.faultmask
[M_REG_S
], ARMCPU
),
507 VMSTATE_UINT32(env
.v7m
.control
[M_REG_S
], ARMCPU
),
508 VMSTATE_UINT32(env
.v7m
.vecbase
[M_REG_S
], ARMCPU
),
509 VMSTATE_UINT32(env
.pmsav8
.mair0
[M_REG_S
], ARMCPU
),
510 VMSTATE_UINT32(env
.pmsav8
.mair1
[M_REG_S
], ARMCPU
),
511 VMSTATE_VARRAY_UINT32(env
.pmsav8
.rbar
[M_REG_S
], ARMCPU
, pmsav7_dregion
,
512 0, vmstate_info_uint32
, uint32_t),
513 VMSTATE_VARRAY_UINT32(env
.pmsav8
.rlar
[M_REG_S
], ARMCPU
, pmsav7_dregion
,
514 0, vmstate_info_uint32
, uint32_t),
515 VMSTATE_UINT32(env
.pmsav7
.rnr
[M_REG_S
], ARMCPU
),
516 VMSTATE_VALIDATE("secure MPU_RNR is valid", s_rnr_vmstate_validate
),
517 VMSTATE_UINT32(env
.v7m
.mpu_ctrl
[M_REG_S
], ARMCPU
),
518 VMSTATE_UINT32(env
.v7m
.ccr
[M_REG_S
], ARMCPU
),
519 VMSTATE_UINT32(env
.v7m
.mmfar
[M_REG_S
], ARMCPU
),
520 VMSTATE_UINT32(env
.v7m
.cfsr
[M_REG_S
], ARMCPU
),
521 VMSTATE_UINT32(env
.v7m
.sfsr
, ARMCPU
),
522 VMSTATE_UINT32(env
.v7m
.sfar
, ARMCPU
),
523 VMSTATE_VARRAY_UINT32(env
.sau
.rbar
, ARMCPU
, sau_sregion
, 0,
524 vmstate_info_uint32
, uint32_t),
525 VMSTATE_VARRAY_UINT32(env
.sau
.rlar
, ARMCPU
, sau_sregion
, 0,
526 vmstate_info_uint32
, uint32_t),
527 VMSTATE_UINT32(env
.sau
.rnr
, ARMCPU
),
528 VMSTATE_VALIDATE("SAU_RNR is valid", sau_rnr_vmstate_validate
),
529 VMSTATE_UINT32(env
.sau
.ctrl
, ARMCPU
),
530 VMSTATE_UINT32(env
.v7m
.scr
[M_REG_S
], ARMCPU
),
531 /* AIRCR is not secure-only, but our implementation is R/O if the
532 * security extension is unimplemented, so we migrate it here.
534 VMSTATE_UINT32(env
.v7m
.aircr
, ARMCPU
),
535 VMSTATE_END_OF_LIST()
539 static int get_cpsr(QEMUFile
*f
, void *opaque
, size_t size
,
540 const VMStateField
*field
)
542 ARMCPU
*cpu
= opaque
;
543 CPUARMState
*env
= &cpu
->env
;
544 uint32_t val
= qemu_get_be32(f
);
546 if (arm_feature(env
, ARM_FEATURE_M
)) {
547 if (val
& XPSR_EXCP
) {
548 /* This is a CPSR format value from an older QEMU. (We can tell
549 * because values transferred in XPSR format always have zero
550 * for the EXCP field, and CPSR format will always have bit 4
551 * set in CPSR_M.) Rearrange it into XPSR format. The significant
552 * differences are that the T bit is not in the same place, the
553 * primask/faultmask info may be in the CPSR I and F bits, and
554 * we do not want the mode bits.
555 * We know that this cleanup happened before v8M, so there
556 * is no complication with banked primask/faultmask.
558 uint32_t newval
= val
;
560 assert(!arm_feature(env
, ARM_FEATURE_M_SECURITY
));
562 newval
&= (CPSR_NZCV
| CPSR_Q
| CPSR_IT
| CPSR_GE
);
566 /* If the I or F bits are set then this is a migration from
567 * an old QEMU which still stored the M profile FAULTMASK
568 * and PRIMASK in env->daif. For a new QEMU, the data is
569 * transferred using the vmstate_m_faultmask_primask subsection.
572 env
->v7m
.faultmask
[M_REG_NS
] = 1;
575 env
->v7m
.primask
[M_REG_NS
] = 1;
579 /* Ignore the low bits, they are handled by vmstate_m. */
580 xpsr_write(env
, val
, ~XPSR_EXCP
);
584 env
->aarch64
= ((val
& PSTATE_nRW
) == 0);
587 pstate_write(env
, val
);
591 cpsr_write(env
, val
, 0xffffffff, CPSRWriteRaw
);
595 static int put_cpsr(QEMUFile
*f
, void *opaque
, size_t size
,
596 const VMStateField
*field
, JSONWriter
*vmdesc
)
598 ARMCPU
*cpu
= opaque
;
599 CPUARMState
*env
= &cpu
->env
;
602 if (arm_feature(env
, ARM_FEATURE_M
)) {
603 /* The low 9 bits are v7m.exception, which is handled by vmstate_m. */
604 val
= xpsr_read(env
) & ~XPSR_EXCP
;
605 } else if (is_a64(env
)) {
606 val
= pstate_read(env
);
608 val
= cpsr_read(env
);
611 qemu_put_be32(f
, val
);
615 static const VMStateInfo vmstate_cpsr
= {
621 static int get_power(QEMUFile
*f
, void *opaque
, size_t size
,
622 const VMStateField
*field
)
624 ARMCPU
*cpu
= opaque
;
625 bool powered_off
= qemu_get_byte(f
);
626 cpu
->power_state
= powered_off
? PSCI_OFF
: PSCI_ON
;
630 static int put_power(QEMUFile
*f
, void *opaque
, size_t size
,
631 const VMStateField
*field
, JSONWriter
*vmdesc
)
633 ARMCPU
*cpu
= opaque
;
635 /* Migration should never happen while we transition power states */
637 if (cpu
->power_state
== PSCI_ON
||
638 cpu
->power_state
== PSCI_OFF
) {
639 bool powered_off
= (cpu
->power_state
== PSCI_OFF
) ? true : false;
640 qemu_put_byte(f
, powered_off
);
647 static const VMStateInfo vmstate_powered_off
= {
648 .name
= "powered_off",
653 static int cpu_pre_save(void *opaque
)
655 ARMCPU
*cpu
= opaque
;
657 if (!kvm_enabled()) {
658 pmu_op_start(&cpu
->env
);
662 if (!write_kvmstate_to_list(cpu
)) {
663 /* This should never fail */
668 * kvm_arm_cpu_pre_save() must be called after
669 * write_kvmstate_to_list()
671 kvm_arm_cpu_pre_save(cpu
);
673 if (!write_cpustate_to_list(cpu
, false)) {
674 /* This should never fail. */
679 cpu
->cpreg_vmstate_array_len
= cpu
->cpreg_array_len
;
680 memcpy(cpu
->cpreg_vmstate_indexes
, cpu
->cpreg_indexes
,
681 cpu
->cpreg_array_len
* sizeof(uint64_t));
682 memcpy(cpu
->cpreg_vmstate_values
, cpu
->cpreg_values
,
683 cpu
->cpreg_array_len
* sizeof(uint64_t));
688 static int cpu_post_save(void *opaque
)
690 ARMCPU
*cpu
= opaque
;
692 if (!kvm_enabled()) {
693 pmu_op_finish(&cpu
->env
);
699 static int cpu_pre_load(void *opaque
)
701 ARMCPU
*cpu
= opaque
;
702 CPUARMState
*env
= &cpu
->env
;
705 * Pre-initialize irq_line_state to a value that's never valid as
706 * real data, so cpu_post_load() can tell whether we've seen the
707 * irq-line-state subsection in the incoming migration state.
709 env
->irq_line_state
= UINT32_MAX
;
711 if (!kvm_enabled()) {
712 pmu_op_start(&cpu
->env
);
718 static int cpu_post_load(void *opaque
, int version_id
)
720 ARMCPU
*cpu
= opaque
;
721 CPUARMState
*env
= &cpu
->env
;
725 * Handle migration compatibility from old QEMU which didn't
726 * send the irq-line-state subsection. A QEMU without it did not
727 * implement the HCR_EL2.{VI,VF} bits as generating interrupts,
728 * so for TCG the line state matches the bits set in cs->interrupt_request.
729 * For KVM the line state is not stored in cs->interrupt_request
730 * and so this will leave irq_line_state as 0, but this is OK because
731 * we only need to care about it for TCG.
733 if (env
->irq_line_state
== UINT32_MAX
) {
734 CPUState
*cs
= CPU(cpu
);
736 env
->irq_line_state
= cs
->interrupt_request
&
737 (CPU_INTERRUPT_HARD
| CPU_INTERRUPT_FIQ
|
738 CPU_INTERRUPT_VIRQ
| CPU_INTERRUPT_VFIQ
);
741 /* Update the values list from the incoming migration data.
742 * Anything in the incoming data which we don't know about is
743 * a migration failure; anything we know about but the incoming
744 * data doesn't specify retains its current (reset) value.
745 * The indexes list remains untouched -- we only inspect the
746 * incoming migration index list so we can match the values array
747 * entries with the right slots in our own values array.
750 for (i
= 0, v
= 0; i
< cpu
->cpreg_array_len
751 && v
< cpu
->cpreg_vmstate_array_len
; i
++) {
752 if (cpu
->cpreg_vmstate_indexes
[v
] > cpu
->cpreg_indexes
[i
]) {
753 /* register in our list but not incoming : skip it */
756 if (cpu
->cpreg_vmstate_indexes
[v
] < cpu
->cpreg_indexes
[i
]) {
757 /* register in their list but not ours: fail migration */
760 /* matching register, copy the value over */
761 cpu
->cpreg_values
[i
] = cpu
->cpreg_vmstate_values
[v
];
766 if (!write_list_to_kvmstate(cpu
, KVM_PUT_FULL_STATE
)) {
769 /* Note that it's OK for the TCG side not to know about
770 * every register in the list; KVM is authoritative if
773 write_list_to_cpustate(cpu
);
774 kvm_arm_cpu_post_load(cpu
);
776 if (!write_list_to_cpustate(cpu
)) {
781 hw_breakpoint_update_all(cpu
);
782 hw_watchpoint_update_all(cpu
);
785 * TCG gen_update_fp_context() relies on the invariant that
786 * FPDSCR.LTPSIZE is constant 4 for M-profile with the LOB extension;
787 * forbid bogus incoming data with some other value.
789 if (arm_feature(env
, ARM_FEATURE_M
) && cpu_isar_feature(aa32_lob
, cpu
)) {
790 if (extract32(env
->v7m
.fpdscr
[M_REG_NS
],
791 FPCR_LTPSIZE_SHIFT
, FPCR_LTPSIZE_LENGTH
) != 4 ||
792 extract32(env
->v7m
.fpdscr
[M_REG_S
],
793 FPCR_LTPSIZE_SHIFT
, FPCR_LTPSIZE_LENGTH
) != 4) {
797 if (!kvm_enabled()) {
798 pmu_op_finish(&cpu
->env
);
800 arm_rebuild_hflags(&cpu
->env
);
805 const VMStateDescription vmstate_arm_cpu
= {
808 .minimum_version_id
= 22,
809 .pre_save
= cpu_pre_save
,
810 .post_save
= cpu_post_save
,
811 .pre_load
= cpu_pre_load
,
812 .post_load
= cpu_post_load
,
813 .fields
= (VMStateField
[]) {
814 VMSTATE_UINT32_ARRAY(env
.regs
, ARMCPU
, 16),
815 VMSTATE_UINT64_ARRAY(env
.xregs
, ARMCPU
, 32),
816 VMSTATE_UINT64(env
.pc
, ARMCPU
),
820 .size
= sizeof(uint32_t),
821 .info
= &vmstate_cpsr
,
825 VMSTATE_UINT32(env
.spsr
, ARMCPU
),
826 VMSTATE_UINT64_ARRAY(env
.banked_spsr
, ARMCPU
, 8),
827 VMSTATE_UINT32_ARRAY(env
.banked_r13
, ARMCPU
, 8),
828 VMSTATE_UINT32_ARRAY(env
.banked_r14
, ARMCPU
, 8),
829 VMSTATE_UINT32_ARRAY(env
.usr_regs
, ARMCPU
, 5),
830 VMSTATE_UINT32_ARRAY(env
.fiq_regs
, ARMCPU
, 5),
831 VMSTATE_UINT64_ARRAY(env
.elr_el
, ARMCPU
, 4),
832 VMSTATE_UINT64_ARRAY(env
.sp_el
, ARMCPU
, 4),
833 /* The length-check must come before the arrays to avoid
834 * incoming data possibly overflowing the array.
836 VMSTATE_INT32_POSITIVE_LE(cpreg_vmstate_array_len
, ARMCPU
),
837 VMSTATE_VARRAY_INT32(cpreg_vmstate_indexes
, ARMCPU
,
838 cpreg_vmstate_array_len
,
839 0, vmstate_info_uint64
, uint64_t),
840 VMSTATE_VARRAY_INT32(cpreg_vmstate_values
, ARMCPU
,
841 cpreg_vmstate_array_len
,
842 0, vmstate_info_uint64
, uint64_t),
843 VMSTATE_UINT64(env
.exclusive_addr
, ARMCPU
),
844 VMSTATE_UINT64(env
.exclusive_val
, ARMCPU
),
845 VMSTATE_UINT64(env
.exclusive_high
, ARMCPU
),
846 VMSTATE_UNUSED(sizeof(uint64_t)),
847 VMSTATE_UINT32(env
.exception
.syndrome
, ARMCPU
),
848 VMSTATE_UINT32(env
.exception
.fsr
, ARMCPU
),
849 VMSTATE_UINT64(env
.exception
.vaddress
, ARMCPU
),
850 VMSTATE_TIMER_PTR(gt_timer
[GTIMER_PHYS
], ARMCPU
),
851 VMSTATE_TIMER_PTR(gt_timer
[GTIMER_VIRT
], ARMCPU
),
853 .name
= "power_state",
855 .size
= sizeof(bool),
856 .info
= &vmstate_powered_off
,
860 VMSTATE_END_OF_LIST()
862 .subsections
= (const VMStateDescription
*[]) {
867 /* pmsav7_rnr must come before pmsav7 so that we have the
868 * region number before we test it in the VMSTATE_VALIDATE
875 #ifdef TARGET_AARCH64
879 &vmstate_irq_line_state
,