2 * RISC-V Emulation Helpers for QEMU.
4 * Copyright (c) 2016-2017 Sagar Karandikar, sagark@eecs.berkeley.edu
5 * Copyright (c) 2017-2018 SiFive, Inc.
6 * Copyright (c) 2022 VRULL GmbH
8 * This program is free software; you can redistribute it and/or modify it
9 * under the terms and conditions of the GNU General Public License,
10 * version 2 or later, as published by the Free Software Foundation.
12 * This program is distributed in the hope it will be useful, but WITHOUT
13 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
14 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
17 * You should have received a copy of the GNU General Public License along with
18 * this program. If not, see <http://www.gnu.org/licenses/>.
21 #include "qemu/osdep.h"
23 #include "internals.h"
24 #include "qemu/main-loop.h"
25 #include "exec/exec-all.h"
26 #include "exec/cpu_ldst.h"
27 #include "exec/helper-proto.h"
29 /* Exceptions processing helpers */
30 G_NORETURN
void riscv_raise_exception(CPURISCVState
*env
,
31 uint32_t exception
, uintptr_t pc
)
33 CPUState
*cs
= env_cpu(env
);
34 cs
->exception_index
= exception
;
35 cpu_loop_exit_restore(cs
, pc
);
38 void helper_raise_exception(CPURISCVState
*env
, uint32_t exception
)
40 riscv_raise_exception(env
, exception
, 0);
43 target_ulong
helper_csrr(CPURISCVState
*env
, int csr
)
46 * The seed CSR must be accessed with a read-write instruction. A
47 * read-only instruction such as CSRRS/CSRRC with rs1=x0 or CSRRSI/
48 * CSRRCI with uimm=0 will raise an illegal instruction exception.
50 if (csr
== CSR_SEED
) {
51 riscv_raise_exception(env
, RISCV_EXCP_ILLEGAL_INST
, GETPC());
55 RISCVException ret
= riscv_csrrw(env
, csr
, &val
, 0, 0);
57 if (ret
!= RISCV_EXCP_NONE
) {
58 riscv_raise_exception(env
, ret
, GETPC());
63 void helper_csrw(CPURISCVState
*env
, int csr
, target_ulong src
)
65 target_ulong mask
= env
->xl
== MXL_RV32
? UINT32_MAX
: (target_ulong
)-1;
66 RISCVException ret
= riscv_csrrw(env
, csr
, NULL
, src
, mask
);
68 if (ret
!= RISCV_EXCP_NONE
) {
69 riscv_raise_exception(env
, ret
, GETPC());
73 target_ulong
helper_csrrw(CPURISCVState
*env
, int csr
,
74 target_ulong src
, target_ulong write_mask
)
77 RISCVException ret
= riscv_csrrw(env
, csr
, &val
, src
, write_mask
);
79 if (ret
!= RISCV_EXCP_NONE
) {
80 riscv_raise_exception(env
, ret
, GETPC());
85 target_ulong
helper_csrr_i128(CPURISCVState
*env
, int csr
)
87 Int128 rv
= int128_zero();
88 RISCVException ret
= riscv_csrrw_i128(env
, csr
, &rv
,
92 if (ret
!= RISCV_EXCP_NONE
) {
93 riscv_raise_exception(env
, ret
, GETPC());
96 env
->retxh
= int128_gethi(rv
);
97 return int128_getlo(rv
);
100 void helper_csrw_i128(CPURISCVState
*env
, int csr
,
101 target_ulong srcl
, target_ulong srch
)
103 RISCVException ret
= riscv_csrrw_i128(env
, csr
, NULL
,
104 int128_make128(srcl
, srch
),
107 if (ret
!= RISCV_EXCP_NONE
) {
108 riscv_raise_exception(env
, ret
, GETPC());
112 target_ulong
helper_csrrw_i128(CPURISCVState
*env
, int csr
,
113 target_ulong srcl
, target_ulong srch
,
114 target_ulong maskl
, target_ulong maskh
)
116 Int128 rv
= int128_zero();
117 RISCVException ret
= riscv_csrrw_i128(env
, csr
, &rv
,
118 int128_make128(srcl
, srch
),
119 int128_make128(maskl
, maskh
));
121 if (ret
!= RISCV_EXCP_NONE
) {
122 riscv_raise_exception(env
, ret
, GETPC());
125 env
->retxh
= int128_gethi(rv
);
126 return int128_getlo(rv
);
133 * Raise virtual exceptions and illegal instruction exceptions for
134 * Zicbo[mz] instructions based on the settings of [mhs]envcfg as
135 * specified in section 2.5.1 of the CMO specification.
137 static void check_zicbo_envcfg(CPURISCVState
*env
, target_ulong envbits
,
140 #ifndef CONFIG_USER_ONLY
141 if ((env
->priv
< PRV_M
) && !get_field(env
->menvcfg
, envbits
)) {
142 riscv_raise_exception(env
, RISCV_EXCP_ILLEGAL_INST
, ra
);
145 if (env
->virt_enabled
&&
146 (((env
->priv
<= PRV_S
) && !get_field(env
->henvcfg
, envbits
)) ||
147 ((env
->priv
< PRV_S
) && !get_field(env
->senvcfg
, envbits
)))) {
148 riscv_raise_exception(env
, RISCV_EXCP_VIRT_INSTRUCTION_FAULT
, ra
);
151 if ((env
->priv
< PRV_S
) && !get_field(env
->senvcfg
, envbits
)) {
152 riscv_raise_exception(env
, RISCV_EXCP_ILLEGAL_INST
, ra
);
157 void helper_cbo_zero(CPURISCVState
*env
, target_ulong address
)
159 RISCVCPU
*cpu
= env_archcpu(env
);
160 uint16_t cbozlen
= cpu
->cfg
.cboz_blocksize
;
161 int mmu_idx
= cpu_mmu_index(env
, false);
162 uintptr_t ra
= GETPC();
165 check_zicbo_envcfg(env
, MENVCFG_CBZE
, ra
);
167 /* Mask off low-bits to align-down to the cache-block. */
168 address
&= ~(cbozlen
- 1);
171 * cbo.zero requires MMU_DATA_STORE access. Do a probe_write()
172 * to raise any exceptions, including PMP.
174 mem
= probe_write(env
, address
, cbozlen
, mmu_idx
, ra
);
177 memset(mem
, 0, cbozlen
);
180 * This means that we're dealing with an I/O page. Section 4.2
181 * of cmobase v1.0.1 says:
183 * "Cache-block zero instructions store zeros independently
184 * of whether data from the underlying memory locations are
187 * Write zeros in address + cbozlen regardless of not being
190 for (int i
= 0; i
< cbozlen
; i
++) {
191 cpu_stb_mmuidx_ra(env
, address
+ i
, 0, mmu_idx
, ra
);
197 * check_zicbom_access
199 * Check access permissions (LOAD, STORE or FETCH as specified in
200 * section 2.5.2 of the CMO specification) for Zicbom, raising
201 * either store page-fault (non-virtualized) or store guest-page
202 * fault (virtualized).
204 static void check_zicbom_access(CPURISCVState
*env
,
205 target_ulong address
,
208 RISCVCPU
*cpu
= env_archcpu(env
);
209 int mmu_idx
= cpu_mmu_index(env
, false);
210 uint16_t cbomlen
= cpu
->cfg
.cbom_blocksize
;
214 /* Mask off low-bits to align-down to the cache-block. */
215 address
&= ~(cbomlen
- 1);
218 * Section 2.5.2 of cmobase v1.0.1:
220 * "A cache-block management instruction is permitted to
221 * access the specified cache block whenever a load instruction
222 * or store instruction is permitted to access the corresponding
223 * physical addresses. If neither a load instruction nor store
224 * instruction is permitted to access the physical addresses,
225 * but an instruction fetch is permitted to access the physical
226 * addresses, whether a cache-block management instruction is
227 * permitted to access the cache block is UNSPECIFIED."
229 ret
= probe_access_flags(env
, address
, cbomlen
, MMU_DATA_LOAD
,
230 mmu_idx
, true, &phost
, ra
);
231 if (ret
!= TLB_INVALID_MASK
) {
232 /* Success: readable */
237 * Since not readable, must be writable. On failure, store
238 * fault/store guest amo fault will be raised by
239 * riscv_cpu_tlb_fill(). PMP exceptions will be caught
242 probe_write(env
, address
, cbomlen
, mmu_idx
, ra
);
245 void helper_cbo_clean_flush(CPURISCVState
*env
, target_ulong address
)
247 uintptr_t ra
= GETPC();
248 check_zicbo_envcfg(env
, MENVCFG_CBCFE
, ra
);
249 check_zicbom_access(env
, address
, ra
);
251 /* We don't emulate the cache-hierarchy, so we're done. */
254 void helper_cbo_inval(CPURISCVState
*env
, target_ulong address
)
256 uintptr_t ra
= GETPC();
257 check_zicbo_envcfg(env
, MENVCFG_CBIE
, ra
);
258 check_zicbom_access(env
, address
, ra
);
260 /* We don't emulate the cache-hierarchy, so we're done. */
263 #ifndef CONFIG_USER_ONLY
265 target_ulong
helper_sret(CPURISCVState
*env
)
268 target_ulong prev_priv
, prev_virt
;
270 if (!(env
->priv
>= PRV_S
)) {
271 riscv_raise_exception(env
, RISCV_EXCP_ILLEGAL_INST
, GETPC());
274 target_ulong retpc
= env
->sepc
;
275 if (!riscv_has_ext(env
, RVC
) && (retpc
& 0x3)) {
276 riscv_raise_exception(env
, RISCV_EXCP_INST_ADDR_MIS
, GETPC());
279 if (get_field(env
->mstatus
, MSTATUS_TSR
) && !(env
->priv
>= PRV_M
)) {
280 riscv_raise_exception(env
, RISCV_EXCP_ILLEGAL_INST
, GETPC());
283 if (env
->virt_enabled
&& get_field(env
->hstatus
, HSTATUS_VTSR
)) {
284 riscv_raise_exception(env
, RISCV_EXCP_VIRT_INSTRUCTION_FAULT
, GETPC());
287 mstatus
= env
->mstatus
;
288 prev_priv
= get_field(mstatus
, MSTATUS_SPP
);
289 mstatus
= set_field(mstatus
, MSTATUS_SIE
,
290 get_field(mstatus
, MSTATUS_SPIE
));
291 mstatus
= set_field(mstatus
, MSTATUS_SPIE
, 1);
292 mstatus
= set_field(mstatus
, MSTATUS_SPP
, PRV_U
);
293 if (env
->priv_ver
>= PRIV_VERSION_1_12_0
) {
294 mstatus
= set_field(mstatus
, MSTATUS_MPRV
, 0);
296 env
->mstatus
= mstatus
;
298 if (riscv_has_ext(env
, RVH
) && !env
->virt_enabled
) {
299 /* We support Hypervisor extensions and virtulisation is disabled */
300 target_ulong hstatus
= env
->hstatus
;
302 prev_virt
= get_field(hstatus
, HSTATUS_SPV
);
304 hstatus
= set_field(hstatus
, HSTATUS_SPV
, 0);
306 env
->hstatus
= hstatus
;
309 riscv_cpu_swap_hypervisor_regs(env
);
312 riscv_cpu_set_virt_enabled(env
, prev_virt
);
315 riscv_cpu_set_mode(env
, prev_priv
);
320 target_ulong
helper_mret(CPURISCVState
*env
)
322 if (!(env
->priv
>= PRV_M
)) {
323 riscv_raise_exception(env
, RISCV_EXCP_ILLEGAL_INST
, GETPC());
326 target_ulong retpc
= env
->mepc
;
327 if (!riscv_has_ext(env
, RVC
) && (retpc
& 0x3)) {
328 riscv_raise_exception(env
, RISCV_EXCP_INST_ADDR_MIS
, GETPC());
331 uint64_t mstatus
= env
->mstatus
;
332 target_ulong prev_priv
= get_field(mstatus
, MSTATUS_MPP
);
334 if (riscv_cpu_cfg(env
)->pmp
&&
335 !pmp_get_num_rules(env
) && (prev_priv
!= PRV_M
)) {
336 riscv_raise_exception(env
, RISCV_EXCP_INST_ACCESS_FAULT
, GETPC());
339 target_ulong prev_virt
= get_field(env
->mstatus
, MSTATUS_MPV
) &&
340 (prev_priv
!= PRV_M
);
341 mstatus
= set_field(mstatus
, MSTATUS_MIE
,
342 get_field(mstatus
, MSTATUS_MPIE
));
343 mstatus
= set_field(mstatus
, MSTATUS_MPIE
, 1);
344 mstatus
= set_field(mstatus
, MSTATUS_MPP
,
345 riscv_has_ext(env
, RVU
) ? PRV_U
: PRV_M
);
346 mstatus
= set_field(mstatus
, MSTATUS_MPV
, 0);
347 if ((env
->priv_ver
>= PRIV_VERSION_1_12_0
) && (prev_priv
!= PRV_M
)) {
348 mstatus
= set_field(mstatus
, MSTATUS_MPRV
, 0);
350 env
->mstatus
= mstatus
;
351 riscv_cpu_set_mode(env
, prev_priv
);
353 if (riscv_has_ext(env
, RVH
)) {
355 riscv_cpu_swap_hypervisor_regs(env
);
358 riscv_cpu_set_virt_enabled(env
, prev_virt
);
364 void helper_wfi(CPURISCVState
*env
)
366 CPUState
*cs
= env_cpu(env
);
367 bool rvs
= riscv_has_ext(env
, RVS
);
368 bool prv_u
= env
->priv
== PRV_U
;
369 bool prv_s
= env
->priv
== PRV_S
;
371 if (((prv_s
|| (!rvs
&& prv_u
)) && get_field(env
->mstatus
, MSTATUS_TW
)) ||
372 (rvs
&& prv_u
&& !env
->virt_enabled
)) {
373 riscv_raise_exception(env
, RISCV_EXCP_ILLEGAL_INST
, GETPC());
374 } else if (env
->virt_enabled
&&
375 (prv_u
|| (prv_s
&& get_field(env
->hstatus
, HSTATUS_VTW
)))) {
376 riscv_raise_exception(env
, RISCV_EXCP_VIRT_INSTRUCTION_FAULT
, GETPC());
379 cs
->exception_index
= EXCP_HLT
;
384 void helper_tlb_flush(CPURISCVState
*env
)
386 CPUState
*cs
= env_cpu(env
);
387 if (!env
->virt_enabled
&&
388 (env
->priv
== PRV_U
||
389 (env
->priv
== PRV_S
&& get_field(env
->mstatus
, MSTATUS_TVM
)))) {
390 riscv_raise_exception(env
, RISCV_EXCP_ILLEGAL_INST
, GETPC());
391 } else if (env
->virt_enabled
&&
392 (env
->priv
== PRV_U
|| get_field(env
->hstatus
, HSTATUS_VTVM
))) {
393 riscv_raise_exception(env
, RISCV_EXCP_VIRT_INSTRUCTION_FAULT
, GETPC());
399 void helper_tlb_flush_all(CPURISCVState
*env
)
401 CPUState
*cs
= env_cpu(env
);
402 tlb_flush_all_cpus_synced(cs
);
405 void helper_hyp_tlb_flush(CPURISCVState
*env
)
407 CPUState
*cs
= env_cpu(env
);
409 if (env
->virt_enabled
) {
410 riscv_raise_exception(env
, RISCV_EXCP_VIRT_INSTRUCTION_FAULT
, GETPC());
413 if (env
->priv
== PRV_M
||
414 (env
->priv
== PRV_S
&& !env
->virt_enabled
)) {
419 riscv_raise_exception(env
, RISCV_EXCP_ILLEGAL_INST
, GETPC());
422 void helper_hyp_gvma_tlb_flush(CPURISCVState
*env
)
424 if (env
->priv
== PRV_S
&& !env
->virt_enabled
&&
425 get_field(env
->mstatus
, MSTATUS_TVM
)) {
426 riscv_raise_exception(env
, RISCV_EXCP_ILLEGAL_INST
, GETPC());
429 helper_hyp_tlb_flush(env
);
432 static int check_access_hlsv(CPURISCVState
*env
, bool x
, uintptr_t ra
)
434 if (env
->priv
== PRV_M
) {
436 } else if (env
->virt_enabled
) {
437 riscv_raise_exception(env
, RISCV_EXCP_VIRT_INSTRUCTION_FAULT
, ra
);
438 } else if (env
->priv
== PRV_U
&& !get_field(env
->hstatus
, HSTATUS_HU
)) {
439 riscv_raise_exception(env
, RISCV_EXCP_ILLEGAL_INST
, ra
);
442 int mode
= get_field(env
->hstatus
, HSTATUS_SPVP
);
443 if (!x
&& mode
== PRV_S
&& get_field(env
->vsstatus
, MSTATUS_SUM
)) {
446 return mode
| MMU_2STAGE_BIT
;
449 target_ulong
helper_hyp_hlv_bu(CPURISCVState
*env
, target_ulong addr
)
451 uintptr_t ra
= GETPC();
452 int mmu_idx
= check_access_hlsv(env
, false, ra
);
453 MemOpIdx oi
= make_memop_idx(MO_UB
, mmu_idx
);
455 return cpu_ldb_mmu(env
, addr
, oi
, ra
);
458 target_ulong
helper_hyp_hlv_hu(CPURISCVState
*env
, target_ulong addr
)
460 uintptr_t ra
= GETPC();
461 int mmu_idx
= check_access_hlsv(env
, false, ra
);
462 MemOpIdx oi
= make_memop_idx(MO_TEUW
, mmu_idx
);
464 return cpu_ldw_mmu(env
, addr
, oi
, ra
);
467 target_ulong
helper_hyp_hlv_wu(CPURISCVState
*env
, target_ulong addr
)
469 uintptr_t ra
= GETPC();
470 int mmu_idx
= check_access_hlsv(env
, false, ra
);
471 MemOpIdx oi
= make_memop_idx(MO_TEUL
, mmu_idx
);
473 return cpu_ldl_mmu(env
, addr
, oi
, ra
);
476 target_ulong
helper_hyp_hlv_d(CPURISCVState
*env
, target_ulong addr
)
478 uintptr_t ra
= GETPC();
479 int mmu_idx
= check_access_hlsv(env
, false, ra
);
480 MemOpIdx oi
= make_memop_idx(MO_TEUQ
, mmu_idx
);
482 return cpu_ldq_mmu(env
, addr
, oi
, ra
);
485 void helper_hyp_hsv_b(CPURISCVState
*env
, target_ulong addr
, target_ulong val
)
487 uintptr_t ra
= GETPC();
488 int mmu_idx
= check_access_hlsv(env
, false, ra
);
489 MemOpIdx oi
= make_memop_idx(MO_UB
, mmu_idx
);
491 cpu_stb_mmu(env
, addr
, val
, oi
, ra
);
494 void helper_hyp_hsv_h(CPURISCVState
*env
, target_ulong addr
, target_ulong val
)
496 uintptr_t ra
= GETPC();
497 int mmu_idx
= check_access_hlsv(env
, false, ra
);
498 MemOpIdx oi
= make_memop_idx(MO_TEUW
, mmu_idx
);
500 cpu_stw_mmu(env
, addr
, val
, oi
, ra
);
503 void helper_hyp_hsv_w(CPURISCVState
*env
, target_ulong addr
, target_ulong val
)
505 uintptr_t ra
= GETPC();
506 int mmu_idx
= check_access_hlsv(env
, false, ra
);
507 MemOpIdx oi
= make_memop_idx(MO_TEUL
, mmu_idx
);
509 cpu_stl_mmu(env
, addr
, val
, oi
, ra
);
512 void helper_hyp_hsv_d(CPURISCVState
*env
, target_ulong addr
, target_ulong val
)
514 uintptr_t ra
= GETPC();
515 int mmu_idx
= check_access_hlsv(env
, false, ra
);
516 MemOpIdx oi
= make_memop_idx(MO_TEUQ
, mmu_idx
);
518 cpu_stq_mmu(env
, addr
, val
, oi
, ra
);
522 * TODO: These implementations are not quite correct. They perform the
523 * access using execute permission just fine, but the final PMP check
524 * is supposed to have read permission as well. Without replicating
525 * a fair fraction of cputlb.c, fixing this requires adding new mmu_idx
526 * which would imply that exact check in tlb_fill.
528 target_ulong
helper_hyp_hlvx_hu(CPURISCVState
*env
, target_ulong addr
)
530 uintptr_t ra
= GETPC();
531 int mmu_idx
= check_access_hlsv(env
, true, ra
);
532 MemOpIdx oi
= make_memop_idx(MO_TEUW
, mmu_idx
);
534 return cpu_ldw_code_mmu(env
, addr
, oi
, GETPC());
537 target_ulong
helper_hyp_hlvx_wu(CPURISCVState
*env
, target_ulong addr
)
539 uintptr_t ra
= GETPC();
540 int mmu_idx
= check_access_hlsv(env
, true, ra
);
541 MemOpIdx oi
= make_memop_idx(MO_TEUL
, mmu_idx
);
543 return cpu_ldl_code_mmu(env
, addr
, oi
, ra
);
546 #endif /* !CONFIG_USER_ONLY */