target/translate: Include missing 'exec/cpu_ldst.h' header
[qemu/ar7.git] / target / riscv / op_helper.c
blob7e2f1908ee39febfcda9ee2b621cdb415f484e0d
1 /*
2 * RISC-V Emulation Helpers for QEMU.
4 * Copyright (c) 2016-2017 Sagar Karandikar, sagark@eecs.berkeley.edu
5 * Copyright (c) 2017-2018 SiFive, Inc.
6 * Copyright (c) 2022 VRULL GmbH
8 * This program is free software; you can redistribute it and/or modify it
9 * under the terms and conditions of the GNU General Public License,
10 * version 2 or later, as published by the Free Software Foundation.
12 * This program is distributed in the hope it will be useful, but WITHOUT
13 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
14 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
15 * more details.
17 * You should have received a copy of the GNU General Public License along with
18 * this program. If not, see <http://www.gnu.org/licenses/>.
21 #include "qemu/osdep.h"
22 #include "cpu.h"
23 #include "internals.h"
24 #include "qemu/main-loop.h"
25 #include "exec/exec-all.h"
26 #include "exec/cpu_ldst.h"
27 #include "exec/helper-proto.h"
29 /* Exceptions processing helpers */
30 G_NORETURN void riscv_raise_exception(CPURISCVState *env,
31 uint32_t exception, uintptr_t pc)
33 CPUState *cs = env_cpu(env);
34 cs->exception_index = exception;
35 cpu_loop_exit_restore(cs, pc);
38 void helper_raise_exception(CPURISCVState *env, uint32_t exception)
40 riscv_raise_exception(env, exception, 0);
43 target_ulong helper_csrr(CPURISCVState *env, int csr)
46 * The seed CSR must be accessed with a read-write instruction. A
47 * read-only instruction such as CSRRS/CSRRC with rs1=x0 or CSRRSI/
48 * CSRRCI with uimm=0 will raise an illegal instruction exception.
50 if (csr == CSR_SEED) {
51 riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC());
54 target_ulong val = 0;
55 RISCVException ret = riscv_csrrw(env, csr, &val, 0, 0);
57 if (ret != RISCV_EXCP_NONE) {
58 riscv_raise_exception(env, ret, GETPC());
60 return val;
63 void helper_csrw(CPURISCVState *env, int csr, target_ulong src)
65 target_ulong mask = env->xl == MXL_RV32 ? UINT32_MAX : (target_ulong)-1;
66 RISCVException ret = riscv_csrrw(env, csr, NULL, src, mask);
68 if (ret != RISCV_EXCP_NONE) {
69 riscv_raise_exception(env, ret, GETPC());
73 target_ulong helper_csrrw(CPURISCVState *env, int csr,
74 target_ulong src, target_ulong write_mask)
76 target_ulong val = 0;
77 RISCVException ret = riscv_csrrw(env, csr, &val, src, write_mask);
79 if (ret != RISCV_EXCP_NONE) {
80 riscv_raise_exception(env, ret, GETPC());
82 return val;
85 target_ulong helper_csrr_i128(CPURISCVState *env, int csr)
87 Int128 rv = int128_zero();
88 RISCVException ret = riscv_csrrw_i128(env, csr, &rv,
89 int128_zero(),
90 int128_zero());
92 if (ret != RISCV_EXCP_NONE) {
93 riscv_raise_exception(env, ret, GETPC());
96 env->retxh = int128_gethi(rv);
97 return int128_getlo(rv);
100 void helper_csrw_i128(CPURISCVState *env, int csr,
101 target_ulong srcl, target_ulong srch)
103 RISCVException ret = riscv_csrrw_i128(env, csr, NULL,
104 int128_make128(srcl, srch),
105 UINT128_MAX);
107 if (ret != RISCV_EXCP_NONE) {
108 riscv_raise_exception(env, ret, GETPC());
112 target_ulong helper_csrrw_i128(CPURISCVState *env, int csr,
113 target_ulong srcl, target_ulong srch,
114 target_ulong maskl, target_ulong maskh)
116 Int128 rv = int128_zero();
117 RISCVException ret = riscv_csrrw_i128(env, csr, &rv,
118 int128_make128(srcl, srch),
119 int128_make128(maskl, maskh));
121 if (ret != RISCV_EXCP_NONE) {
122 riscv_raise_exception(env, ret, GETPC());
125 env->retxh = int128_gethi(rv);
126 return int128_getlo(rv);
131 * check_zicbo_envcfg
133 * Raise virtual exceptions and illegal instruction exceptions for
134 * Zicbo[mz] instructions based on the settings of [mhs]envcfg as
135 * specified in section 2.5.1 of the CMO specification.
137 static void check_zicbo_envcfg(CPURISCVState *env, target_ulong envbits,
138 uintptr_t ra)
140 #ifndef CONFIG_USER_ONLY
141 if ((env->priv < PRV_M) && !get_field(env->menvcfg, envbits)) {
142 riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, ra);
145 if (env->virt_enabled &&
146 (((env->priv <= PRV_S) && !get_field(env->henvcfg, envbits)) ||
147 ((env->priv < PRV_S) && !get_field(env->senvcfg, envbits)))) {
148 riscv_raise_exception(env, RISCV_EXCP_VIRT_INSTRUCTION_FAULT, ra);
151 if ((env->priv < PRV_S) && !get_field(env->senvcfg, envbits)) {
152 riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, ra);
154 #endif
157 void helper_cbo_zero(CPURISCVState *env, target_ulong address)
159 RISCVCPU *cpu = env_archcpu(env);
160 uint16_t cbozlen = cpu->cfg.cboz_blocksize;
161 int mmu_idx = cpu_mmu_index(env, false);
162 uintptr_t ra = GETPC();
163 void *mem;
165 check_zicbo_envcfg(env, MENVCFG_CBZE, ra);
167 /* Mask off low-bits to align-down to the cache-block. */
168 address &= ~(cbozlen - 1);
171 * cbo.zero requires MMU_DATA_STORE access. Do a probe_write()
172 * to raise any exceptions, including PMP.
174 mem = probe_write(env, address, cbozlen, mmu_idx, ra);
176 if (likely(mem)) {
177 memset(mem, 0, cbozlen);
178 } else {
180 * This means that we're dealing with an I/O page. Section 4.2
181 * of cmobase v1.0.1 says:
183 * "Cache-block zero instructions store zeros independently
184 * of whether data from the underlying memory locations are
185 * cacheable."
187 * Write zeros in address + cbozlen regardless of not being
188 * a RAM page.
190 for (int i = 0; i < cbozlen; i++) {
191 cpu_stb_mmuidx_ra(env, address + i, 0, mmu_idx, ra);
197 * check_zicbom_access
199 * Check access permissions (LOAD, STORE or FETCH as specified in
200 * section 2.5.2 of the CMO specification) for Zicbom, raising
201 * either store page-fault (non-virtualized) or store guest-page
202 * fault (virtualized).
204 static void check_zicbom_access(CPURISCVState *env,
205 target_ulong address,
206 uintptr_t ra)
208 RISCVCPU *cpu = env_archcpu(env);
209 int mmu_idx = cpu_mmu_index(env, false);
210 uint16_t cbomlen = cpu->cfg.cbom_blocksize;
211 void *phost;
212 int ret;
214 /* Mask off low-bits to align-down to the cache-block. */
215 address &= ~(cbomlen - 1);
218 * Section 2.5.2 of cmobase v1.0.1:
220 * "A cache-block management instruction is permitted to
221 * access the specified cache block whenever a load instruction
222 * or store instruction is permitted to access the corresponding
223 * physical addresses. If neither a load instruction nor store
224 * instruction is permitted to access the physical addresses,
225 * but an instruction fetch is permitted to access the physical
226 * addresses, whether a cache-block management instruction is
227 * permitted to access the cache block is UNSPECIFIED."
229 ret = probe_access_flags(env, address, cbomlen, MMU_DATA_LOAD,
230 mmu_idx, true, &phost, ra);
231 if (ret != TLB_INVALID_MASK) {
232 /* Success: readable */
233 return;
237 * Since not readable, must be writable. On failure, store
238 * fault/store guest amo fault will be raised by
239 * riscv_cpu_tlb_fill(). PMP exceptions will be caught
240 * there as well.
242 probe_write(env, address, cbomlen, mmu_idx, ra);
245 void helper_cbo_clean_flush(CPURISCVState *env, target_ulong address)
247 uintptr_t ra = GETPC();
248 check_zicbo_envcfg(env, MENVCFG_CBCFE, ra);
249 check_zicbom_access(env, address, ra);
251 /* We don't emulate the cache-hierarchy, so we're done. */
254 void helper_cbo_inval(CPURISCVState *env, target_ulong address)
256 uintptr_t ra = GETPC();
257 check_zicbo_envcfg(env, MENVCFG_CBIE, ra);
258 check_zicbom_access(env, address, ra);
260 /* We don't emulate the cache-hierarchy, so we're done. */
263 #ifndef CONFIG_USER_ONLY
265 target_ulong helper_sret(CPURISCVState *env)
267 uint64_t mstatus;
268 target_ulong prev_priv, prev_virt;
270 if (!(env->priv >= PRV_S)) {
271 riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC());
274 target_ulong retpc = env->sepc;
275 if (!riscv_has_ext(env, RVC) && (retpc & 0x3)) {
276 riscv_raise_exception(env, RISCV_EXCP_INST_ADDR_MIS, GETPC());
279 if (get_field(env->mstatus, MSTATUS_TSR) && !(env->priv >= PRV_M)) {
280 riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC());
283 if (env->virt_enabled && get_field(env->hstatus, HSTATUS_VTSR)) {
284 riscv_raise_exception(env, RISCV_EXCP_VIRT_INSTRUCTION_FAULT, GETPC());
287 mstatus = env->mstatus;
288 prev_priv = get_field(mstatus, MSTATUS_SPP);
289 mstatus = set_field(mstatus, MSTATUS_SIE,
290 get_field(mstatus, MSTATUS_SPIE));
291 mstatus = set_field(mstatus, MSTATUS_SPIE, 1);
292 mstatus = set_field(mstatus, MSTATUS_SPP, PRV_U);
293 if (env->priv_ver >= PRIV_VERSION_1_12_0) {
294 mstatus = set_field(mstatus, MSTATUS_MPRV, 0);
296 env->mstatus = mstatus;
298 if (riscv_has_ext(env, RVH) && !env->virt_enabled) {
299 /* We support Hypervisor extensions and virtulisation is disabled */
300 target_ulong hstatus = env->hstatus;
302 prev_virt = get_field(hstatus, HSTATUS_SPV);
304 hstatus = set_field(hstatus, HSTATUS_SPV, 0);
306 env->hstatus = hstatus;
308 if (prev_virt) {
309 riscv_cpu_swap_hypervisor_regs(env);
312 riscv_cpu_set_virt_enabled(env, prev_virt);
315 riscv_cpu_set_mode(env, prev_priv);
317 return retpc;
320 target_ulong helper_mret(CPURISCVState *env)
322 if (!(env->priv >= PRV_M)) {
323 riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC());
326 target_ulong retpc = env->mepc;
327 if (!riscv_has_ext(env, RVC) && (retpc & 0x3)) {
328 riscv_raise_exception(env, RISCV_EXCP_INST_ADDR_MIS, GETPC());
331 uint64_t mstatus = env->mstatus;
332 target_ulong prev_priv = get_field(mstatus, MSTATUS_MPP);
334 if (riscv_cpu_cfg(env)->pmp &&
335 !pmp_get_num_rules(env) && (prev_priv != PRV_M)) {
336 riscv_raise_exception(env, RISCV_EXCP_INST_ACCESS_FAULT, GETPC());
339 target_ulong prev_virt = get_field(env->mstatus, MSTATUS_MPV) &&
340 (prev_priv != PRV_M);
341 mstatus = set_field(mstatus, MSTATUS_MIE,
342 get_field(mstatus, MSTATUS_MPIE));
343 mstatus = set_field(mstatus, MSTATUS_MPIE, 1);
344 mstatus = set_field(mstatus, MSTATUS_MPP,
345 riscv_has_ext(env, RVU) ? PRV_U : PRV_M);
346 mstatus = set_field(mstatus, MSTATUS_MPV, 0);
347 if ((env->priv_ver >= PRIV_VERSION_1_12_0) && (prev_priv != PRV_M)) {
348 mstatus = set_field(mstatus, MSTATUS_MPRV, 0);
350 env->mstatus = mstatus;
351 riscv_cpu_set_mode(env, prev_priv);
353 if (riscv_has_ext(env, RVH)) {
354 if (prev_virt) {
355 riscv_cpu_swap_hypervisor_regs(env);
358 riscv_cpu_set_virt_enabled(env, prev_virt);
361 return retpc;
364 void helper_wfi(CPURISCVState *env)
366 CPUState *cs = env_cpu(env);
367 bool rvs = riscv_has_ext(env, RVS);
368 bool prv_u = env->priv == PRV_U;
369 bool prv_s = env->priv == PRV_S;
371 if (((prv_s || (!rvs && prv_u)) && get_field(env->mstatus, MSTATUS_TW)) ||
372 (rvs && prv_u && !env->virt_enabled)) {
373 riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC());
374 } else if (env->virt_enabled &&
375 (prv_u || (prv_s && get_field(env->hstatus, HSTATUS_VTW)))) {
376 riscv_raise_exception(env, RISCV_EXCP_VIRT_INSTRUCTION_FAULT, GETPC());
377 } else {
378 cs->halted = 1;
379 cs->exception_index = EXCP_HLT;
380 cpu_loop_exit(cs);
384 void helper_tlb_flush(CPURISCVState *env)
386 CPUState *cs = env_cpu(env);
387 if (!env->virt_enabled &&
388 (env->priv == PRV_U ||
389 (env->priv == PRV_S && get_field(env->mstatus, MSTATUS_TVM)))) {
390 riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC());
391 } else if (env->virt_enabled &&
392 (env->priv == PRV_U || get_field(env->hstatus, HSTATUS_VTVM))) {
393 riscv_raise_exception(env, RISCV_EXCP_VIRT_INSTRUCTION_FAULT, GETPC());
394 } else {
395 tlb_flush(cs);
399 void helper_tlb_flush_all(CPURISCVState *env)
401 CPUState *cs = env_cpu(env);
402 tlb_flush_all_cpus_synced(cs);
405 void helper_hyp_tlb_flush(CPURISCVState *env)
407 CPUState *cs = env_cpu(env);
409 if (env->virt_enabled) {
410 riscv_raise_exception(env, RISCV_EXCP_VIRT_INSTRUCTION_FAULT, GETPC());
413 if (env->priv == PRV_M ||
414 (env->priv == PRV_S && !env->virt_enabled)) {
415 tlb_flush(cs);
416 return;
419 riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC());
422 void helper_hyp_gvma_tlb_flush(CPURISCVState *env)
424 if (env->priv == PRV_S && !env->virt_enabled &&
425 get_field(env->mstatus, MSTATUS_TVM)) {
426 riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC());
429 helper_hyp_tlb_flush(env);
432 static int check_access_hlsv(CPURISCVState *env, bool x, uintptr_t ra)
434 if (env->priv == PRV_M) {
435 /* always allowed */
436 } else if (env->virt_enabled) {
437 riscv_raise_exception(env, RISCV_EXCP_VIRT_INSTRUCTION_FAULT, ra);
438 } else if (env->priv == PRV_U && !get_field(env->hstatus, HSTATUS_HU)) {
439 riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, ra);
442 int mode = get_field(env->hstatus, HSTATUS_SPVP);
443 if (!x && mode == PRV_S && get_field(env->vsstatus, MSTATUS_SUM)) {
444 mode = MMUIdx_S_SUM;
446 return mode | MMU_2STAGE_BIT;
449 target_ulong helper_hyp_hlv_bu(CPURISCVState *env, target_ulong addr)
451 uintptr_t ra = GETPC();
452 int mmu_idx = check_access_hlsv(env, false, ra);
453 MemOpIdx oi = make_memop_idx(MO_UB, mmu_idx);
455 return cpu_ldb_mmu(env, addr, oi, ra);
458 target_ulong helper_hyp_hlv_hu(CPURISCVState *env, target_ulong addr)
460 uintptr_t ra = GETPC();
461 int mmu_idx = check_access_hlsv(env, false, ra);
462 MemOpIdx oi = make_memop_idx(MO_TEUW, mmu_idx);
464 return cpu_ldw_mmu(env, addr, oi, ra);
467 target_ulong helper_hyp_hlv_wu(CPURISCVState *env, target_ulong addr)
469 uintptr_t ra = GETPC();
470 int mmu_idx = check_access_hlsv(env, false, ra);
471 MemOpIdx oi = make_memop_idx(MO_TEUL, mmu_idx);
473 return cpu_ldl_mmu(env, addr, oi, ra);
476 target_ulong helper_hyp_hlv_d(CPURISCVState *env, target_ulong addr)
478 uintptr_t ra = GETPC();
479 int mmu_idx = check_access_hlsv(env, false, ra);
480 MemOpIdx oi = make_memop_idx(MO_TEUQ, mmu_idx);
482 return cpu_ldq_mmu(env, addr, oi, ra);
485 void helper_hyp_hsv_b(CPURISCVState *env, target_ulong addr, target_ulong val)
487 uintptr_t ra = GETPC();
488 int mmu_idx = check_access_hlsv(env, false, ra);
489 MemOpIdx oi = make_memop_idx(MO_UB, mmu_idx);
491 cpu_stb_mmu(env, addr, val, oi, ra);
494 void helper_hyp_hsv_h(CPURISCVState *env, target_ulong addr, target_ulong val)
496 uintptr_t ra = GETPC();
497 int mmu_idx = check_access_hlsv(env, false, ra);
498 MemOpIdx oi = make_memop_idx(MO_TEUW, mmu_idx);
500 cpu_stw_mmu(env, addr, val, oi, ra);
503 void helper_hyp_hsv_w(CPURISCVState *env, target_ulong addr, target_ulong val)
505 uintptr_t ra = GETPC();
506 int mmu_idx = check_access_hlsv(env, false, ra);
507 MemOpIdx oi = make_memop_idx(MO_TEUL, mmu_idx);
509 cpu_stl_mmu(env, addr, val, oi, ra);
512 void helper_hyp_hsv_d(CPURISCVState *env, target_ulong addr, target_ulong val)
514 uintptr_t ra = GETPC();
515 int mmu_idx = check_access_hlsv(env, false, ra);
516 MemOpIdx oi = make_memop_idx(MO_TEUQ, mmu_idx);
518 cpu_stq_mmu(env, addr, val, oi, ra);
522 * TODO: These implementations are not quite correct. They perform the
523 * access using execute permission just fine, but the final PMP check
524 * is supposed to have read permission as well. Without replicating
525 * a fair fraction of cputlb.c, fixing this requires adding new mmu_idx
526 * which would imply that exact check in tlb_fill.
528 target_ulong helper_hyp_hlvx_hu(CPURISCVState *env, target_ulong addr)
530 uintptr_t ra = GETPC();
531 int mmu_idx = check_access_hlsv(env, true, ra);
532 MemOpIdx oi = make_memop_idx(MO_TEUW, mmu_idx);
534 return cpu_ldw_code_mmu(env, addr, oi, GETPC());
537 target_ulong helper_hyp_hlvx_wu(CPURISCVState *env, target_ulong addr)
539 uintptr_t ra = GETPC();
540 int mmu_idx = check_access_hlsv(env, true, ra);
541 MemOpIdx oi = make_memop_idx(MO_TEUL, mmu_idx);
543 return cpu_ldl_code_mmu(env, addr, oi, ra);
546 #endif /* !CONFIG_USER_ONLY */