2 * S/390 misc helper routines
4 * Copyright (c) 2009 Ulrich Hecht
5 * Copyright (c) 2009 Alexander Graf
7 * This library is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU Lesser General Public
9 * License as published by the Free Software Foundation; either
10 * version 2 of the License, or (at your option) any later version.
12 * This library is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15 * Lesser General Public License for more details.
17 * You should have received a copy of the GNU Lesser General Public
18 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
21 #include "qemu/osdep.h"
22 #include "qemu/main-loop.h"
25 #include "exec/memory.h"
26 #include "qemu/host-utils.h"
27 #include "exec/helper-proto.h"
28 #include "qemu/timer.h"
29 #include "exec/address-spaces.h"
30 #include "exec/exec-all.h"
31 #include "exec/cpu_ldst.h"
33 #if !defined(CONFIG_USER_ONLY)
34 #include "sysemu/cpus.h"
35 #include "sysemu/sysemu.h"
36 #include "hw/s390x/ebcdic.h"
37 #include "hw/s390x/s390-virtio-hcall.h"
38 #include "hw/s390x/sclp.h"
41 /* #define DEBUG_HELPER */
43 #define HELPER_LOG(x...) qemu_log(x)
45 #define HELPER_LOG(x...)
48 /* Raise an exception dynamically from a helper function. */
49 void QEMU_NORETURN
runtime_exception(CPUS390XState
*env
, int excp
,
52 CPUState
*cs
= CPU(s390_env_get_cpu(env
));
54 cs
->exception_index
= EXCP_PGM
;
55 env
->int_pgm_code
= excp
;
56 env
->int_pgm_ilen
= ILEN_AUTO
;
58 /* Use the (ultimate) callers address to find the insn that trapped. */
59 cpu_restore_state(cs
, retaddr
);
64 /* Raise an exception statically from a TB. */
65 void HELPER(exception
)(CPUS390XState
*env
, uint32_t excp
)
67 CPUState
*cs
= CPU(s390_env_get_cpu(env
));
69 HELPER_LOG("%s: exception %d\n", __func__
, excp
);
70 cs
->exception_index
= excp
;
74 #ifndef CONFIG_USER_ONLY
76 /* SCLP service call */
77 uint32_t HELPER(servc
)(CPUS390XState
*env
, uint64_t r1
, uint64_t r2
)
79 qemu_mutex_lock_iothread();
80 int r
= sclp_service_call(env
, r1
, r2
);
82 program_interrupt(env
, -r
, 4);
85 qemu_mutex_unlock_iothread();
89 void HELPER(diag
)(CPUS390XState
*env
, uint32_t r1
, uint32_t r3
, uint32_t num
)
96 qemu_mutex_lock_iothread();
97 r
= s390_virtio_hypercall(env
);
98 qemu_mutex_unlock_iothread();
106 handle_diag_308(env
, r1
, r3
);
110 /* time bomb (watchdog) */
111 r
= handle_diag_288(env
, r1
, r3
);
119 program_interrupt(env
, PGM_SPECIFICATION
, ILEN_AUTO
);
124 void HELPER(spx
)(CPUS390XState
*env
, uint64_t a1
)
126 CPUState
*cs
= CPU(s390_env_get_cpu(env
));
127 uint32_t prefix
= a1
& 0x7fffe000;
130 HELPER_LOG("prefix: %#x\n", prefix
);
131 tlb_flush_page(cs
, 0);
132 tlb_flush_page(cs
, TARGET_PAGE_SIZE
);
136 uint64_t HELPER(stck
)(CPUS390XState
*env
)
140 time
= env
->tod_offset
+
141 time2tod(qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL
) - env
->tod_basetime
);
146 /* Set Clock Comparator */
147 void HELPER(sckc
)(CPUS390XState
*env
, uint64_t time
)
155 /* difference between origins */
156 time
-= env
->tod_offset
;
159 time
= tod2time(time
);
161 timer_mod(env
->tod_timer
, env
->tod_basetime
+ time
);
164 /* Store Clock Comparator */
165 uint64_t HELPER(stckc
)(CPUS390XState
*env
)
171 void HELPER(spt
)(CPUS390XState
*env
, uint64_t time
)
178 time
= tod2time(time
);
180 env
->cputm
= qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL
) + time
;
182 timer_mod(env
->cpu_timer
, env
->cputm
);
185 /* Store CPU Timer */
186 uint64_t HELPER(stpt
)(CPUS390XState
*env
)
188 return time2tod(env
->cputm
- qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL
));
191 /* Store System Information */
192 uint32_t HELPER(stsi
)(CPUS390XState
*env
, uint64_t a0
,
193 uint64_t r0
, uint64_t r1
)
195 S390CPU
*cpu
= s390_env_get_cpu(env
);
199 if ((r0
& STSI_LEVEL_MASK
) <= STSI_LEVEL_3
&&
200 ((r0
& STSI_R0_RESERVED_MASK
) || (r1
& STSI_R1_RESERVED_MASK
))) {
201 /* valid function code, invalid reserved bits */
202 program_interrupt(env
, PGM_SPECIFICATION
, 4);
205 sel1
= r0
& STSI_R0_SEL1_MASK
;
206 sel2
= r1
& STSI_R1_SEL2_MASK
;
208 /* XXX: spec exception if sysib is not 4k-aligned */
210 switch (r0
& STSI_LEVEL_MASK
) {
212 if ((sel1
== 1) && (sel2
== 1)) {
213 /* Basic Machine Configuration */
214 struct sysib_111 sysib
;
217 memset(&sysib
, 0, sizeof(sysib
));
218 ebcdic_put(sysib
.manuf
, "QEMU ", 16);
219 /* same as machine type number in STORE CPU ID, but in EBCDIC */
220 snprintf(type
, ARRAY_SIZE(type
), "%X", cpu
->model
->def
->type
);
221 ebcdic_put(sysib
.type
, type
, 4);
222 /* model number (not stored in STORE CPU ID for z/Architecure) */
223 ebcdic_put(sysib
.model
, "QEMU ", 16);
224 ebcdic_put(sysib
.sequence
, "QEMU ", 16);
225 ebcdic_put(sysib
.plant
, "QEMU", 4);
226 cpu_physical_memory_write(a0
, &sysib
, sizeof(sysib
));
227 } else if ((sel1
== 2) && (sel2
== 1)) {
228 /* Basic Machine CPU */
229 struct sysib_121 sysib
;
231 memset(&sysib
, 0, sizeof(sysib
));
232 /* XXX make different for different CPUs? */
233 ebcdic_put(sysib
.sequence
, "QEMUQEMUQEMUQEMU", 16);
234 ebcdic_put(sysib
.plant
, "QEMU", 4);
235 stw_p(&sysib
.cpu_addr
, env
->core_id
);
236 cpu_physical_memory_write(a0
, &sysib
, sizeof(sysib
));
237 } else if ((sel1
== 2) && (sel2
== 2)) {
238 /* Basic Machine CPUs */
239 struct sysib_122 sysib
;
241 memset(&sysib
, 0, sizeof(sysib
));
242 stl_p(&sysib
.capability
, 0x443afc29);
243 /* XXX change when SMP comes */
244 stw_p(&sysib
.total_cpus
, 1);
245 stw_p(&sysib
.active_cpus
, 1);
246 stw_p(&sysib
.standby_cpus
, 0);
247 stw_p(&sysib
.reserved_cpus
, 0);
248 cpu_physical_memory_write(a0
, &sysib
, sizeof(sysib
));
255 if ((sel1
== 2) && (sel2
== 1)) {
257 struct sysib_221 sysib
;
259 memset(&sysib
, 0, sizeof(sysib
));
260 /* XXX make different for different CPUs? */
261 ebcdic_put(sysib
.sequence
, "QEMUQEMUQEMUQEMU", 16);
262 ebcdic_put(sysib
.plant
, "QEMU", 4);
263 stw_p(&sysib
.cpu_addr
, env
->core_id
);
264 stw_p(&sysib
.cpu_id
, 0);
265 cpu_physical_memory_write(a0
, &sysib
, sizeof(sysib
));
266 } else if ((sel1
== 2) && (sel2
== 2)) {
268 struct sysib_222 sysib
;
270 memset(&sysib
, 0, sizeof(sysib
));
271 stw_p(&sysib
.lpar_num
, 0);
273 /* XXX change when SMP comes */
274 stw_p(&sysib
.total_cpus
, 1);
275 stw_p(&sysib
.conf_cpus
, 1);
276 stw_p(&sysib
.standby_cpus
, 0);
277 stw_p(&sysib
.reserved_cpus
, 0);
278 ebcdic_put(sysib
.name
, "QEMU ", 8);
279 stl_p(&sysib
.caf
, 1000);
280 stw_p(&sysib
.dedicated_cpus
, 0);
281 stw_p(&sysib
.shared_cpus
, 0);
282 cpu_physical_memory_write(a0
, &sysib
, sizeof(sysib
));
290 if ((sel1
== 2) && (sel2
== 2)) {
292 struct sysib_322 sysib
;
294 memset(&sysib
, 0, sizeof(sysib
));
296 /* XXX change when SMP comes */
297 stw_p(&sysib
.vm
[0].total_cpus
, 1);
298 stw_p(&sysib
.vm
[0].conf_cpus
, 1);
299 stw_p(&sysib
.vm
[0].standby_cpus
, 0);
300 stw_p(&sysib
.vm
[0].reserved_cpus
, 0);
301 ebcdic_put(sysib
.vm
[0].name
, "KVMguest", 8);
302 stl_p(&sysib
.vm
[0].caf
, 1000);
303 ebcdic_put(sysib
.vm
[0].cpi
, "KVM/Linux ", 16);
304 cpu_physical_memory_write(a0
, &sysib
, sizeof(sysib
));
310 case STSI_LEVEL_CURRENT
:
311 env
->regs
[0] = STSI_LEVEL_3
;
321 uint32_t HELPER(sigp
)(CPUS390XState
*env
, uint64_t order_code
, uint32_t r1
,
324 int cc
= SIGP_CC_ORDER_CODE_ACCEPTED
;
326 HELPER_LOG("%s: %016" PRIx64
" %08x %016" PRIx64
"\n",
327 __func__
, order_code
, r1
, cpu_addr
);
329 /* Remember: Use "R1 or R1 + 1, whichever is the odd-numbered register"
330 as parameter (input). Status (output) is always R1. */
332 switch (order_code
& SIGP_ORDER_MASK
) {
337 /* enumerate CPU status */
339 /* XXX implement when SMP comes */
342 env
->regs
[r1
] &= 0xffffffff00000000ULL
;
345 #if !defined(CONFIG_USER_ONLY)
347 qemu_system_reset_request(SHUTDOWN_CAUSE_GUEST_RESET
);
348 cpu_loop_exit(CPU(s390_env_get_cpu(env
)));
351 qemu_system_shutdown_request(SHUTDOWN_CAUSE_GUEST_SHUTDOWN
);
352 cpu_loop_exit(CPU(s390_env_get_cpu(env
)));
357 fprintf(stderr
, "XXX unknown sigp: 0x%" PRIx64
"\n", order_code
);
358 cc
= SIGP_CC_NOT_OPERATIONAL
;
365 #ifndef CONFIG_USER_ONLY
366 void HELPER(xsch
)(CPUS390XState
*env
, uint64_t r1
)
368 S390CPU
*cpu
= s390_env_get_cpu(env
);
369 qemu_mutex_lock_iothread();
370 ioinst_handle_xsch(cpu
, r1
);
371 qemu_mutex_unlock_iothread();
374 void HELPER(csch
)(CPUS390XState
*env
, uint64_t r1
)
376 S390CPU
*cpu
= s390_env_get_cpu(env
);
377 qemu_mutex_lock_iothread();
378 ioinst_handle_csch(cpu
, r1
);
379 qemu_mutex_unlock_iothread();
382 void HELPER(hsch
)(CPUS390XState
*env
, uint64_t r1
)
384 S390CPU
*cpu
= s390_env_get_cpu(env
);
385 qemu_mutex_lock_iothread();
386 ioinst_handle_hsch(cpu
, r1
);
387 qemu_mutex_unlock_iothread();
390 void HELPER(msch
)(CPUS390XState
*env
, uint64_t r1
, uint64_t inst
)
392 S390CPU
*cpu
= s390_env_get_cpu(env
);
393 qemu_mutex_lock_iothread();
394 ioinst_handle_msch(cpu
, r1
, inst
>> 16);
395 qemu_mutex_unlock_iothread();
398 void HELPER(rchp
)(CPUS390XState
*env
, uint64_t r1
)
400 S390CPU
*cpu
= s390_env_get_cpu(env
);
401 qemu_mutex_lock_iothread();
402 ioinst_handle_rchp(cpu
, r1
);
403 qemu_mutex_unlock_iothread();
406 void HELPER(rsch
)(CPUS390XState
*env
, uint64_t r1
)
408 S390CPU
*cpu
= s390_env_get_cpu(env
);
409 qemu_mutex_lock_iothread();
410 ioinst_handle_rsch(cpu
, r1
);
411 qemu_mutex_unlock_iothread();
414 void HELPER(ssch
)(CPUS390XState
*env
, uint64_t r1
, uint64_t inst
)
416 S390CPU
*cpu
= s390_env_get_cpu(env
);
417 qemu_mutex_lock_iothread();
418 ioinst_handle_ssch(cpu
, r1
, inst
>> 16);
419 qemu_mutex_unlock_iothread();
422 void HELPER(stsch
)(CPUS390XState
*env
, uint64_t r1
, uint64_t inst
)
424 S390CPU
*cpu
= s390_env_get_cpu(env
);
425 qemu_mutex_lock_iothread();
426 ioinst_handle_stsch(cpu
, r1
, inst
>> 16);
427 qemu_mutex_unlock_iothread();
430 void HELPER(tsch
)(CPUS390XState
*env
, uint64_t r1
, uint64_t inst
)
432 S390CPU
*cpu
= s390_env_get_cpu(env
);
433 qemu_mutex_lock_iothread();
434 ioinst_handle_tsch(cpu
, r1
, inst
>> 16);
435 qemu_mutex_unlock_iothread();
438 void HELPER(chsc
)(CPUS390XState
*env
, uint64_t inst
)
440 S390CPU
*cpu
= s390_env_get_cpu(env
);
441 qemu_mutex_lock_iothread();
442 ioinst_handle_chsc(cpu
, inst
>> 16);
443 qemu_mutex_unlock_iothread();
447 #ifndef CONFIG_USER_ONLY
448 void HELPER(per_check_exception
)(CPUS390XState
*env
)
452 if (env
->per_perc_atmid
) {
454 * FIXME: ILEN_AUTO is most probably the right thing to use. ilen
455 * always has to match the instruction referenced in the PSW. E.g.
456 * if a PER interrupt is triggered via EXECUTE, we have to use ilen
457 * of EXECUTE, while per_address contains the target of EXECUTE.
459 ilen
= get_ilen(cpu_ldub_code(env
, env
->per_address
));
460 program_interrupt(env
, PGM_PER
, ilen
);
464 /* Check if an address is within the PER starting address and the PER
465 ending address. The address range might loop. */
466 static inline bool get_per_in_range(CPUS390XState
*env
, uint64_t addr
)
468 if (env
->cregs
[10] <= env
->cregs
[11]) {
469 return env
->cregs
[10] <= addr
&& addr
<= env
->cregs
[11];
471 return env
->cregs
[10] <= addr
|| addr
<= env
->cregs
[11];
475 void HELPER(per_branch
)(CPUS390XState
*env
, uint64_t from
, uint64_t to
)
477 if ((env
->cregs
[9] & PER_CR9_EVENT_BRANCH
)) {
478 if (!(env
->cregs
[9] & PER_CR9_CONTROL_BRANCH_ADDRESS
)
479 || get_per_in_range(env
, to
)) {
480 env
->per_address
= from
;
481 env
->per_perc_atmid
= PER_CODE_EVENT_BRANCH
| get_per_atmid(env
);
486 void HELPER(per_ifetch
)(CPUS390XState
*env
, uint64_t addr
)
488 if ((env
->cregs
[9] & PER_CR9_EVENT_IFETCH
) && get_per_in_range(env
, addr
)) {
489 env
->per_address
= addr
;
490 env
->per_perc_atmid
= PER_CODE_EVENT_IFETCH
| get_per_atmid(env
);
492 /* If the instruction has to be nullified, trigger the
493 exception immediately. */
494 if (env
->cregs
[9] & PER_CR9_EVENT_NULLIFICATION
) {
495 CPUState
*cs
= CPU(s390_env_get_cpu(env
));
497 env
->per_perc_atmid
|= PER_CODE_EVENT_NULLIFICATION
;
498 env
->int_pgm_code
= PGM_PER
;
499 env
->int_pgm_ilen
= get_ilen(cpu_ldub_code(env
, addr
));
501 cs
->exception_index
= EXCP_PGM
;
508 /* The maximum bit defined at the moment is 129. */
509 #define MAX_STFL_WORDS 3
511 /* Canonicalize the current cpu's features into the 64-bit words required
512 by STFLE. Return the index-1 of the max word that is non-zero. */
513 static unsigned do_stfle(CPUS390XState
*env
, uint64_t words
[MAX_STFL_WORDS
])
515 S390CPU
*cpu
= s390_env_get_cpu(env
);
516 const unsigned long *features
= cpu
->model
->features
;
517 unsigned max_bit
= 0;
520 memset(words
, 0, sizeof(uint64_t) * MAX_STFL_WORDS
);
522 if (test_bit(S390_FEAT_ZARCH
, features
)) {
523 /* z/Architecture is always active if around */
524 words
[0] = 1ull << (63 - 2);
527 for (feat
= find_first_bit(features
, S390_FEAT_MAX
);
528 feat
< S390_FEAT_MAX
;
529 feat
= find_next_bit(features
, S390_FEAT_MAX
, feat
+ 1)) {
530 const S390FeatDef
*def
= s390_feat_def(feat
);
531 if (def
->type
== S390_FEAT_TYPE_STFL
) {
532 unsigned bit
= def
->bit
;
536 assert(bit
/ 64 < MAX_STFL_WORDS
);
537 words
[bit
/ 64] |= 1ULL << (63 - bit
% 64);
544 void HELPER(stfl
)(CPUS390XState
*env
)
546 uint64_t words
[MAX_STFL_WORDS
];
548 do_stfle(env
, words
);
549 cpu_stl_data(env
, 200, words
[0] >> 32);
552 uint32_t HELPER(stfle
)(CPUS390XState
*env
, uint64_t addr
)
554 uint64_t words
[MAX_STFL_WORDS
];
555 unsigned count_m1
= env
->regs
[0] & 0xff;
556 unsigned max_m1
= do_stfle(env
, words
);
559 for (i
= 0; i
<= count_m1
; ++i
) {
560 cpu_stq_data(env
, addr
+ 8 * i
, words
[i
]);
563 env
->regs
[0] = deposit64(env
->regs
[0], 0, 8, max_m1
);
564 return (count_m1
>= max_m1
? 0 : 3);