2 * S/390 misc helper routines
4 * Copyright (c) 2009 Ulrich Hecht
5 * Copyright (c) 2009 Alexander Graf
7 * This library is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU Lesser General Public
9 * License as published by the Free Software Foundation; either
10 * version 2 of the License, or (at your option) any later version.
12 * This library is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15 * Lesser General Public License for more details.
17 * You should have received a copy of the GNU Lesser General Public
18 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
21 #include "qemu/osdep.h"
23 #include "exec/memory.h"
24 #include "qemu/host-utils.h"
25 #include "exec/helper-proto.h"
26 #include "sysemu/kvm.h"
27 #include "qemu/timer.h"
28 #include "exec/address-spaces.h"
30 #include <linux/kvm.h>
32 #include "exec/cpu_ldst.h"
33 #include "hw/watchdog/wdt_diag288.h"
35 #if !defined(CONFIG_USER_ONLY)
36 #include "sysemu/cpus.h"
37 #include "sysemu/sysemu.h"
38 #include "hw/s390x/ebcdic.h"
39 #include "hw/s390x/ipl.h"
42 /* #define DEBUG_HELPER */
44 #define HELPER_LOG(x...) qemu_log(x)
46 #define HELPER_LOG(x...)
49 /* Raise an exception dynamically from a helper function. */
50 void QEMU_NORETURN
runtime_exception(CPUS390XState
*env
, int excp
,
53 CPUState
*cs
= CPU(s390_env_get_cpu(env
));
56 cs
->exception_index
= EXCP_PGM
;
57 env
->int_pgm_code
= excp
;
59 /* Use the (ultimate) callers address to find the insn that trapped. */
60 cpu_restore_state(cs
, retaddr
);
62 /* Advance past the insn. */
63 t
= cpu_ldub_code(env
, env
->psw
.addr
);
64 env
->int_pgm_ilen
= t
= get_ilen(t
);
70 /* Raise an exception statically from a TB. */
71 void HELPER(exception
)(CPUS390XState
*env
, uint32_t excp
)
73 CPUState
*cs
= CPU(s390_env_get_cpu(env
));
75 HELPER_LOG("%s: exception %d\n", __func__
, excp
);
76 cs
->exception_index
= excp
;
80 #ifndef CONFIG_USER_ONLY
82 void program_interrupt(CPUS390XState
*env
, uint32_t code
, int ilen
)
84 S390CPU
*cpu
= s390_env_get_cpu(env
);
86 qemu_log_mask(CPU_LOG_INT
, "program interrupt at %#" PRIx64
"\n",
91 struct kvm_s390_irq irq
= {
92 .type
= KVM_S390_PROGRAM_INT
,
96 kvm_s390_vcpu_interrupt(cpu
, &irq
);
99 CPUState
*cs
= CPU(cpu
);
101 env
->int_pgm_code
= code
;
102 env
->int_pgm_ilen
= ilen
;
103 cs
->exception_index
= EXCP_PGM
;
108 /* SCLP service call */
109 uint32_t HELPER(servc
)(CPUS390XState
*env
, uint64_t r1
, uint64_t r2
)
111 int r
= sclp_service_call(env
, r1
, r2
);
113 program_interrupt(env
, -r
, 4);
119 #ifndef CONFIG_USER_ONLY
120 static int modified_clear_reset(S390CPU
*cpu
)
122 S390CPUClass
*scc
= S390_CPU_GET_CLASS(cpu
);
126 cpu_synchronize_all_states();
128 run_on_cpu(t
, s390_do_cpu_full_reset
, t
);
133 scc
->load_normal(CPU(cpu
));
134 cpu_synchronize_all_post_reset();
139 static int load_normal_reset(S390CPU
*cpu
)
141 S390CPUClass
*scc
= S390_CPU_GET_CLASS(cpu
);
145 cpu_synchronize_all_states();
147 run_on_cpu(t
, s390_do_cpu_reset
, t
);
151 scc
->initial_cpu_reset(CPU(cpu
));
152 scc
->load_normal(CPU(cpu
));
153 cpu_synchronize_all_post_reset();
158 int handle_diag_288(CPUS390XState
*env
, uint64_t r1
, uint64_t r3
)
160 uint64_t func
= env
->regs
[r1
];
161 uint64_t timeout
= env
->regs
[r1
+ 1];
162 uint64_t action
= env
->regs
[r3
];
164 DIAG288State
*diag288
;
165 DIAG288Class
*diag288_class
;
167 if (r1
% 2 || action
!= 0) {
171 /* Timeout must be more than 15 seconds except for timer deletion */
172 if (func
!= WDT_DIAG288_CANCEL
&& timeout
< 15) {
176 obj
= object_resolve_path_type("", TYPE_WDT_DIAG288
, NULL
);
181 diag288
= DIAG288(obj
);
182 diag288_class
= DIAG288_GET_CLASS(diag288
);
183 return diag288_class
->handle_timer(diag288
, func
, timeout
);
186 #define DIAG_308_RC_OK 0x0001
187 #define DIAG_308_RC_NO_CONF 0x0102
188 #define DIAG_308_RC_INVALID 0x0402
190 void handle_diag_308(CPUS390XState
*env
, uint64_t r1
, uint64_t r3
)
192 uint64_t addr
= env
->regs
[r1
];
193 uint64_t subcode
= env
->regs
[r3
];
194 IplParameterBlock
*iplb
;
196 if (env
->psw
.mask
& PSW_MASK_PSTATE
) {
197 program_interrupt(env
, PGM_PRIVILEGED
, ILEN_LATER_INC
);
201 if ((subcode
& ~0x0ffffULL
) || (subcode
> 6)) {
202 program_interrupt(env
, PGM_SPECIFICATION
, ILEN_LATER_INC
);
208 modified_clear_reset(s390_env_get_cpu(env
));
210 cpu_loop_exit(CPU(s390_env_get_cpu(env
)));
214 load_normal_reset(s390_env_get_cpu(env
));
216 cpu_loop_exit(CPU(s390_env_get_cpu(env
)));
220 s390_reipl_request();
222 cpu_loop_exit(CPU(s390_env_get_cpu(env
)));
226 if ((r1
& 1) || (addr
& 0x0fffULL
)) {
227 program_interrupt(env
, PGM_SPECIFICATION
, ILEN_LATER_INC
);
230 if (!address_space_access_valid(&address_space_memory
, addr
,
231 sizeof(IplParameterBlock
), false)) {
232 program_interrupt(env
, PGM_ADDRESSING
, ILEN_LATER_INC
);
235 iplb
= g_malloc0(sizeof(IplParameterBlock
));
236 cpu_physical_memory_read(addr
, iplb
, sizeof(iplb
->len
));
237 if (!iplb_valid_len(iplb
)) {
238 env
->regs
[r1
+ 1] = DIAG_308_RC_INVALID
;
242 cpu_physical_memory_read(addr
, iplb
, be32_to_cpu(iplb
->len
));
244 if (!iplb_valid_ccw(iplb
) && !iplb_valid_fcp(iplb
)) {
245 env
->regs
[r1
+ 1] = DIAG_308_RC_INVALID
;
249 s390_ipl_update_diag308(iplb
);
250 env
->regs
[r1
+ 1] = DIAG_308_RC_OK
;
255 if ((r1
& 1) || (addr
& 0x0fffULL
)) {
256 program_interrupt(env
, PGM_SPECIFICATION
, ILEN_LATER_INC
);
259 if (!address_space_access_valid(&address_space_memory
, addr
,
260 sizeof(IplParameterBlock
), true)) {
261 program_interrupt(env
, PGM_ADDRESSING
, ILEN_LATER_INC
);
264 iplb
= s390_ipl_get_iplb();
266 cpu_physical_memory_write(addr
, iplb
, be32_to_cpu(iplb
->len
));
267 env
->regs
[r1
+ 1] = DIAG_308_RC_OK
;
269 env
->regs
[r1
+ 1] = DIAG_308_RC_NO_CONF
;
273 hw_error("Unhandled diag308 subcode %" PRIx64
, subcode
);
279 void HELPER(diag
)(CPUS390XState
*env
, uint32_t r1
, uint32_t r3
, uint32_t num
)
286 r
= s390_virtio_hypercall(env
);
294 handle_diag_308(env
, r1
, r3
);
303 program_interrupt(env
, PGM_OPERATION
, ILEN_LATER_INC
);
308 void HELPER(spx
)(CPUS390XState
*env
, uint64_t a1
)
310 CPUState
*cs
= CPU(s390_env_get_cpu(env
));
311 uint32_t prefix
= a1
& 0x7fffe000;
314 HELPER_LOG("prefix: %#x\n", prefix
);
315 tlb_flush_page(cs
, 0);
316 tlb_flush_page(cs
, TARGET_PAGE_SIZE
);
320 uint64_t HELPER(stck
)(CPUS390XState
*env
)
324 time
= env
->tod_offset
+
325 time2tod(qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL
) - env
->tod_basetime
);
330 /* Set Clock Comparator */
331 void HELPER(sckc
)(CPUS390XState
*env
, uint64_t time
)
339 /* difference between origins */
340 time
-= env
->tod_offset
;
343 time
= tod2time(time
);
345 timer_mod(env
->tod_timer
, env
->tod_basetime
+ time
);
348 /* Store Clock Comparator */
349 uint64_t HELPER(stckc
)(CPUS390XState
*env
)
355 void HELPER(spt
)(CPUS390XState
*env
, uint64_t time
)
362 time
= tod2time(time
);
364 env
->cputm
= qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL
) + time
;
366 timer_mod(env
->cpu_timer
, env
->cputm
);
369 /* Store CPU Timer */
370 uint64_t HELPER(stpt
)(CPUS390XState
*env
)
372 return time2tod(env
->cputm
- qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL
));
375 /* Store System Information */
376 uint32_t HELPER(stsi
)(CPUS390XState
*env
, uint64_t a0
,
377 uint64_t r0
, uint64_t r1
)
382 if ((r0
& STSI_LEVEL_MASK
) <= STSI_LEVEL_3
&&
383 ((r0
& STSI_R0_RESERVED_MASK
) || (r1
& STSI_R1_RESERVED_MASK
))) {
384 /* valid function code, invalid reserved bits */
385 program_interrupt(env
, PGM_SPECIFICATION
, 2);
388 sel1
= r0
& STSI_R0_SEL1_MASK
;
389 sel2
= r1
& STSI_R1_SEL2_MASK
;
391 /* XXX: spec exception if sysib is not 4k-aligned */
393 switch (r0
& STSI_LEVEL_MASK
) {
395 if ((sel1
== 1) && (sel2
== 1)) {
396 /* Basic Machine Configuration */
397 struct sysib_111 sysib
;
399 memset(&sysib
, 0, sizeof(sysib
));
400 ebcdic_put(sysib
.manuf
, "QEMU ", 16);
401 /* same as machine type number in STORE CPU ID */
402 ebcdic_put(sysib
.type
, "QEMU", 4);
403 /* same as model number in STORE CPU ID */
404 ebcdic_put(sysib
.model
, "QEMU ", 16);
405 ebcdic_put(sysib
.sequence
, "QEMU ", 16);
406 ebcdic_put(sysib
.plant
, "QEMU", 4);
407 cpu_physical_memory_write(a0
, &sysib
, sizeof(sysib
));
408 } else if ((sel1
== 2) && (sel2
== 1)) {
409 /* Basic Machine CPU */
410 struct sysib_121 sysib
;
412 memset(&sysib
, 0, sizeof(sysib
));
413 /* XXX make different for different CPUs? */
414 ebcdic_put(sysib
.sequence
, "QEMUQEMUQEMUQEMU", 16);
415 ebcdic_put(sysib
.plant
, "QEMU", 4);
416 stw_p(&sysib
.cpu_addr
, env
->cpu_num
);
417 cpu_physical_memory_write(a0
, &sysib
, sizeof(sysib
));
418 } else if ((sel1
== 2) && (sel2
== 2)) {
419 /* Basic Machine CPUs */
420 struct sysib_122 sysib
;
422 memset(&sysib
, 0, sizeof(sysib
));
423 stl_p(&sysib
.capability
, 0x443afc29);
424 /* XXX change when SMP comes */
425 stw_p(&sysib
.total_cpus
, 1);
426 stw_p(&sysib
.active_cpus
, 1);
427 stw_p(&sysib
.standby_cpus
, 0);
428 stw_p(&sysib
.reserved_cpus
, 0);
429 cpu_physical_memory_write(a0
, &sysib
, sizeof(sysib
));
436 if ((sel1
== 2) && (sel2
== 1)) {
438 struct sysib_221 sysib
;
440 memset(&sysib
, 0, sizeof(sysib
));
441 /* XXX make different for different CPUs? */
442 ebcdic_put(sysib
.sequence
, "QEMUQEMUQEMUQEMU", 16);
443 ebcdic_put(sysib
.plant
, "QEMU", 4);
444 stw_p(&sysib
.cpu_addr
, env
->cpu_num
);
445 stw_p(&sysib
.cpu_id
, 0);
446 cpu_physical_memory_write(a0
, &sysib
, sizeof(sysib
));
447 } else if ((sel1
== 2) && (sel2
== 2)) {
449 struct sysib_222 sysib
;
451 memset(&sysib
, 0, sizeof(sysib
));
452 stw_p(&sysib
.lpar_num
, 0);
454 /* XXX change when SMP comes */
455 stw_p(&sysib
.total_cpus
, 1);
456 stw_p(&sysib
.conf_cpus
, 1);
457 stw_p(&sysib
.standby_cpus
, 0);
458 stw_p(&sysib
.reserved_cpus
, 0);
459 ebcdic_put(sysib
.name
, "QEMU ", 8);
460 stl_p(&sysib
.caf
, 1000);
461 stw_p(&sysib
.dedicated_cpus
, 0);
462 stw_p(&sysib
.shared_cpus
, 0);
463 cpu_physical_memory_write(a0
, &sysib
, sizeof(sysib
));
471 if ((sel1
== 2) && (sel2
== 2)) {
473 struct sysib_322 sysib
;
475 memset(&sysib
, 0, sizeof(sysib
));
477 /* XXX change when SMP comes */
478 stw_p(&sysib
.vm
[0].total_cpus
, 1);
479 stw_p(&sysib
.vm
[0].conf_cpus
, 1);
480 stw_p(&sysib
.vm
[0].standby_cpus
, 0);
481 stw_p(&sysib
.vm
[0].reserved_cpus
, 0);
482 ebcdic_put(sysib
.vm
[0].name
, "KVMguest", 8);
483 stl_p(&sysib
.vm
[0].caf
, 1000);
484 ebcdic_put(sysib
.vm
[0].cpi
, "KVM/Linux ", 16);
485 cpu_physical_memory_write(a0
, &sysib
, sizeof(sysib
));
491 case STSI_LEVEL_CURRENT
:
492 env
->regs
[0] = STSI_LEVEL_3
;
502 uint32_t HELPER(sigp
)(CPUS390XState
*env
, uint64_t order_code
, uint32_t r1
,
505 int cc
= SIGP_CC_ORDER_CODE_ACCEPTED
;
507 HELPER_LOG("%s: %016" PRIx64
" %08x %016" PRIx64
"\n",
508 __func__
, order_code
, r1
, cpu_addr
);
510 /* Remember: Use "R1 or R1 + 1, whichever is the odd-numbered register"
511 as parameter (input). Status (output) is always R1. */
513 switch (order_code
) {
518 /* enumerate CPU status */
520 /* XXX implement when SMP comes */
523 env
->regs
[r1
] &= 0xffffffff00000000ULL
;
526 #if !defined(CONFIG_USER_ONLY)
528 qemu_system_reset_request();
529 cpu_loop_exit(CPU(s390_env_get_cpu(env
)));
532 qemu_system_shutdown_request();
533 cpu_loop_exit(CPU(s390_env_get_cpu(env
)));
538 fprintf(stderr
, "XXX unknown sigp: 0x%" PRIx64
"\n", order_code
);
539 cc
= SIGP_CC_NOT_OPERATIONAL
;
546 #ifndef CONFIG_USER_ONLY
547 void HELPER(xsch
)(CPUS390XState
*env
, uint64_t r1
)
549 S390CPU
*cpu
= s390_env_get_cpu(env
);
550 ioinst_handle_xsch(cpu
, r1
);
553 void HELPER(csch
)(CPUS390XState
*env
, uint64_t r1
)
555 S390CPU
*cpu
= s390_env_get_cpu(env
);
556 ioinst_handle_csch(cpu
, r1
);
559 void HELPER(hsch
)(CPUS390XState
*env
, uint64_t r1
)
561 S390CPU
*cpu
= s390_env_get_cpu(env
);
562 ioinst_handle_hsch(cpu
, r1
);
565 void HELPER(msch
)(CPUS390XState
*env
, uint64_t r1
, uint64_t inst
)
567 S390CPU
*cpu
= s390_env_get_cpu(env
);
568 ioinst_handle_msch(cpu
, r1
, inst
>> 16);
571 void HELPER(rchp
)(CPUS390XState
*env
, uint64_t r1
)
573 S390CPU
*cpu
= s390_env_get_cpu(env
);
574 ioinst_handle_rchp(cpu
, r1
);
577 void HELPER(rsch
)(CPUS390XState
*env
, uint64_t r1
)
579 S390CPU
*cpu
= s390_env_get_cpu(env
);
580 ioinst_handle_rsch(cpu
, r1
);
583 void HELPER(ssch
)(CPUS390XState
*env
, uint64_t r1
, uint64_t inst
)
585 S390CPU
*cpu
= s390_env_get_cpu(env
);
586 ioinst_handle_ssch(cpu
, r1
, inst
>> 16);
589 void HELPER(stsch
)(CPUS390XState
*env
, uint64_t r1
, uint64_t inst
)
591 S390CPU
*cpu
= s390_env_get_cpu(env
);
592 ioinst_handle_stsch(cpu
, r1
, inst
>> 16);
595 void HELPER(tsch
)(CPUS390XState
*env
, uint64_t r1
, uint64_t inst
)
597 S390CPU
*cpu
= s390_env_get_cpu(env
);
598 ioinst_handle_tsch(cpu
, r1
, inst
>> 16);
601 void HELPER(chsc
)(CPUS390XState
*env
, uint64_t inst
)
603 S390CPU
*cpu
= s390_env_get_cpu(env
);
604 ioinst_handle_chsc(cpu
, inst
>> 16);
608 #ifndef CONFIG_USER_ONLY
609 void HELPER(per_check_exception
)(CPUS390XState
*env
)
611 CPUState
*cs
= CPU(s390_env_get_cpu(env
));
613 if (env
->per_perc_atmid
) {
614 env
->int_pgm_code
= PGM_PER
;
615 env
->int_pgm_ilen
= get_ilen(cpu_ldub_code(env
, env
->per_address
));
617 cs
->exception_index
= EXCP_PGM
;
622 void HELPER(per_branch
)(CPUS390XState
*env
, uint64_t from
, uint64_t to
)
624 if ((env
->cregs
[9] & PER_CR9_EVENT_BRANCH
)) {
625 if (!(env
->cregs
[9] & PER_CR9_CONTROL_BRANCH_ADDRESS
)
626 || get_per_in_range(env
, to
)) {
627 env
->per_address
= from
;
628 env
->per_perc_atmid
= PER_CODE_EVENT_BRANCH
| get_per_atmid(env
);
633 void HELPER(per_ifetch
)(CPUS390XState
*env
, uint64_t addr
)
635 if ((env
->cregs
[9] & PER_CR9_EVENT_IFETCH
) && get_per_in_range(env
, addr
)) {
636 env
->per_address
= addr
;
637 env
->per_perc_atmid
= PER_CODE_EVENT_IFETCH
| get_per_atmid(env
);
639 /* If the instruction has to be nullified, trigger the
640 exception immediately. */
641 if (env
->cregs
[9] & PER_CR9_EVENT_NULLIFICATION
) {
642 CPUState
*cs
= CPU(s390_env_get_cpu(env
));
644 env
->int_pgm_code
= PGM_PER
;
645 env
->int_pgm_ilen
= get_ilen(cpu_ldub_code(env
, addr
));
647 cs
->exception_index
= EXCP_PGM
;