4 * Copyright (c) 2009 Ulrich Hecht
5 * Copyright (c) 2011 Alexander Graf
7 * This library is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU Lesser General Public
9 * License as published by the Free Software Foundation; either
10 * version 2 of the License, or (at your option) any later version.
12 * This library is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15 * Lesser General Public License for more details.
17 * You should have received a copy of the GNU Lesser General Public
18 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
22 #include "exec/gdbstub.h"
23 #include "qemu/timer.h"
24 #include "exec/cpu_ldst.h"
25 #ifndef CONFIG_USER_ONLY
26 #include "sysemu/sysemu.h"
30 //#define DEBUG_S390_STDOUT
33 #ifdef DEBUG_S390_STDOUT
34 #define DPRINTF(fmt, ...) \
35 do { fprintf(stderr, fmt, ## __VA_ARGS__); \
36 qemu_log(fmt, ##__VA_ARGS__); } while (0)
38 #define DPRINTF(fmt, ...) \
39 do { qemu_log(fmt, ## __VA_ARGS__); } while (0)
42 #define DPRINTF(fmt, ...) \
47 #ifndef CONFIG_USER_ONLY
48 void s390x_tod_timer(void *opaque
)
50 S390CPU
*cpu
= opaque
;
51 CPUS390XState
*env
= &cpu
->env
;
53 env
->pending_int
|= INTERRUPT_TOD
;
54 cpu_interrupt(CPU(cpu
), CPU_INTERRUPT_HARD
);
57 void s390x_cpu_timer(void *opaque
)
59 S390CPU
*cpu
= opaque
;
60 CPUS390XState
*env
= &cpu
->env
;
62 env
->pending_int
|= INTERRUPT_CPUTIMER
;
63 cpu_interrupt(CPU(cpu
), CPU_INTERRUPT_HARD
);
67 S390CPU
*cpu_s390x_init(const char *cpu_model
)
71 cpu
= S390_CPU(object_new(TYPE_S390_CPU
));
73 object_property_set_bool(OBJECT(cpu
), true, "realized", NULL
);
78 #if defined(CONFIG_USER_ONLY)
80 void s390_cpu_do_interrupt(CPUState
*cs
)
82 cs
->exception_index
= -1;
85 int s390_cpu_handle_mmu_fault(CPUState
*cs
, vaddr address
,
88 S390CPU
*cpu
= S390_CPU(cs
);
90 cs
->exception_index
= EXCP_PGM
;
91 cpu
->env
.int_pgm_code
= PGM_ADDRESSING
;
92 /* On real machines this value is dropped into LowMem. Since this
93 is userland, simply put this someplace that cpu_loop can find it. */
94 cpu
->env
.__excp_addr
= address
;
98 #else /* !CONFIG_USER_ONLY */
100 /* Ensure to exit the TB after this call! */
101 void trigger_pgm_exception(CPUS390XState
*env
, uint32_t code
, uint32_t ilen
)
103 CPUState
*cs
= CPU(s390_env_get_cpu(env
));
105 cs
->exception_index
= EXCP_PGM
;
106 env
->int_pgm_code
= code
;
107 env
->int_pgm_ilen
= ilen
;
110 int s390_cpu_handle_mmu_fault(CPUState
*cs
, vaddr orig_vaddr
,
113 S390CPU
*cpu
= S390_CPU(cs
);
114 CPUS390XState
*env
= &cpu
->env
;
115 uint64_t asc
= cpu_mmu_idx_to_asc(mmu_idx
);
116 target_ulong vaddr
, raddr
;
119 DPRINTF("%s: address 0x%" VADDR_PRIx
" rw %d mmu_idx %d\n",
120 __func__
, orig_vaddr
, rw
, mmu_idx
);
122 orig_vaddr
&= TARGET_PAGE_MASK
;
126 if (!(env
->psw
.mask
& PSW_MASK_64
)) {
130 if (mmu_translate(env
, vaddr
, rw
, asc
, &raddr
, &prot
, true)) {
131 /* Translation ended in exception */
135 /* check out of RAM access */
136 if (raddr
> (ram_size
+ virtio_size
)) {
137 DPRINTF("%s: raddr %" PRIx64
" > ram_size %" PRIx64
"\n", __func__
,
138 (uint64_t)raddr
, (uint64_t)ram_size
);
139 trigger_pgm_exception(env
, PGM_ADDRESSING
, ILEN_LATER
);
143 qemu_log_mask(CPU_LOG_MMU
, "%s: set tlb %" PRIx64
" -> %" PRIx64
" (%x)\n",
144 __func__
, (uint64_t)vaddr
, (uint64_t)raddr
, prot
);
146 tlb_set_page(cs
, orig_vaddr
, raddr
, prot
,
147 mmu_idx
, TARGET_PAGE_SIZE
);
152 hwaddr
s390_cpu_get_phys_page_debug(CPUState
*cs
, vaddr vaddr
)
154 S390CPU
*cpu
= S390_CPU(cs
);
155 CPUS390XState
*env
= &cpu
->env
;
158 uint64_t asc
= env
->psw
.mask
& PSW_MASK_ASC
;
161 if (!(env
->psw
.mask
& PSW_MASK_64
)) {
165 mmu_translate(env
, vaddr
, MMU_INST_FETCH
, asc
, &raddr
, &prot
, false);
170 hwaddr
s390_cpu_get_phys_addr_debug(CPUState
*cs
, vaddr vaddr
)
175 page
= vaddr
& TARGET_PAGE_MASK
;
176 phys_addr
= cpu_get_phys_page_debug(cs
, page
);
177 phys_addr
+= (vaddr
& ~TARGET_PAGE_MASK
);
182 void load_psw(CPUS390XState
*env
, uint64_t mask
, uint64_t addr
)
184 uint64_t old_mask
= env
->psw
.mask
;
186 env
->psw
.addr
= addr
;
187 env
->psw
.mask
= mask
;
189 env
->cc_op
= (mask
>> 44) & 3;
192 if ((old_mask
^ mask
) & PSW_MASK_PER
) {
193 s390_cpu_recompute_watchpoints(CPU(s390_env_get_cpu(env
)));
196 if (mask
& PSW_MASK_WAIT
) {
197 S390CPU
*cpu
= s390_env_get_cpu(env
);
198 if (s390_cpu_halt(cpu
) == 0) {
199 #ifndef CONFIG_USER_ONLY
200 qemu_system_shutdown_request();
206 static uint64_t get_psw_mask(CPUS390XState
*env
)
208 uint64_t r
= env
->psw
.mask
;
211 env
->cc_op
= calc_cc(env
, env
->cc_op
, env
->cc_src
, env
->cc_dst
,
215 assert(!(env
->cc_op
& ~3));
216 r
|= (uint64_t)env
->cc_op
<< 44;
222 static LowCore
*cpu_map_lowcore(CPUS390XState
*env
)
224 S390CPU
*cpu
= s390_env_get_cpu(env
);
226 hwaddr len
= sizeof(LowCore
);
228 lowcore
= cpu_physical_memory_map(env
->psa
, &len
, 1);
230 if (len
< sizeof(LowCore
)) {
231 cpu_abort(CPU(cpu
), "Could not map lowcore\n");
237 static void cpu_unmap_lowcore(LowCore
*lowcore
)
239 cpu_physical_memory_unmap(lowcore
, sizeof(LowCore
), 1, sizeof(LowCore
));
242 void do_restart_interrupt(CPUS390XState
*env
)
247 lowcore
= cpu_map_lowcore(env
);
249 lowcore
->restart_old_psw
.mask
= cpu_to_be64(get_psw_mask(env
));
250 lowcore
->restart_old_psw
.addr
= cpu_to_be64(env
->psw
.addr
);
251 mask
= be64_to_cpu(lowcore
->restart_new_psw
.mask
);
252 addr
= be64_to_cpu(lowcore
->restart_new_psw
.addr
);
254 cpu_unmap_lowcore(lowcore
);
256 load_psw(env
, mask
, addr
);
259 static void do_program_interrupt(CPUS390XState
*env
)
263 int ilen
= env
->int_pgm_ilen
;
267 ilen
= get_ilen(cpu_ldub_code(env
, env
->psw
.addr
));
270 ilen
= get_ilen(cpu_ldub_code(env
, env
->psw
.addr
));
271 env
->psw
.addr
+= ilen
;
274 assert(ilen
== 2 || ilen
== 4 || ilen
== 6);
277 qemu_log_mask(CPU_LOG_INT
, "%s: code=0x%x ilen=%d\n",
278 __func__
, env
->int_pgm_code
, ilen
);
280 lowcore
= cpu_map_lowcore(env
);
282 /* Signal PER events with the exception. */
283 if (env
->per_perc_atmid
) {
284 env
->int_pgm_code
|= PGM_PER
;
285 lowcore
->per_address
= cpu_to_be64(env
->per_address
);
286 lowcore
->per_perc_atmid
= cpu_to_be16(env
->per_perc_atmid
);
287 env
->per_perc_atmid
= 0;
290 lowcore
->pgm_ilen
= cpu_to_be16(ilen
);
291 lowcore
->pgm_code
= cpu_to_be16(env
->int_pgm_code
);
292 lowcore
->program_old_psw
.mask
= cpu_to_be64(get_psw_mask(env
));
293 lowcore
->program_old_psw
.addr
= cpu_to_be64(env
->psw
.addr
);
294 mask
= be64_to_cpu(lowcore
->program_new_psw
.mask
);
295 addr
= be64_to_cpu(lowcore
->program_new_psw
.addr
);
296 lowcore
->per_breaking_event_addr
= cpu_to_be64(env
->gbea
);
298 cpu_unmap_lowcore(lowcore
);
300 DPRINTF("%s: %x %x %" PRIx64
" %" PRIx64
"\n", __func__
,
301 env
->int_pgm_code
, ilen
, env
->psw
.mask
,
304 load_psw(env
, mask
, addr
);
307 static void do_svc_interrupt(CPUS390XState
*env
)
312 lowcore
= cpu_map_lowcore(env
);
314 lowcore
->svc_code
= cpu_to_be16(env
->int_svc_code
);
315 lowcore
->svc_ilen
= cpu_to_be16(env
->int_svc_ilen
);
316 lowcore
->svc_old_psw
.mask
= cpu_to_be64(get_psw_mask(env
));
317 lowcore
->svc_old_psw
.addr
= cpu_to_be64(env
->psw
.addr
+ env
->int_svc_ilen
);
318 mask
= be64_to_cpu(lowcore
->svc_new_psw
.mask
);
319 addr
= be64_to_cpu(lowcore
->svc_new_psw
.addr
);
321 cpu_unmap_lowcore(lowcore
);
323 load_psw(env
, mask
, addr
);
325 /* When a PER event is pending, the PER exception has to happen
326 immediately after the SERVICE CALL one. */
327 if (env
->per_perc_atmid
) {
328 env
->int_pgm_code
= PGM_PER
;
329 env
->int_pgm_ilen
= env
->int_svc_ilen
;
330 do_program_interrupt(env
);
334 #define VIRTIO_SUBCODE_64 0x0D00
336 static void do_ext_interrupt(CPUS390XState
*env
)
338 S390CPU
*cpu
= s390_env_get_cpu(env
);
343 if (!(env
->psw
.mask
& PSW_MASK_EXT
)) {
344 cpu_abort(CPU(cpu
), "Ext int w/o ext mask\n");
347 if (env
->ext_index
< 0 || env
->ext_index
>= MAX_EXT_QUEUE
) {
348 cpu_abort(CPU(cpu
), "Ext queue overrun: %d\n", env
->ext_index
);
351 q
= &env
->ext_queue
[env
->ext_index
];
352 lowcore
= cpu_map_lowcore(env
);
354 lowcore
->ext_int_code
= cpu_to_be16(q
->code
);
355 lowcore
->ext_params
= cpu_to_be32(q
->param
);
356 lowcore
->ext_params2
= cpu_to_be64(q
->param64
);
357 lowcore
->external_old_psw
.mask
= cpu_to_be64(get_psw_mask(env
));
358 lowcore
->external_old_psw
.addr
= cpu_to_be64(env
->psw
.addr
);
359 lowcore
->cpu_addr
= cpu_to_be16(env
->cpu_num
| VIRTIO_SUBCODE_64
);
360 mask
= be64_to_cpu(lowcore
->external_new_psw
.mask
);
361 addr
= be64_to_cpu(lowcore
->external_new_psw
.addr
);
363 cpu_unmap_lowcore(lowcore
);
366 if (env
->ext_index
== -1) {
367 env
->pending_int
&= ~INTERRUPT_EXT
;
370 DPRINTF("%s: %" PRIx64
" %" PRIx64
"\n", __func__
,
371 env
->psw
.mask
, env
->psw
.addr
);
373 load_psw(env
, mask
, addr
);
376 static void do_io_interrupt(CPUS390XState
*env
)
378 S390CPU
*cpu
= s390_env_get_cpu(env
);
385 if (!(env
->psw
.mask
& PSW_MASK_IO
)) {
386 cpu_abort(CPU(cpu
), "I/O int w/o I/O mask\n");
389 for (isc
= 0; isc
< ARRAY_SIZE(env
->io_index
); isc
++) {
392 if (env
->io_index
[isc
] < 0) {
395 if (env
->io_index
[isc
] >= MAX_IO_QUEUE
) {
396 cpu_abort(CPU(cpu
), "I/O queue overrun for isc %d: %d\n",
397 isc
, env
->io_index
[isc
]);
400 q
= &env
->io_queue
[env
->io_index
[isc
]][isc
];
401 isc_bits
= ISC_TO_ISC_BITS(IO_INT_WORD_ISC(q
->word
));
402 if (!(env
->cregs
[6] & isc_bits
)) {
410 lowcore
= cpu_map_lowcore(env
);
412 lowcore
->subchannel_id
= cpu_to_be16(q
->id
);
413 lowcore
->subchannel_nr
= cpu_to_be16(q
->nr
);
414 lowcore
->io_int_parm
= cpu_to_be32(q
->parm
);
415 lowcore
->io_int_word
= cpu_to_be32(q
->word
);
416 lowcore
->io_old_psw
.mask
= cpu_to_be64(get_psw_mask(env
));
417 lowcore
->io_old_psw
.addr
= cpu_to_be64(env
->psw
.addr
);
418 mask
= be64_to_cpu(lowcore
->io_new_psw
.mask
);
419 addr
= be64_to_cpu(lowcore
->io_new_psw
.addr
);
421 cpu_unmap_lowcore(lowcore
);
423 env
->io_index
[isc
]--;
425 DPRINTF("%s: %" PRIx64
" %" PRIx64
"\n", __func__
,
426 env
->psw
.mask
, env
->psw
.addr
);
427 load_psw(env
, mask
, addr
);
429 if (env
->io_index
[isc
] >= 0) {
436 env
->pending_int
&= ~INTERRUPT_IO
;
441 static void do_mchk_interrupt(CPUS390XState
*env
)
443 S390CPU
*cpu
= s390_env_get_cpu(env
);
449 if (!(env
->psw
.mask
& PSW_MASK_MCHECK
)) {
450 cpu_abort(CPU(cpu
), "Machine check w/o mchk mask\n");
453 if (env
->mchk_index
< 0 || env
->mchk_index
>= MAX_MCHK_QUEUE
) {
454 cpu_abort(CPU(cpu
), "Mchk queue overrun: %d\n", env
->mchk_index
);
457 q
= &env
->mchk_queue
[env
->mchk_index
];
460 /* Don't know how to handle this... */
461 cpu_abort(CPU(cpu
), "Unknown machine check type %d\n", q
->type
);
463 if (!(env
->cregs
[14] & (1 << 28))) {
464 /* CRW machine checks disabled */
468 lowcore
= cpu_map_lowcore(env
);
470 for (i
= 0; i
< 16; i
++) {
471 lowcore
->floating_pt_save_area
[i
] = cpu_to_be64(get_freg(env
, i
)->ll
);
472 lowcore
->gpregs_save_area
[i
] = cpu_to_be64(env
->regs
[i
]);
473 lowcore
->access_regs_save_area
[i
] = cpu_to_be32(env
->aregs
[i
]);
474 lowcore
->cregs_save_area
[i
] = cpu_to_be64(env
->cregs
[i
]);
476 lowcore
->prefixreg_save_area
= cpu_to_be32(env
->psa
);
477 lowcore
->fpt_creg_save_area
= cpu_to_be32(env
->fpc
);
478 lowcore
->tod_progreg_save_area
= cpu_to_be32(env
->todpr
);
479 lowcore
->cpu_timer_save_area
[0] = cpu_to_be32(env
->cputm
>> 32);
480 lowcore
->cpu_timer_save_area
[1] = cpu_to_be32((uint32_t)env
->cputm
);
481 lowcore
->clock_comp_save_area
[0] = cpu_to_be32(env
->ckc
>> 32);
482 lowcore
->clock_comp_save_area
[1] = cpu_to_be32((uint32_t)env
->ckc
);
484 lowcore
->mcck_interruption_code
[0] = cpu_to_be32(0x00400f1d);
485 lowcore
->mcck_interruption_code
[1] = cpu_to_be32(0x40330000);
486 lowcore
->mcck_old_psw
.mask
= cpu_to_be64(get_psw_mask(env
));
487 lowcore
->mcck_old_psw
.addr
= cpu_to_be64(env
->psw
.addr
);
488 mask
= be64_to_cpu(lowcore
->mcck_new_psw
.mask
);
489 addr
= be64_to_cpu(lowcore
->mcck_new_psw
.addr
);
491 cpu_unmap_lowcore(lowcore
);
494 if (env
->mchk_index
== -1) {
495 env
->pending_int
&= ~INTERRUPT_MCHK
;
498 DPRINTF("%s: %" PRIx64
" %" PRIx64
"\n", __func__
,
499 env
->psw
.mask
, env
->psw
.addr
);
501 load_psw(env
, mask
, addr
);
504 void s390_cpu_do_interrupt(CPUState
*cs
)
506 S390CPU
*cpu
= S390_CPU(cs
);
507 CPUS390XState
*env
= &cpu
->env
;
509 qemu_log_mask(CPU_LOG_INT
, "%s: %d at pc=%" PRIx64
"\n",
510 __func__
, cs
->exception_index
, env
->psw
.addr
);
512 s390_cpu_set_state(CPU_STATE_OPERATING
, cpu
);
513 /* handle machine checks */
514 if ((env
->psw
.mask
& PSW_MASK_MCHECK
) &&
515 (cs
->exception_index
== -1)) {
516 if (env
->pending_int
& INTERRUPT_MCHK
) {
517 cs
->exception_index
= EXCP_MCHK
;
520 /* handle external interrupts */
521 if ((env
->psw
.mask
& PSW_MASK_EXT
) &&
522 cs
->exception_index
== -1) {
523 if (env
->pending_int
& INTERRUPT_EXT
) {
524 /* code is already in env */
525 cs
->exception_index
= EXCP_EXT
;
526 } else if (env
->pending_int
& INTERRUPT_TOD
) {
527 cpu_inject_ext(cpu
, 0x1004, 0, 0);
528 cs
->exception_index
= EXCP_EXT
;
529 env
->pending_int
&= ~INTERRUPT_EXT
;
530 env
->pending_int
&= ~INTERRUPT_TOD
;
531 } else if (env
->pending_int
& INTERRUPT_CPUTIMER
) {
532 cpu_inject_ext(cpu
, 0x1005, 0, 0);
533 cs
->exception_index
= EXCP_EXT
;
534 env
->pending_int
&= ~INTERRUPT_EXT
;
535 env
->pending_int
&= ~INTERRUPT_TOD
;
538 /* handle I/O interrupts */
539 if ((env
->psw
.mask
& PSW_MASK_IO
) &&
540 (cs
->exception_index
== -1)) {
541 if (env
->pending_int
& INTERRUPT_IO
) {
542 cs
->exception_index
= EXCP_IO
;
546 switch (cs
->exception_index
) {
548 do_program_interrupt(env
);
551 do_svc_interrupt(env
);
554 do_ext_interrupt(env
);
557 do_io_interrupt(env
);
560 do_mchk_interrupt(env
);
563 cs
->exception_index
= -1;
565 if (!env
->pending_int
) {
566 cs
->interrupt_request
&= ~CPU_INTERRUPT_HARD
;
570 bool s390_cpu_exec_interrupt(CPUState
*cs
, int interrupt_request
)
572 if (interrupt_request
& CPU_INTERRUPT_HARD
) {
573 S390CPU
*cpu
= S390_CPU(cs
);
574 CPUS390XState
*env
= &cpu
->env
;
576 if (env
->psw
.mask
& PSW_MASK_EXT
) {
577 s390_cpu_do_interrupt(cs
);
584 void s390_cpu_recompute_watchpoints(CPUState
*cs
)
586 const int wp_flags
= BP_CPU
| BP_MEM_WRITE
| BP_STOP_BEFORE_ACCESS
;
587 S390CPU
*cpu
= S390_CPU(cs
);
588 CPUS390XState
*env
= &cpu
->env
;
590 /* We are called when the watchpoints have changed. First
592 cpu_watchpoint_remove_all(cs
, BP_CPU
);
594 /* Return if PER is not enabled */
595 if (!(env
->psw
.mask
& PSW_MASK_PER
)) {
599 /* Return if storage-alteration event is not enabled. */
600 if (!(env
->cregs
[9] & PER_CR9_EVENT_STORE
)) {
604 if (env
->cregs
[10] == 0 && env
->cregs
[11] == -1LL) {
605 /* We can't create a watchoint spanning the whole memory range, so
606 split it in two parts. */
607 cpu_watchpoint_insert(cs
, 0, 1ULL << 63, wp_flags
, NULL
);
608 cpu_watchpoint_insert(cs
, 1ULL << 63, 1ULL << 63, wp_flags
, NULL
);
609 } else if (env
->cregs
[10] > env
->cregs
[11]) {
610 /* The address range loops, create two watchpoints. */
611 cpu_watchpoint_insert(cs
, env
->cregs
[10], -env
->cregs
[10],
613 cpu_watchpoint_insert(cs
, 0, env
->cregs
[11] + 1, wp_flags
, NULL
);
616 /* Default case, create a single watchpoint. */
617 cpu_watchpoint_insert(cs
, env
->cregs
[10],
618 env
->cregs
[11] - env
->cregs
[10] + 1,
623 void s390x_cpu_debug_excp_handler(CPUState
*cs
)
625 S390CPU
*cpu
= S390_CPU(cs
);
626 CPUS390XState
*env
= &cpu
->env
;
627 CPUWatchpoint
*wp_hit
= cs
->watchpoint_hit
;
629 if (wp_hit
&& wp_hit
->flags
& BP_CPU
) {
630 /* FIXME: When the storage-alteration-space control bit is set,
631 the exception should only be triggered if the memory access
632 is done using an address space with the storage-alteration-event
633 bit set. We have no way to detect that with the current
635 cs
->watchpoint_hit
= NULL
;
637 env
->per_address
= env
->psw
.addr
;
638 env
->per_perc_atmid
|= PER_CODE_EVENT_STORE
| get_per_atmid(env
);
639 /* FIXME: We currently no way to detect the address space used
640 to trigger the watchpoint. For now just consider it is the
641 current default ASC. This turn to be true except when MVCP
642 and MVCS instrutions are not used. */
643 env
->per_perc_atmid
|= env
->psw
.mask
& (PSW_MASK_ASC
) >> 46;
645 /* Remove all watchpoints to re-execute the code. A PER exception
646 will be triggered, it will call load_psw which will recompute
648 cpu_watchpoint_remove_all(cs
, BP_CPU
);
649 cpu_resume_from_signal(cs
, NULL
);
652 #endif /* CONFIG_USER_ONLY */