2 * common defines for all CPUs
4 * Copyright (c) 2003 Fabrice Bellard
6 * This library is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU Lesser General Public
8 * License as published by the Free Software Foundation; either
9 * version 2 of the License, or (at your option) any later version.
11 * This library is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14 * Lesser General Public License for more details.
16 * You should have received a copy of the GNU Lesser General Public
17 * License along with this library; if not, write to the Free Software
18 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston MA 02110-1301 USA
24 #error cpu.h included from common code
33 #include "sys-queue.h"
36 #ifndef TARGET_LONG_BITS
37 #error TARGET_LONG_BITS must be defined before including this header
40 #define TARGET_LONG_SIZE (TARGET_LONG_BITS / 8)
42 /* target_ulong is the type of a virtual address */
43 #if TARGET_LONG_SIZE == 4
44 typedef int32_t target_long
;
45 typedef uint32_t target_ulong
;
46 #define TARGET_FMT_lx "%08x"
47 #define TARGET_FMT_ld "%d"
48 #define TARGET_FMT_lu "%u"
49 #elif TARGET_LONG_SIZE == 8
50 typedef int64_t target_long
;
51 typedef uint64_t target_ulong
;
52 #define TARGET_FMT_lx "%016" PRIx64
53 #define TARGET_FMT_ld "%" PRId64
54 #define TARGET_FMT_lu "%" PRIu64
56 #error TARGET_LONG_SIZE undefined
59 #define HOST_LONG_SIZE (HOST_LONG_BITS / 8)
61 #define EXCP_INTERRUPT 0x10000 /* async interruption */
62 #define EXCP_HLT 0x10001 /* hlt instruction reached */
63 #define EXCP_DEBUG 0x10002 /* cpu stopped after a breakpoint or singlestep */
64 #define EXCP_HALTED 0x10003 /* cpu is halted (waiting for external event) */
66 #define TB_JMP_CACHE_BITS 12
67 #define TB_JMP_CACHE_SIZE (1 << TB_JMP_CACHE_BITS)
69 /* Only the bottom TB_JMP_PAGE_BITS of the jump cache hash bits vary for
70 addresses on the same page. The top bits are the same. This allows
71 TLB invalidation to quickly clear a subset of the hash table. */
72 #define TB_JMP_PAGE_BITS (TB_JMP_CACHE_BITS / 2)
73 #define TB_JMP_PAGE_SIZE (1 << TB_JMP_PAGE_BITS)
74 #define TB_JMP_ADDR_MASK (TB_JMP_PAGE_SIZE - 1)
75 #define TB_JMP_PAGE_MASK (TB_JMP_CACHE_SIZE - TB_JMP_PAGE_SIZE)
77 #define CPU_TLB_BITS 8
78 #define CPU_TLB_SIZE (1 << CPU_TLB_BITS)
80 #if TARGET_PHYS_ADDR_BITS == 32 && TARGET_LONG_BITS == 32
81 #define CPU_TLB_ENTRY_BITS 4
83 #define CPU_TLB_ENTRY_BITS 5
86 typedef struct CPUTLBEntry
{
87 /* bit TARGET_LONG_BITS to TARGET_PAGE_BITS : virtual address
88 bit TARGET_PAGE_BITS-1..4 : Nonzero for accesses that should not
90 bit 3 : indicates that the entry is invalid
93 target_ulong addr_read
;
94 target_ulong addr_write
;
95 target_ulong addr_code
;
96 /* Addend to virtual address to get physical address. IO accesses
97 use the corresponding iotlb value. */
98 #if TARGET_PHYS_ADDR_BITS == 64
99 /* on i386 Linux make sure it is aligned */
100 target_phys_addr_t addend
__attribute__((aligned(8)));
102 target_phys_addr_t addend
;
104 /* padding to get a power of two size */
105 uint8_t dummy
[(1 << CPU_TLB_ENTRY_BITS
) -
106 (sizeof(target_ulong
) * 3 +
107 ((-sizeof(target_ulong
) * 3) & (sizeof(target_phys_addr_t
) - 1)) +
108 sizeof(target_phys_addr_t
))];
111 #ifdef WORDS_BIGENDIAN
112 typedef struct icount_decr_u16
{
117 typedef struct icount_decr_u16
{
126 typedef struct CPUBreakpoint
{
128 int flags
; /* BP_* */
129 TAILQ_ENTRY(CPUBreakpoint
) entry
;
132 typedef struct CPUWatchpoint
{
134 target_ulong len_mask
;
135 int flags
; /* BP_* */
136 TAILQ_ENTRY(CPUWatchpoint
) entry
;
139 /* forward decleration */
140 struct qemu_work_item
;
150 struct qemu_work_item
*queued_work_first
, *queued_work_last
;
153 #define CPU_TEMP_BUF_NLONGS 128
155 struct TranslationBlock *current_tb; /* currently executing TB */ \
156 /* soft mmu support */ \
157 /* in order to avoid passing too many arguments to the MMIO \
158 helpers, we store some rarely used information in the CPU \
160 unsigned long mem_io_pc; /* host pc at which the memory was \
162 target_ulong mem_io_vaddr; /* target virtual addr at which the \
163 memory was accessed */ \
164 uint32_t halted; /* Nonzero if the CPU is in suspend state */ \
165 uint32_t stop; /* Stop request */ \
166 uint32_t stopped; /* Artificially stopped */ \
167 uint32_t interrupt_request; \
168 volatile sig_atomic_t exit_request; \
169 /* The meaning of the MMU modes is defined in the target code. */ \
170 CPUTLBEntry tlb_table[NB_MMU_MODES][CPU_TLB_SIZE]; \
171 target_phys_addr_t iotlb[NB_MMU_MODES][CPU_TLB_SIZE]; \
172 struct TranslationBlock *tb_jmp_cache[TB_JMP_CACHE_SIZE]; \
173 /* buffer for temporaries in the code generator */ \
174 long temp_buf[CPU_TEMP_BUF_NLONGS]; \
176 int64_t icount_extra; /* Instructions until next timer event. */ \
177 /* Number of cycles left, with interrupt flag in high bit. \
178 This allows a single read-compare-cbranch-write sequence to test \
179 for both decrementer underflow and exceptions. */ \
182 icount_decr_u16 u16; \
184 uint32_t can_do_io; /* nonzero if memory mapped IO is safe. */ \
186 /* from this point: preserved by CPU reset */ \
187 /* ice debug support */ \
188 TAILQ_HEAD(breakpoints_head, CPUBreakpoint) breakpoints; \
189 int singlestep_enabled; \
191 TAILQ_HEAD(watchpoints_head, CPUWatchpoint) watchpoints; \
192 CPUWatchpoint *watchpoint_hit; \
194 struct GDBRegisterState *gdb_regs; \
196 /* Core interrupt code */ \
198 int exception_index; \
200 CPUState *next_cpu; /* next CPU sharing TB cache */ \
201 int cpu_index; /* CPU index (informative) */ \
202 int numa_node; /* NUMA node this cpu is belonging to */ \
203 int running; /* Nonzero if cpu is currently running(usermode). */ \
209 struct QemuThread *thread; \
210 struct QemuCond *halt_cond; \
211 const char *cpu_model_str; \
212 struct KVMState *kvm_state; \
213 struct kvm_run *kvm_run; \
215 struct KVMCPUState kvm_cpu_state;