]> Git Repo - qemu.git/blob - cpu-defs.h
cpu: Move kvm_state field into CPUState
[qemu.git] / cpu-defs.h
1 /*
2  * common defines for all CPUs
3  *
4  * Copyright (c) 2003 Fabrice Bellard
5  *
6  * This library is free software; you can redistribute it and/or
7  * modify it under the terms of the GNU Lesser General Public
8  * License as published by the Free Software Foundation; either
9  * version 2 of the License, or (at your option) any later version.
10  *
11  * This library is distributed in the hope that it will be useful,
12  * but WITHOUT ANY WARRANTY; without even the implied warranty of
13  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
14  * Lesser General Public License for more details.
15  *
16  * You should have received a copy of the GNU Lesser General Public
17  * License along with this library; if not, see <http://www.gnu.org/licenses/>.
18  */
19 #ifndef CPU_DEFS_H
20 #define CPU_DEFS_H
21
22 #ifndef NEED_CPU_H
23 #error cpu.h included from common code
24 #endif
25
26 #include "config.h"
27 #include <setjmp.h>
28 #include <inttypes.h>
29 #include <signal.h>
30 #include "osdep.h"
31 #include "qemu-queue.h"
32 #include "hwaddr.h"
33
34 #ifndef TARGET_LONG_BITS
35 #error TARGET_LONG_BITS must be defined before including this header
36 #endif
37
38 #define TARGET_LONG_SIZE (TARGET_LONG_BITS / 8)
39
40 typedef int16_t target_short __attribute__ ((aligned(TARGET_SHORT_ALIGNMENT)));
41 typedef uint16_t target_ushort __attribute__((aligned(TARGET_SHORT_ALIGNMENT)));
42 typedef int32_t target_int __attribute__((aligned(TARGET_INT_ALIGNMENT)));
43 typedef uint32_t target_uint __attribute__((aligned(TARGET_INT_ALIGNMENT)));
44 typedef int64_t target_llong __attribute__((aligned(TARGET_LLONG_ALIGNMENT)));
45 typedef uint64_t target_ullong __attribute__((aligned(TARGET_LLONG_ALIGNMENT)));
46 /* target_ulong is the type of a virtual address */
47 #if TARGET_LONG_SIZE == 4
48 typedef int32_t target_long __attribute__((aligned(TARGET_LONG_ALIGNMENT)));
49 typedef uint32_t target_ulong __attribute__((aligned(TARGET_LONG_ALIGNMENT)));
50 #define TARGET_FMT_lx "%08x"
51 #define TARGET_FMT_ld "%d"
52 #define TARGET_FMT_lu "%u"
53 #elif TARGET_LONG_SIZE == 8
54 typedef int64_t target_long __attribute__((aligned(TARGET_LONG_ALIGNMENT)));
55 typedef uint64_t target_ulong __attribute__((aligned(TARGET_LONG_ALIGNMENT)));
56 #define TARGET_FMT_lx "%016" PRIx64
57 #define TARGET_FMT_ld "%" PRId64
58 #define TARGET_FMT_lu "%" PRIu64
59 #else
60 #error TARGET_LONG_SIZE undefined
61 #endif
62
63 #define EXCP_INTERRUPT  0x10000 /* async interruption */
64 #define EXCP_HLT        0x10001 /* hlt instruction reached */
65 #define EXCP_DEBUG      0x10002 /* cpu stopped after a breakpoint or singlestep */
66 #define EXCP_HALTED     0x10003 /* cpu is halted (waiting for external event) */
67
68 #define TB_JMP_CACHE_BITS 12
69 #define TB_JMP_CACHE_SIZE (1 << TB_JMP_CACHE_BITS)
70
71 /* Only the bottom TB_JMP_PAGE_BITS of the jump cache hash bits vary for
72    addresses on the same page.  The top bits are the same.  This allows
73    TLB invalidation to quickly clear a subset of the hash table.  */
74 #define TB_JMP_PAGE_BITS (TB_JMP_CACHE_BITS / 2)
75 #define TB_JMP_PAGE_SIZE (1 << TB_JMP_PAGE_BITS)
76 #define TB_JMP_ADDR_MASK (TB_JMP_PAGE_SIZE - 1)
77 #define TB_JMP_PAGE_MASK (TB_JMP_CACHE_SIZE - TB_JMP_PAGE_SIZE)
78
79 #if !defined(CONFIG_USER_ONLY)
80 #define CPU_TLB_BITS 8
81 #define CPU_TLB_SIZE (1 << CPU_TLB_BITS)
82
83 #if HOST_LONG_BITS == 32 && TARGET_LONG_BITS == 32
84 #define CPU_TLB_ENTRY_BITS 4
85 #else
86 #define CPU_TLB_ENTRY_BITS 5
87 #endif
88
89 typedef struct CPUTLBEntry {
90     /* bit TARGET_LONG_BITS to TARGET_PAGE_BITS : virtual address
91        bit TARGET_PAGE_BITS-1..4  : Nonzero for accesses that should not
92                                     go directly to ram.
93        bit 3                      : indicates that the entry is invalid
94        bit 2..0                   : zero
95     */
96     target_ulong addr_read;
97     target_ulong addr_write;
98     target_ulong addr_code;
99     /* Addend to virtual address to get host address.  IO accesses
100        use the corresponding iotlb value.  */
101     uintptr_t addend;
102     /* padding to get a power of two size */
103     uint8_t dummy[(1 << CPU_TLB_ENTRY_BITS) -
104                   (sizeof(target_ulong) * 3 +
105                    ((-sizeof(target_ulong) * 3) & (sizeof(uintptr_t) - 1)) +
106                    sizeof(uintptr_t))];
107 } CPUTLBEntry;
108
109 extern int CPUTLBEntry_wrong_size[sizeof(CPUTLBEntry) == (1 << CPU_TLB_ENTRY_BITS) ? 1 : -1];
110
111 #define CPU_COMMON_TLB \
112     /* The meaning of the MMU modes is defined in the target code. */   \
113     CPUTLBEntry tlb_table[NB_MMU_MODES][CPU_TLB_SIZE];                  \
114     hwaddr iotlb[NB_MMU_MODES][CPU_TLB_SIZE];               \
115     target_ulong tlb_flush_addr;                                        \
116     target_ulong tlb_flush_mask;
117
118 #else
119
120 #define CPU_COMMON_TLB
121
122 #endif
123
124
125 #ifdef HOST_WORDS_BIGENDIAN
126 typedef struct icount_decr_u16 {
127     uint16_t high;
128     uint16_t low;
129 } icount_decr_u16;
130 #else
131 typedef struct icount_decr_u16 {
132     uint16_t low;
133     uint16_t high;
134 } icount_decr_u16;
135 #endif
136
137 struct kvm_run;
138 struct qemu_work_item;
139
140 typedef struct CPUBreakpoint {
141     target_ulong pc;
142     int flags; /* BP_* */
143     QTAILQ_ENTRY(CPUBreakpoint) entry;
144 } CPUBreakpoint;
145
146 typedef struct CPUWatchpoint {
147     target_ulong vaddr;
148     target_ulong len_mask;
149     int flags; /* BP_* */
150     QTAILQ_ENTRY(CPUWatchpoint) entry;
151 } CPUWatchpoint;
152
153 #define CPU_TEMP_BUF_NLONGS 128
154 #define CPU_COMMON                                                      \
155     struct TranslationBlock *current_tb; /* currently executing TB  */  \
156     /* soft mmu support */                                              \
157     /* in order to avoid passing too many arguments to the MMIO         \
158        helpers, we store some rarely used information in the CPU        \
159        context) */                                                      \
160     uintptr_t mem_io_pc; /* host pc at which the memory was             \
161                             accessed */                                 \
162     target_ulong mem_io_vaddr; /* target virtual addr at which the      \
163                                      memory was accessed */             \
164     uint32_t halted; /* Nonzero if the CPU is in suspend state */       \
165     uint32_t interrupt_request;                                         \
166     volatile sig_atomic_t exit_request;                                 \
167     CPU_COMMON_TLB                                                      \
168     struct TranslationBlock *tb_jmp_cache[TB_JMP_CACHE_SIZE];           \
169     /* buffer for temporaries in the code generator */                  \
170     long temp_buf[CPU_TEMP_BUF_NLONGS];                                 \
171                                                                         \
172     int64_t icount_extra; /* Instructions until next timer event.  */   \
173     /* Number of cycles left, with interrupt flag in high bit.          \
174        This allows a single read-compare-cbranch-write sequence to test \
175        for both decrementer underflow and exceptions.  */               \
176     union {                                                             \
177         uint32_t u32;                                                   \
178         icount_decr_u16 u16;                                            \
179     } icount_decr;                                                      \
180     uint32_t can_do_io; /* nonzero if memory mapped IO is safe.  */     \
181                                                                         \
182     /* from this point: preserved by CPU reset */                       \
183     /* ice debug support */                                             \
184     QTAILQ_HEAD(breakpoints_head, CPUBreakpoint) breakpoints;            \
185     int singlestep_enabled;                                             \
186                                                                         \
187     QTAILQ_HEAD(watchpoints_head, CPUWatchpoint) watchpoints;            \
188     CPUWatchpoint *watchpoint_hit;                                      \
189                                                                         \
190     struct GDBRegisterState *gdb_regs;                                  \
191                                                                         \
192     /* Core interrupt code */                                           \
193     jmp_buf jmp_env;                                                    \
194     int exception_index;                                                \
195                                                                         \
196     CPUArchState *next_cpu; /* next CPU sharing TB cache */                 \
197     int cpu_index; /* CPU index (informative) */                        \
198     uint32_t host_tid; /* host thread ID */                             \
199     int numa_node; /* NUMA node this cpu is belonging to  */            \
200     int nr_cores;  /* number of cores within this CPU package */        \
201     int nr_threads;/* number of threads within this CPU */              \
202     int running; /* Nonzero if cpu is currently running(usermode).  */  \
203     /* user data */                                                     \
204     void *opaque;                                                       \
205                                                                         \
206     const char *cpu_model_str;                                          \
207     struct kvm_run *kvm_run;
208
209 #endif
This page took 0.036973 seconds and 4 git commands to generate.