2 * i386 helpers (without register variable usage)
4 * Copyright (c) 2003 Fabrice Bellard
6 * This library is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU Lesser General Public
8 * License as published by the Free Software Foundation; either
9 * version 2 of the License, or (at your option) any later version.
11 * This library is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14 * Lesser General Public License for more details.
16 * You should have received a copy of the GNU Lesser General Public
17 * License along with this library; if not, write to the Free Software
18 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
34 CPUX86State *cpu_x86_init(void)
42 env = malloc(sizeof(CPUX86State));
45 memset(env, 0, sizeof(CPUX86State));
50 /* flags setup : we activate the IRQs by default as in user mode */
51 env->eflags = 0x2 | IF_MASK;
55 env->hflags |= HF_SOFTMMU_MASK;
57 /* init various static tables */
60 optimize_flags_init();
65 void cpu_x86_close(CPUX86State *env)
70 /***********************************************************/
73 static const char *cc_op_str[] = {
108 void cpu_x86_dump_state(CPUX86State *env, FILE *f, int flags)
112 static const char *seg_name[6] = { "ES", "CS", "SS", "DS", "FS", "GS" };
114 eflags = env->eflags;
115 fprintf(f, "EAX=%08x EBX=%08x ECX=%08x EDX=%08x\n"
116 "ESI=%08x EDI=%08x EBP=%08x ESP=%08x\n"
117 "EIP=%08x EFL=%08x [%c%c%c%c%c%c%c] CPL=%d II=%d\n",
118 env->regs[R_EAX], env->regs[R_EBX], env->regs[R_ECX], env->regs[R_EDX],
119 env->regs[R_ESI], env->regs[R_EDI], env->regs[R_EBP], env->regs[R_ESP],
121 eflags & DF_MASK ? 'D' : '-',
122 eflags & CC_O ? 'O' : '-',
123 eflags & CC_S ? 'S' : '-',
124 eflags & CC_Z ? 'Z' : '-',
125 eflags & CC_A ? 'A' : '-',
126 eflags & CC_P ? 'P' : '-',
127 eflags & CC_C ? 'C' : '-',
128 env->hflags & HF_CPL_MASK,
129 (env->hflags >> HF_INHIBIT_IRQ_SHIFT) & 1);
130 for(i = 0; i < 6; i++) {
131 SegmentCache *sc = &env->segs[i];
132 fprintf(f, "%s =%04x %08x %08x %08x\n",
139 fprintf(f, "LDT=%04x %08x %08x %08x\n",
144 fprintf(f, "TR =%04x %08x %08x %08x\n",
149 fprintf(f, "GDT= %08x %08x\n",
150 (int)env->gdt.base, env->gdt.limit);
151 fprintf(f, "IDT= %08x %08x\n",
152 (int)env->idt.base, env->idt.limit);
153 fprintf(f, "CR0=%08x CR2=%08x CR3=%08x CR4=%08x\n",
154 env->cr[0], env->cr[2], env->cr[3], env->cr[4]);
156 if (flags & X86_DUMP_CCOP) {
157 if ((unsigned)env->cc_op < CC_OP_NB)
158 strcpy(cc_op_name, cc_op_str[env->cc_op]);
160 snprintf(cc_op_name, sizeof(cc_op_name), "[%d]", env->cc_op);
161 fprintf(f, "CCS=%08x CCD=%08x CCO=%-8s\n",
162 env->cc_src, env->cc_dst, cc_op_name);
164 if (flags & X86_DUMP_FPU) {
165 fprintf(f, "ST0=%f ST1=%f ST2=%f ST3=%f\n",
166 (double)env->fpregs[0],
167 (double)env->fpregs[1],
168 (double)env->fpregs[2],
169 (double)env->fpregs[3]);
170 fprintf(f, "ST4=%f ST5=%f ST6=%f ST7=%f\n",
171 (double)env->fpregs[4],
172 (double)env->fpregs[5],
173 (double)env->fpregs[7],
174 (double)env->fpregs[8]);
178 /***********************************************************/
180 /* XXX: add PGE support */
182 /* called when cr3 or PG bit are modified */
183 static int last_pg_state = -1;
184 static uint32_t a20_mask;
187 void cpu_x86_set_a20(CPUX86State *env, int a20_state)
189 a20_state = (a20_state != 0);
190 if (a20_state != a20_enabled) {
191 #if defined(DEBUG_MMU)
192 printf("A20 update: a20=%d\n", a20_state);
194 /* if the cpu is currently executing code, we must unlink it and
195 all the potentially executing TB */
196 cpu_interrupt(env, 0);
198 /* when a20 is changed, all the MMU mappings are invalid, so
199 we must flush everything */
201 a20_enabled = a20_state;
203 a20_mask = 0xffffffff;
205 a20_mask = 0xffefffff;
209 void cpu_x86_update_cr0(CPUX86State *env)
211 int pg_state, pe_state;
213 #if defined(DEBUG_MMU)
214 printf("CR0 update: CR0=0x%08x\n", env->cr[0]);
216 pg_state = env->cr[0] & CR0_PG_MASK;
217 if (pg_state != last_pg_state) {
219 last_pg_state = pg_state;
221 /* update PE flag in hidden flags */
222 pe_state = (env->cr[0] & CR0_PE_MASK);
223 env->hflags = (env->hflags & ~HF_PE_MASK) | (pe_state << HF_PE_SHIFT);
224 /* ensure that ADDSEG is always set in real mode */
225 env->hflags |= ((pe_state ^ 1) << HF_ADDSEG_SHIFT);
228 void cpu_x86_update_cr3(CPUX86State *env)
230 if (env->cr[0] & CR0_PG_MASK) {
231 #if defined(DEBUG_MMU)
232 printf("CR3 update: CR3=%08x\n", env->cr[3]);
238 void cpu_x86_init_mmu(CPUX86State *env)
241 a20_mask = 0xffffffff;
244 cpu_x86_update_cr0(env);
247 /* XXX: also flush 4MB pages */
248 void cpu_x86_flush_tlb(CPUX86State *env, uint32_t addr)
250 tlb_flush_page(env, addr);
254 -1 = cannot handle fault
255 0 = nothing more to do
256 1 = generate PF fault
257 2 = soft MMU activation required for this block
259 int cpu_x86_handle_mmu_fault(CPUX86State *env, uint32_t addr,
260 int is_write, int is_user, int is_softmmu)
262 uint8_t *pde_ptr, *pte_ptr;
263 uint32_t pde, pte, virt_addr;
264 int error_code, is_dirty, prot, page_size, ret;
265 unsigned long paddr, vaddr, page_offset;
267 #if defined(DEBUG_MMU)
268 printf("MMU fault: addr=0x%08x w=%d u=%d eip=%08x\n",
269 addr, is_write, is_user, env->eip);
272 if (env->user_mode_only) {
273 /* user mode only emulation */
278 if (!(env->cr[0] & CR0_PG_MASK)) {
280 virt_addr = addr & TARGET_PAGE_MASK;
281 prot = PROT_READ | PROT_WRITE;
286 /* page directory entry */
287 pde_ptr = phys_ram_base +
288 (((env->cr[3] & ~0xfff) + ((addr >> 20) & ~3)) & a20_mask);
289 pde = ldl_raw(pde_ptr);
290 if (!(pde & PG_PRESENT_MASK)) {
295 if (!(pde & PG_USER_MASK))
296 goto do_fault_protect;
297 if (is_write && !(pde & PG_RW_MASK))
298 goto do_fault_protect;
300 if ((env->cr[0] & CR0_WP_MASK) && (pde & PG_USER_MASK) &&
301 is_write && !(pde & PG_RW_MASK))
302 goto do_fault_protect;
304 /* if PSE bit is set, then we use a 4MB page */
305 if ((pde & PG_PSE_MASK) && (env->cr[4] & CR4_PSE_MASK)) {
306 is_dirty = is_write && !(pde & PG_DIRTY_MASK);
307 if (!(pde & PG_ACCESSED_MASK)) {
308 pde |= PG_ACCESSED_MASK;
310 pde |= PG_DIRTY_MASK;
311 stl_raw(pde_ptr, pde);
314 pte = pde & ~0x003ff000; /* align to 4MB */
315 page_size = 4096 * 1024;
316 virt_addr = addr & ~0x003fffff;
318 if (!(pde & PG_ACCESSED_MASK)) {
319 pde |= PG_ACCESSED_MASK;
320 stl_raw(pde_ptr, pde);
323 /* page directory entry */
324 pte_ptr = phys_ram_base +
325 (((pde & ~0xfff) + ((addr >> 10) & 0xffc)) & a20_mask);
326 pte = ldl_raw(pte_ptr);
327 if (!(pte & PG_PRESENT_MASK)) {
332 if (!(pte & PG_USER_MASK))
333 goto do_fault_protect;
334 if (is_write && !(pte & PG_RW_MASK))
335 goto do_fault_protect;
337 if ((env->cr[0] & CR0_WP_MASK) && (pte & PG_USER_MASK) &&
338 is_write && !(pte & PG_RW_MASK))
339 goto do_fault_protect;
341 is_dirty = is_write && !(pte & PG_DIRTY_MASK);
342 if (!(pte & PG_ACCESSED_MASK) || is_dirty) {
343 pte |= PG_ACCESSED_MASK;
345 pte |= PG_DIRTY_MASK;
346 stl_raw(pte_ptr, pte);
349 virt_addr = addr & ~0xfff;
352 /* the page can be put in the TLB */
354 if (pte & PG_DIRTY_MASK) {
355 /* only set write access if already dirty... otherwise wait
358 if (pte & PG_RW_MASK)
361 if (!(env->cr[0] & CR0_WP_MASK) || !(pte & PG_USER_MASK) ||
368 pte = pte & a20_mask;
370 /* Even if 4MB pages, we map only one 4KB page in the cache to
371 avoid filling it too fast */
372 page_offset = (addr & TARGET_PAGE_MASK) & (page_size - 1);
373 paddr = (pte & TARGET_PAGE_MASK) + page_offset;
374 vaddr = virt_addr + page_offset;
376 ret = tlb_set_page(env, vaddr, paddr, prot, is_user, is_softmmu);
379 error_code = PG_ERROR_P_MASK;
382 env->error_code = (is_write << PG_ERROR_W_BIT) | error_code;
384 env->error_code |= PG_ERROR_U_MASK;