2 * i386 helpers (without register variable usage)
4 * Copyright (c) 2003 Fabrice Bellard
6 * This library is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU Lesser General Public
8 * License as published by the Free Software Foundation; either
9 * version 2 of the License, or (at your option) any later version.
11 * This library is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14 * Lesser General Public License for more details.
16 * You should have received a copy of the GNU Lesser General Public
17 * License along with this library; if not, write to the Free Software
18 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
34 CPUX86State *cpu_x86_init(void)
42 env = malloc(sizeof(CPUX86State));
45 memset(env, 0, sizeof(CPUX86State));
50 /* flags setup : we activate the IRQs by default as in user mode */
51 env->eflags = 0x2 | IF_MASK;
55 env->hflags |= HF_SOFTMMU_MASK;
57 /* init various static tables */
60 optimize_flags_init();
65 void cpu_x86_close(CPUX86State *env)
70 /***********************************************************/
73 static const char *cc_op_str[] = {
106 void cpu_x86_dump_state(CPUX86State *env, FILE *f, int flags)
111 eflags = env->eflags;
112 fprintf(f, "EAX=%08x EBX=%08x ECX=%08x EDX=%08x\n"
113 "ESI=%08x EDI=%08x EBP=%08x ESP=%08x\n"
114 "EIP=%08x EFL=%08x [%c%c%c%c%c%c%c]\n",
115 env->regs[R_EAX], env->regs[R_EBX], env->regs[R_ECX], env->regs[R_EDX],
116 env->regs[R_ESI], env->regs[R_EDI], env->regs[R_EBP], env->regs[R_ESP],
118 eflags & DF_MASK ? 'D' : '-',
119 eflags & CC_O ? 'O' : '-',
120 eflags & CC_S ? 'S' : '-',
121 eflags & CC_Z ? 'Z' : '-',
122 eflags & CC_A ? 'A' : '-',
123 eflags & CC_P ? 'P' : '-',
124 eflags & CC_C ? 'C' : '-');
125 fprintf(f, "CS=%04x SS=%04x DS=%04x ES=%04x FS=%04x GS=%04x\n",
126 env->segs[R_CS].selector,
127 env->segs[R_SS].selector,
128 env->segs[R_DS].selector,
129 env->segs[R_ES].selector,
130 env->segs[R_FS].selector,
131 env->segs[R_GS].selector);
132 if (flags & X86_DUMP_CCOP) {
133 if ((unsigned)env->cc_op < CC_OP_NB)
134 strcpy(cc_op_name, cc_op_str[env->cc_op]);
136 snprintf(cc_op_name, sizeof(cc_op_name), "[%d]", env->cc_op);
137 fprintf(f, "CCS=%08x CCD=%08x CCO=%-8s\n",
138 env->cc_src, env->cc_dst, cc_op_name);
140 if (flags & X86_DUMP_FPU) {
141 fprintf(f, "ST0=%f ST1=%f ST2=%f ST3=%f\n",
142 (double)env->fpregs[0],
143 (double)env->fpregs[1],
144 (double)env->fpregs[2],
145 (double)env->fpregs[3]);
146 fprintf(f, "ST4=%f ST5=%f ST6=%f ST7=%f\n",
147 (double)env->fpregs[4],
148 (double)env->fpregs[5],
149 (double)env->fpregs[7],
150 (double)env->fpregs[8]);
154 /***********************************************************/
156 /* XXX: add PGE support */
158 /* called when cr3 or PG bit are modified */
159 static int last_pg_state = -1;
160 static int last_pe_state = 0;
163 uint8_t *phys_ram_base;
165 void cpu_x86_update_cr0(CPUX86State *env)
167 int pg_state, pe_state;
170 printf("CR0 update: CR0=0x%08x\n", env->cr[0]);
172 pg_state = env->cr[0] & CR0_PG_MASK;
173 if (pg_state != last_pg_state) {
176 last_pg_state = pg_state;
178 pe_state = env->cr[0] & CR0_PE_MASK;
179 if (last_pe_state != pe_state) {
181 last_pe_state = pe_state;
185 void cpu_x86_update_cr3(CPUX86State *env)
187 if (env->cr[0] & CR0_PG_MASK) {
188 #if defined(DEBUG_MMU)
189 printf("CR3 update: CR3=%08x\n", env->cr[3]);
196 void cpu_x86_init_mmu(CPUX86State *env)
199 cpu_x86_update_cr0(env);
202 /* XXX: also flush 4MB pages */
203 void cpu_x86_flush_tlb(CPUX86State *env, uint32_t addr)
206 unsigned long virt_addr;
208 tlb_flush_page(env, addr);
210 flags = page_get_flags(addr);
211 if (flags & PAGE_VALID) {
212 virt_addr = addr & ~0xfff;
213 munmap((void *)virt_addr, 4096);
214 page_set_flags(virt_addr, virt_addr + 4096, 0);
219 -1 = cannot handle fault
220 0 = nothing more to do
221 1 = generate PF fault
222 2 = soft MMU activation required for this block
224 int cpu_x86_handle_mmu_fault(CPUX86State *env, uint32_t addr, int is_write)
226 uint8_t *pde_ptr, *pte_ptr;
227 uint32_t pde, pte, virt_addr;
228 int cpl, error_code, is_dirty, is_user, prot, page_size, ret;
231 cpl = env->hflags & HF_CPL_MASK;
232 is_user = (cpl == 3);
235 printf("MMU fault: addr=0x%08x w=%d u=%d eip=%08x\n",
236 addr, is_write, is_user, env->eip);
239 if (env->user_mode_only) {
240 /* user mode only emulation */
245 if (!(env->cr[0] & CR0_PG_MASK)) {
247 virt_addr = addr & ~0xfff;
248 prot = PROT_READ | PROT_WRITE;
253 /* page directory entry */
254 pde_ptr = phys_ram_base + ((env->cr[3] & ~0xfff) + ((addr >> 20) & ~3));
256 if (!(pde & PG_PRESENT_MASK)) {
261 if (!(pde & PG_USER_MASK))
262 goto do_fault_protect;
263 if (is_write && !(pde & PG_RW_MASK))
264 goto do_fault_protect;
266 if ((env->cr[0] & CR0_WP_MASK) && (pde & PG_USER_MASK) &&
267 is_write && !(pde & PG_RW_MASK))
268 goto do_fault_protect;
270 /* if PSE bit is set, then we use a 4MB page */
271 if ((pde & PG_PSE_MASK) && (env->cr[4] & CR4_PSE_MASK)) {
272 is_dirty = is_write && !(pde & PG_DIRTY_MASK);
273 if (!(pde & PG_ACCESSED_MASK)) {
274 pde |= PG_ACCESSED_MASK;
276 pde |= PG_DIRTY_MASK;
280 pte = pde & ~0x003ff000; /* align to 4MB */
281 page_size = 4096 * 1024;
282 virt_addr = addr & ~0x003fffff;
284 if (!(pde & PG_ACCESSED_MASK)) {
285 pde |= PG_ACCESSED_MASK;
289 /* page directory entry */
290 pte_ptr = phys_ram_base + ((pde & ~0xfff) + ((addr >> 10) & 0xffc));
292 if (!(pte & PG_PRESENT_MASK)) {
297 if (!(pte & PG_USER_MASK))
298 goto do_fault_protect;
299 if (is_write && !(pte & PG_RW_MASK))
300 goto do_fault_protect;
302 if ((env->cr[0] & CR0_WP_MASK) && (pte & PG_USER_MASK) &&
303 is_write && !(pte & PG_RW_MASK))
304 goto do_fault_protect;
306 is_dirty = is_write && !(pte & PG_DIRTY_MASK);
307 if (!(pte & PG_ACCESSED_MASK) || is_dirty) {
308 pte |= PG_ACCESSED_MASK;
310 pte |= PG_DIRTY_MASK;
314 virt_addr = addr & ~0xfff;
316 /* the page can be put in the TLB */
319 if (pte & PG_RW_MASK)
322 if (!(env->cr[0] & CR0_WP_MASK) || !(pte & PG_USER_MASK) ||
328 if (env->hflags & HF_SOFTMMU_MASK) {
329 unsigned long paddr, vaddr, address, addend, page_offset;
332 /* software MMU case. Even if 4MB pages, we map only one 4KB
333 page in the cache to avoid filling it too fast */
334 page_offset = (addr & ~0xfff) & (page_size - 1);
335 paddr = (pte & ~0xfff) + page_offset;
336 vaddr = virt_addr + page_offset;
337 index = (addr >> 12) & (CPU_TLB_SIZE - 1);
338 pd = physpage_find(paddr);
341 address = vaddr | pd;
344 /* standard memory */
346 addend = (unsigned long)phys_ram_base + pd;
349 env->tlb_read[is_user][index].address = address;
350 env->tlb_read[is_user][index].addend = addend;
351 if (prot & PROT_WRITE) {
352 env->tlb_write[is_user][index].address = address;
353 env->tlb_write[is_user][index].addend = addend;
357 /* XXX: incorrect for 4MB pages */
358 pd = physpage_find(pte & ~0xfff);
359 if ((pd & 0xfff) != 0) {
360 /* IO access: no mapping is done as it will be handled by the
362 if (!(env->hflags & HF_SOFTMMU_MASK))
366 map_addr = mmap((void *)virt_addr, page_size, prot,
367 MAP_SHARED | MAP_FIXED, phys_ram_fd, pd);
368 if (map_addr == MAP_FAILED) {
370 "mmap failed when mapped physical address 0x%08x to virtual address 0x%08x\n",
371 pte & ~0xfff, virt_addr);
375 printf("mmaping 0x%08x to virt 0x%08x pse=%d\n",
376 pte & ~0xfff, virt_addr, (page_size != 4096));
378 page_set_flags(virt_addr, virt_addr + page_size,
379 PAGE_VALID | PAGE_EXEC | prot);
383 error_code = PG_ERROR_P_MASK;
386 env->error_code = (is_write << PG_ERROR_W_BIT) | error_code;
388 env->error_code |= PG_ERROR_U_MASK;