4 * Copyright Fujitsu, Corp. 2011, 2012
9 * This work is licensed under the terms of the GNU GPL, version 2 or later.
10 * See the COPYING file in the top-level directory.
15 #include "exec/cpu-all.h"
16 #include "sysemu/memory_mapping.h"
18 /* PAE Paging or IA-32e Paging */
19 static void walk_pte(MemoryMappingList *list, hwaddr pte_start_addr,
20 int32_t a20_mask, target_ulong start_line_addr)
22 hwaddr pte_addr, start_paddr;
24 target_ulong start_vaddr;
27 for (i = 0; i < 512; i++) {
28 pte_addr = (pte_start_addr + i * 8) & a20_mask;
29 pte = ldq_phys(pte_addr);
30 if (!(pte & PG_PRESENT_MASK)) {
35 start_paddr = (pte & ~0xfff) & ~(0x1ULL << 63);
36 if (cpu_physical_memory_is_io(start_paddr)) {
41 start_vaddr = start_line_addr | ((i & 0x1fff) << 12);
42 memory_mapping_list_add_merge_sorted(list, start_paddr,
43 start_vaddr, 1 << 12);
48 static void walk_pte2(MemoryMappingList *list,
49 hwaddr pte_start_addr, int32_t a20_mask,
50 target_ulong start_line_addr)
52 hwaddr pte_addr, start_paddr;
54 target_ulong start_vaddr;
57 for (i = 0; i < 1024; i++) {
58 pte_addr = (pte_start_addr + i * 4) & a20_mask;
59 pte = ldl_phys(pte_addr);
60 if (!(pte & PG_PRESENT_MASK)) {
65 start_paddr = pte & ~0xfff;
66 if (cpu_physical_memory_is_io(start_paddr)) {
71 start_vaddr = start_line_addr | ((i & 0x3ff) << 12);
72 memory_mapping_list_add_merge_sorted(list, start_paddr,
73 start_vaddr, 1 << 12);
77 /* PAE Paging or IA-32e Paging */
78 static void walk_pde(MemoryMappingList *list, hwaddr pde_start_addr,
79 int32_t a20_mask, target_ulong start_line_addr)
81 hwaddr pde_addr, pte_start_addr, start_paddr;
83 target_ulong line_addr, start_vaddr;
86 for (i = 0; i < 512; i++) {
87 pde_addr = (pde_start_addr + i * 8) & a20_mask;
88 pde = ldq_phys(pde_addr);
89 if (!(pde & PG_PRESENT_MASK)) {
94 line_addr = start_line_addr | ((i & 0x1ff) << 21);
95 if (pde & PG_PSE_MASK) {
97 start_paddr = (pde & ~0x1fffff) & ~(0x1ULL << 63);
98 if (cpu_physical_memory_is_io(start_paddr)) {
102 start_vaddr = line_addr;
103 memory_mapping_list_add_merge_sorted(list, start_paddr,
104 start_vaddr, 1 << 21);
108 pte_start_addr = (pde & ~0xfff) & a20_mask;
109 walk_pte(list, pte_start_addr, a20_mask, line_addr);
114 static void walk_pde2(MemoryMappingList *list,
115 hwaddr pde_start_addr, int32_t a20_mask,
118 hwaddr pde_addr, pte_start_addr, start_paddr, high_paddr;
120 target_ulong line_addr, start_vaddr;
123 for (i = 0; i < 1024; i++) {
124 pde_addr = (pde_start_addr + i * 4) & a20_mask;
125 pde = ldl_phys(pde_addr);
126 if (!(pde & PG_PRESENT_MASK)) {
131 line_addr = (((unsigned int)i & 0x3ff) << 22);
132 if ((pde & PG_PSE_MASK) && pse) {
135 * bits 39:32 are bits 20:13 of the PDE
136 * bit3 31:22 are bits 31:22 of the PDE
138 high_paddr = ((hwaddr)(pde & 0x1fe000) << 19);
139 start_paddr = (pde & ~0x3fffff) | high_paddr;
140 if (cpu_physical_memory_is_io(start_paddr)) {
144 start_vaddr = line_addr;
145 memory_mapping_list_add_merge_sorted(list, start_paddr,
146 start_vaddr, 1 << 22);
150 pte_start_addr = (pde & ~0xfff) & a20_mask;
151 walk_pte2(list, pte_start_addr, a20_mask, line_addr);
156 static void walk_pdpe2(MemoryMappingList *list,
157 hwaddr pdpe_start_addr, int32_t a20_mask)
159 hwaddr pdpe_addr, pde_start_addr;
161 target_ulong line_addr;
164 for (i = 0; i < 4; i++) {
165 pdpe_addr = (pdpe_start_addr + i * 8) & a20_mask;
166 pdpe = ldq_phys(pdpe_addr);
167 if (!(pdpe & PG_PRESENT_MASK)) {
172 line_addr = (((unsigned int)i & 0x3) << 30);
173 pde_start_addr = (pdpe & ~0xfff) & a20_mask;
174 walk_pde(list, pde_start_addr, a20_mask, line_addr);
180 static void walk_pdpe(MemoryMappingList *list,
181 hwaddr pdpe_start_addr, int32_t a20_mask,
182 target_ulong start_line_addr)
184 hwaddr pdpe_addr, pde_start_addr, start_paddr;
186 target_ulong line_addr, start_vaddr;
189 for (i = 0; i < 512; i++) {
190 pdpe_addr = (pdpe_start_addr + i * 8) & a20_mask;
191 pdpe = ldq_phys(pdpe_addr);
192 if (!(pdpe & PG_PRESENT_MASK)) {
197 line_addr = start_line_addr | ((i & 0x1ffULL) << 30);
198 if (pdpe & PG_PSE_MASK) {
200 start_paddr = (pdpe & ~0x3fffffff) & ~(0x1ULL << 63);
201 if (cpu_physical_memory_is_io(start_paddr)) {
205 start_vaddr = line_addr;
206 memory_mapping_list_add_merge_sorted(list, start_paddr,
207 start_vaddr, 1 << 30);
211 pde_start_addr = (pdpe & ~0xfff) & a20_mask;
212 walk_pde(list, pde_start_addr, a20_mask, line_addr);
217 static void walk_pml4e(MemoryMappingList *list,
218 hwaddr pml4e_start_addr, int32_t a20_mask)
220 hwaddr pml4e_addr, pdpe_start_addr;
222 target_ulong line_addr;
225 for (i = 0; i < 512; i++) {
226 pml4e_addr = (pml4e_start_addr + i * 8) & a20_mask;
227 pml4e = ldq_phys(pml4e_addr);
228 if (!(pml4e & PG_PRESENT_MASK)) {
233 line_addr = ((i & 0x1ffULL) << 39) | (0xffffULL << 48);
234 pdpe_start_addr = (pml4e & ~0xfff) & a20_mask;
235 walk_pdpe(list, pdpe_start_addr, a20_mask, line_addr);
240 int cpu_get_memory_mapping(MemoryMappingList *list, CPUArchState *env)
242 if (!cpu_paging_enabled(env)) {
243 /* paging is disabled */
247 if (env->cr[4] & CR4_PAE_MASK) {
249 if (env->hflags & HF_LMA_MASK) {
252 pml4e_addr = (env->cr[3] & ~0xfff) & env->a20_mask;
253 walk_pml4e(list, pml4e_addr, env->a20_mask);
259 pdpe_addr = (env->cr[3] & ~0x1f) & env->a20_mask;
260 walk_pdpe2(list, pdpe_addr, env->a20_mask);
266 pde_addr = (env->cr[3] & ~0xfff) & env->a20_mask;
267 pse = !!(env->cr[4] & CR4_PSE_MASK);
268 walk_pde2(list, pde_addr, env->a20_mask, pse);
274 bool cpu_paging_enabled(CPUArchState *env)
276 return env->cr[0] & CR0_PG_MASK;