4 * Copyright Fujitsu, Corp. 2011, 2012
7 * Wen Congyang <wency@cn.fujitsu.com>
9 * This work is licensed under the terms of the GNU GPL, version 2 or later.
10 * See the COPYING file in the top-level directory.
14 #include "qemu/osdep.h"
15 #include "qapi/error.h"
17 #include "sysemu/memory_mapping.h"
18 #include "exec/memory.h"
19 #include "exec/address-spaces.h"
20 #include "hw/core/cpu.h"
22 //#define DEBUG_GUEST_PHYS_REGION_ADD
24 static void memory_mapping_list_add_mapping_sorted(MemoryMappingList
*list
,
25 MemoryMapping
*mapping
)
29 QTAILQ_FOREACH(p
, &list
->head
, next
) {
30 if (p
->phys_addr
>= mapping
->phys_addr
) {
31 QTAILQ_INSERT_BEFORE(p
, mapping
, next
);
35 QTAILQ_INSERT_TAIL(&list
->head
, mapping
, next
);
38 static void create_new_memory_mapping(MemoryMappingList
*list
,
43 MemoryMapping
*memory_mapping
;
45 memory_mapping
= g_new(MemoryMapping
, 1);
46 memory_mapping
->phys_addr
= phys_addr
;
47 memory_mapping
->virt_addr
= virt_addr
;
48 memory_mapping
->length
= length
;
49 list
->last_mapping
= memory_mapping
;
51 memory_mapping_list_add_mapping_sorted(list
, memory_mapping
);
54 static inline bool mapping_contiguous(MemoryMapping
*map
,
58 return phys_addr
== map
->phys_addr
+ map
->length
&&
59 virt_addr
== map
->virt_addr
+ map
->length
;
63 * [map->phys_addr, map->phys_addr + map->length) and
64 * [phys_addr, phys_addr + length) have intersection?
66 static inline bool mapping_have_same_region(MemoryMapping
*map
,
70 return !(phys_addr
+ length
< map
->phys_addr
||
71 phys_addr
>= map
->phys_addr
+ map
->length
);
75 * [map->phys_addr, map->phys_addr + map->length) and
76 * [phys_addr, phys_addr + length) have intersection. The virtual address in the
77 * intersection are the same?
79 static inline bool mapping_conflict(MemoryMapping
*map
,
83 return virt_addr
- map
->virt_addr
!= phys_addr
- map
->phys_addr
;
87 * [map->virt_addr, map->virt_addr + map->length) and
88 * [virt_addr, virt_addr + length) have intersection. And the physical address
89 * in the intersection are the same.
91 static inline void mapping_merge(MemoryMapping
*map
,
95 if (virt_addr
< map
->virt_addr
) {
96 map
->length
+= map
->virt_addr
- virt_addr
;
97 map
->virt_addr
= virt_addr
;
100 if ((virt_addr
+ length
) >
101 (map
->virt_addr
+ map
->length
)) {
102 map
->length
= virt_addr
+ length
- map
->virt_addr
;
106 void memory_mapping_list_add_merge_sorted(MemoryMappingList
*list
,
111 MemoryMapping
*memory_mapping
, *last_mapping
;
113 if (QTAILQ_EMPTY(&list
->head
)) {
114 create_new_memory_mapping(list
, phys_addr
, virt_addr
, length
);
118 last_mapping
= list
->last_mapping
;
120 if (mapping_contiguous(last_mapping
, phys_addr
, virt_addr
)) {
121 last_mapping
->length
+= length
;
126 QTAILQ_FOREACH(memory_mapping
, &list
->head
, next
) {
127 if (mapping_contiguous(memory_mapping
, phys_addr
, virt_addr
)) {
128 memory_mapping
->length
+= length
;
129 list
->last_mapping
= memory_mapping
;
133 if (phys_addr
+ length
< memory_mapping
->phys_addr
) {
134 /* create a new region before memory_mapping */
138 if (mapping_have_same_region(memory_mapping
, phys_addr
, length
)) {
139 if (mapping_conflict(memory_mapping
, phys_addr
, virt_addr
)) {
143 /* merge this region into memory_mapping */
144 mapping_merge(memory_mapping
, virt_addr
, length
);
145 list
->last_mapping
= memory_mapping
;
150 /* this region can not be merged into any existed memory mapping. */
151 create_new_memory_mapping(list
, phys_addr
, virt_addr
, length
);
154 void memory_mapping_list_free(MemoryMappingList
*list
)
156 MemoryMapping
*p
, *q
;
158 QTAILQ_FOREACH_SAFE(p
, &list
->head
, next
, q
) {
159 QTAILQ_REMOVE(&list
->head
, p
, next
);
164 list
->last_mapping
= NULL
;
167 void memory_mapping_list_init(MemoryMappingList
*list
)
170 list
->last_mapping
= NULL
;
171 QTAILQ_INIT(&list
->head
);
174 void guest_phys_blocks_free(GuestPhysBlockList
*list
)
176 GuestPhysBlock
*p
, *q
;
178 QTAILQ_FOREACH_SAFE(p
, &list
->head
, next
, q
) {
179 QTAILQ_REMOVE(&list
->head
, p
, next
);
180 memory_region_unref(p
->mr
);
186 void guest_phys_blocks_init(GuestPhysBlockList
*list
)
189 QTAILQ_INIT(&list
->head
);
192 typedef struct GuestPhysListener
{
193 GuestPhysBlockList
*list
;
194 MemoryListener listener
;
197 static void guest_phys_block_add_section(GuestPhysListener
*g
,
198 MemoryRegionSection
*section
)
200 const hwaddr target_start
= section
->offset_within_address_space
;
201 const hwaddr target_end
= target_start
+ int128_get64(section
->size
);
202 uint8_t *host_addr
= memory_region_get_ram_ptr(section
->mr
) +
203 section
->offset_within_region
;
204 GuestPhysBlock
*predecessor
= NULL
;
206 /* find continuity in guest physical address space */
207 if (!QTAILQ_EMPTY(&g
->list
->head
)) {
208 hwaddr predecessor_size
;
210 predecessor
= QTAILQ_LAST(&g
->list
->head
);
211 predecessor_size
= predecessor
->target_end
- predecessor
->target_start
;
213 /* the memory API guarantees monotonically increasing traversal */
214 g_assert(predecessor
->target_end
<= target_start
);
216 /* we want continuity in both guest-physical and host-virtual memory */
217 if (predecessor
->target_end
< target_start
||
218 predecessor
->host_addr
+ predecessor_size
!= host_addr
||
219 predecessor
->mr
!= section
->mr
) {
224 if (predecessor
== NULL
) {
225 /* isolated mapping, allocate it and add it to the list */
226 GuestPhysBlock
*block
= g_malloc0(sizeof *block
);
228 block
->target_start
= target_start
;
229 block
->target_end
= target_end
;
230 block
->host_addr
= host_addr
;
231 block
->mr
= section
->mr
;
232 memory_region_ref(section
->mr
);
234 QTAILQ_INSERT_TAIL(&g
->list
->head
, block
, next
);
237 /* expand predecessor until @target_end; predecessor's start doesn't
240 predecessor
->target_end
= target_end
;
243 #ifdef DEBUG_GUEST_PHYS_REGION_ADD
244 fprintf(stderr
, "%s: target_start=" HWADDR_FMT_plx
" target_end="
245 HWADDR_FMT_plx
": %s (count: %u)\n", __func__
, target_start
,
246 target_end
, predecessor
? "joined" : "added", g
->list
->num
);
250 static int guest_phys_ram_populate_cb(MemoryRegionSection
*section
,
253 GuestPhysListener
*g
= opaque
;
255 guest_phys_block_add_section(g
, section
);
259 static void guest_phys_blocks_region_add(MemoryListener
*listener
,
260 MemoryRegionSection
*section
)
262 GuestPhysListener
*g
= container_of(listener
, GuestPhysListener
, listener
);
264 /* we only care about RAM */
265 if (!memory_region_is_ram(section
->mr
) ||
266 memory_region_is_ram_device(section
->mr
) ||
267 memory_region_is_nonvolatile(section
->mr
)) {
271 /* for special sparse regions, only add populated parts */
272 if (memory_region_has_ram_discard_manager(section
->mr
)) {
273 RamDiscardManager
*rdm
;
275 rdm
= memory_region_get_ram_discard_manager(section
->mr
);
276 ram_discard_manager_replay_populated(rdm
, section
,
277 guest_phys_ram_populate_cb
, g
);
281 guest_phys_block_add_section(g
, section
);
284 void guest_phys_blocks_append(GuestPhysBlockList
*list
)
286 GuestPhysListener g
= { 0 };
289 g
.listener
.region_add
= &guest_phys_blocks_region_add
;
290 memory_listener_register(&g
.listener
, &address_space_memory
);
291 memory_listener_unregister(&g
.listener
);
294 static CPUState
*find_paging_enabled_cpu(void)
299 if (cpu_paging_enabled(cpu
)) {
307 bool qemu_get_guest_memory_mapping(MemoryMappingList
*list
,
308 const GuestPhysBlockList
*guest_phys_blocks
,
312 CPUState
*cpu
, *first_paging_enabled_cpu
;
313 GuestPhysBlock
*block
;
314 ram_addr_t offset
, length
;
316 first_paging_enabled_cpu
= find_paging_enabled_cpu();
317 if (first_paging_enabled_cpu
) {
318 for (cpu
= first_paging_enabled_cpu
; cpu
!= NULL
;
319 cpu
= CPU_NEXT(cpu
)) {
320 if (!cpu_get_memory_mapping(cpu
, list
, errp
)) {
328 * If the guest doesn't use paging, the virtual address is equal to physical
331 QTAILQ_FOREACH(block
, &guest_phys_blocks
->head
, next
) {
332 offset
= block
->target_start
;
333 length
= block
->target_end
- block
->target_start
;
334 create_new_memory_mapping(list
, offset
, offset
, length
);
339 void qemu_get_guest_simple_memory_mapping(MemoryMappingList
*list
,
340 const GuestPhysBlockList
*guest_phys_blocks
)
342 GuestPhysBlock
*block
;
344 QTAILQ_FOREACH(block
, &guest_phys_blocks
->head
, next
) {
345 create_new_memory_mapping(list
, block
->target_start
, 0,
346 block
->target_end
- block
->target_start
);
350 void memory_mapping_filter(MemoryMappingList
*list
, int64_t begin
,
353 MemoryMapping
*cur
, *next
;
355 QTAILQ_FOREACH_SAFE(cur
, &list
->head
, next
, next
) {
356 if (cur
->phys_addr
>= begin
+ length
||
357 cur
->phys_addr
+ cur
->length
<= begin
) {
358 QTAILQ_REMOVE(&list
->head
, cur
, next
);
364 if (cur
->phys_addr
< begin
) {
365 cur
->length
-= begin
- cur
->phys_addr
;
366 if (cur
->virt_addr
) {
367 cur
->virt_addr
+= begin
- cur
->phys_addr
;
369 cur
->phys_addr
= begin
;
372 if (cur
->phys_addr
+ cur
->length
> begin
+ length
) {
373 cur
->length
-= cur
->phys_addr
+ cur
->length
- begin
- length
;