 05871c72be
			
		
	
	
		05871c72be
		
	
	
	
	
		
			
			This commit was created with scripts/clean-includes: ./scripts/clean-includes --git hyperv hw/hyperv/*.[ch] All .c should include qemu/osdep.h first. The script performs three related cleanups: * Ensure .c files include qemu/osdep.h first. * Including it in a .h is redundant, since the .c already includes it. Drop such inclusions. * Likewise, including headers qemu/osdep.h includes is redundant. Drop these, too. Signed-off-by: Peter Maydell <peter.maydell@linaro.org> Reviewed-by: Zhao Liu <zhao1.liu@intel.com> Reviewed-by: Richard Henderson <richard.henderson@linaro.org> Signed-off-by: Michael Tokarev <mjt@tls.msk.ru>
		
			
				
	
	
		
			203 lines
		
	
	
		
			6.7 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			203 lines
		
	
	
		
			6.7 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * QEMU Hyper-V Dynamic Memory Protocol driver
 | |
|  *
 | |
|  * Copyright (C) 2020-2023 Oracle and/or its affiliates.
 | |
|  *
 | |
|  * This work is licensed under the terms of the GNU GPL, version 2 or later.
 | |
|  * See the COPYING file in the top-level directory.
 | |
|  */
 | |
| 
 | |
| #include "qemu/osdep.h"
 | |
| #include "hv-balloon-internal.h"
 | |
| #include "hv-balloon-our_range_memslots.h"
 | |
| #include "trace.h"
 | |
| 
 | |
| /* OurRange */
 | |
| static void our_range_init(OurRange *our_range, uint64_t start, uint64_t count)
 | |
| {
 | |
|     assert(count <= UINT64_MAX - start);
 | |
|     our_range->range.start = start;
 | |
|     our_range->range.count = count;
 | |
| 
 | |
|     hvb_page_range_tree_init(&our_range->removed_guest);
 | |
|     hvb_page_range_tree_init(&our_range->removed_both);
 | |
| 
 | |
|     /* mark the whole range as unused but for potential use */
 | |
|     our_range->added = 0;
 | |
|     our_range->unusable_tail = 0;
 | |
| }
 | |
| 
 | |
| static void our_range_destroy(OurRange *our_range)
 | |
| {
 | |
|     hvb_page_range_tree_destroy(&our_range->removed_guest);
 | |
|     hvb_page_range_tree_destroy(&our_range->removed_both);
 | |
| }
 | |
| 
 | |
| void hvb_our_range_clear_removed_trees(OurRange *our_range)
 | |
| {
 | |
|     hvb_page_range_tree_destroy(&our_range->removed_guest);
 | |
|     hvb_page_range_tree_destroy(&our_range->removed_both);
 | |
|     hvb_page_range_tree_init(&our_range->removed_guest);
 | |
|     hvb_page_range_tree_init(&our_range->removed_both);
 | |
| }
 | |
| 
 | |
| void hvb_our_range_mark_added(OurRange *our_range, uint64_t additional_size)
 | |
| {
 | |
|     assert(additional_size <= UINT64_MAX - our_range->added);
 | |
| 
 | |
|     our_range->added += additional_size;
 | |
| 
 | |
|     assert(our_range->added <= UINT64_MAX - our_range->unusable_tail);
 | |
|     assert(our_range->added + our_range->unusable_tail <=
 | |
|            our_range->range.count);
 | |
| }
 | |
| 
 | |
| /* OurRangeMemslots */
 | |
| static void our_range_memslots_init_slots(OurRangeMemslots *our_range,
 | |
|                                           MemoryRegion *backing_mr,
 | |
|                                           Object *memslot_owner)
 | |
| {
 | |
|     OurRangeMemslotsSlots *memslots = &our_range->slots;
 | |
|     unsigned int idx;
 | |
|     uint64_t memslot_offset;
 | |
| 
 | |
|     assert(memslots->count > 0);
 | |
|     memslots->slots = g_new0(MemoryRegion, memslots->count);
 | |
| 
 | |
|     /* Initialize our memslots, but don't map them yet. */
 | |
|     assert(memslots->size_each > 0);
 | |
|     for (idx = 0, memslot_offset = 0; idx < memslots->count;
 | |
|          idx++, memslot_offset += memslots->size_each) {
 | |
|         uint64_t memslot_size;
 | |
|         g_autofree char *name = NULL;
 | |
| 
 | |
|         /* The size of the last memslot might be smaller. */
 | |
|         if (idx == memslots->count - 1) {
 | |
|             uint64_t region_size;
 | |
| 
 | |
|             assert(our_range->mr);
 | |
|             region_size = memory_region_size(our_range->mr);
 | |
|             memslot_size = region_size - memslot_offset;
 | |
|         } else {
 | |
|             memslot_size = memslots->size_each;
 | |
|         }
 | |
| 
 | |
|         name = g_strdup_printf("memslot-%u", idx);
 | |
|         memory_region_init_alias(&memslots->slots[idx], memslot_owner, name,
 | |
|                                  backing_mr, memslot_offset, memslot_size);
 | |
|         /*
 | |
|          * We want to be able to atomically and efficiently activate/deactivate
 | |
|          * individual memslots without affecting adjacent memslots in memory
 | |
|          * notifiers.
 | |
|          */
 | |
|         memory_region_set_unmergeable(&memslots->slots[idx], true);
 | |
|     }
 | |
| 
 | |
|     memslots->mapped_count = 0;
 | |
| }
 | |
| 
 | |
| OurRangeMemslots *hvb_our_range_memslots_new(uint64_t addr,
 | |
|                                              MemoryRegion *parent_mr,
 | |
|                                              MemoryRegion *backing_mr,
 | |
|                                              Object *memslot_owner,
 | |
|                                              unsigned int memslot_count,
 | |
|                                              uint64_t memslot_size)
 | |
| {
 | |
|     OurRangeMemslots *our_range;
 | |
| 
 | |
|     our_range = g_malloc(sizeof(*our_range));
 | |
|     our_range_init(&our_range->range,
 | |
|                    addr / HV_BALLOON_PAGE_SIZE,
 | |
|                    memory_region_size(parent_mr) / HV_BALLOON_PAGE_SIZE);
 | |
|     our_range->slots.size_each = memslot_size;
 | |
|     our_range->slots.count = memslot_count;
 | |
|     our_range->mr = parent_mr;
 | |
|     our_range_memslots_init_slots(our_range, backing_mr, memslot_owner);
 | |
| 
 | |
|     return our_range;
 | |
| }
 | |
| 
 | |
| static void our_range_memslots_free_memslots(OurRangeMemslots *our_range)
 | |
| {
 | |
|     OurRangeMemslotsSlots *memslots = &our_range->slots;
 | |
|     unsigned int idx;
 | |
|     uint64_t offset;
 | |
| 
 | |
|     memory_region_transaction_begin();
 | |
|     for (idx = 0, offset = 0; idx < memslots->mapped_count;
 | |
|          idx++, offset += memslots->size_each) {
 | |
|         trace_hv_balloon_unmap_slot(idx, memslots->count, offset);
 | |
|         assert(memory_region_is_mapped(&memslots->slots[idx]));
 | |
|         memory_region_del_subregion(our_range->mr, &memslots->slots[idx]);
 | |
|     }
 | |
|     memory_region_transaction_commit();
 | |
| 
 | |
|     for (idx = 0; idx < memslots->count; idx++) {
 | |
|         object_unparent(OBJECT(&memslots->slots[idx]));
 | |
|     }
 | |
| 
 | |
|     g_clear_pointer(&our_range->slots.slots, g_free);
 | |
| }
 | |
| 
 | |
| void hvb_our_range_memslots_free(OurRangeMemslots *our_range)
 | |
| {
 | |
|     OurRangeMemslotsSlots *memslots = &our_range->slots;
 | |
|     MemoryRegion *hostmem_mr;
 | |
|     RAMBlock *rb;
 | |
| 
 | |
|     assert(our_range->slots.count > 0);
 | |
|     assert(our_range->slots.slots);
 | |
| 
 | |
|     hostmem_mr = memslots->slots[0].alias;
 | |
|     rb = hostmem_mr->ram_block;
 | |
|     ram_block_discard_range(rb, 0, qemu_ram_get_used_length(rb));
 | |
| 
 | |
|     our_range_memslots_free_memslots(our_range);
 | |
|     our_range_destroy(&our_range->range);
 | |
|     g_free(our_range);
 | |
| }
 | |
| 
 | |
| void hvb_our_range_memslots_ensure_mapped_additional(OurRangeMemslots *our_range,
 | |
|                                                      uint64_t additional_map_size)
 | |
| {
 | |
|     OurRangeMemslotsSlots *memslots = &our_range->slots;
 | |
|     uint64_t total_map_size;
 | |
|     unsigned int idx;
 | |
|     uint64_t offset;
 | |
| 
 | |
|     total_map_size = (our_range->range.added + additional_map_size) *
 | |
|         HV_BALLOON_PAGE_SIZE;
 | |
|     idx = memslots->mapped_count;
 | |
|     assert(memslots->size_each > 0);
 | |
|     offset = idx * memslots->size_each;
 | |
| 
 | |
|     /*
 | |
|      * Activate all memslots covered by the newly added region in a single
 | |
|      * transaction.
 | |
|      */
 | |
|     memory_region_transaction_begin();
 | |
|     for ( ; idx < memslots->count;
 | |
|           idx++, offset += memslots->size_each) {
 | |
|         /*
 | |
|          * If this memslot starts beyond or at the end of the range to map so
 | |
|          * does every next one.
 | |
|          */
 | |
|         if (offset >= total_map_size) {
 | |
|             break;
 | |
|         }
 | |
| 
 | |
|         /*
 | |
|          * Instead of enabling/disabling memslot, we add/remove them. This
 | |
|          * should make address space updates faster, because we don't have to
 | |
|          * loop over many disabled subregions.
 | |
|          */
 | |
|         trace_hv_balloon_map_slot(idx, memslots->count, offset);
 | |
|         assert(!memory_region_is_mapped(&memslots->slots[idx]));
 | |
|         memory_region_add_subregion(our_range->mr, offset,
 | |
|                                     &memslots->slots[idx]);
 | |
| 
 | |
|         memslots->mapped_count++;
 | |
|     }
 | |
|     memory_region_transaction_commit();
 | |
| }
 |