Commit 22721a56 authored by Maciej S. Szmigiero's avatar Maciej S. Szmigiero Committed by Paolo Bonzini

KVM: selftests: Keep track of memslots more efficiently

The KVM selftest framework was using a simple list for keeping track of
the memslots currently in use.
This resulted in lookups and adding a single memslot being O(n), the
later due to linear scanning of the existing memslot set to check for
the presence of any conflicting entries.

Before this change, benchmarking high count of memslots was more or less
impossible as pretty much all the benchmark time was spent in the
selftest framework code.

We can simply use a rbtree for keeping track of both of gfn and hva.
We don't need an interval tree for hva here as we can't have overlapping
memslots because we allocate a completely new memory chunk for each new
memslot.
Signed-off-by: default avatarMaciej S. Szmigiero <maciej.szmigiero@oracle.com>
Reviewed-by: default avatarAndrew Jones <drjones@redhat.com>
Message-Id: <b12749d47ee860468240cf027412c91b76dbe3db.1618253574.git.maciej.szmigiero@oracle.com>
Signed-off-by: default avatarPaolo Bonzini <pbonzini@redhat.com>
parent a13534d6
...@@ -33,7 +33,7 @@ ifeq ($(ARCH),s390) ...@@ -33,7 +33,7 @@ ifeq ($(ARCH),s390)
UNAME_M := s390x UNAME_M := s390x
endif endif
LIBKVM = lib/assert.c lib/elf.c lib/io.c lib/kvm_util.c lib/sparsebit.c lib/test_util.c lib/guest_modes.c lib/perf_test_util.c LIBKVM = lib/assert.c lib/elf.c lib/io.c lib/kvm_util.c lib/rbtree.c lib/sparsebit.c lib/test_util.c lib/guest_modes.c lib/perf_test_util.c
LIBKVM_x86_64 = lib/x86_64/processor.c lib/x86_64/vmx.c lib/x86_64/svm.c lib/x86_64/ucall.c lib/x86_64/handlers.S LIBKVM_x86_64 = lib/x86_64/processor.c lib/x86_64/vmx.c lib/x86_64/svm.c lib/x86_64/ucall.c lib/x86_64/handlers.S
LIBKVM_aarch64 = lib/aarch64/processor.c lib/aarch64/ucall.c LIBKVM_aarch64 = lib/aarch64/processor.c lib/aarch64/ucall.c
LIBKVM_s390x = lib/s390x/processor.c lib/s390x/ucall.c lib/s390x/diag318_test_handler.c LIBKVM_s390x = lib/s390x/processor.c lib/s390x/ucall.c lib/s390x/diag318_test_handler.c
......
...@@ -203,7 +203,9 @@ struct kvm_vm *vm_create(enum vm_guest_mode mode, uint64_t phy_pages, int perm) ...@@ -203,7 +203,9 @@ struct kvm_vm *vm_create(enum vm_guest_mode mode, uint64_t phy_pages, int perm)
TEST_ASSERT(vm != NULL, "Insufficient Memory"); TEST_ASSERT(vm != NULL, "Insufficient Memory");
INIT_LIST_HEAD(&vm->vcpus); INIT_LIST_HEAD(&vm->vcpus);
INIT_LIST_HEAD(&vm->userspace_mem_regions); vm->regions.gpa_tree = RB_ROOT;
vm->regions.hva_tree = RB_ROOT;
hash_init(vm->regions.slot_hash);
vm->mode = mode; vm->mode = mode;
vm->type = 0; vm->type = 0;
...@@ -355,13 +357,14 @@ struct kvm_vm *vm_create_default(uint32_t vcpuid, uint64_t extra_mem_pages, ...@@ -355,13 +357,14 @@ struct kvm_vm *vm_create_default(uint32_t vcpuid, uint64_t extra_mem_pages,
*/ */
void kvm_vm_restart(struct kvm_vm *vmp, int perm) void kvm_vm_restart(struct kvm_vm *vmp, int perm)
{ {
int ctr;
struct userspace_mem_region *region; struct userspace_mem_region *region;
vm_open(vmp, perm); vm_open(vmp, perm);
if (vmp->has_irqchip) if (vmp->has_irqchip)
vm_create_irqchip(vmp); vm_create_irqchip(vmp);
list_for_each_entry(region, &vmp->userspace_mem_regions, list) { hash_for_each(vmp->regions.slot_hash, ctr, region, slot_node) {
int ret = ioctl(vmp->fd, KVM_SET_USER_MEMORY_REGION, &region->region); int ret = ioctl(vmp->fd, KVM_SET_USER_MEMORY_REGION, &region->region);
TEST_ASSERT(ret == 0, "KVM_SET_USER_MEMORY_REGION IOCTL failed,\n" TEST_ASSERT(ret == 0, "KVM_SET_USER_MEMORY_REGION IOCTL failed,\n"
" rc: %i errno: %i\n" " rc: %i errno: %i\n"
...@@ -424,14 +427,21 @@ uint32_t kvm_vm_reset_dirty_ring(struct kvm_vm *vm) ...@@ -424,14 +427,21 @@ uint32_t kvm_vm_reset_dirty_ring(struct kvm_vm *vm)
static struct userspace_mem_region * static struct userspace_mem_region *
userspace_mem_region_find(struct kvm_vm *vm, uint64_t start, uint64_t end) userspace_mem_region_find(struct kvm_vm *vm, uint64_t start, uint64_t end)
{ {
struct userspace_mem_region *region; struct rb_node *node;
list_for_each_entry(region, &vm->userspace_mem_regions, list) { for (node = vm->regions.gpa_tree.rb_node; node; ) {
struct userspace_mem_region *region =
container_of(node, struct userspace_mem_region, gpa_node);
uint64_t existing_start = region->region.guest_phys_addr; uint64_t existing_start = region->region.guest_phys_addr;
uint64_t existing_end = region->region.guest_phys_addr uint64_t existing_end = region->region.guest_phys_addr
+ region->region.memory_size - 1; + region->region.memory_size - 1;
if (start <= existing_end && end >= existing_start) if (start <= existing_end && end >= existing_start)
return region; return region;
if (start < existing_start)
node = node->rb_left;
else
node = node->rb_right;
} }
return NULL; return NULL;
...@@ -546,11 +556,16 @@ void kvm_vm_release(struct kvm_vm *vmp) ...@@ -546,11 +556,16 @@ void kvm_vm_release(struct kvm_vm *vmp)
} }
static void __vm_mem_region_delete(struct kvm_vm *vm, static void __vm_mem_region_delete(struct kvm_vm *vm,
struct userspace_mem_region *region) struct userspace_mem_region *region,
bool unlink)
{ {
int ret; int ret;
list_del(&region->list); if (unlink) {
rb_erase(&region->gpa_node, &vm->regions.gpa_tree);
rb_erase(&region->hva_node, &vm->regions.hva_tree);
hash_del(&region->slot_node);
}
region->region.memory_size = 0; region->region.memory_size = 0;
ret = ioctl(vm->fd, KVM_SET_USER_MEMORY_REGION, &region->region); ret = ioctl(vm->fd, KVM_SET_USER_MEMORY_REGION, &region->region);
...@@ -569,14 +584,16 @@ static void __vm_mem_region_delete(struct kvm_vm *vm, ...@@ -569,14 +584,16 @@ static void __vm_mem_region_delete(struct kvm_vm *vm,
*/ */
void kvm_vm_free(struct kvm_vm *vmp) void kvm_vm_free(struct kvm_vm *vmp)
{ {
struct userspace_mem_region *region, *tmp; int ctr;
struct hlist_node *node;
struct userspace_mem_region *region;
if (vmp == NULL) if (vmp == NULL)
return; return;
/* Free userspace_mem_regions. */ /* Free userspace_mem_regions. */
list_for_each_entry_safe(region, tmp, &vmp->userspace_mem_regions, list) hash_for_each_safe(vmp->regions.slot_hash, ctr, node, region, slot_node)
__vm_mem_region_delete(vmp, region); __vm_mem_region_delete(vmp, region, false);
/* Free sparsebit arrays. */ /* Free sparsebit arrays. */
sparsebit_free(&vmp->vpages_valid); sparsebit_free(&vmp->vpages_valid);
...@@ -658,6 +675,57 @@ int kvm_memcmp_hva_gva(void *hva, struct kvm_vm *vm, vm_vaddr_t gva, size_t len) ...@@ -658,6 +675,57 @@ int kvm_memcmp_hva_gva(void *hva, struct kvm_vm *vm, vm_vaddr_t gva, size_t len)
return 0; return 0;
} }
static void vm_userspace_mem_region_gpa_insert(struct rb_root *gpa_tree,
struct userspace_mem_region *region)
{
struct rb_node **cur, *parent;
for (cur = &gpa_tree->rb_node, parent = NULL; *cur; ) {
struct userspace_mem_region *cregion;
cregion = container_of(*cur, typeof(*cregion), gpa_node);
parent = *cur;
if (region->region.guest_phys_addr <
cregion->region.guest_phys_addr)
cur = &(*cur)->rb_left;
else {
TEST_ASSERT(region->region.guest_phys_addr !=
cregion->region.guest_phys_addr,
"Duplicate GPA in region tree");
cur = &(*cur)->rb_right;
}
}
rb_link_node(&region->gpa_node, parent, cur);
rb_insert_color(&region->gpa_node, gpa_tree);
}
static void vm_userspace_mem_region_hva_insert(struct rb_root *hva_tree,
struct userspace_mem_region *region)
{
struct rb_node **cur, *parent;
for (cur = &hva_tree->rb_node, parent = NULL; *cur; ) {
struct userspace_mem_region *cregion;
cregion = container_of(*cur, typeof(*cregion), hva_node);
parent = *cur;
if (region->host_mem < cregion->host_mem)
cur = &(*cur)->rb_left;
else {
TEST_ASSERT(region->host_mem !=
cregion->host_mem,
"Duplicate HVA in region tree");
cur = &(*cur)->rb_right;
}
}
rb_link_node(&region->hva_node, parent, cur);
rb_insert_color(&region->hva_node, hva_tree);
}
/* /*
* VM Userspace Memory Region Add * VM Userspace Memory Region Add
* *
...@@ -722,7 +790,8 @@ void vm_userspace_mem_region_add(struct kvm_vm *vm, ...@@ -722,7 +790,8 @@ void vm_userspace_mem_region_add(struct kvm_vm *vm,
(uint64_t) region->region.memory_size); (uint64_t) region->region.memory_size);
/* Confirm no region with the requested slot already exists. */ /* Confirm no region with the requested slot already exists. */
list_for_each_entry(region, &vm->userspace_mem_regions, list) { hash_for_each_possible(vm->regions.slot_hash, region, slot_node,
slot) {
if (region->region.slot != slot) if (region->region.slot != slot)
continue; continue;
...@@ -793,8 +862,10 @@ void vm_userspace_mem_region_add(struct kvm_vm *vm, ...@@ -793,8 +862,10 @@ void vm_userspace_mem_region_add(struct kvm_vm *vm,
ret, errno, slot, flags, ret, errno, slot, flags,
guest_paddr, (uint64_t) region->region.memory_size); guest_paddr, (uint64_t) region->region.memory_size);
/* Add to linked-list of memory regions. */ /* Add to quick lookup data structures */
list_add(&region->list, &vm->userspace_mem_regions); vm_userspace_mem_region_gpa_insert(&vm->regions.gpa_tree, region);
vm_userspace_mem_region_hva_insert(&vm->regions.hva_tree, region);
hash_add(vm->regions.slot_hash, &region->slot_node, slot);
} }
/* /*
...@@ -817,10 +888,10 @@ memslot2region(struct kvm_vm *vm, uint32_t memslot) ...@@ -817,10 +888,10 @@ memslot2region(struct kvm_vm *vm, uint32_t memslot)
{ {
struct userspace_mem_region *region; struct userspace_mem_region *region;
list_for_each_entry(region, &vm->userspace_mem_regions, list) { hash_for_each_possible(vm->regions.slot_hash, region, slot_node,
memslot)
if (region->region.slot == memslot) if (region->region.slot == memslot)
return region; return region;
}
fprintf(stderr, "No mem region with the requested slot found,\n" fprintf(stderr, "No mem region with the requested slot found,\n"
" requested slot: %u\n", memslot); " requested slot: %u\n", memslot);
...@@ -905,7 +976,7 @@ void vm_mem_region_move(struct kvm_vm *vm, uint32_t slot, uint64_t new_gpa) ...@@ -905,7 +976,7 @@ void vm_mem_region_move(struct kvm_vm *vm, uint32_t slot, uint64_t new_gpa)
*/ */
void vm_mem_region_delete(struct kvm_vm *vm, uint32_t slot) void vm_mem_region_delete(struct kvm_vm *vm, uint32_t slot)
{ {
__vm_mem_region_delete(vm, memslot2region(vm, slot)); __vm_mem_region_delete(vm, memslot2region(vm, slot), true);
} }
/* /*
...@@ -1176,16 +1247,14 @@ void *addr_gpa2hva(struct kvm_vm *vm, vm_paddr_t gpa) ...@@ -1176,16 +1247,14 @@ void *addr_gpa2hva(struct kvm_vm *vm, vm_paddr_t gpa)
{ {
struct userspace_mem_region *region; struct userspace_mem_region *region;
list_for_each_entry(region, &vm->userspace_mem_regions, list) { region = userspace_mem_region_find(vm, gpa, gpa);
if ((gpa >= region->region.guest_phys_addr) if (!region) {
&& (gpa <= (region->region.guest_phys_addr TEST_FAIL("No vm physical memory at 0x%lx", gpa);
+ region->region.memory_size - 1))) return NULL;
return (void *) ((uintptr_t) region->host_mem
+ (gpa - region->region.guest_phys_addr));
} }
TEST_FAIL("No vm physical memory at 0x%lx", gpa); return (void *)((uintptr_t)region->host_mem
return NULL; + (gpa - region->region.guest_phys_addr));
} }
/* /*
...@@ -1207,15 +1276,22 @@ void *addr_gpa2hva(struct kvm_vm *vm, vm_paddr_t gpa) ...@@ -1207,15 +1276,22 @@ void *addr_gpa2hva(struct kvm_vm *vm, vm_paddr_t gpa)
*/ */
vm_paddr_t addr_hva2gpa(struct kvm_vm *vm, void *hva) vm_paddr_t addr_hva2gpa(struct kvm_vm *vm, void *hva)
{ {
struct userspace_mem_region *region; struct rb_node *node;
for (node = vm->regions.hva_tree.rb_node; node; ) {
struct userspace_mem_region *region =
container_of(node, struct userspace_mem_region, hva_node);
if (hva >= region->host_mem) {
if (hva <= (region->host_mem
+ region->region.memory_size - 1))
return (vm_paddr_t)((uintptr_t)
region->region.guest_phys_addr
+ (hva - (uintptr_t)region->host_mem));
list_for_each_entry(region, &vm->userspace_mem_regions, list) { node = node->rb_right;
if ((hva >= region->host_mem) } else
&& (hva <= (region->host_mem node = node->rb_left;
+ region->region.memory_size - 1)))
return (vm_paddr_t) ((uintptr_t)
region->region.guest_phys_addr
+ (hva - (uintptr_t) region->host_mem));
} }
TEST_FAIL("No mapping to a guest physical address, hva: %p", hva); TEST_FAIL("No mapping to a guest physical address, hva: %p", hva);
...@@ -1821,6 +1897,7 @@ int kvm_device_access(int dev_fd, uint32_t group, uint64_t attr, ...@@ -1821,6 +1897,7 @@ int kvm_device_access(int dev_fd, uint32_t group, uint64_t attr,
*/ */
void vm_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent) void vm_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent)
{ {
int ctr;
struct userspace_mem_region *region; struct userspace_mem_region *region;
struct vcpu *vcpu; struct vcpu *vcpu;
...@@ -1828,7 +1905,7 @@ void vm_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent) ...@@ -1828,7 +1905,7 @@ void vm_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent)
fprintf(stream, "%*sfd: %i\n", indent, "", vm->fd); fprintf(stream, "%*sfd: %i\n", indent, "", vm->fd);
fprintf(stream, "%*spage_size: 0x%x\n", indent, "", vm->page_size); fprintf(stream, "%*spage_size: 0x%x\n", indent, "", vm->page_size);
fprintf(stream, "%*sMem Regions:\n", indent, ""); fprintf(stream, "%*sMem Regions:\n", indent, "");
list_for_each_entry(region, &vm->userspace_mem_regions, list) { hash_for_each(vm->regions.slot_hash, ctr, region, slot_node) {
fprintf(stream, "%*sguest_phys: 0x%lx size: 0x%lx " fprintf(stream, "%*sguest_phys: 0x%lx size: 0x%lx "
"host_virt: %p\n", indent + 2, "", "host_virt: %p\n", indent + 2, "",
(uint64_t) region->region.guest_phys_addr, (uint64_t) region->region.guest_phys_addr,
......
...@@ -8,6 +8,9 @@ ...@@ -8,6 +8,9 @@
#ifndef SELFTEST_KVM_UTIL_INTERNAL_H #ifndef SELFTEST_KVM_UTIL_INTERNAL_H
#define SELFTEST_KVM_UTIL_INTERNAL_H #define SELFTEST_KVM_UTIL_INTERNAL_H
#include "linux/hashtable.h"
#include "linux/rbtree.h"
#include "sparsebit.h" #include "sparsebit.h"
struct userspace_mem_region { struct userspace_mem_region {
...@@ -18,7 +21,9 @@ struct userspace_mem_region { ...@@ -18,7 +21,9 @@ struct userspace_mem_region {
void *host_mem; void *host_mem;
void *mmap_start; void *mmap_start;
size_t mmap_size; size_t mmap_size;
struct list_head list; struct rb_node gpa_node;
struct rb_node hva_node;
struct hlist_node slot_node;
}; };
struct vcpu { struct vcpu {
...@@ -31,6 +36,12 @@ struct vcpu { ...@@ -31,6 +36,12 @@ struct vcpu {
uint32_t dirty_gfns_count; uint32_t dirty_gfns_count;
}; };
struct userspace_mem_regions {
struct rb_root gpa_tree;
struct rb_root hva_tree;
DECLARE_HASHTABLE(slot_hash, 9);
};
struct kvm_vm { struct kvm_vm {
int mode; int mode;
unsigned long type; unsigned long type;
...@@ -43,7 +54,7 @@ struct kvm_vm { ...@@ -43,7 +54,7 @@ struct kvm_vm {
unsigned int va_bits; unsigned int va_bits;
uint64_t max_gfn; uint64_t max_gfn;
struct list_head vcpus; struct list_head vcpus;
struct list_head userspace_mem_regions; struct userspace_mem_regions regions;
struct sparsebit *vpages_valid; struct sparsebit *vpages_valid;
struct sparsebit *vpages_mapped; struct sparsebit *vpages_mapped;
bool has_irqchip; bool has_irqchip;
......
#include "../../../../lib/rbtree.c"
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment