KVM: selftests: Use kernel's list instead of homebrewed replacement
Replace the KVM selftests' homebrewed linked lists for vCPUs and memory regions with the kernel's 'struct list_head'. Signed-off-by: Sean Christopherson <sean.j.christopherson@intel.com> Reviewed-by: Andrew Jones <drjones@redhat.com> Message-Id: <20200410231707.7128-3-sean.j.christopherson@intel.com> Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
This commit is contained in:
parent
238022ff5d
commit
4d9bba9007
@ -10,6 +10,7 @@
|
||||
#include "test_util.h"
|
||||
|
||||
#include "asm/kvm.h"
|
||||
#include "linux/list.h"
|
||||
#include "linux/kvm.h"
|
||||
#include <sys/ioctl.h>
|
||||
|
||||
|
@ -161,6 +161,9 @@ struct kvm_vm *_vm_create(enum vm_guest_mode mode, uint64_t phy_pages, int perm)
|
||||
vm = calloc(1, sizeof(*vm));
|
||||
TEST_ASSERT(vm != NULL, "Insufficient Memory");
|
||||
|
||||
INIT_LIST_HEAD(&vm->vcpus);
|
||||
INIT_LIST_HEAD(&vm->userspace_mem_regions);
|
||||
|
||||
vm->mode = mode;
|
||||
vm->type = 0;
|
||||
|
||||
@ -258,8 +261,7 @@ void kvm_vm_restart(struct kvm_vm *vmp, int perm)
|
||||
if (vmp->has_irqchip)
|
||||
vm_create_irqchip(vmp);
|
||||
|
||||
for (region = vmp->userspace_mem_region_head; region;
|
||||
region = region->next) {
|
||||
list_for_each_entry(region, &vmp->userspace_mem_regions, list) {
|
||||
int ret = ioctl(vmp->fd, KVM_SET_USER_MEMORY_REGION, ®ion->region);
|
||||
TEST_ASSERT(ret == 0, "KVM_SET_USER_MEMORY_REGION IOCTL failed,\n"
|
||||
" rc: %i errno: %i\n"
|
||||
@ -319,8 +321,7 @@ userspace_mem_region_find(struct kvm_vm *vm, uint64_t start, uint64_t end)
|
||||
{
|
||||
struct userspace_mem_region *region;
|
||||
|
||||
for (region = vm->userspace_mem_region_head; region;
|
||||
region = region->next) {
|
||||
list_for_each_entry(region, &vm->userspace_mem_regions, list) {
|
||||
uint64_t existing_start = region->region.guest_phys_addr;
|
||||
uint64_t existing_end = region->region.guest_phys_addr
|
||||
+ region->region.memory_size - 1;
|
||||
@ -378,11 +379,11 @@ kvm_userspace_memory_region_find(struct kvm_vm *vm, uint64_t start,
|
||||
*/
|
||||
struct vcpu *vcpu_find(struct kvm_vm *vm, uint32_t vcpuid)
|
||||
{
|
||||
struct vcpu *vcpup;
|
||||
struct vcpu *vcpu;
|
||||
|
||||
for (vcpup = vm->vcpu_head; vcpup; vcpup = vcpup->next) {
|
||||
if (vcpup->id == vcpuid)
|
||||
return vcpup;
|
||||
list_for_each_entry(vcpu, &vm->vcpus, list) {
|
||||
if (vcpu->id == vcpuid)
|
||||
return vcpu;
|
||||
}
|
||||
|
||||
return NULL;
|
||||
@ -392,16 +393,15 @@ struct vcpu *vcpu_find(struct kvm_vm *vm, uint32_t vcpuid)
|
||||
* VM VCPU Remove
|
||||
*
|
||||
* Input Args:
|
||||
* vm - Virtual Machine
|
||||
* vcpu - VCPU to remove
|
||||
*
|
||||
* Output Args: None
|
||||
*
|
||||
* Return: None, TEST_ASSERT failures for all error conditions
|
||||
*
|
||||
* Within the VM specified by vm, removes the VCPU given by vcpuid.
|
||||
* Removes a vCPU from a VM and frees its resources.
|
||||
*/
|
||||
static void vm_vcpu_rm(struct kvm_vm *vm, struct vcpu *vcpu)
|
||||
static void vm_vcpu_rm(struct vcpu *vcpu)
|
||||
{
|
||||
int ret;
|
||||
|
||||
@ -412,21 +412,17 @@ static void vm_vcpu_rm(struct kvm_vm *vm, struct vcpu *vcpu)
|
||||
TEST_ASSERT(ret == 0, "Close of VCPU fd failed, rc: %i "
|
||||
"errno: %i", ret, errno);
|
||||
|
||||
if (vcpu->next)
|
||||
vcpu->next->prev = vcpu->prev;
|
||||
if (vcpu->prev)
|
||||
vcpu->prev->next = vcpu->next;
|
||||
else
|
||||
vm->vcpu_head = vcpu->next;
|
||||
list_del(&vcpu->list);
|
||||
free(vcpu);
|
||||
}
|
||||
|
||||
void kvm_vm_release(struct kvm_vm *vmp)
|
||||
{
|
||||
struct vcpu *vcpu, *tmp;
|
||||
int ret;
|
||||
|
||||
while (vmp->vcpu_head)
|
||||
vm_vcpu_rm(vmp, vmp->vcpu_head);
|
||||
list_for_each_entry_safe(vcpu, tmp, &vmp->vcpus, list)
|
||||
vm_vcpu_rm(vcpu);
|
||||
|
||||
ret = close(vmp->fd);
|
||||
TEST_ASSERT(ret == 0, "Close of vm fd failed,\n"
|
||||
@ -442,15 +438,15 @@ void kvm_vm_release(struct kvm_vm *vmp)
|
||||
*/
|
||||
void kvm_vm_free(struct kvm_vm *vmp)
|
||||
{
|
||||
struct userspace_mem_region *region, *tmp;
|
||||
int ret;
|
||||
|
||||
if (vmp == NULL)
|
||||
return;
|
||||
|
||||
/* Free userspace_mem_regions. */
|
||||
while (vmp->userspace_mem_region_head) {
|
||||
struct userspace_mem_region *region
|
||||
= vmp->userspace_mem_region_head;
|
||||
list_for_each_entry_safe(region, tmp, &vmp->userspace_mem_regions, list) {
|
||||
list_del(®ion->list);
|
||||
|
||||
region->region.memory_size = 0;
|
||||
ret = ioctl(vmp->fd, KVM_SET_USER_MEMORY_REGION,
|
||||
@ -458,7 +454,6 @@ void kvm_vm_free(struct kvm_vm *vmp)
|
||||
TEST_ASSERT(ret == 0, "KVM_SET_USER_MEMORY_REGION IOCTL failed, "
|
||||
"rc: %i errno: %i", ret, errno);
|
||||
|
||||
vmp->userspace_mem_region_head = region->next;
|
||||
sparsebit_free(®ion->unused_phy_pages);
|
||||
ret = munmap(region->mmap_start, region->mmap_size);
|
||||
TEST_ASSERT(ret == 0, "munmap failed, rc: %i errno: %i",
|
||||
@ -611,12 +606,10 @@ void vm_userspace_mem_region_add(struct kvm_vm *vm,
|
||||
(uint64_t) region->region.memory_size);
|
||||
|
||||
/* Confirm no region with the requested slot already exists. */
|
||||
for (region = vm->userspace_mem_region_head; region;
|
||||
region = region->next) {
|
||||
if (region->region.slot == slot)
|
||||
break;
|
||||
}
|
||||
if (region != NULL)
|
||||
list_for_each_entry(region, &vm->userspace_mem_regions, list) {
|
||||
if (region->region.slot != slot)
|
||||
continue;
|
||||
|
||||
TEST_FAIL("A mem region with the requested slot "
|
||||
"already exists.\n"
|
||||
" requested slot: %u paddr: 0x%lx npages: 0x%lx\n"
|
||||
@ -625,6 +618,7 @@ void vm_userspace_mem_region_add(struct kvm_vm *vm,
|
||||
region->region.slot,
|
||||
(uint64_t) region->region.guest_phys_addr,
|
||||
(uint64_t) region->region.memory_size);
|
||||
}
|
||||
|
||||
/* Allocate and initialize new mem region structure. */
|
||||
region = calloc(1, sizeof(*region));
|
||||
@ -685,10 +679,7 @@ void vm_userspace_mem_region_add(struct kvm_vm *vm,
|
||||
guest_paddr, (uint64_t) region->region.memory_size);
|
||||
|
||||
/* Add to linked-list of memory regions. */
|
||||
if (vm->userspace_mem_region_head)
|
||||
vm->userspace_mem_region_head->prev = region;
|
||||
region->next = vm->userspace_mem_region_head;
|
||||
vm->userspace_mem_region_head = region;
|
||||
list_add(®ion->list, &vm->userspace_mem_regions);
|
||||
}
|
||||
|
||||
/*
|
||||
@ -711,20 +702,17 @@ memslot2region(struct kvm_vm *vm, uint32_t memslot)
|
||||
{
|
||||
struct userspace_mem_region *region;
|
||||
|
||||
for (region = vm->userspace_mem_region_head; region;
|
||||
region = region->next) {
|
||||
list_for_each_entry(region, &vm->userspace_mem_regions, list) {
|
||||
if (region->region.slot == memslot)
|
||||
break;
|
||||
}
|
||||
if (region == NULL) {
|
||||
fprintf(stderr, "No mem region with the requested slot found,\n"
|
||||
" requested slot: %u\n", memslot);
|
||||
fputs("---- vm dump ----\n", stderr);
|
||||
vm_dump(stderr, vm, 2);
|
||||
TEST_FAIL("Mem region not found");
|
||||
return region;
|
||||
}
|
||||
|
||||
return region;
|
||||
fprintf(stderr, "No mem region with the requested slot found,\n"
|
||||
" requested slot: %u\n", memslot);
|
||||
fputs("---- vm dump ----\n", stderr);
|
||||
vm_dump(stderr, vm, 2);
|
||||
TEST_FAIL("Mem region not found");
|
||||
return NULL;
|
||||
}
|
||||
|
||||
/*
|
||||
@ -862,10 +850,7 @@ void vm_vcpu_add(struct kvm_vm *vm, uint32_t vcpuid)
|
||||
"vcpu id: %u errno: %i", vcpuid, errno);
|
||||
|
||||
/* Add to linked-list of VCPUs. */
|
||||
if (vm->vcpu_head)
|
||||
vm->vcpu_head->prev = vcpu;
|
||||
vcpu->next = vm->vcpu_head;
|
||||
vm->vcpu_head = vcpu;
|
||||
list_add(&vcpu->list, &vm->vcpus);
|
||||
}
|
||||
|
||||
/*
|
||||
@ -1058,8 +1043,8 @@ void virt_map(struct kvm_vm *vm, uint64_t vaddr, uint64_t paddr,
|
||||
void *addr_gpa2hva(struct kvm_vm *vm, vm_paddr_t gpa)
|
||||
{
|
||||
struct userspace_mem_region *region;
|
||||
for (region = vm->userspace_mem_region_head; region;
|
||||
region = region->next) {
|
||||
|
||||
list_for_each_entry(region, &vm->userspace_mem_regions, list) {
|
||||
if ((gpa >= region->region.guest_phys_addr)
|
||||
&& (gpa <= (region->region.guest_phys_addr
|
||||
+ region->region.memory_size - 1)))
|
||||
@ -1091,8 +1076,8 @@ void *addr_gpa2hva(struct kvm_vm *vm, vm_paddr_t gpa)
|
||||
vm_paddr_t addr_hva2gpa(struct kvm_vm *vm, void *hva)
|
||||
{
|
||||
struct userspace_mem_region *region;
|
||||
for (region = vm->userspace_mem_region_head; region;
|
||||
region = region->next) {
|
||||
|
||||
list_for_each_entry(region, &vm->userspace_mem_regions, list) {
|
||||
if ((hva >= region->host_mem)
|
||||
&& (hva <= (region->host_mem
|
||||
+ region->region.memory_size - 1)))
|
||||
@ -1519,8 +1504,7 @@ void vm_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent)
|
||||
fprintf(stream, "%*sfd: %i\n", indent, "", vm->fd);
|
||||
fprintf(stream, "%*spage_size: 0x%x\n", indent, "", vm->page_size);
|
||||
fprintf(stream, "%*sMem Regions:\n", indent, "");
|
||||
for (region = vm->userspace_mem_region_head; region;
|
||||
region = region->next) {
|
||||
list_for_each_entry(region, &vm->userspace_mem_regions, list) {
|
||||
fprintf(stream, "%*sguest_phys: 0x%lx size: 0x%lx "
|
||||
"host_virt: %p\n", indent + 2, "",
|
||||
(uint64_t) region->region.guest_phys_addr,
|
||||
@ -1539,7 +1523,7 @@ void vm_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent)
|
||||
virt_dump(stream, vm, indent + 4);
|
||||
}
|
||||
fprintf(stream, "%*sVCPUs:\n", indent, "");
|
||||
for (vcpu = vm->vcpu_head; vcpu; vcpu = vcpu->next)
|
||||
list_for_each_entry(vcpu, &vm->vcpus, list)
|
||||
vcpu_dump(stream, vm, vcpu->id, indent + 2);
|
||||
}
|
||||
|
||||
|
@ -13,7 +13,6 @@
|
||||
#define KVM_DEV_PATH "/dev/kvm"
|
||||
|
||||
struct userspace_mem_region {
|
||||
struct userspace_mem_region *next, *prev;
|
||||
struct kvm_userspace_memory_region region;
|
||||
struct sparsebit *unused_phy_pages;
|
||||
int fd;
|
||||
@ -21,10 +20,11 @@ struct userspace_mem_region {
|
||||
void *host_mem;
|
||||
void *mmap_start;
|
||||
size_t mmap_size;
|
||||
struct list_head list;
|
||||
};
|
||||
|
||||
struct vcpu {
|
||||
struct vcpu *next, *prev;
|
||||
struct list_head list;
|
||||
uint32_t id;
|
||||
int fd;
|
||||
struct kvm_run *state;
|
||||
@ -41,8 +41,8 @@ struct kvm_vm {
|
||||
unsigned int pa_bits;
|
||||
unsigned int va_bits;
|
||||
uint64_t max_gfn;
|
||||
struct vcpu *vcpu_head;
|
||||
struct userspace_mem_region *userspace_mem_region_head;
|
||||
struct list_head vcpus;
|
||||
struct list_head userspace_mem_regions;
|
||||
struct sparsebit *vpages_valid;
|
||||
struct sparsebit *vpages_mapped;
|
||||
bool has_irqchip;
|
||||
|
@ -233,7 +233,10 @@ void vcpu_args_set(struct kvm_vm *vm, uint32_t vcpuid, unsigned int num, ...)
|
||||
|
||||
void vcpu_dump(FILE *stream, struct kvm_vm *vm, uint32_t vcpuid, uint8_t indent)
|
||||
{
|
||||
struct vcpu *vcpu = vm->vcpu_head;
|
||||
struct vcpu *vcpu = vcpu_find(vm, vcpuid);
|
||||
|
||||
if (!vcpu)
|
||||
return;
|
||||
|
||||
fprintf(stream, "%*spstate: psw: 0x%.16llx:0x%.16llx\n",
|
||||
indent, "", vcpu->state->psw_mask, vcpu->state->psw_addr);
|
||||
|
Loading…
x
Reference in New Issue
Block a user