[PATCH v3 5/8] KVM: selftests: Pass args to vCPU instead of using globals

From: Ben Gardon
Date: Mon Dec 16 2019 - 16:39:23 EST


In preparation for supporting multiple vCPUs in the demand paging test,
pass arguments to the vCPU instead of syncing globals to it.

Signed-off-by: Ben Gardon <bgardon@xxxxxxxxxx>
---
.../selftests/kvm/demand_paging_test.c | 61 +++++++++++--------
1 file changed, 37 insertions(+), 24 deletions(-)

diff --git a/tools/testing/selftests/kvm/demand_paging_test.c b/tools/testing/selftests/kvm/demand_paging_test.c
index 4aa90a3fce99c..8ede26e088ab6 100644
--- a/tools/testing/selftests/kvm/demand_paging_test.c
+++ b/tools/testing/selftests/kvm/demand_paging_test.c
@@ -42,7 +42,6 @@
*/
static uint64_t host_page_size;
static uint64_t guest_page_size;
-static uint64_t guest_num_pages;

static char *guest_data_prototype;

@@ -63,14 +62,13 @@ static uint64_t guest_test_virt_mem = DEFAULT_GUEST_TEST_MEM;
* Continuously write to the first 8 bytes of each page in the demand paging
* memory region.
*/
-static void guest_code(void)
+static void guest_code(uint64_t gva, uint64_t pages)
{
int i;

- for (i = 0; i < guest_num_pages; i++) {
- uint64_t addr = guest_test_virt_mem;
+ for (i = 0; i < pages; i++) {
+ uint64_t addr = gva + (i * guest_page_size);

- addr += i * guest_page_size;
addr &= ~(host_page_size - 1);
*(uint64_t *)addr = 0x0123456789ABCDEF;
}
@@ -82,18 +80,31 @@ static void guest_code(void)
static void *host_test_mem;
static uint64_t host_num_pages;

+struct vcpu_thread_args {
+ uint64_t gva;
+ uint64_t pages;
+ struct kvm_vm *vm;
+ int vcpu_id;
+};
+
static void *vcpu_worker(void *data)
{
int ret;
- struct kvm_vm *vm = data;
+ struct vcpu_thread_args *args = (struct vcpu_thread_args *)data;
+ struct kvm_vm *vm = args->vm;
+ int vcpu_id = args->vcpu_id;
+ uint64_t gva = args->gva;
+ uint64_t pages = args->pages;
struct kvm_run *run;

- run = vcpu_state(vm, VCPU_ID);
+ vcpu_args_set(vm, vcpu_id, 2, gva, pages);
+
+ run = vcpu_state(vm, vcpu_id);

/* Let the guest access its memory */
- ret = _vcpu_run(vm, VCPU_ID);
+ ret = _vcpu_run(vm, vcpu_id);
TEST_ASSERT(ret == 0, "vcpu_run failed: %d\n", ret);
- if (get_ucall(vm, VCPU_ID, NULL) != UCALL_SYNC) {
+ if (get_ucall(vm, vcpu_id, NULL) != UCALL_SYNC) {
TEST_ASSERT(false,
"Invalid guest sync status: exit_reason=%s\n",
exit_reason_str(run->exit_reason));
@@ -269,11 +280,13 @@ static int setup_demand_paging(struct kvm_vm *vm,
#define PAGE_SHIFT_4K 12

static void run_test(enum vm_guest_mode mode, bool use_uffd,
- useconds_t uffd_delay, uint64_t guest_memory_bytes)
+ useconds_t uffd_delay, uint64_t vcpu_wss)
{
pthread_t vcpu_thread;
pthread_t uffd_handler_thread;
struct kvm_vm *vm;
+ struct vcpu_thread_args vcpu_args;
+ uint64_t guest_num_pages;
int r;

/*
@@ -283,16 +296,15 @@ static void run_test(enum vm_guest_mode mode, bool use_uffd,
* number will be enough for all archs. (e.g., 64K page size guest
* will need even less memory for page tables).
*/
- vm = create_vm(mode, VCPU_ID,
- (2 * guest_memory_bytes) >> PAGE_SHIFT_4K,
+ vm = create_vm(mode, VCPU_ID, (2 * vcpu_wss) >> PAGE_SHIFT_4K,
guest_code);

guest_page_size = vm_get_page_size(vm);

- TEST_ASSERT(guest_memory_bytes % guest_page_size == 0,
+ TEST_ASSERT(vcpu_wss % guest_page_size == 0,
"Guest memory size is not guest page size aligned.");

- guest_num_pages = guest_memory_bytes / guest_page_size;
+ guest_num_pages = vcpu_wss / guest_page_size;

#ifdef __s390x__
/* Round up to multiple of 1M (segment size) */
@@ -308,9 +320,9 @@ static void run_test(enum vm_guest_mode mode, bool use_uffd,
guest_num_pages, vm_get_max_gfn(vm));

host_page_size = getpagesize();
- TEST_ASSERT(guest_memory_bytes % host_page_size == 0,
+ TEST_ASSERT(vcpu_wss % host_page_size == 0,
"Guest memory size is not host page size aligned.");
- host_num_pages = guest_memory_bytes / host_page_size;
+ host_num_pages = vcpu_wss / host_page_size;

guest_test_phys_mem = (vm_get_max_gfn(vm) - guest_num_pages) *
guest_page_size;
@@ -354,10 +366,12 @@ static void run_test(enum vm_guest_mode mode, bool use_uffd,
/* Export the shared variables to the guest */
sync_global_to_guest(vm, host_page_size);
sync_global_to_guest(vm, guest_page_size);
- sync_global_to_guest(vm, guest_test_virt_mem);
- sync_global_to_guest(vm, guest_num_pages);

- pthread_create(&vcpu_thread, NULL, vcpu_worker, vm);
+ vcpu_args.vm = vm;
+ vcpu_args.vcpu_id = VCPU_ID;
+ vcpu_args.gva = guest_test_virt_mem;
+ vcpu_args.pages = guest_num_pages;
+ pthread_create(&vcpu_thread, NULL, vcpu_worker, &vcpu_args);

/* Wait for the vcpu thread to quit */
pthread_join(vcpu_thread, NULL);
@@ -404,8 +418,7 @@ static void help(char *name)
printf(" -d: add a delay in usec to the User Fault\n"
" FD handler to simulate demand paging\n"
" overheads. Ignored without -u.\n");
- printf(" -b: specify the number of bytes of memory which should be\n"
- " allocated to the guest.\n");
+ printf(" -b: specify the working set size, in bytes for each vCPU.\n");
puts("");
exit(0);
}
@@ -413,7 +426,7 @@ static void help(char *name)
int main(int argc, char *argv[])
{
bool mode_selected = false;
- uint64_t guest_memory_bytes = DEFAULT_GUEST_TEST_MEM_SIZE;
+ uint64_t vcpu_wss = DEFAULT_GUEST_TEST_MEM_SIZE;
unsigned int mode;
int opt, i;
bool use_uffd = false;
@@ -448,7 +461,7 @@ int main(int argc, char *argv[])
"A negative UFFD delay is not supported.");
break;
case 'b':
- guest_memory_bytes = strtoull(optarg, NULL, 0);
+ vcpu_wss = strtoull(optarg, NULL, 0);
case 'h':
default:
help(argv[0]);
@@ -462,7 +475,7 @@ int main(int argc, char *argv[])
TEST_ASSERT(vm_guest_mode_params[i].supported,
"Guest mode ID %d (%s) not supported.",
i, vm_guest_mode_string(i));
- run_test(i, use_uffd, uffd_delay, guest_memory_bytes);
+ run_test(i, use_uffd, uffd_delay, vcpu_wss);
}

return 0;
--
2.24.1.735.g03f4e72817-goog