1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * KVM page table test 4 * 5 * Copyright (C) 2021, Huawei, Inc. 6 * 7 * Make sure that THP has been enabled or enough HUGETLB pages with specific 8 * page size have been pre-allocated on your system, if you are planning to 9 * use hugepages to back the guest memory for testing. 10 */ 11 #include <stdio.h> 12 #include <stdlib.h> 13 #include <time.h> 14 #include <pthread.h> 15 #include <semaphore.h> 16 17 #include "test_util.h" 18 #include "kvm_util.h" 19 #include "processor.h" 20 #include "guest_modes.h" 21 #include "ucall_common.h" 22 23 #define TEST_MEM_SLOT_INDEX 1 24 25 /* Default size(1GB) of the memory for testing */ 26 #define DEFAULT_TEST_MEM_SIZE (1 << 30) 27 28 /* Default guest test virtual memory offset */ 29 #define DEFAULT_GUEST_TEST_MEM 0xc0000000 30 31 /* Different guest memory accessing stages */ 32 enum test_stage { 33 KVM_BEFORE_MAPPINGS, 34 KVM_CREATE_MAPPINGS, 35 KVM_UPDATE_MAPPINGS, 36 KVM_ADJUST_MAPPINGS, 37 NUM_TEST_STAGES, 38 }; 39 40 static const char * const test_stage_string[] = { 41 "KVM_BEFORE_MAPPINGS", 42 "KVM_CREATE_MAPPINGS", 43 "KVM_UPDATE_MAPPINGS", 44 "KVM_ADJUST_MAPPINGS", 45 }; 46 47 struct test_args { 48 struct kvm_vm *vm; 49 uint64_t guest_test_virt_mem; 50 uint64_t host_page_size; 51 uint64_t host_num_pages; 52 uint64_t large_page_size; 53 uint64_t large_num_pages; 54 uint64_t host_pages_per_lpage; 55 enum vm_mem_backing_src_type src_type; 56 struct kvm_vcpu *vcpus[KVM_MAX_VCPUS]; 57 }; 58 59 /* 60 * Guest variables. Use addr_gva2hva() if these variables need 61 * to be changed in host. 62 */ 63 static enum test_stage guest_test_stage; 64 65 /* Host variables */ 66 static uint32_t nr_vcpus = 1; 67 static struct test_args test_args; 68 static enum test_stage *current_stage; 69 static bool host_quit; 70 71 /* Whether the test stage is updated, or completed */ 72 static sem_t test_stage_updated; 73 static sem_t test_stage_completed; 74 75 /* 76 * Guest physical memory offset of the testing memory slot. 77 * This will be set to the topmost valid physical address minus 78 * the test memory size. 79 */ 80 static uint64_t guest_test_phys_mem; 81 82 /* 83 * Guest virtual memory offset of the testing memory slot. 84 * Must not conflict with identity mapped test code. 85 */ 86 static uint64_t guest_test_virt_mem = DEFAULT_GUEST_TEST_MEM; 87 88 static void guest_code(bool do_write) 89 { 90 struct test_args *p = &test_args; 91 enum test_stage *current_stage = &guest_test_stage; 92 uint64_t addr; 93 int i, j; 94 95 while (true) { 96 addr = p->guest_test_virt_mem; 97 98 switch (READ_ONCE(*current_stage)) { 99 /* 100 * All vCPU threads will be started in this stage, 101 * where guest code of each vCPU will do nothing. 102 */ 103 case KVM_BEFORE_MAPPINGS: 104 break; 105 106 /* 107 * Before dirty logging, vCPUs concurrently access the first 108 * 8 bytes of each page (host page/large page) within the same 109 * memory region with different accessing types (read/write). 110 * Then KVM will create normal page mappings or huge block 111 * mappings for them. 112 */ 113 case KVM_CREATE_MAPPINGS: 114 for (i = 0; i < p->large_num_pages; i++) { 115 if (do_write) 116 *(uint64_t *)addr = 0x0123456789ABCDEF; 117 else 118 READ_ONCE(*(uint64_t *)addr); 119 120 addr += p->large_page_size; 121 } 122 break; 123 124 /* 125 * During dirty logging, KVM will only update attributes of the 126 * normal page mappings from RO to RW if memory backing src type 127 * is anonymous. In other cases, KVM will split the huge block 128 * mappings into normal page mappings if memory backing src type 129 * is THP or HUGETLB. 130 */ 131 case KVM_UPDATE_MAPPINGS: 132 if (p->src_type == VM_MEM_SRC_ANONYMOUS) { 133 for (i = 0; i < p->host_num_pages; i++) { 134 *(uint64_t *)addr = 0x0123456789ABCDEF; 135 addr += p->host_page_size; 136 } 137 break; 138 } 139 140 for (i = 0; i < p->large_num_pages; i++) { 141 /* 142 * Write to the first host page in each large 143 * page region, and triger break of large pages. 144 */ 145 *(uint64_t *)addr = 0x0123456789ABCDEF; 146 147 /* 148 * Access the middle host pages in each large 149 * page region. Since dirty logging is enabled, 150 * this will create new mappings at the smallest 151 * granularity. 152 */ 153 addr += p->large_page_size / 2; 154 for (j = 0; j < p->host_pages_per_lpage / 2; j++) { 155 READ_ONCE(*(uint64_t *)addr); 156 addr += p->host_page_size; 157 } 158 } 159 break; 160 161 /* 162 * After dirty logging is stopped, vCPUs concurrently read 163 * from every single host page. Then KVM will coalesce the 164 * split page mappings back to block mappings. And a TLB 165 * conflict abort could occur here if TLB entries of the 166 * page mappings are not fully invalidated. 167 */ 168 case KVM_ADJUST_MAPPINGS: 169 for (i = 0; i < p->host_num_pages; i++) { 170 READ_ONCE(*(uint64_t *)addr); 171 addr += p->host_page_size; 172 } 173 break; 174 175 default: 176 GUEST_ASSERT(0); 177 } 178 179 GUEST_SYNC(1); 180 } 181 } 182 183 static void *vcpu_worker(void *data) 184 { 185 struct kvm_vcpu *vcpu = data; 186 bool do_write = !(vcpu->id % 2); 187 struct timespec start; 188 struct timespec ts_diff; 189 enum test_stage stage; 190 int ret; 191 192 vcpu_args_set(vcpu, 1, do_write); 193 194 while (!READ_ONCE(host_quit)) { 195 ret = sem_wait(&test_stage_updated); 196 TEST_ASSERT(ret == 0, "Error in sem_wait"); 197 198 if (READ_ONCE(host_quit)) 199 return NULL; 200 201 clock_gettime(CLOCK_MONOTONIC, &start); 202 ret = _vcpu_run(vcpu); 203 ts_diff = timespec_elapsed(start); 204 205 TEST_ASSERT(ret == 0, "vcpu_run failed: %d", ret); 206 TEST_ASSERT(get_ucall(vcpu, NULL) == UCALL_SYNC, 207 "Invalid guest sync status: exit_reason=%s", 208 exit_reason_str(vcpu->run->exit_reason)); 209 210 pr_debug("Got sync event from vCPU %d\n", vcpu->id); 211 stage = READ_ONCE(*current_stage); 212 213 /* 214 * Here we can know the execution time of every 215 * single vcpu running in different test stages. 216 */ 217 pr_debug("vCPU %d has completed stage %s\n" 218 "execution time is: %ld.%.9lds\n\n", 219 vcpu->id, test_stage_string[stage], 220 ts_diff.tv_sec, ts_diff.tv_nsec); 221 222 ret = sem_post(&test_stage_completed); 223 TEST_ASSERT(ret == 0, "Error in sem_post"); 224 } 225 226 return NULL; 227 } 228 229 struct test_params { 230 uint64_t phys_offset; 231 uint64_t test_mem_size; 232 enum vm_mem_backing_src_type src_type; 233 }; 234 235 static struct kvm_vm *pre_init_before_test(enum vm_guest_mode mode, void *arg) 236 { 237 int ret; 238 struct test_params *p = arg; 239 enum vm_mem_backing_src_type src_type = p->src_type; 240 uint64_t large_page_size = get_backing_src_pagesz(src_type); 241 uint64_t guest_page_size = vm_guest_mode_params[mode].page_size; 242 uint64_t host_page_size = getpagesize(); 243 uint64_t test_mem_size = p->test_mem_size; 244 uint64_t guest_num_pages; 245 uint64_t alignment; 246 void *host_test_mem; 247 struct kvm_vm *vm; 248 249 /* Align up the test memory size */ 250 alignment = max(large_page_size, guest_page_size); 251 test_mem_size = (test_mem_size + alignment - 1) & ~(alignment - 1); 252 253 /* Create a VM with enough guest pages */ 254 guest_num_pages = test_mem_size / guest_page_size; 255 vm = __vm_create_with_vcpus(VM_SHAPE(mode), nr_vcpus, guest_num_pages, 256 guest_code, test_args.vcpus); 257 258 /* Align down GPA of the testing memslot */ 259 if (!p->phys_offset) 260 guest_test_phys_mem = (vm->max_gfn - guest_num_pages) * 261 guest_page_size; 262 else 263 guest_test_phys_mem = p->phys_offset; 264 #ifdef __s390x__ 265 alignment = max(0x100000UL, alignment); 266 #endif 267 guest_test_phys_mem = align_down(guest_test_phys_mem, alignment); 268 269 /* Set up the shared data structure test_args */ 270 test_args.vm = vm; 271 test_args.guest_test_virt_mem = guest_test_virt_mem; 272 test_args.host_page_size = host_page_size; 273 test_args.host_num_pages = test_mem_size / host_page_size; 274 test_args.large_page_size = large_page_size; 275 test_args.large_num_pages = test_mem_size / large_page_size; 276 test_args.host_pages_per_lpage = large_page_size / host_page_size; 277 test_args.src_type = src_type; 278 279 /* Add an extra memory slot with specified backing src type */ 280 vm_userspace_mem_region_add(vm, src_type, guest_test_phys_mem, 281 TEST_MEM_SLOT_INDEX, guest_num_pages, 0); 282 283 /* Do mapping(GVA->GPA) for the testing memory slot */ 284 virt_map(vm, guest_test_virt_mem, guest_test_phys_mem, guest_num_pages); 285 286 /* Cache the HVA pointer of the region */ 287 host_test_mem = addr_gpa2hva(vm, (vm_paddr_t)guest_test_phys_mem); 288 289 /* Export shared structure test_args to guest */ 290 sync_global_to_guest(vm, test_args); 291 292 ret = sem_init(&test_stage_updated, 0, 0); 293 TEST_ASSERT(ret == 0, "Error in sem_init"); 294 295 ret = sem_init(&test_stage_completed, 0, 0); 296 TEST_ASSERT(ret == 0, "Error in sem_init"); 297 298 current_stage = addr_gva2hva(vm, (vm_vaddr_t)(&guest_test_stage)); 299 *current_stage = NUM_TEST_STAGES; 300 301 pr_info("Testing guest mode: %s\n", vm_guest_mode_string(mode)); 302 pr_info("Testing memory backing src type: %s\n", 303 vm_mem_backing_src_alias(src_type)->name); 304 pr_info("Testing memory backing src granularity: 0x%lx\n", 305 large_page_size); 306 pr_info("Testing memory size(aligned): 0x%lx\n", test_mem_size); 307 pr_info("Guest physical test memory offset: 0x%lx\n", 308 guest_test_phys_mem); 309 pr_info("Host virtual test memory offset: 0x%lx\n", 310 (uint64_t)host_test_mem); 311 pr_info("Number of testing vCPUs: %d\n", nr_vcpus); 312 313 return vm; 314 } 315 316 static void vcpus_complete_new_stage(enum test_stage stage) 317 { 318 int ret; 319 int vcpus; 320 321 /* Wake up all the vcpus to run new test stage */ 322 for (vcpus = 0; vcpus < nr_vcpus; vcpus++) { 323 ret = sem_post(&test_stage_updated); 324 TEST_ASSERT(ret == 0, "Error in sem_post"); 325 } 326 pr_debug("All vcpus have been notified to continue\n"); 327 328 /* Wait for all the vcpus to complete new test stage */ 329 for (vcpus = 0; vcpus < nr_vcpus; vcpus++) { 330 ret = sem_wait(&test_stage_completed); 331 TEST_ASSERT(ret == 0, "Error in sem_wait"); 332 333 pr_debug("%d vcpus have completed stage %s\n", 334 vcpus + 1, test_stage_string[stage]); 335 } 336 337 pr_debug("All vcpus have completed stage %s\n", 338 test_stage_string[stage]); 339 } 340 341 static void run_test(enum vm_guest_mode mode, void *arg) 342 { 343 pthread_t *vcpu_threads; 344 struct kvm_vm *vm; 345 struct timespec start; 346 struct timespec ts_diff; 347 int ret, i; 348 349 /* Create VM with vCPUs and make some pre-initialization */ 350 vm = pre_init_before_test(mode, arg); 351 352 vcpu_threads = malloc(nr_vcpus * sizeof(*vcpu_threads)); 353 TEST_ASSERT(vcpu_threads, "Memory allocation failed"); 354 355 host_quit = false; 356 *current_stage = KVM_BEFORE_MAPPINGS; 357 358 for (i = 0; i < nr_vcpus; i++) 359 pthread_create(&vcpu_threads[i], NULL, vcpu_worker, 360 test_args.vcpus[i]); 361 362 vcpus_complete_new_stage(*current_stage); 363 pr_info("Started all vCPUs successfully\n"); 364 365 /* Test the stage of KVM creating mappings */ 366 *current_stage = KVM_CREATE_MAPPINGS; 367 368 clock_gettime(CLOCK_MONOTONIC, &start); 369 vcpus_complete_new_stage(*current_stage); 370 ts_diff = timespec_elapsed(start); 371 372 pr_info("KVM_CREATE_MAPPINGS: total execution time: %ld.%.9lds\n\n", 373 ts_diff.tv_sec, ts_diff.tv_nsec); 374 375 /* Test the stage of KVM updating mappings */ 376 vm_mem_region_set_flags(vm, TEST_MEM_SLOT_INDEX, 377 KVM_MEM_LOG_DIRTY_PAGES); 378 379 *current_stage = KVM_UPDATE_MAPPINGS; 380 381 clock_gettime(CLOCK_MONOTONIC, &start); 382 vcpus_complete_new_stage(*current_stage); 383 ts_diff = timespec_elapsed(start); 384 385 pr_info("KVM_UPDATE_MAPPINGS: total execution time: %ld.%.9lds\n\n", 386 ts_diff.tv_sec, ts_diff.tv_nsec); 387 388 /* Test the stage of KVM adjusting mappings */ 389 vm_mem_region_set_flags(vm, TEST_MEM_SLOT_INDEX, 0); 390 391 *current_stage = KVM_ADJUST_MAPPINGS; 392 393 clock_gettime(CLOCK_MONOTONIC, &start); 394 vcpus_complete_new_stage(*current_stage); 395 ts_diff = timespec_elapsed(start); 396 397 pr_info("KVM_ADJUST_MAPPINGS: total execution time: %ld.%.9lds\n\n", 398 ts_diff.tv_sec, ts_diff.tv_nsec); 399 400 /* Tell the vcpu thread to quit */ 401 host_quit = true; 402 for (i = 0; i < nr_vcpus; i++) { 403 ret = sem_post(&test_stage_updated); 404 TEST_ASSERT(ret == 0, "Error in sem_post"); 405 } 406 407 for (i = 0; i < nr_vcpus; i++) 408 pthread_join(vcpu_threads[i], NULL); 409 410 ret = sem_destroy(&test_stage_updated); 411 TEST_ASSERT(ret == 0, "Error in sem_destroy"); 412 413 ret = sem_destroy(&test_stage_completed); 414 TEST_ASSERT(ret == 0, "Error in sem_destroy"); 415 416 free(vcpu_threads); 417 kvm_vm_free(vm); 418 } 419 420 static void help(char *name) 421 { 422 puts(""); 423 printf("usage: %s [-h] [-p offset] [-m mode] " 424 "[-b mem-size] [-v vcpus] [-s mem-type]\n", name); 425 puts(""); 426 printf(" -p: specify guest physical test memory offset\n" 427 " Warning: a low offset can conflict with the loaded test code.\n"); 428 guest_modes_help(); 429 printf(" -b: specify size of the memory region for testing. e.g. 10M or 3G.\n" 430 " (default: 1G)\n"); 431 printf(" -v: specify the number of vCPUs to run\n" 432 " (default: 1)\n"); 433 backing_src_help("-s"); 434 puts(""); 435 } 436 437 int main(int argc, char *argv[]) 438 { 439 int max_vcpus = kvm_check_cap(KVM_CAP_MAX_VCPUS); 440 struct test_params p = { 441 .test_mem_size = DEFAULT_TEST_MEM_SIZE, 442 .src_type = DEFAULT_VM_MEM_SRC, 443 }; 444 int opt; 445 446 guest_modes_append_default(); 447 448 while ((opt = getopt(argc, argv, "hp:m:b:v:s:")) != -1) { 449 switch (opt) { 450 case 'p': 451 p.phys_offset = strtoull(optarg, NULL, 0); 452 break; 453 case 'm': 454 guest_modes_cmdline(optarg); 455 break; 456 case 'b': 457 p.test_mem_size = parse_size(optarg); 458 break; 459 case 'v': 460 nr_vcpus = atoi_positive("Number of vCPUs", optarg); 461 TEST_ASSERT(nr_vcpus <= max_vcpus, 462 "Invalid number of vcpus, must be between 1 and %d", max_vcpus); 463 break; 464 case 's': 465 p.src_type = parse_backing_src_type(optarg); 466 break; 467 case 'h': 468 default: 469 help(argv[0]); 470 exit(0); 471 } 472 } 473 474 for_each_guest_mode(run_test, &p); 475 476 return 0; 477 } 478