| /linux/tools/testing/selftests/kvm/ |
| H A D | arch_timer.c | 49 unsigned int vcpu_idx = (unsigned long)arg; in test_vcpu_run() local 51 struct kvm_vcpu *vcpu = vcpus[vcpu_idx]; in test_vcpu_run() 53 struct test_vcpu_shared_data *shared_data = &vcpu_shared_data[vcpu_idx]; in test_vcpu_run() 59 __set_bit(vcpu_idx, vcpu_done_map); in test_vcpu_run() 69 vcpu_idx, shared_data->guest_stage, shared_data->nr_iter); in test_vcpu_run() 76 pr_info("PASS(vCPU-%d).\n", vcpu_idx); in test_vcpu_run() 98 static int test_migrate_vcpu(unsigned int vcpu_idx) in test_migrate_vcpu() argument 103 pr_debug("Migrating vCPU: %u to pCPU: %u\n", vcpu_idx, new_pcpu); in test_migrate_vcpu() 105 ret = __pin_task_to_cpu(pt_vcpu_run[vcpu_idx], new_pcpu); in test_migrate_vcpu() 110 vcpu_idx, new_pcpu, ret); in test_migrate_vcpu()
|
| H A D | access_tracking_perf_test.c | 157 static void too_many_idle_pages(long idle_pages, long total_pages, int vcpu_idx) in too_many_idle_pages() argument 161 if (vcpu_idx >= 0) in too_many_idle_pages() 162 snprintf(prefix, 18, "vCPU%d: ", vcpu_idx); in too_many_idle_pages() 176 int vcpu_idx = vcpu_args->vcpu_idx; in pageidle_mark_vcpu_memory_idle() local 186 if (overlap_memory_access && vcpu_idx) in pageidle_mark_vcpu_memory_idle() 218 vcpu_idx, no_pfn, pages); in pageidle_mark_vcpu_memory_idle() 230 overlap_memory_access ? -1 : vcpu_idx); in pageidle_mark_vcpu_memory_idle() 329 int vcpu_idx = vcpu_args->vcpu_idx; in vcpu_thread_main() local 343 vcpu_last_completed_iteration[vcpu_idx] = current_iteration; in vcpu_thread_main() 347 static void spin_wait_for_vcpu(int vcpu_idx, int target_iteration) in spin_wait_for_vcpu() argument [all …]
|
| H A D | dirty_log_perf_test.c | 39 int vcpu_idx = vcpu_args->vcpu_idx; in vcpu_worker() local 62 pr_debug("Got sync event from vCPU %d\n", vcpu_idx); in vcpu_worker() 63 vcpu_last_completed_iteration[vcpu_idx] = current_iteration; in vcpu_worker() 65 vcpu_idx, vcpu_last_completed_iteration[vcpu_idx]); in vcpu_worker() 71 vcpu_idx, current_iteration, ts_diff.tv_sec, in vcpu_worker() 75 vcpu_idx, current_iteration, ts_diff.tv_sec, in vcpu_worker() 88 avg = timespec_div(total, vcpu_last_completed_iteration[vcpu_idx]); in vcpu_worker() 90 vcpu_idx, pages_count, vcpu_last_completed_iteration[vcpu_idx], in vcpu_worker()
|
| H A D | steal_time.c | 79 static void steal_time_dump(struct kvm_vm *vm, u32 vcpu_idx) in steal_time_dump() argument 81 struct kvm_steal_time *st = addr_gva2hva(vm, (ulong)st_gva[vcpu_idx]); in steal_time_dump() 83 ksft_print_msg("VCPU%d:\n", vcpu_idx); in steal_time_dump() 197 static void steal_time_dump(struct kvm_vm *vm, u32 vcpu_idx) in steal_time_dump() argument 199 struct st_time *st = addr_gva2hva(vm, (ulong)st_gva[vcpu_idx]); in steal_time_dump() 201 ksft_print_msg("VCPU%d:\n", vcpu_idx); in steal_time_dump() 317 static void steal_time_dump(struct kvm_vm *vm, u32 vcpu_idx) in steal_time_dump() argument 319 struct sta_struct *st = addr_gva2hva(vm, (ulong)st_gva[vcpu_idx]); in steal_time_dump() 322 pr_info("VCPU%d:\n", vcpu_idx); in steal_time_dump() 454 static void steal_time_dump(struct kvm_vm *vm, u32 vcpu_idx) in steal_time_dump() argument [all …]
|
| H A D | demand_paging_test.c | 35 int vcpu_idx = vcpu_args->vcpu_idx; in vcpu_worker() local 53 PER_VCPU_DEBUG("vCPU %d execution time: %ld.%.9lds\n", vcpu_idx, in vcpu_worker()
|
| /linux/tools/testing/selftests/kvm/lib/ |
| H A D | memstress.c | 23 int vcpu_idx; member 47 void memstress_guest_code(u32 vcpu_idx) in memstress_guest_code() argument 50 struct memstress_vcpu_args *vcpu_args = &args->vcpu_args[vcpu_idx]; in memstress_guest_code() 58 rand_state = new_guest_random_state(guest_random_seed + vcpu_idx); in memstress_guest_code() 64 GUEST_ASSERT(vcpu_args->vcpu_idx == vcpu_idx); in memstress_guest_code() 101 vcpu_args->vcpu_idx = i; in memstress_setup_vcpus() 261 int vcpu_idx = vcpu->vcpu_idx; in vcpu_thread_main() local 264 pin_self_to_cpu(memstress_args.vcpu_to_pcpu[vcpu_idx]); in vcpu_thread_main() 277 vcpu_thread_fn(&memstress_args.vcpu_args[vcpu_idx]); in vcpu_thread_main() 294 vcpu->vcpu_idx = i; in memstress_start_vcpu_threads()
|
| /linux/arch/arm64/kvm/ |
| H A D | trace_arm.h | 94 TP_PROTO(unsigned int type, int vcpu_idx, int irq_num, int level), 95 TP_ARGS(type, vcpu_idx, irq_num, level), 99 __field( int, vcpu_idx ) 106 __entry->vcpu_idx = vcpu_idx; 115 __entry->type, __entry->vcpu_idx, __entry->irq_num, __entry->level)
|
| H A D | arm.c | 680 if (*last_ran != vcpu->vcpu_idx) { in kvm_arch_vcpu_load() 682 *last_ran = vcpu->vcpu_idx; in kvm_arch_vcpu_load() 718 vcpu->vcpu_idx, vcpu->arch.hcr_el2); in kvm_arch_vcpu_load()
|
| /linux/arch/x86/kvm/svm/ |
| H A D | avic.c | 55 #define __AVIC_GATAG(vm_id, vcpu_idx) ((((vm_id) & AVIC_VM_ID_MASK) << AVIC_VM_ID_SHIFT) | \ argument 56 ((vcpu_idx) & AVIC_VCPU_IDX_MASK)) 57 #define AVIC_GATAG(vm_id, vcpu_idx) \ argument 59 u32 ga_tag = __AVIC_GATAG(vm_id, vcpu_idx); \ 61 WARN_ON_ONCE(AVIC_GATAG_TO_VCPUIDX(ga_tag) != (vcpu_idx)); \ 265 u32 vcpu_idx = AVIC_GATAG_TO_VCPUIDX(ga_tag); in avic_ga_log_notifier() local 267 pr_debug("SVM: %s: vm_id=%#x, vcpu_idx=%#x\n", __func__, vm_id, vcpu_idx); in avic_ga_log_notifier() 268 trace_kvm_avic_ga_log(vm_id, vcpu_idx); in avic_ga_log_notifier() 274 vcpu = kvm_get_vcpu(&kvm_svm->kvm, vcpu_idx); in avic_ga_log_notifier() 904 vcpu->vcpu_idx), in avic_pi_update_irte()
|
| /linux/tools/testing/selftests/kvm/x86/ |
| H A D | dirty_log_page_splitting_test.c | 69 int vcpu_idx = vcpu_args->vcpu_idx; in vcpu_worker() local 78 vcpu_last_completed_iteration[vcpu_idx] = current_iteration; in vcpu_worker()
|
| /linux/arch/x86/kvm/ |
| H A D | xen.c | 114 e.vcpu_idx = vcpu->vcpu_idx; in kvm_xen_inject_timer_irqs() 136 e.vcpu_idx = vcpu->vcpu_idx; in xen_timer_callback() 1543 set_bit(vcpu->vcpu_idx, vcpu->kvm->arch.xen.poll_mask); in kvm_xen_schedop_poll() 1564 clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.xen.poll_mask); in kvm_xen_schedop_poll() 1776 test_and_clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.xen.poll_mask)) { in kvm_xen_check_poller() 1800 int vcpu_idx, idx, rc; in kvm_xen_set_evtchn_fast() local 1802 vcpu_idx = READ_ONCE(xe->vcpu_idx); in kvm_xen_set_evtchn_fast() 1803 if (vcpu_idx >= 0) in kvm_xen_set_evtchn_fast() 1804 vcpu = kvm_get_vcpu(kvm, vcpu_idx); in kvm_xen_set_evtchn_fast() 1809 WRITE_ONCE(xe->vcpu_idx, vcpu->vcpu_idx); in kvm_xen_set_evtchn_fast() [all …]
|
| /linux/arch/arm64/kvm/hyp/nvhe/ |
| H A D | pkvm.c | 252 unsigned int vcpu_idx) in pkvm_load_hyp_vcpu() argument 266 if (hyp_vm->kvm.created_vcpus <= vcpu_idx) in pkvm_load_hyp_vcpu() 270 hyp_vcpu = smp_load_acquire(&hyp_vm->vcpus[vcpu_idx]); in pkvm_load_hyp_vcpu() 526 hyp_vcpu->vcpu.vcpu_idx = READ_ONCE(host_vcpu->vcpu_idx); in init_pkvm_hyp_vcpu() 867 unsigned int idx = hyp_vcpu->vcpu.vcpu_idx; in register_hyp_vcpu()
|
| H A D | hyp-main.c | 172 DECLARE_REG(unsigned int, vcpu_idx, host_ctxt, 2); in handle___pkvm_vcpu_load() 176 hyp_vcpu = pkvm_load_hyp_vcpu(handle, vcpu_idx); in handle___pkvm_vcpu_load()
|
| /linux/arch/arm64/kvm/vgic/ |
| H A D | vgic-debug.c | 173 id = vcpu->vcpu_idx; in print_header() 219 (irq->target_vcpu) ? irq->target_vcpu->vcpu_idx : -1, in print_irq_state() 231 (irq->vcpu) ? irq->vcpu->vcpu_idx : -1); in print_irq_state()
|
| H A D | vgic.c | 546 trace_vgic_update_irq_pending(vcpu ? vcpu->vcpu_idx : 0, intid, level); in kvm_vgic_inject_irq()
|
| /linux/tools/testing/selftests/kvm/include/ |
| H A D | memstress.h | 29 int vcpu_idx; member
|
| /linux/arch/riscv/kvm/ |
| H A D | aia_device.c | 165 unsigned long vcpu_idx, bool write) in aia_imsic_addr() argument 170 vcpu = kvm_get_vcpu(kvm, vcpu_idx); in aia_imsic_addr() 538 vaia->hart_index = vcpu->vcpu_idx; in kvm_riscv_vcpu_aia_init()
|
| H A D | vcpu.c | 189 if (vcpu->vcpu_idx != 0) in kvm_arch_vcpu_postcreate()
|
| /linux/arch/s390/kvm/ |
| H A D | interrupt.c | 364 set_bit(vcpu->vcpu_idx, vcpu->kvm->arch.idle_mask); in __set_cpu_idle() 370 clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.idle_mask); in __unset_cpu_idle() 1282 vcpu->kvm->arch.float_int.last_sleep_cpu = vcpu->vcpu_idx; in kvm_s390_handle_wait() 3052 int vcpu_idx, online_vcpus = atomic_read(&kvm->online_vcpus); in __airqs_kick_single_vcpu() local 3057 for_each_set_bit(vcpu_idx, kvm->arch.idle_mask, online_vcpus) { in __airqs_kick_single_vcpu() 3058 vcpu = kvm_get_vcpu(kvm, vcpu_idx); in __airqs_kick_single_vcpu() 3064 if (test_and_set_bit(vcpu_idx, gi->kicked_mask)) in __airqs_kick_single_vcpu()
|
| H A D | kvm-s390.c | 3790 clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.gisa_int.kicked_mask); in kvm_arch_vcpu_runnable() 4481 clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.gisa_int.kicked_mask); in vcpu_pre_run()
|
| /linux/virt/kvm/ |
| H A D | kvm_main.c | 3997 if (idx == me->vcpu_idx) in kvm_vcpu_on_spin() 4218 vcpu->vcpu_idx = atomic_read(&kvm->online_vcpus); in kvm_vm_ioctl_create_vcpu() 4219 r = xa_insert(&kvm->vcpu_array, vcpu->vcpu_idx, vcpu, GFP_KERNEL_ACCOUNT); in kvm_vm_ioctl_create_vcpu() 4254 xa_erase(&kvm->vcpu_array, vcpu->vcpu_idx); in kvm_vm_ioctl_create_vcpu() 4386 if (likely(vcpu->vcpu_idx < atomic_read(&kvm->online_vcpus))) in kvm_wait_for_vcpu_online() 4399 if (WARN_ON_ONCE(!kvm_get_vcpu(kvm, vcpu->vcpu_idx))) in kvm_wait_for_vcpu_online()
|
| /linux/include/linux/ |
| H A D | kvm_host.h | 332 int vcpu_idx; /* index into kvm->vcpu_array */ member 663 int vcpu_idx; member
|