| /linux/tools/testing/selftests/kvm/ |
| H A D | arch_timer.c | 49 unsigned int vcpu_idx = (unsigned long)arg; in test_vcpu_run() local 51 struct kvm_vcpu *vcpu = vcpus[vcpu_idx]; in test_vcpu_run() 53 struct test_vcpu_shared_data *shared_data = &vcpu_shared_data[vcpu_idx]; in test_vcpu_run() 59 __set_bit(vcpu_idx, vcpu_done_map); in test_vcpu_run() 69 vcpu_idx, shared_data->guest_stage, shared_data->nr_iter); in test_vcpu_run() 76 pr_info("PASS(vCPU-%d).\n", vcpu_idx); in test_vcpu_run() 98 static int test_migrate_vcpu(unsigned int vcpu_idx) in test_migrate_vcpu() argument 103 pr_debug("Migrating vCPU: %u to pCPU: %u\n", vcpu_idx, new_pcpu); in test_migrate_vcpu() 105 ret = __pin_task_to_cpu(pt_vcpu_run[vcpu_idx], new_pcpu); in test_migrate_vcpu() 110 vcpu_idx, new_pcpu, ret); in test_migrate_vcpu()
|
| H A D | access_tracking_perf_test.c | 157 static void too_many_idle_pages(long idle_pages, long total_pages, int vcpu_idx) in too_many_idle_pages() 161 if (vcpu_idx >= 0) in too_many_idle_pages() 162 snprintf(prefix, 18, "vCPU%d: ", vcpu_idx); in too_many_idle_pages() 176 int vcpu_idx = vcpu_args->vcpu_idx; in pageidle_mark_vcpu_memory_idle() 186 if (overlap_memory_access && vcpu_idx) in pageidle_mark_vcpu_memory_idle() 218 vcpu_idx, no_pfn, pages); in pageidle_mark_vcpu_memory_idle() 230 overlap_memory_access ? -1 : vcpu_idx); in pageidle_mark_vcpu_memory_idle() 329 int vcpu_idx = vcpu_args->vcpu_idx; in vcpu_thread_main() 156 too_many_idle_pages(long idle_pages,long total_pages,int vcpu_idx) too_many_idle_pages() argument 175 int vcpu_idx = vcpu_args->vcpu_idx; pageidle_mark_vcpu_memory_idle() local 328 int vcpu_idx = vcpu_args->vcpu_idx; vcpu_thread_main() local 346 spin_wait_for_vcpu(int vcpu_idx,int target_iteration) spin_wait_for_vcpu() argument [all...] |
| H A D | demand_paging_test.c | 35 int vcpu_idx = vcpu_args->vcpu_idx; in vcpu_worker() local 53 PER_VCPU_DEBUG("vCPU %d execution time: %ld.%.9lds\n", vcpu_idx, in vcpu_worker()
|
| /linux/tools/testing/selftests/kvm/lib/ |
| H A D | memstress.c | 23 int vcpu_idx; member 47 void memstress_guest_code(uint32_t vcpu_idx) in memstress_guest_code() argument 50 struct memstress_vcpu_args *vcpu_args = &args->vcpu_args[vcpu_idx]; in memstress_guest_code() 58 rand_state = new_guest_random_state(guest_random_seed + vcpu_idx); in memstress_guest_code() 64 GUEST_ASSERT(vcpu_args->vcpu_idx == vcpu_idx); in memstress_guest_code() 101 vcpu_args->vcpu_idx = i; in memstress_setup_vcpus() 265 int vcpu_idx = vcpu->vcpu_idx; in vcpu_thread_main() local 268 pin_self_to_cpu(memstress_args.vcpu_to_pcpu[vcpu_idx]); in vcpu_thread_main() 281 vcpu_thread_fn(&memstress_args.vcpu_args[vcpu_idx]); in vcpu_thread_main() 298 vcpu->vcpu_idx = i; in memstress_start_vcpu_threads()
|
| /linux/arch/arm64/kvm/ |
| H A D | trace_arm.h | 94 TP_PROTO(unsigned int type, int vcpu_idx, int irq_num, int level), 95 TP_ARGS(type, vcpu_idx, irq_num, level), 99 __field( int, vcpu_idx ) 106 __entry->vcpu_idx = vcpu_idx; 115 __entry->type, __entry->vcpu_idx, __entry->irq_num, __entry->level)
|
| H A D | arm.c | 617 if (*last_ran != vcpu->vcpu_idx) { in kvm_arch_vcpu_load() 619 *last_ran = vcpu->vcpu_idx; in kvm_arch_vcpu_load() 655 vcpu->vcpu_idx, vcpu->arch.hcr_el2); in kvm_arch_vcpu_load()
|
| /linux/arch/x86/kvm/svm/ |
| H A D | avic.c | 54 #define __AVIC_GATAG(vm_id, vcpu_idx) ((((vm_id) & AVIC_VM_ID_MASK) << AVIC_VM_ID_SHIFT) | \ argument 55 ((vcpu_idx) & AVIC_VCPU_IDX_MASK)) 56 #define AVIC_GATAG(vm_id, vcpu_idx) \ argument 58 u32 ga_tag = __AVIC_GATAG(vm_id, vcpu_idx); \ 60 WARN_ON_ONCE(AVIC_GATAG_TO_VCPUIDX(ga_tag) != (vcpu_idx)); \ 251 u32 vcpu_idx = AVIC_GATAG_TO_VCPUIDX(ga_tag); in avic_ga_log_notifier() local 253 pr_debug("SVM: %s: vm_id=%#x, vcpu_idx=%#x\n", __func__, vm_id, vcpu_idx); in avic_ga_log_notifier() 254 trace_kvm_avic_ga_log(vm_id, vcpu_idx); in avic_ga_log_notifier() 260 vcpu = kvm_get_vcpu(&kvm_svm->kvm, vcpu_idx); in avic_ga_log_notifier() 890 vcpu->vcpu_idx), in avic_pi_update_irte()
|
| /linux/arch/x86/kvm/ |
| H A D | xen.c | 114 e.vcpu_idx = vcpu->vcpu_idx; in kvm_xen_inject_timer_irqs() 136 e.vcpu_idx = vcpu->vcpu_idx; in xen_timer_callback() 1544 set_bit(vcpu->vcpu_idx, vcpu->kvm->arch.xen.poll_mask); in kvm_xen_schedop_poll() 1565 clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.xen.poll_mask); in kvm_xen_schedop_poll() 1777 test_and_clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.xen.poll_mask)) { in kvm_xen_check_poller() 1801 int vcpu_idx, idx, rc; in kvm_xen_set_evtchn_fast() local 1803 vcpu_idx = READ_ONCE(xe->vcpu_idx); in kvm_xen_set_evtchn_fast() 1804 if (vcpu_idx >= 0) in kvm_xen_set_evtchn_fast() 1805 vcpu = kvm_get_vcpu(kvm, vcpu_idx); in kvm_xen_set_evtchn_fast() 1810 WRITE_ONCE(xe->vcpu_idx, vcpu->vcpu_idx); in kvm_xen_set_evtchn_fast() [all …]
|
| H A D | hyperv.h | 92 return hv_vcpu ? hv_vcpu->vp_index : vcpu->vcpu_idx; in kvm_hv_get_vpindex() 322 return vcpu->vcpu_idx; in kvm_hv_get_vpindex()
|
| H A D | lapic.c | 1467 int vcpu_idx; in kvm_bitmap_or_dest_vcpus() local 1479 vcpu_idx = dest_vcpu[i]->vcpu->vcpu_idx; in kvm_bitmap_or_dest_vcpus() 1480 __set_bit(vcpu_idx, vcpu_bitmap); in kvm_bitmap_or_dest_vcpus()
|
| H A D | hyperv.c | 984 hv_vcpu->vp_index = vcpu->vcpu_idx; in kvm_hv_vcpu_init() 1543 if (hv_vcpu->vp_index == vcpu->vcpu_idx) in kvm_hv_set_msr() 1545 else if (new_vp_index == vcpu->vcpu_idx) in kvm_hv_set_msr()
|
| /linux/arch/arm64/kvm/hyp/nvhe/ |
| H A D | pkvm.c | 248 unsigned int vcpu_idx) in pkvm_load_hyp_vcpu() argument 259 if (!hyp_vm || hyp_vm->kvm.created_vcpus <= vcpu_idx) in pkvm_load_hyp_vcpu() 262 hyp_vcpu = hyp_vm->vcpus[vcpu_idx]; in pkvm_load_hyp_vcpu() 488 hyp_vcpu->vcpu.vcpu_idx = READ_ONCE(host_vcpu->vcpu_idx); in init_pkvm_hyp_vcpu() 800 idx = hyp_vcpu->vcpu.vcpu_idx; in __pkvm_init_vcpu()
|
| H A D | hyp-main.c | 168 DECLARE_REG(unsigned int, vcpu_idx, host_ctxt, 2); in handle___pkvm_vcpu_load() 175 hyp_vcpu = pkvm_load_hyp_vcpu(handle, vcpu_idx); in handle___pkvm_vcpu_load()
|
| /linux/arch/arm64/kvm/vgic/ |
| H A D | vgic-debug.c | 214 id = vcpu->vcpu_idx; in print_header() 260 (irq->target_vcpu) ? irq->target_vcpu->vcpu_idx : -1, in print_irq_state() 272 (irq->vcpu) ? irq->vcpu->vcpu_idx : -1); in print_irq_state()
|
| H A D | vgic.c | 527 trace_vgic_update_irq_pending(vcpu ? vcpu->vcpu_idx : 0, intid, level); in kvm_vgic_inject_irq()
|
| /linux/tools/testing/selftests/kvm/include/ |
| H A D | memstress.h | 29 int vcpu_idx; member
|
| /linux/arch/riscv/kvm/ |
| H A D | aia_device.c | 161 unsigned long vcpu_idx, bool write) in aia_imsic_addr() argument 166 vcpu = kvm_get_vcpu(kvm, vcpu_idx); in aia_imsic_addr() 528 vaia->hart_index = vcpu->vcpu_idx; in kvm_riscv_vcpu_aia_init()
|
| H A D | vcpu.c | 185 if (vcpu->vcpu_idx != 0) in kvm_arch_vcpu_postcreate()
|
| /linux/arch/s390/kvm/ |
| H A D | interrupt.c | 364 set_bit(vcpu->vcpu_idx, vcpu->kvm->arch.idle_mask); in __set_cpu_idle() 370 clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.idle_mask); in __unset_cpu_idle() 1279 vcpu->kvm->arch.float_int.last_sleep_cpu = vcpu->vcpu_idx; in kvm_s390_handle_wait() 3029 int vcpu_idx, online_vcpus = atomic_read(&kvm->online_vcpus); in __airqs_kick_single_vcpu() local 3034 for_each_set_bit(vcpu_idx, kvm->arch.idle_mask, online_vcpus) { in __airqs_kick_single_vcpu() 3035 vcpu = kvm_get_vcpu(kvm, vcpu_idx); in __airqs_kick_single_vcpu() 3041 if (test_and_set_bit(vcpu_idx, gi->kicked_mask)) in __airqs_kick_single_vcpu()
|
| H A D | kvm-s390.h | 103 return test_bit(vcpu->vcpu_idx, vcpu->kvm->arch.idle_mask); in is_vcpu_idle()
|
| H A D | kvm-s390.c | 3944 clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.gisa_int.kicked_mask); in kvm_arch_vcpu_runnable() 4693 clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.gisa_int.kicked_mask); in vcpu_pre_run()
|
| /linux/virt/kvm/ |
| H A D | kvm_main.c | 4020 if (idx == me->vcpu_idx) in kvm_vcpu_on_spin() 4241 vcpu->vcpu_idx = atomic_read(&kvm->online_vcpus); in kvm_vm_ioctl_create_vcpu() 4242 r = xa_insert(&kvm->vcpu_array, vcpu->vcpu_idx, vcpu, GFP_KERNEL_ACCOUNT); in kvm_vm_ioctl_create_vcpu() 4277 xa_erase(&kvm->vcpu_array, vcpu->vcpu_idx); in kvm_vm_ioctl_create_vcpu() 4409 if (likely(vcpu->vcpu_idx < atomic_read(&kvm->online_vcpus))) in kvm_wait_for_vcpu_online() 4422 if (WARN_ON_ONCE(!kvm_get_vcpu(kvm, vcpu->vcpu_idx))) in kvm_vcpu_ioctl()
|
| /linux/include/linux/ |
| H A D | kvm_host.h | 333 int vcpu_idx; /* index into kvm->vcpu_array */ member 662 int vcpu_idx; member
|