Searched refs:hv_vcpu (Results 1 – 5 of 5) sorted by relevance
71 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in to_hv_synic() local73 return &hv_vcpu->synic; in to_hv_synic()78 struct kvm_vcpu_hv *hv_vcpu = container_of(synic, struct kvm_vcpu_hv, synic); in hv_synic_to_vcpu() local80 return hv_vcpu->vcpu; in hv_synic_to_vcpu()90 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in kvm_hv_get_vpindex() local92 return hv_vcpu ? hv_vcpu->vp_index : vcpu->vcpu_idx; in kvm_hv_get_vpindex()135 struct kvm_vcpu_hv *hv_vcpu; in hv_stimer_to_vcpu() local137 hv_vcpu = container_of(stimer - stimer->index, struct kvm_vcpu_hv, in hv_stimer_to_vcpu()139 return hv_vcpu->vcpu; in hv_stimer_to_vcpu()144 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in kvm_hv_has_stimer_pending() local[all …]
223 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in kvm_hv_notify_acked_sint() local230 for (idx = 0; idx < ARRAY_SIZE(hv_vcpu->stimer); idx++) { in kvm_hv_notify_acked_sint()231 stimer = &hv_vcpu->stimer[idx]; in kvm_hv_notify_acked_sint()248 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in synic_exit() local250 hv_vcpu->exit.type = KVM_EXIT_HYPERV_SYNIC; in synic_exit()251 hv_vcpu->exit.u.synic.msr = msr; in synic_exit()252 hv_vcpu->exit.u.synic.control = synic->control; in synic_exit()253 hv_vcpu->exit.u.synic.evt_page = synic->evt_page; in synic_exit()254 hv_vcpu->exit.u.synic.msg_page = synic->msg_page; in synic_exit()330 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in kvm_hv_is_syndbg_enabled() local[all …]
11214 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in vcpu_enter_guest() local11217 vcpu->run->hyperv = hv_vcpu->exit; in vcpu_enter_guest()
19 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in nested_svm_hv_update_vm_vp_ids() local21 if (!hv_vcpu) in nested_svm_hv_update_vm_vp_ids()24 hv_vcpu->nested.pa_page_gpa = hve->partition_assist_page; in nested_svm_hv_update_vm_vp_ids()25 hv_vcpu->nested.vm_id = hve->hv_vm_id; in nested_svm_hv_update_vm_vp_ids()26 hv_vcpu->nested.vp_id = hve->hv_vp_id; in nested_svm_hv_update_vm_vp_ids()33 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in nested_svm_l2_tlb_flush_enabled() local35 if (!hv_vcpu) in nested_svm_l2_tlb_flush_enabled()41 return hv_vcpu->vp_assist_page.nested_control.features.directhypercall; in nested_svm_l2_tlb_flush_enabled()
236 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in nested_release_evmcs() local243 if (hv_vcpu) { in nested_release_evmcs()244 hv_vcpu->nested.pa_page_gpa = INVALID_GPA; in nested_release_evmcs()245 hv_vcpu->nested.vm_id = 0; in nested_release_evmcs()246 hv_vcpu->nested.vp_id = 0; in nested_release_evmcs()1694 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(&vmx->vcpu); in copy_enlightened_to_vmcs12() local1702 hv_vcpu->nested.pa_page_gpa = evmcs->partition_assist_page; in copy_enlightened_to_vmcs12()1703 hv_vcpu->nested.vm_id = evmcs->hv_vm_id; in copy_enlightened_to_vmcs12()1704 hv_vcpu->nested.vp_id = evmcs->hv_vp_id; in copy_enlightened_to_vmcs12()