Home
last modified time | relevance | path

Searched refs:svm (Results 1 – 25 of 35) sorted by relevance

12

/linux/arch/x86/kvm/svm/
H A Dnested.c39 struct vcpu_svm *svm = to_svm(vcpu); in nested_svm_inject_npf_exit() local
40 struct vmcb *vmcb = svm->vmcb; in nested_svm_inject_npf_exit()
55 nested_svm_vmexit(svm); in nested_svm_inject_npf_exit()
60 struct vcpu_svm *svm = to_svm(vcpu); in nested_svm_get_tdp_pdptr() local
61 u64 cr3 = svm->nested.ctl.nested_cr3; in nested_svm_get_tdp_pdptr()
78 struct vcpu_svm *svm = to_svm(vcpu); in nested_svm_get_tdp_cr3() local
80 return svm->nested.ctl.nested_cr3; in nested_svm_get_tdp_cr3()
85 struct vcpu_svm *svm = to_svm(vcpu); in nested_svm_init_mmu_context() local
96 kvm_init_shadow_npt_mmu(vcpu, X86_CR0_PG, svm->vmcb01.ptr->save.cr4, in nested_svm_init_mmu_context()
97 svm->vmcb01.ptr->save.efer, in nested_svm_init_mmu_context()
[all …]
H A Dsvm.c205 struct vcpu_svm *svm = to_svm(vcpu); in svm_set_efer() local
222 clr_exception_intercept(svm, GP_VECTOR); in svm_set_efer()
230 svm_free_nested(svm); in svm_set_efer()
233 int ret = svm_allocate_nested(svm); in svm_set_efer()
245 set_exception_intercept(svm, GP_VECTOR); in svm_set_efer()
249 svm->vmcb->save.efer = efer | EFER_SVME; in svm_set_efer()
250 vmcb_mark_dirty(svm->vmcb, VMCB_CR); in svm_set_efer()
256 struct vcpu_svm *svm = to_svm(vcpu); in svm_get_interrupt_shadow() local
259 if (svm->vmcb->control.int_state & SVM_INTERRUPT_SHADOW_MASK) in svm_get_interrupt_shadow()
266 struct vcpu_svm *svm = to_svm(vcpu); in svm_set_interrupt_shadow() local
[all …]
H A Dsvm.h360 void recalc_intercepts(struct vcpu_svm *svm);
397 static inline bool ghcb_gpa_is_registered(struct vcpu_svm *svm, u64 val) in ghcb_gpa_is_registered() argument
399 return svm->sev_es.ghcb_registered_gpa == val; in ghcb_gpa_is_registered()
489 static inline void set_exception_intercept(struct vcpu_svm *svm, u32 bit) in set_exception_intercept() argument
491 struct vmcb *vmcb = svm->vmcb01.ptr; in set_exception_intercept()
496 recalc_intercepts(svm); in set_exception_intercept()
499 static inline void clr_exception_intercept(struct vcpu_svm *svm, u32 bit) in clr_exception_intercept() argument
501 struct vmcb *vmcb = svm->vmcb01.ptr; in clr_exception_intercept()
506 recalc_intercepts(svm); in clr_exception_intercept()
509 static inline void svm_set_intercept(struct vcpu_svm *svm, int bit) in svm_set_intercept() argument
[all …]
H A Dsev.c159 static bool sev_vcpu_has_debug_swap(struct vcpu_svm *svm) in sev_vcpu_has_debug_swap() argument
161 struct kvm_vcpu *vcpu = &svm->vcpu; in sev_vcpu_has_debug_swap()
875 static int sev_es_sync_vmsa(struct vcpu_svm *svm) in sev_es_sync_vmsa() argument
877 struct kvm_vcpu *vcpu = &svm->vcpu; in sev_es_sync_vmsa()
879 struct sev_es_save_area *save = svm->sev_es.vmsa; in sev_es_sync_vmsa()
886 if (svm->vcpu.guest_debug || (svm->vmcb->save.dr7 & ~DR7_FIXED_1)) in sev_es_sync_vmsa()
895 memcpy(save, &svm->vmcb->save, sizeof(svm->vmcb->save)); in sev_es_sync_vmsa()
898 save->rax = svm->vcpu.arch.regs[VCPU_REGS_RAX]; in sev_es_sync_vmsa()
899 save->rbx = svm->vcpu.arch.regs[VCPU_REGS_RBX]; in sev_es_sync_vmsa()
900 save->rcx = svm->vcpu.arch.regs[VCPU_REGS_RCX]; in sev_es_sync_vmsa()
[all …]
H A Davic.c111 static void avic_set_x2apic_msr_interception(struct vcpu_svm *svm, in avic_set_x2apic_msr_interception() argument
148 if (intercept == svm->x2avic_msrs_intercepted) in avic_set_x2apic_msr_interception()
155 svm_set_intercept_for_msr(&svm->vcpu, x2avic_passthrough_msrs[i], in avic_set_x2apic_msr_interception()
158 svm->x2avic_msrs_intercepted = intercept; in avic_set_x2apic_msr_interception()
186 static void avic_activate_vmcb(struct vcpu_svm *svm) in avic_activate_vmcb() argument
188 struct vmcb *vmcb = svm->vmcb01.ptr; in avic_activate_vmcb()
189 struct kvm_vcpu *vcpu = &svm->vcpu; in avic_activate_vmcb()
205 if (x2avic_enabled && apic_x2apic_mode(svm->vcpu.arch.apic)) { in avic_activate_vmcb()
209 avic_set_x2apic_msr_interception(svm, false); in avic_activate_vmcb()
215 kvm_make_request(KVM_REQ_TLB_FLUSH_CURRENT, &svm->vcpu); in avic_activate_vmcb()
[all …]
H A Dhyperv.h17 struct vcpu_svm *svm = to_svm(vcpu); in nested_svm_hv_update_vm_vp_ids() local
18 struct hv_vmcb_enlightenments *hve = &svm->nested.ctl.hv_enlightenments; in nested_svm_hv_update_vm_vp_ids()
31 struct vcpu_svm *svm = to_svm(vcpu); in nested_svm_l2_tlb_flush_enabled() local
32 struct hv_vmcb_enlightenments *hve = &svm->nested.ctl.hv_enlightenments; in nested_svm_l2_tlb_flush_enabled()
/linux/drivers/gpu/drm/nouveau/
H A Dnouveau_svm.c94 nouveau_ivmm_find(struct nouveau_svm *svm, u64 inst) in nouveau_ivmm_find() argument
97 list_for_each_entry(ivmm, &svm->inst, head) { in nouveau_ivmm_find()
169 if (!cli->svm.svmm) { in nouveau_svmm_bind()
186 nouveau_dmem_migrate_vma(cli->drm, cli->svm.svmm, vma, addr, in nouveau_svmm_bind()
210 mutex_lock(&svmm->vmm->cli->drm->svm->mutex); in nouveau_svmm_part()
211 ivmm = nouveau_ivmm_find(svmm->vmm->cli->drm->svm, inst); in nouveau_svmm_part()
216 mutex_unlock(&svmm->vmm->cli->drm->svm->mutex); in nouveau_svmm_part()
231 mutex_lock(&svmm->vmm->cli->drm->svm->mutex); in nouveau_svmm_join()
232 list_add(&ivmm->head, &svmm->vmm->cli->drm->svm->inst); in nouveau_svmm_join()
233 mutex_unlock(&svmm->vmm->cli->drm->svm->mutex); in nouveau_svmm_join()
[all …]
/linux/tools/testing/selftests/kvm/x86/
H A Dnested_invalid_cr3_test.c21 static void l1_svm_code(struct svm_test_data *svm) in l1_svm_code() argument
26 generic_svm_setup(svm, l2_guest_code, in l1_svm_code()
30 save_cr3 = svm->vmcb->save.cr3; in l1_svm_code()
31 svm->vmcb->save.cr3 = -1ull; in l1_svm_code()
32 run_guest(svm->vmcb, svm->vmcb_gpa); in l1_svm_code()
33 GUEST_ASSERT(svm->vmcb->control.exit_code == SVM_EXIT_ERR); in l1_svm_code()
36 svm->vmcb->save.cr3 = save_cr3; in l1_svm_code()
37 run_guest(svm->vmcb, svm->vmcb_gpa); in l1_svm_code()
38 GUEST_ASSERT(svm->vmcb->control.exit_code == SVM_EXIT_VMMCALL); in l1_svm_code()
H A Dnested_exceptions_test.c74 static void svm_run_l2(struct svm_test_data *svm, void *l2_code, int vector, in svm_run_l2() argument
77 struct vmcb *vmcb = svm->vmcb; in svm_run_l2()
81 run_guest(vmcb, svm->vmcb_gpa); in svm_run_l2()
91 static void l1_svm_code(struct svm_test_data *svm) in l1_svm_code() argument
93 struct vmcb_control_area *ctrl = &svm->vmcb->control; in l1_svm_code()
96 generic_svm_setup(svm, NULL, &l2_guest_stack[L2_GUEST_STACK_SIZE]); in l1_svm_code()
97 svm->vmcb->save.idtr.limit = 0; in l1_svm_code()
101 svm_run_l2(svm, l2_ss_pending_test, SS_VECTOR, SS_ERROR_CODE); in l1_svm_code()
102 svm_run_l2(svm, l2_ss_injected_gp_test, GP_VECTOR, GP_ERROR_CODE_AMD); in l1_svm_code()
105 svm_run_l2(svm, l2_ss_injected_df_test, DF_VECTOR, DF_ERROR_CODE); in l1_svm_code()
[all …]
H A Dnested_tsc_adjust_test.c106 struct svm_test_data *svm = data; in l1_guest_code() local
108 generic_svm_setup(svm, l2_guest_code, in l1_guest_code()
111 svm->vmcb->control.tsc_offset = TSC_OFFSET_VALUE; in l1_guest_code()
112 run_guest(svm->vmcb, svm->vmcb_gpa); in l1_guest_code()
113 GUEST_ASSERT(svm->vmcb->control.exit_code == SVM_EXIT_VMMCALL); in l1_guest_code()
H A Dnested_close_kvm_test.c48 static void l1_svm_code(struct svm_test_data *svm) in l1_svm_code() argument
53 generic_svm_setup(svm, l2_guest_code, in l1_svm_code()
56 run_guest(svm->vmcb, svm->vmcb_gpa); in l1_svm_code()
H A Dnested_tsc_scaling_test.c83 static void l1_svm_code(struct svm_test_data *svm) in l1_svm_code() argument
90 generic_svm_setup(svm, l2_guest_code, in l1_svm_code()
97 run_guest(svm->vmcb, svm->vmcb_gpa); in l1_svm_code()
98 GUEST_ASSERT(svm->vmcb->control.exit_code == SVM_EXIT_VMMCALL); in l1_svm_code()
H A Dstate_test.c34 static void svm_l1_guest_code(struct svm_test_data *svm) in svm_l1_guest_code() argument
37 struct vmcb *vmcb = svm->vmcb; in svm_l1_guest_code()
39 GUEST_ASSERT(svm->vmcb_gpa); in svm_l1_guest_code()
41 generic_svm_setup(svm, svm_l2_guest_code, in svm_l1_guest_code()
45 run_guest(vmcb, svm->vmcb_gpa); in svm_l1_guest_code()
49 run_guest(vmcb, svm->vmcb_gpa); in svm_l1_guest_code()
H A Daperfmperf_test.c65 static void l1_svm_code(struct svm_test_data *svm) in l1_svm_code() argument
68 struct vmcb *vmcb = svm->vmcb; in l1_svm_code()
70 generic_svm_setup(svm, l2_guest_code, &l2_guest_stack[L2_GUEST_STACK_SIZE]); in l1_svm_code()
71 run_guest(vmcb, svm->vmcb_gpa); in l1_svm_code()
/linux/arch/arm/mm/
H A Dioremap.c53 struct static_vm *svm; in find_static_vm_paddr() local
56 list_for_each_entry(svm, &static_vmlist, list) { in find_static_vm_paddr()
57 vm = &svm->vm; in find_static_vm_paddr()
67 return svm; in find_static_vm_paddr()
75 struct static_vm *svm; in find_static_vm_vaddr() local
78 list_for_each_entry(svm, &static_vmlist, list) { in find_static_vm_vaddr()
79 vm = &svm->vm; in find_static_vm_vaddr()
86 return svm; in find_static_vm_vaddr()
92 void __init add_static_vm_early(struct static_vm *svm) in add_static_vm_early() argument
98 vm = &svm->vm; in add_static_vm_early()
[all …]
H A Dmm.h75 extern __init void add_static_vm_early(struct static_vm *svm);
/linux/drivers/gpu/drm/xe/
H A Dxe_svm.c56 return force_smem ? NULL : vm->svm.peer.owner; in xe_svm_private_page_owner()
82 return container_of(gpusvm, struct xe_vm, svm.gpusvm); in gpusvm_to_vm()
138 spin_lock(&vm->svm.garbage_collector.lock); in xe_svm_garbage_collector_add_range()
141 &vm->svm.garbage_collector.range_list); in xe_svm_garbage_collector_add_range()
142 spin_unlock(&vm->svm.garbage_collector.lock); in xe_svm_garbage_collector_add_range()
144 queue_work(xe->usm.pf_wq, &vm->svm.garbage_collector.work); in xe_svm_garbage_collector_add_range()
210 drm_gpusvm_range_unmap_pages(&vm->svm.gpusvm, r, &ctx); in xe_svm_range_notifier_event_end()
309 drm_gpusvm_range_remove(&vm->svm.gpusvm, &range->base); in __xe_svm_garbage_collector()
387 spin_lock(&vm->svm.garbage_collector.lock); in xe_svm_garbage_collector()
388 range = list_first_entry_or_null(&vm->svm.garbage_collector.range_list, in xe_svm_garbage_collector()
[all …]
H A Dxe_svm.h236 return drm_gpusvm_init(&vm->svm.gpusvm, "Xe SVM (simple)", &vm->xe->drm, in xe_svm_init()
248 drm_gpusvm_fini(&vm->svm.gpusvm); in xe_svm_fini()
395 lockdep_assert_held_write(&(vm__)->svm.gpusvm.notifier_lock)
398 lockdep_assert_held_read(&(vm__)->svm.gpusvm.notifier_lock)
401 drm_gpusvm_notifier_lock(&(vm__)->svm.gpusvm)
404 down_read_interruptible(&(vm__)->svm.gpusvm.notifier_lock)
407 drm_gpusvm_notifier_unlock(&(vm__)->svm.gpusvm)
H A Dxe_vm_types.h207 } svm; member
/linux/tools/testing/vsock/
H A Dvsock_perf.c104 struct sockaddr_vm svm; in vsock_connect() member
106 .svm = { in vsock_connect()
121 if (connect(fd, &addr.sa, sizeof(addr.svm)) < 0) { in vsock_connect()
147 struct sockaddr_vm svm; in run_receiver() member
149 .svm = { in run_receiver()
157 struct sockaddr_vm svm; in run_receiver() member
160 socklen_t clientaddr_len = sizeof(clientaddr.svm); in run_receiver()
173 if (bind(fd, &addr.sa, sizeof(addr.svm)) < 0) in run_receiver()
H A Dvsock_diag_test.c343 struct sockaddr_vm svm; in test_listen_socket_server() member
345 .svm = { in test_listen_socket_server()
357 if (bind(fd, &addr.sa, sizeof(addr.svm)) < 0) { in test_listen_socket_server()
H A Dutil.c273 struct sockaddr_vm svm; in vsock_accept() member
275 socklen_t clientaddr_len = sizeof(clientaddr.svm); in vsock_accept()
296 if (clientaddr_len != sizeof(clientaddr.svm)) { in vsock_accept()
308 *clientaddrp = clientaddr.svm; in vsock_accept()
H A Dvsock_test.c42 struct sockaddr_vm svm; in test_stream_connection_reset() member
44 .svm = { in test_stream_connection_reset()
57 ret = connect(fd, &addr.sa, sizeof(addr.svm)); in test_stream_connection_reset()
78 struct sockaddr_vm svm; in test_stream_bind_only_client() member
80 .svm = { in test_stream_bind_only_client()
96 ret = connect(fd, &addr.sa, sizeof(addr.svm)); in test_stream_bind_only_client()
/linux/arch/x86/virt/
H A DMakefile2 obj-y += svm/ vmx/
/linux/drivers/iommu/intel/
H A DMakefile6 obj-$(CONFIG_INTEL_IOMMU_SVM) += svm.o

12