/linux/arch/x86/kvm/vmx/ |
H A D | nested.h | 33 int vmx_get_vmx_msr(struct nested_vmx_msrs *msrs, u32 msr_index, u64 *pdata); 107 return vmx_misc_cr3_count(to_vmx(vcpu)->nested.msrs.misc_low); in nested_cpu_vmx_misc_cr3_count() 117 return to_vmx(vcpu)->nested.msrs.misc_low & in nested_cpu_has_vmwrite_any_field() 123 return to_vmx(vcpu)->nested.msrs.misc_low & VMX_MISC_ZERO_LEN_INS; in nested_cpu_has_zero_length_injection() 128 return to_vmx(vcpu)->nested.msrs.procbased_ctls_high & in nested_cpu_supports_monitor_trap_flag() 134 return to_vmx(vcpu)->nested.msrs.secondary_ctls_high & in nested_cpu_has_vmx_shadow_vmcs() 265 u64 fixed0 = to_vmx(vcpu)->nested.msrs.cr0_fixed0; in nested_guest_cr0_valid() 266 u64 fixed1 = to_vmx(vcpu)->nested.msrs.cr0_fixed1; in nested_guest_cr0_valid() 269 if (to_vmx(vcpu)->nested.msrs.secondary_ctls_high & in nested_guest_cr0_valid() 279 u64 fixed0 = to_vmx(vcpu)->nested.msrs.cr0_fixed0; in nested_host_cr0_valid() [all …]
|
H A D | nested.c | 466 bool execonly = vmx->nested.msrs.ept_caps & VMX_EPT_EXECUTE_ONLY_BIT; in nested_ept_new_eptp() 467 int ept_lpage_level = ept_caps_to_lpage_level(vmx->nested.msrs.ept_caps); in nested_ept_new_eptp() 946 u64 vmx_misc = vmx_control_msr(vmx->nested.msrs.misc_low, in nested_vmx_max_atomic_switch_msrs() 947 vmx->nested.msrs.misc_high); in nested_vmx_max_atomic_switch_msrs() 1292 vmx->nested.msrs.basic = data; in vmx_restore_vmx_basic() 1296 static void vmx_get_control_msr(struct nested_vmx_msrs *msrs, u32 msr_index, in vmx_get_control_msr() argument 1301 *low = &msrs->pinbased_ctls_low; in vmx_get_control_msr() 1302 *high = &msrs->pinbased_ctls_high; in vmx_get_control_msr() 1305 *low = &msrs->procbased_ctls_low; in vmx_get_control_msr() 1306 *high = &msrs->procbased_ctls_high; in vmx_get_control_msr() [all …]
|
H A D | vmx.c | 2092 if (vmx_get_vmx_msr(&vmx->nested.msrs, msr_info->index, in vmx_get_msr() 2295 ((vmx->nested.msrs.entry_ctls_high & VM_ENTRY_LOAD_BNDCFGS) || in vmx_set_msr() 2296 (vmx->nested.msrs.exit_ctls_high & VM_EXIT_CLEAR_BNDCFGS))) in vmx_set_msr() 4579 vmx->nested.msrs.secondary_ctls_high |= control; in vmx_adjust_secondary_exec_control() 4581 vmx->nested.msrs.secondary_ctls_high &= ~control; in vmx_adjust_secondary_exec_control() 4859 memcpy(&vmx->nested.msrs, &vmcs_config.nested, sizeof(vmx->nested.msrs)); in __vmx_vcpu_reset() 7198 struct perf_guest_switch_msr *msrs; in atomic_switch_perf_msrs() local 7206 msrs = perf_guest_get_msrs(&nr_msrs, (void *)pmu); in atomic_switch_perf_msrs() 7207 if (!msrs) in atomic_switch_perf_msrs() 7211 if (msrs[i].host == msrs[i].guest) in atomic_switch_perf_msrs() [all …]
|
H A D | vmx.h | 232 struct nested_vmx_msrs msrs; member
|
/linux/drivers/platform/x86/intel/ifs/ |
H A D | load.c | 121 const struct ifs_test_msrs *msrs; in copy_hashes_authenticate_chunks() local 128 msrs = ifs_get_test_msrs(dev); in copy_hashes_authenticate_chunks() 130 wrmsrl(msrs->copy_hashes, ifs_hash_ptr); in copy_hashes_authenticate_chunks() 131 rdmsrl(msrs->copy_hashes_status, hashes_status.data); in copy_hashes_authenticate_chunks() 152 wrmsrl(msrs->copy_chunks, linear_addr); in copy_hashes_authenticate_chunks() 153 rdmsrl(msrs->copy_chunks_status, chunk_status.data); in copy_hashes_authenticate_chunks() 185 const struct ifs_test_msrs *msrs; in copy_hashes_authenticate_chunks_gen2() local 195 msrs = ifs_get_test_msrs(dev); in copy_hashes_authenticate_chunks_gen2() 198 wrmsrl(msrs->copy_hashes, ifs_hash_ptr); in copy_hashes_authenticate_chunks_gen2() 199 rdmsrl(msrs->copy_hashes_status, hashes_status.data); in copy_hashes_authenticate_chunks_gen2() [all …]
|
/linux/arch/x86/lib/ |
H A D | msr-smp.c | 13 if (rv->msrs) in __rdmsr_on_cpu() 14 reg = this_cpu_ptr(rv->msrs); in __rdmsr_on_cpu() 26 if (rv->msrs) in __wrmsr_on_cpu() 27 reg = this_cpu_ptr(rv->msrs); in __wrmsr_on_cpu() 98 struct msr __percpu *msrs, in __rwmsr_on_cpus() argument 106 rv.msrs = msrs; in __rwmsr_on_cpus() 125 void rdmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr __percpu *msrs) in rdmsr_on_cpus() argument 127 __rwmsr_on_cpus(mask, msr_no, msrs, __rdmsr_on_cpu); in rdmsr_on_cpus() 139 void wrmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr __percpu *msrs) in wrmsr_on_cpus() argument 141 __rwmsr_on_cpus(mask, msr_no, msrs, __wrmsr_on_cpu); in wrmsr_on_cpus()
|
H A D | msr.c | 11 struct msr __percpu *msrs = NULL; in msrs_alloc() local 13 msrs = alloc_percpu(struct msr); in msrs_alloc() 14 if (!msrs) { in msrs_alloc() 19 return msrs; in msrs_alloc() 23 void msrs_free(struct msr __percpu *msrs) in msrs_free() argument 25 free_percpu(msrs); in msrs_free()
|
/linux/Documentation/trace/postprocess/ |
H A D | decode_msr.py | 7 msrs = dict() variable 13 msrs[int(m.group(2), 16)] = m.group(1) 26 if num in msrs: 27 r = msrs[num]
|
/linux/arch/x86/include/asm/ |
H A D | msr.h | 20 struct msr __percpu *msrs; member 326 void msrs_free(struct msr __percpu *msrs); 335 void rdmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr __percpu *msrs); 336 void wrmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr __percpu *msrs); 365 struct msr __percpu *msrs) in rdmsr_on_cpus() argument 367 rdmsr_on_cpu(0, msr_no, raw_cpu_ptr(&msrs->l), raw_cpu_ptr(&msrs->h)); in rdmsr_on_cpus() 370 struct msr __percpu *msrs) in wrmsr_on_cpus() argument 372 wrmsr_on_cpu(0, msr_no, raw_cpu_read(msrs->l), raw_cpu_read(msrs->h)); in wrmsr_on_cpus()
|
/linux/Documentation/virt/kvm/x86/ |
H A D | cpuid.rst | 44 KVM_FEATURE_CLOCKSOURCE 0 kvmclock available at msrs 52 KVM_FEATURE_CLOCKSOURCE2 3 kvmclock available at msrs
|
/linux/drivers/edac/ |
H A D | amd64_edac.c | 16 static struct msr __percpu *msrs; variable 3204 rdmsr_on_cpus(mask, MSR_IA32_MCG_CTL, msrs); in nb_mce_bank_enabled_on_node() 3207 struct msr *reg = per_cpu_ptr(msrs, cpu); in nb_mce_bank_enabled_on_node() 3236 rdmsr_on_cpus(cmask, MSR_IA32_MCG_CTL, msrs); in toggle_ecc_err_reporting() 3240 struct msr *reg = per_cpu_ptr(msrs, cpu); in toggle_ecc_err_reporting() 3255 wrmsr_on_cpus(cmask, MSR_IA32_MCG_CTL, msrs); in toggle_ecc_err_reporting() 4152 msrs = msrs_alloc(); in amd64_edac_init() 4153 if (!msrs) in amd64_edac_init() 4189 msrs_free(msrs); in amd64_edac_init() 4190 msrs = NULL; in amd64_edac_init() [all …]
|
/linux/arch/arm64/boot/dts/qcom/ |
H A D | x1e80100.dtsi | 6414 qcom,cmb-msrs-num = <32>; 6471 qcom,cmb-msrs-num = <32>; 6609 qcom,cmb-msrs-num = <32>; 6628 qcom,dsb-msrs-num = <32>; 6647 qcom,cmb-msrs-num = <32>; 6666 qcom,dsb-msrs-num = <32>; 6717 qcom,dsb-msrs-num = <32>; 6737 qcom,dsb-msrs-num = <32>; 6757 qcom,cmb-msrs-num = <32>; 6991 qcom,cmb-msrs-num = <32>; [all …]
|
H A D | sa8775p.dtsi | 2459 qcom,cmb-msrs-num = <32>; 2518 qcom,cmb-msrs-num = <32>; 2800 qcom,cmb-msrs-num = <32>; 2820 qcom,cmb-msrs-num = <32>; 2840 qcom,cmb-msrs-num = <32>; 2860 qcom,cmb-msrs-num = <32>; 2880 qcom,dsb-msrs-num = <32>; 3192 qcom,cmb-msrs-num = <32>; 3212 qcom,dsb-msrs-num = <32>; 3294 qcom,cmb-msrs-num = <32>; [all …]
|
/linux/arch/x86/kvm/ |
H A D | x86.c | 563 struct kvm_user_return_msrs *msrs in kvm_on_user_return() local 573 if (msrs->registered) { in kvm_on_user_return() 574 msrs->registered = false; in kvm_on_user_return() 579 values = &msrs->values[slot]; in kvm_on_user_return() 628 struct kvm_user_return_msrs *msrs = this_cpu_ptr(user_return_msrs); in kvm_user_return_msr_cpu_online() local 634 msrs->values[i].host = value; in kvm_user_return_msr_cpu_online() 635 msrs->values[i].curr = value; in kvm_user_return_msr_cpu_online() 641 struct kvm_user_return_msrs *msrs = this_cpu_ptr(user_return_msrs); in kvm_set_user_return_msr() local 644 value = (value & mask) | (msrs->values[slot].host & ~mask); in kvm_set_user_return_msr() 645 if (value == msrs->values[slot].curr) in kvm_set_user_return_msr() [all …]
|
/linux/Documentation/virt/kvm/ |
H A D | api.rst | 238 __u32 nmsrs; /* number of msrs in entries */ 243 kvm adjusts nmsrs to reflect the actual number of msrs and fills in the 246 KVM_GET_MSR_INDEX_LIST returns the guest msrs that are supported. The list 650 :Returns: number of msrs successfully returned; 666 __u32 nmsrs; /* number of msrs in entries */ 690 :Returns: number of msrs successfully set (see below), -1 on error 4172 __u32 nmsrs; /* number of msrs in bitmap */
|