Lines Matching refs:edx
161 xfeatures = best->eax | ((u64)best->edx << 32); in kvm_check_cpuid()
197 e2[i].ecx != orig->ecx || e2[i].edx != orig->edx) in kvm_cpuid_check_equal()
219 signature[2] = entry->edx; in kvm_get_hypervisor_cpuid()
263 return (best->eax | ((u64)best->edx << 32)) & kvm_caps.supported_xcr0; in cpuid_get_supported_xcr0()
331 return is_guest_vendor_amd(entry->ebx, entry->ecx, entry->edx) || in guest_cpuid_is_amd_or_hygon()
332 is_guest_vendor_hygon(entry->ebx, entry->ecx, entry->edx); in guest_cpuid_is_amd_or_hygon()
350 return entry->edx; in cpuid_get_reg_unsafe()
588 e2[i].edx = e[i].edx; in kvm_vcpu_ioctl_set_cpuid()
665 &entry.eax, &entry.ebx, &entry.ecx, &entry.edx); in raw_cpuid_get()
1286 &entry->eax, &entry->ebx, &entry->ecx, &entry->edx); in do_host_cpuid()
1398 entry->edx = 0; in __do_cpuid_func()
1431 union cpuid10_edx edx = { }; in __do_cpuid_func() local
1434 entry->eax = entry->ebx = entry->ecx = entry->edx = 0; in __do_cpuid_func()
1442 edx.split.num_counters_fixed = kvm_pmu_cap.num_counters_fixed; in __do_cpuid_func()
1443 edx.split.bit_width_fixed = kvm_pmu_cap.bit_width_fixed; in __do_cpuid_func()
1446 edx.split.anythread_deprecated = 1; in __do_cpuid_func()
1451 entry->edx = edx.full; in __do_cpuid_func()
1469 entry->edx &= permitted_xcr0 >> 32; in __do_cpuid_func()
1486 entry->edx &= permitted_xss >> 32; in __do_cpuid_func()
1516 entry->edx = 0; in __do_cpuid_func()
1523 entry->eax = entry->ebx = entry->ecx = entry->edx = 0; in __do_cpuid_func()
1553 entry->eax = entry->ebx = entry->ecx = entry->edx = 0; in __do_cpuid_func()
1565 entry->eax = entry->ebx = entry->ecx = entry->edx = 0; in __do_cpuid_func()
1576 entry->eax = entry->ebx = entry->ecx = entry->edx = 0; in __do_cpuid_func()
1584 entry->eax = entry->ebx = entry->ecx = entry->edx = 0; in __do_cpuid_func()
1599 entry->edx = 0; in __do_cpuid_func()
1607 entry->edx = sigptr[2]; in __do_cpuid_func()
1630 entry->edx = 0; in __do_cpuid_func()
1662 entry->edx &= ~GENMASK(17, 16); in __do_cpuid_func()
1668 entry->edx &= boot_cpu_data.x86_power; in __do_cpuid_func()
1714 entry->edx = 0; in __do_cpuid_func()
1720 entry->eax = entry->ebx = entry->ecx = entry->edx = 0; in __do_cpuid_func()
1730 entry->ecx = entry->edx = 0; in __do_cpuid_func()
1734 entry->ebx = entry->ecx = entry->edx = 0; in __do_cpuid_func()
1739 entry->edx = 0; /* reserved */ in __do_cpuid_func()
1743 entry->eax = entry->ebx = entry->ecx = entry->edx = 0; in __do_cpuid_func()
1756 entry->eax = entry->ebx = entry->ecx = entry->edx = 0; in __do_cpuid_func()
1759 entry->ebx = entry->edx = 0; in __do_cpuid_func()
1767 entry->ecx = entry->edx = 0; in __do_cpuid_func()
1793 entry->eax = entry->ebx = entry->ecx = entry->edx = 0; in __do_cpuid_func()
1948 if (is_guest_vendor_amd(basic->ebx, basic->ecx, basic->edx) || in get_out_of_range_cpuid_entry()
1949 is_guest_vendor_hygon(basic->ebx, basic->ecx, basic->edx)) in get_out_of_range_cpuid_entry()
1979 u32 *ecx, u32 *edx, bool exact_only) in kvm_cpuid() argument
2000 *edx = entry->edx; in kvm_cpuid()
2009 *edx &= ~feature_bit(CONSTANT_TSC); in kvm_cpuid()
2022 *edx = vcpu->arch.pvclock_tsc_shift; in kvm_cpuid()
2028 *eax = *ebx = *ecx = *edx = 0; in kvm_cpuid()
2040 *edx = entry->edx; in kvm_cpuid()
2044 trace_kvm_cpuid(orig_function, index, *eax, *ebx, *ecx, *edx, exact, in kvm_cpuid()
2052 u32 eax, ebx, ecx, edx; in kvm_emulate_cpuid() local
2059 kvm_cpuid(vcpu, &eax, &ebx, &ecx, &edx, false); in kvm_emulate_cpuid()
2063 kvm_rdx_write(vcpu, edx); in kvm_emulate_cpuid()