Home
last modified time | relevance | path

Searched refs:msr (Results 1 – 25 of 279) sorted by relevance

12345678910>>...12

/linux/tools/testing/selftests/kvm/x86/
H A Dhyperv_features.c37 static bool is_write_only_msr(uint32_t msr) in is_write_only_msr() argument
39 return msr == HV_X64_MSR_EOI; in is_write_only_msr()
42 static void guest_msr(struct msr_data *msr) in guest_msr() argument
47 GUEST_ASSERT(msr->idx); in guest_msr()
49 if (msr->write) in guest_msr()
50 vector = wrmsr_safe(msr->idx, msr->write_val); in guest_msr()
52 if (!vector && (!msr->write || !is_write_only_msr(msr->idx))) in guest_msr()
53 vector = rdmsr_safe(msr->idx, &msr_val); in guest_msr()
55 if (msr->fault_expected) in guest_msr()
58 msr->write ? "WR" : "RD", msr->idx, ex_str(vector)); in guest_msr()
[all …]
H A Dmsrs_test.c23 #define ____MSR_TEST(msr, str, val, rsvd, reset, feat, f2, is_kvm) \ argument
25 .index = msr, \
35 #define __MSR_TEST(msr, str, val, rsvd, reset, feat) \ argument
36 ____MSR_TEST(msr, str, val, rsvd, reset, feat, feat, false)
38 #define MSR_TEST_NON_ZERO(msr, val, rsvd, reset, feat) \ argument
39 __MSR_TEST(msr, #msr, val, rsvd, reset, feat)
41 #define MSR_TEST(msr, val, rsvd, feat) \ argument
42 __MSR_TEST(msr, #msr, val, rsvd, 0, feat)
44 #define MSR_TEST2(msr, val, rsvd, feat, f2) \ argument
45 ____MSR_TEST(msr, #msr, val, rsvd, 0, feat, f2, false)
[all …]
H A Duserspace_msr_exit_test.c80 static void deny_msr(uint8_t *bitmap, u32 msr) in deny_msr() argument
82 u32 idx = msr & (KVM_MSR_FILTER_MAX_BITMAP_SIZE - 1); in deny_msr()
145 static noinline uint64_t test_rdmsr(uint32_t msr) in test_rdmsr() argument
152 "=a"(a), "=d"(d) : "c"(msr) : "memory"); in test_rdmsr()
161 static noinline void test_wrmsr(uint32_t msr, uint64_t value) in test_wrmsr() argument
169 "a"(a), "d"(d), "c"(msr) : "memory"); in test_wrmsr()
179 static noinline uint64_t test_em_rdmsr(uint32_t msr) in test_em_rdmsr() argument
186 "=a"(a), "=d"(d) : "c"(msr) : "memory"); in test_em_rdmsr()
195 static noinline void test_em_wrmsr(uint32_t msr, uint64_t value) in test_em_wrmsr() argument
203 "a"(a), "d"(d), "c"(msr) : "memory"); in test_em_wrmsr()
[all …]
/linux/arch/x86/kernel/cpu/
H A Dperfctr-watchdog.c45 static inline unsigned int nmi_perfctr_msr_to_bit(unsigned int msr) in nmi_perfctr_msr_to_bit() argument
51 if (msr >= MSR_F15H_PERF_CTR) in nmi_perfctr_msr_to_bit()
52 return (msr - MSR_F15H_PERF_CTR) >> 1; in nmi_perfctr_msr_to_bit()
53 return msr - MSR_K7_PERFCTR0; in nmi_perfctr_msr_to_bit()
56 return msr - MSR_ARCH_PERFMON_PERFCTR0; in nmi_perfctr_msr_to_bit()
60 return msr - MSR_P6_PERFCTR0; in nmi_perfctr_msr_to_bit()
62 return msr - MSR_KNC_PERFCTR0; in nmi_perfctr_msr_to_bit()
64 return msr - MSR_P4_BPU_PERFCTR0; in nmi_perfctr_msr_to_bit()
69 return msr - MSR_ARCH_PERFMON_PERFCTR0; in nmi_perfctr_msr_to_bit()
78 static inline unsigned int nmi_evntsel_msr_to_bit(unsigned int msr) in nmi_evntsel_msr_to_bit() argument
[all …]
/linux/arch/powerpc/kvm/
H A Dbook3s_hv_tm.c19 u64 msr = vcpu->arch.shregs.msr; in emulate_tx_failure() local
23 if (MSR_TM_SUSPENDED(vcpu->arch.shregs.msr)) in emulate_tx_failure()
25 if (msr & MSR_PR) { in emulate_tx_failure()
45 u64 msr = vcpu->arch.shregs.msr; in kvmhv_p9_tm_emulation() local
74 WARN_ON_ONCE(!(MSR_TM_SUSPENDED(msr) && in kvmhv_p9_tm_emulation()
78 vcpu->arch.shregs.msr = newmsr; in kvmhv_p9_tm_emulation()
84 if ((msr & MSR_PR) && (vcpu->arch.vcore->pcr & PCR_ARCH_206)) { in kvmhv_p9_tm_emulation()
96 if ((msr & MSR_PR) && !(vcpu->arch.fscr & FSCR_EBB)) { in kvmhv_p9_tm_emulation()
105 WARN_ON_ONCE(!(MSR_TM_SUSPENDED(msr) && in kvmhv_p9_tm_emulation()
111 msr = (msr & ~MSR_TS_MASK) | MSR_TS_T; in kvmhv_p9_tm_emulation()
[all …]
H A Dbook3s_hv_tm_builtin.c23 u64 newmsr, msr, bescr; in kvmhv_p9_tm_emulation_early() local
45 vcpu->arch.shregs.msr = newmsr; in kvmhv_p9_tm_emulation_early()
52 msr = vcpu->arch.shregs.msr; in kvmhv_p9_tm_emulation_early()
53 if ((msr & MSR_PR) && (vcpu->arch.vcore->pcr & PCR_ARCH_206)) in kvmhv_p9_tm_emulation_early()
57 ((msr & MSR_PR) && !(mfspr(SPRN_FSCR) & FSCR_EBB))) in kvmhv_p9_tm_emulation_early()
67 msr = (msr & ~MSR_TS_MASK) | MSR_TS_T; in kvmhv_p9_tm_emulation_early()
68 vcpu->arch.shregs.msr = msr; in kvmhv_p9_tm_emulation_early()
77 msr = vcpu->arch.shregs.msr; in kvmhv_p9_tm_emulation_early()
82 newmsr = (newmsr & ~MSR_LE) | (msr & MSR_LE); in kvmhv_p9_tm_emulation_early()
84 vcpu->arch.shregs.msr = newmsr; in kvmhv_p9_tm_emulation_early()
[all …]
H A Dbook3s_pr.c54 ulong msr);
69 ulong msr = kvmppc_get_msr(vcpu); in kvmppc_is_split_real() local
70 return (msr & (MSR_IR|MSR_DR)) == MSR_DR; in kvmppc_is_split_real()
75 ulong msr = kvmppc_get_msr(vcpu); in kvmppc_fixup_split_real() local
79 if ((msr & (MSR_IR|MSR_DR)) != MSR_DR) in kvmppc_fixup_split_real()
109 unsigned long msr, pc, new_msr, new_pc; in kvmppc_inject_interrupt_pr() local
113 msr = kvmppc_get_msr(vcpu); in kvmppc_inject_interrupt_pr()
120 if (MSR_TM_TRANSACTIONAL(msr)) in kvmppc_inject_interrupt_pr()
123 new_msr |= msr & MSR_TS_MASK; in kvmppc_inject_interrupt_pr()
127 kvmppc_set_srr1(vcpu, (msr & SRR1_MSR_BITS) | srr1_flags); in kvmppc_inject_interrupt_pr()
[all …]
/linux/arch/x86/include/asm/
H A Dmsr-trace.h3 #define TRACE_SYSTEM msr
6 #define TRACE_INCLUDE_FILE msr-trace
22 TP_PROTO(unsigned msr, u64 val, int failed),
23 TP_ARGS(msr, val, failed),
25 __field( unsigned, msr )
30 __entry->msr = msr;
35 __entry->msr,
41 TP_PROTO(unsigned msr, u64 val, int failed),
42 TP_ARGS(msr, val, failed)
46 TP_PROTO(unsigned msr, u64 val, int failed),
[all …]
/linux/arch/m68k/bvme6000/
H A Dconfig.c166 unsigned char msr; in bvme6000_timer_int() local
169 msr = rtc->msr & 0xc0; in bvme6000_timer_int()
170 rtc->msr = msr | 0x20; /* Ack the interrupt */ in bvme6000_timer_int()
191 unsigned char msr = rtc->msr & 0xc0; in bvme6000_sched_init() local
193 rtc->msr = 0; /* Ensure timer registers accessible */ in bvme6000_sched_init()
203 rtc->msr = 0x40; /* Access int.cntrl, etc */ in bvme6000_sched_init()
208 rtc->msr = 0; /* Access timer 1 control */ in bvme6000_sched_init()
211 rtc->msr = msr; in bvme6000_sched_init()
233 unsigned char msr, msb; in bvme6000_read_clk() local
239 msr = rtc->msr & 0xc0; in bvme6000_read_clk()
[all …]
H A Drtc.c42 unsigned char msr; in rtc_ioctl() local
52 msr = rtc->msr & 0xc0; in rtc_ioctl()
53 rtc->msr = 0x40; in rtc_ioctl()
66 rtc->msr = msr; in rtc_ioctl()
108 msr = rtc->msr & 0xc0; in rtc_ioctl()
109 rtc->msr = 0x40; in rtc_ioctl()
123 rtc->msr = msr; in rtc_ioctl()
/linux/drivers/powercap/
H A Dintel_rapl_msr.c40 .reg_unit.msr = MSR_RAPL_POWER_UNIT,
41 .regs[RAPL_DOMAIN_PACKAGE][RAPL_DOMAIN_REG_LIMIT].msr = MSR_PKG_POWER_LIMIT,
42 .regs[RAPL_DOMAIN_PACKAGE][RAPL_DOMAIN_REG_STATUS].msr = MSR_PKG_ENERGY_STATUS,
43 .regs[RAPL_DOMAIN_PACKAGE][RAPL_DOMAIN_REG_PERF].msr = MSR_PKG_PERF_STATUS,
44 .regs[RAPL_DOMAIN_PACKAGE][RAPL_DOMAIN_REG_INFO].msr = MSR_PKG_POWER_INFO,
45 .regs[RAPL_DOMAIN_PP0][RAPL_DOMAIN_REG_LIMIT].msr = MSR_PP0_POWER_LIMIT,
46 .regs[RAPL_DOMAIN_PP0][RAPL_DOMAIN_REG_STATUS].msr = MSR_PP0_ENERGY_STATUS,
47 .regs[RAPL_DOMAIN_PP0][RAPL_DOMAIN_REG_POLICY].msr = MSR_PP0_POLICY,
48 .regs[RAPL_DOMAIN_PP1][RAPL_DOMAIN_REG_LIMIT].msr = MSR_PP1_POWER_LIMIT,
49 .regs[RAPL_DOMAIN_PP1][RAPL_DOMAIN_REG_STATUS].msr = MSR_PP1_ENERGY_STATUS,
[all …]
/linux/arch/microblaze/kernel/
H A Dprocess.c45 regs->msr, regs->ear, regs->esr, regs->fsr); in show_regs()
72 local_save_flags(childregs->msr); in copy_thread()
73 ti->cpu_context.msr = childregs->msr & ~MSR_IE; in copy_thread()
83 childregs->msr |= MSR_UMS; in copy_thread()
95 childregs->msr &= ~MSR_EIP; in copy_thread()
96 childregs->msr |= MSR_IE; in copy_thread()
97 childregs->msr &= ~MSR_VM; in copy_thread()
98 childregs->msr |= MSR_VMS; in copy_thread()
99 childregs->msr |= MSR_EE; /* exceptions will be enabled*/ in copy_thread()
101 ti->cpu_context.msr = (childregs->msr|MSR_VM); in copy_thread()
[all …]
/linux/arch/powerpc/kernel/
H A Dsignal_64.c130 unsigned long msr = regs->msr; in __unsafe_setup_sigcontext() local
147 msr |= MSR_VEC; in __unsafe_setup_sigcontext()
163 msr &= ~MSR_VSX; in __unsafe_setup_sigcontext()
176 msr |= MSR_VSX; in __unsafe_setup_sigcontext()
181 unsafe_put_user(msr, &sc->gp_regs[PT_MSR], efault_out); in __unsafe_setup_sigcontext()
210 unsigned long msr) in setup_tm_sigcontexts() argument
229 BUG_ON(!MSR_TM_ACTIVE(msr)); in setup_tm_sigcontexts()
237 msr |= tsk->thread.ckpt_regs.msr & (MSR_FP | MSR_VEC | MSR_VSX); in setup_tm_sigcontexts()
251 if (msr & MSR_VEC) in setup_tm_sigcontexts()
263 msr |= MSR_VEC; in setup_tm_sigcontexts()
[all …]
H A Dsignal_32.c271 unsigned long msr = regs->msr; in __unsafe_save_user_regs() local
283 msr |= MSR_VEC; in __unsafe_save_user_regs()
302 msr &= ~MSR_VSX; in __unsafe_save_user_regs()
312 msr |= MSR_VSX; in __unsafe_save_user_regs()
322 msr |= MSR_SPE; in __unsafe_save_user_regs()
331 unsafe_put_user(msr, &frame->mc_gregs[PT_MSR], failed); in __unsafe_save_user_regs()
370 struct mcontext __user *tm_frame, unsigned long msr) in save_tm_user_regs_unsafe() argument
382 unsafe_put_user((msr >> 32), &tm_frame->mc_gregs[PT_MSR], failed); in save_tm_user_regs_unsafe()
388 if (msr & MSR_VEC) in save_tm_user_regs_unsafe()
400 msr |= MSR_VEC; in save_tm_user_regs_unsafe()
[all …]
H A Dcpu_setup_power.c19 u64 msr; in init_hvmode_206() local
21 msr = mfmsr(); in init_hvmode_206()
22 if (msr & MSR_HV) in init_hvmode_206()
153 u64 msr; in __restore_cpu_power7() local
155 msr = mfmsr(); in __restore_cpu_power7()
156 if (!(msr & MSR_HV)) in __restore_cpu_power7()
185 u64 msr; in __restore_cpu_power8() local
191 msr = mfmsr(); in __restore_cpu_power8()
192 if (!(msr & MSR_HV)) in __restore_cpu_power8()
225 u64 msr; in __restore_cpu_power9() local
[all …]
/linux/arch/powerpc/kernel/ptrace/
H A Dptrace-tm.c34 return task->thread.ckpt_regs.msr | task->thread.fpexc_mode; in get_user_ckpt_msr()
37 static int set_user_ckpt_msr(struct task_struct *task, unsigned long msr) in set_user_ckpt_msr() argument
39 task->thread.ckpt_regs.msr &= ~MSR_DEBUGCHANGE; in set_user_ckpt_msr()
40 task->thread.ckpt_regs.msr |= msr & MSR_DEBUGCHANGE; in set_user_ckpt_msr()
63 if (!MSR_TM_ACTIVE(target->thread.regs->msr)) in tm_cgpr_active()
89 struct membuf to_msr = membuf_at(&to, offsetof(struct pt_regs, msr)); in tm_cgpr_get()
97 if (!MSR_TM_ACTIVE(target->thread.regs->msr)) in tm_cgpr_get()
144 if (!MSR_TM_ACTIVE(target->thread.regs->msr)) in tm_cgpr_set()
164 offsetof(struct pt_regs, msr) + sizeof(long)); in tm_cgpr_set()
205 if (!MSR_TM_ACTIVE(target->thread.regs->msr)) in tm_cfpr_active()
[all …]
/linux/tools/power/x86/turbostat/
H A Dturbostat.c570 int get_msr(int cpu, off_t offset, unsigned long long *msr);
625 unsigned long long msr = 3; in slm_bclk() local
629 if (get_msr(master_cpu, MSR_FSB_FREQ, &msr)) in slm_bclk()
632 i = msr & 0xf; in slm_bclk()
1446 unsigned long long msr[NUM_RAPL_COUNTERS]; member
1464 unsigned long long msr; member
1479 .msr = MSR_PKG_ENERGY_STATUS,
1492 .msr = MSR_PKG_ENERGY_STATUS,
1505 .msr = MSR_PKG_ENERGY_STAT,
1518 .msr = MSR_PKG_ENERGY_STAT,
[all …]
/linux/tools/power/x86/x86_energy_perf_policy/
H A Dx86_energy_perf_policy.c688 int get_msr(int cpu, int offset, unsigned long long *msr) in get_msr() argument
701 retval = pread(fd, msr, sizeof(*msr), offset); in get_msr()
702 if (retval != sizeof(*msr)) { in get_msr()
708 fprintf(stderr, "get_msr(cpu%d, 0x%X, 0x%llX)\n", cpu, offset, *msr); in get_msr()
790 unsigned long long msr; in read_hwp_cap() local
792 get_msr(cpu, msr_offset, &msr); in read_hwp_cap()
794 cap->highest = msr_perf_2_ratio(HWP_HIGHEST_PERF(msr)); in read_hwp_cap()
795 cap->guaranteed = msr_perf_2_ratio(HWP_GUARANTEED_PERF(msr)); in read_hwp_cap()
796 cap->efficient = msr_perf_2_ratio(HWP_MOSTEFFICIENT_PERF(msr)); in read_hwp_cap()
797 cap->lowest = msr_perf_2_ratio(HWP_LOWEST_PERF(msr)); in read_hwp_cap()
[all …]
/linux/arch/arm64/kernel/
H A Dhyp-stub.S51 msr vbar_el2, x1
112 msr tpidr_el2, x0
116 msr cpacr_el1, x0
118 msr vbar_el1, x0
124 msr mdcr_el2, x0
128 msr tcr_el1, x0
130 msr ttbr0_el1, x0
132 msr ttbr1_el1, x0
134 msr mair_el1, x0
139 msr REG_TCR2_EL1, x0
[all …]
/linux/arch/arm64/mm/
H A Dproc.S139 msr tpidr_el0, x2
140 msr tpidrro_el0, x3
141 msr contextidr_el1, x4
142 msr cpacr_el1, x6
148 msr tcr_el1, x8
149 msr vbar_el1, x9
150 msr mdscr_el1, x10
153 msr REG_TCR2_EL1, x2
156 msr sctlr_el1, x12
158 msr sp_el0, x14
[all …]
/linux/arch/x86/power/
H A Dcpu.c43 struct saved_msr *msr = ctxt->saved_msrs.array; in msr_save_context() local
44 struct saved_msr *end = msr + ctxt->saved_msrs.num; in msr_save_context()
46 while (msr < end) { in msr_save_context()
47 if (msr->valid) in msr_save_context()
48 rdmsrq(msr->info.msr_no, msr->info.reg.q); in msr_save_context()
49 msr++; in msr_save_context()
55 struct saved_msr *msr = ctxt->saved_msrs.array; in msr_restore_context() local
56 struct saved_msr *end = msr + ctxt->saved_msrs.num; in msr_restore_context()
58 while (msr < end) { in msr_restore_context()
59 if (msr->valid) in msr_restore_context()
[all …]
/linux/tools/power/cpupower/debug/i386/
H A Dcentrino-decode.c28 static int rdmsr(unsigned int cpu, unsigned int msr, in rdmsr() argument
47 if (lseek(fd, msr, SEEK_CUR) == -1) in rdmsr()
63 static void decode (unsigned int msr) in decode() argument
68 multiplier = ((msr >> 8) & 0xFF); in decode()
70 mv = (((msr & 0xFF) * 16) + 700); in decode()
72 printf("0x%x means multiplier %d @ %d mV\n", msr, multiplier, mv); in decode()
/linux/sound/pci/ctxfi/
H A Dctatc.c263 desc.msr = atc->msr; in atc_pcm_playback_prepare()
270 (atc->rsr * atc->msr)); in atc_pcm_playback_prepare()
285 mix_dsc.msr = atc->msr; in atc_pcm_playback_prepare()
389 max_cisz = src->multi * src->rsc.msr; in atc_pcm_playback_start()
450 max_cisz = src->multi * src->rsc.msr; in atc_pcm_playback_position()
458 unsigned int msr:8; member
470 pitch = atc_get_pitch((atc->rsr * atc->msr), in setup_src_node_conf()
474 if (1 == atc->msr) { /* FIXME: do we really need SRC here if pitch==1 */ in setup_src_node_conf()
477 conf[0].mix_msr = conf[0].imp_msr = conf[0].msr = 1; in setup_src_node_conf()
479 } else if (2 <= atc->msr) { in setup_src_node_conf()
[all …]
H A Dctdaio.c162 entry = kzalloc_objs(*entry, daio->rscl.msr); in dao_set_left_input()
170 for (i = 0; i < daio->rscl.msr; i++, entry++) { in dao_set_left_input()
191 entry = kzalloc_objs(*entry, daio->rscr.msr); in dao_set_right_input()
199 for (i = 0; i < daio->rscr.msr; i++, entry++) { in dao_set_right_input()
203 dao->imappers[daio->rscl.msr + i] = entry; in dao_set_right_input()
231 return dao_clear_input(dao, 0, dao->daio.rscl.msr); in dao_clear_left_input()
236 return dao_clear_input(dao, dao->daio.rscl.msr, in dao_clear_right_input()
237 dao->daio.rscl.msr + dao->daio.rscr.msr); in dao_clear_right_input()
265 static int dai_set_srt_msr(struct dai *dai, unsigned int msr) in dai_set_srt_msr() argument
269 for (rsr = 0; msr > 1; msr >>= 1) in dai_set_srt_msr()
[all …]
H A Dctsrc.c182 if (src->rsc.msr > 1) { in src_commit_write()
193 for (i = 1; i < src->rsc.msr; i++) { in src_commit_write()
228 unsigned int rsr, msr; in src_default_config_memrd() local
232 for (rsr = 0, msr = src->rsc.msr; msr > 1; msr >>= 1) in src_default_config_memrd()
254 for (msr = 1; msr < src->rsc.msr; msr++) { in src_default_config_memrd()
296 unsigned int rsr, msr; in src_default_config_arcrw() local
301 for (rsr = 0, msr = src->rsc.msr; msr > 1; msr >>= 1) in src_default_config_arcrw()
323 for (msr = 0; msr < src->rsc.msr; msr++) { in src_default_config_arcrw()
365 err = rsc_init(&p->rsc, idx + i, SRC, desc->msr, mgr->mgr.hw); in src_rsc_init()
486 for (i = 0; i < src->rsc.msr; i++) { in src_enable_s()
[all …]

12345678910>>...12