| /linux/arch/powerpc/kvm/ |
| H A D | booke_emulate.c | 26 vcpu->arch.regs.nip = vcpu->arch.shared->srr0; in kvmppc_emul_rfi() 27 kvmppc_set_msr(vcpu, vcpu->arch.shared->srr1); in kvmppc_emul_rfi() 32 vcpu->arch.regs.nip = vcpu->arch.dsrr0; in kvmppc_emul_rfdi() 33 kvmppc_set_msr(vcpu, vcpu->arch.dsrr1); in kvmppc_emul_rfdi() 38 vcpu->arch.regs.nip = vcpu->arch.csrr0; in kvmppc_emul_rfci() 39 kvmppc_set_msr(vcpu, vcpu->arch.csrr1); in kvmppc_emul_rfci() 80 kvmppc_set_gpr(vcpu, rt, vcpu->arch.shared->msr); in kvmppc_booke_emulate_op() 90 vcpu->arch.shared->msr = (vcpu->arch.shared->msr & ~MSR_EE) in kvmppc_booke_emulate_op() 96 vcpu->arch.shared->msr = (vcpu->arch.shared->msr & ~MSR_EE) in kvmppc_booke_emulate_op() 127 vcpu->arch.shared->dar = spr_val; in kvmppc_booke_emulate_mtspr() [all …]
|
| H A D | book3s_hv_tm.c | 19 u64 msr = vcpu->arch.shregs.msr; in emulate_tx_failure() 21 tfiar = vcpu->arch.regs.nip & ~0x3ull; in emulate_tx_failure() 23 if (MSR_TM_SUSPENDED(vcpu->arch.shregs.msr)) in emulate_tx_failure() 29 vcpu->arch.tfiar = tfiar; in emulate_tx_failure() 31 vcpu->arch.texasr = (vcpu->arch.texasr & 0x3ffffff) | texasr; in emulate_tx_failure() 44 u32 instr = vcpu->arch.emul_inst; in kvmhv_p9_tm_emulation() 45 u64 msr = vcpu->arch.shregs.msr; in kvmhv_p9_tm_emulation() 56 vcpu->arch.regs.nip -= 4; in kvmhv_p9_tm_emulation() 72 newmsr = vcpu->arch.shregs.srr1; in kvmhv_p9_tm_emulation() 78 vcpu->arch.shregs.msr = newmsr; in kvmhv_p9_tm_emulation() [all …]
|
| H A D | book3s_hv_p9_entry.c | 14 mtspr(SPRN_TAR, vcpu->arch.tar); in load_spr_state() 18 current->thread.vrsave != vcpu->arch.vrsave) in load_spr_state() 19 mtspr(SPRN_VRSAVE, vcpu->arch.vrsave); in load_spr_state() 22 if (vcpu->arch.hfscr & HFSCR_EBB) { in load_spr_state() 23 if (current->thread.ebbhr != vcpu->arch.ebbhr) in load_spr_state() 24 mtspr(SPRN_EBBHR, vcpu->arch.ebbhr); in load_spr_state() 25 if (current->thread.ebbrr != vcpu->arch.ebbrr) in load_spr_state() 26 mtspr(SPRN_EBBRR, vcpu->arch.ebbrr); in load_spr_state() 27 if (current->thread.bescr != vcpu->arch.bescr) in load_spr_state() 28 mtspr(SPRN_BESCR, vcpu->arch.bescr); in load_spr_state() [all …]
|
| H A D | booke.c | 94 printk("pc: %08lx msr: %08llx\n", vcpu->arch.regs.nip, in kvmppc_dump_vcpu() 95 vcpu->arch.shared->msr); in kvmppc_dump_vcpu() 96 printk("lr: %08lx ctr: %08lx\n", vcpu->arch.regs.link, in kvmppc_dump_vcpu() 97 vcpu->arch.regs.ctr); in kvmppc_dump_vcpu() 98 printk("srr0: %08llx srr1: %08llx\n", vcpu->arch.shared->srr0, in kvmppc_dump_vcpu() 99 vcpu->arch.shared->srr1); in kvmppc_dump_vcpu() 101 printk("exceptions: %08lx\n", vcpu->arch.pending_exceptions); in kvmppc_dump_vcpu() 119 vcpu->arch.shadow_msr &= ~MSR_SPE; in kvmppc_vcpu_disable_spe() 129 vcpu->arch.shadow_msr |= MSR_SPE; in kvmppc_vcpu_enable_spe() 135 if (vcpu->arch.shared->msr & MSR_SPE) { in kvmppc_vcpu_sync_spe() [all …]
|
| H A D | emulate_loadstore.c | 85 vcpu->arch.mmio_vsx_copy_nums = 0; in kvmppc_emulate_loadstore() 86 vcpu->arch.mmio_vsx_offset = 0; in kvmppc_emulate_loadstore() 87 vcpu->arch.mmio_copy_type = KVMPPC_VSX_COPY_NONE; in kvmppc_emulate_loadstore() 88 vcpu->arch.mmio_sp64_extend = 0; in kvmppc_emulate_loadstore() 89 vcpu->arch.mmio_sign_extend = 0; in kvmppc_emulate_loadstore() 90 vcpu->arch.mmio_vmx_copy_nums = 0; in kvmppc_emulate_loadstore() 91 vcpu->arch.mmio_vmx_offset = 0; in kvmppc_emulate_loadstore() 92 vcpu->arch.mmio_host_swabbed = 0; in kvmppc_emulate_loadstore() 95 vcpu->arch.regs.msr = kvmppc_get_msr(vcpu); in kvmppc_emulate_loadstore() 96 if (analyse_instr(&op, &vcpu->arch.regs, inst) == 0) { in kvmppc_emulate_loadstore() [all …]
|
| H A D | timing.c | 27 mutex_lock(&vcpu->arch.exit_timing_lock); in kvmppc_init_timing_stats() 29 vcpu->arch.last_exit_type = 0xDEAD; in kvmppc_init_timing_stats() 31 vcpu->arch.timing_count_type[i] = 0; in kvmppc_init_timing_stats() 32 vcpu->arch.timing_max_duration[i] = 0; in kvmppc_init_timing_stats() 33 vcpu->arch.timing_min_duration[i] = 0xFFFFFFFF; in kvmppc_init_timing_stats() 34 vcpu->arch.timing_sum_duration[i] = 0; in kvmppc_init_timing_stats() 35 vcpu->arch.timing_sum_quad_duration[i] = 0; in kvmppc_init_timing_stats() 37 vcpu->arch.timing_last_exit = 0; in kvmppc_init_timing_stats() 38 vcpu->arch.timing_exit.tv64 = 0; in kvmppc_init_timing_stats() 39 vcpu->arch.timing_last_enter.tv64 = 0; in kvmppc_init_timing_stats() [all …]
|
| H A D | book3s_hv_p9_perf.c | 41 lp = vcpu->arch.vpa.pinned_addr; in switch_pmu_to_guest() 87 if (load_pmu || (vcpu->arch.hfscr & HFSCR_PM)) { in switch_pmu_to_guest() 88 mtspr(SPRN_PMC1, vcpu->arch.pmc[0]); in switch_pmu_to_guest() 89 mtspr(SPRN_PMC2, vcpu->arch.pmc[1]); in switch_pmu_to_guest() 90 mtspr(SPRN_PMC3, vcpu->arch.pmc[2]); in switch_pmu_to_guest() 91 mtspr(SPRN_PMC4, vcpu->arch.pmc[3]); in switch_pmu_to_guest() 92 mtspr(SPRN_PMC5, vcpu->arch.pmc[4]); in switch_pmu_to_guest() 93 mtspr(SPRN_PMC6, vcpu->arch.pmc[5]); in switch_pmu_to_guest() 94 mtspr(SPRN_MMCR1, vcpu->arch.mmcr[1]); in switch_pmu_to_guest() 95 mtspr(SPRN_MMCR2, vcpu->arch.mmcr[2]); in switch_pmu_to_guest() [all …]
|
| H A D | book3s_hv.c | 14 * This file is derived from arch/powerpc/kvm/book3s.c, 243 cpu = READ_ONCE(vcpu->arch.thread_cpu); in kvmppc_fast_vcpu_kick_hv() 280 * Updates to busy_stolen are protected by arch.tbacct_lock; 320 struct kvmppc_vcore *vc = vcpu->arch.vcore; in kvmppc_core_vcpu_load_hv() 325 if (vcpu->arch.busy_preempt != TB_NIL) { in kvmppc_core_vcpu_load_hv() 326 WARN_ON_ONCE(vcpu->arch.state != KVMPPC_VCPU_BUSY_IN_HOST); in kvmppc_core_vcpu_load_hv() 327 vc->stolen_tb += mftb() - vcpu->arch.busy_preempt; in kvmppc_core_vcpu_load_hv() 328 vcpu->arch.busy_preempt = TB_NIL; in kvmppc_core_vcpu_load_hv() 344 spin_lock_irqsave(&vcpu->arch.tbacct_lock, flags); in kvmppc_core_vcpu_load_hv() 345 if (vcpu->arch in kvmppc_core_vcpu_load_hv() 4196 struct kvm_vcpu_arch *arch; kvmhv_get_l1_to_l2_cs_time_vcpu() local 4211 struct kvm_vcpu_arch *arch; kvmhv_get_l2_to_l1_cs_time_vcpu() local 4226 struct kvm_vcpu_arch *arch; kvmhv_get_l2_runtime_agg_vcpu() local [all...] |
| H A D | book3s_emulate.c | 73 if (vcpu->arch.papr_enabled && (level > PRIV_SUPER)) in spr_allowed() 86 memcpy(&vcpu->arch.gpr_tm[0], &vcpu->arch.regs.gpr[0], in kvmppc_copyto_vcpu_tm() 87 sizeof(vcpu->arch.gpr_tm)); in kvmppc_copyto_vcpu_tm() 88 memcpy(&vcpu->arch.fp_tm, &vcpu->arch.fp, in kvmppc_copyto_vcpu_tm() 90 memcpy(&vcpu->arch.vr_tm, &vcpu->arch.vr, in kvmppc_copyto_vcpu_tm() 92 vcpu->arch.ppr_tm = vcpu->arch.ppr; in kvmppc_copyto_vcpu_tm() 93 vcpu->arch.dscr_tm = vcpu->arch.dscr; in kvmppc_copyto_vcpu_tm() 94 vcpu->arch.amr_tm = vcpu->arch.amr; in kvmppc_copyto_vcpu_tm() 95 vcpu->arch.ctr_tm = vcpu->arch.regs.ctr; in kvmppc_copyto_vcpu_tm() 96 vcpu->arch.tar_tm = vcpu->arch.tar; in kvmppc_copyto_vcpu_tm() [all …]
|
| H A D | book3s_hv_tm_builtin.c | 22 u32 instr = vcpu->arch.emul_inst; in kvmhv_p9_tm_emulation_early() 40 newmsr = vcpu->arch.shregs.srr1; in kvmhv_p9_tm_emulation_early() 45 vcpu->arch.shregs.msr = newmsr; in kvmhv_p9_tm_emulation_early() 46 vcpu->arch.cfar = vcpu->arch.regs.nip - 4; in kvmhv_p9_tm_emulation_early() 47 vcpu->arch.regs.nip = vcpu->arch.shregs.srr0; in kvmhv_p9_tm_emulation_early() 52 msr = vcpu->arch.shregs.msr; in kvmhv_p9_tm_emulation_early() 53 if ((msr & MSR_PR) && (vcpu->arch.vcore->pcr & PCR_ARCH_206)) in kvmhv_p9_tm_emulation_early() 56 if (!(vcpu->arch.hfscr & HFSCR_EBB) || in kvmhv_p9_tm_emulation_early() 68 vcpu->arch.shregs.msr = msr; in kvmhv_p9_tm_emulation_early() 69 vcpu->arch.cfar = vcpu->arch.regs.nip - 4; in kvmhv_p9_tm_emulation_early() [all …]
|
| H A D | book3s_pr.c | 15 * This file is derived from arch/powerpc/kvm/44x.c, 83 if (vcpu->arch.hflags & BOOK3S_HFLAG_SPLIT_HACK) in kvmppc_fixup_split_real() 90 vcpu->arch.hflags |= BOOK3S_HFLAG_SPLIT_HACK; in kvmppc_fixup_split_real() 96 if (vcpu->arch.hflags & BOOK3S_HFLAG_SPLIT_HACK) { in kvmppc_unfixup_split_real() 103 vcpu->arch.hflags &= ~BOOK3S_HFLAG_SPLIT_HACK; in kvmppc_unfixup_split_real() 115 new_msr = vcpu->arch.intr_msr; in kvmppc_inject_interrupt_pr() 152 current->thread.kvm_shadow_vcpu = vcpu->arch.shadow_vcpu; in kvmppc_core_vcpu_load_pr() 196 svcpu->gpr[0] = vcpu->arch.regs.gpr[0]; in kvmppc_copy_to_svcpu() 197 svcpu->gpr[1] = vcpu->arch.regs.gpr[1]; in kvmppc_copy_to_svcpu() 198 svcpu->gpr[2] = vcpu->arch in kvmppc_copy_to_svcpu() [all...] |
| H A D | e500_emulate.c | 53 ulong param = vcpu->arch.regs.gpr[rb]; in kvmppc_e500_emul_msgclr() 59 clear_bit(prio, &vcpu->arch.pending_exceptions); in kvmppc_e500_emul_msgclr() 65 ulong param = vcpu->arch.regs.gpr[rb]; in kvmppc_e500_emul_msgsnd() 75 int cpir = cvcpu->arch.shared->pir; in kvmppc_e500_emul_msgsnd() 77 set_bit(prio, &cvcpu->arch.pending_exceptions); in kvmppc_e500_emul_msgsnd() 94 vcpu->run->debug.arch.address = vcpu->arch.regs.nip; in kvmppc_e500_emul_ehpriv() 95 vcpu->run->debug.arch.status = 0; in kvmppc_e500_emul_ehpriv() 225 vcpu->arch.shared->mas0 = spr_val; in kvmppc_core_emulate_mtspr_e500() 228 vcpu->arch.shared->mas1 = spr_val; in kvmppc_core_emulate_mtspr_e500() 231 vcpu->arch.shared->mas2 = spr_val; in kvmppc_core_emulate_mtspr_e500() [all …]
|
| H A D | e500mc.c | 107 vcpu->arch.pid = pid; in kvmppc_set_pid() 124 mtspr(SPRN_EPCR, vcpu->arch.shadow_epcr); in kvmppc_core_vcpu_load_e500mc() 126 mtspr(SPRN_MSRP, vcpu->arch.shadow_msrp); in kvmppc_core_vcpu_load_e500mc() 127 vcpu->arch.eplc = EPC_EGS | (get_lpid(vcpu) << EPC_ELPID_SHIFT); in kvmppc_core_vcpu_load_e500mc() 128 vcpu->arch.epsc = vcpu->arch.eplc; in kvmppc_core_vcpu_load_e500mc() 129 mtspr(SPRN_EPLC, vcpu->arch.eplc); in kvmppc_core_vcpu_load_e500mc() 130 mtspr(SPRN_EPSC, vcpu->arch.epsc); in kvmppc_core_vcpu_load_e500mc() 132 mtspr(SPRN_GIVPR, vcpu->arch.ivpr); in kvmppc_core_vcpu_load_e500mc() 133 mtspr(SPRN_GIVOR2, vcpu->arch.ivor[BOOKE_IRQPRIO_DATA_STORAGE]); in kvmppc_core_vcpu_load_e500mc() 134 mtspr(SPRN_GIVOR8, vcpu->arch.ivor[BOOKE_IRQPRIO_SYSCALL]); in kvmppc_core_vcpu_load_e500mc() [all …]
|
| H A D | book3s_hv_nested.c | 32 struct kvmppc_vcore *vc = vcpu->arch.vcore; in kvmhv_save_hv_regs() 35 hr->dpdes = vcpu->arch.doorbell_request; in kvmhv_save_hv_regs() 36 hr->hfscr = vcpu->arch.hfscr; in kvmhv_save_hv_regs() 38 hr->dawr0 = vcpu->arch.dawr0; in kvmhv_save_hv_regs() 39 hr->dawrx0 = vcpu->arch.dawrx0; in kvmhv_save_hv_regs() 40 hr->ciabr = vcpu->arch.ciabr; in kvmhv_save_hv_regs() 41 hr->purr = vcpu->arch.purr; in kvmhv_save_hv_regs() 42 hr->spurr = vcpu->arch.spurr; in kvmhv_save_hv_regs() 43 hr->ic = vcpu->arch.ic; in kvmhv_save_hv_regs() 45 hr->srr0 = vcpu->arch in kvmhv_save_hv_regs() [all...] |
| /linux/arch/mips/kvm/ |
| H A D | emulate.c | 45 struct kvm_vcpu_arch *arch = &vcpu->arch; in kvm_compute_return_epc() local 65 arch->gprs[insn.r_format.rd] = epc + 8; in kvm_compute_return_epc() 68 nextpc = arch->gprs[insn.r_format.rs]; in kvm_compute_return_epc() 84 if ((long)arch->gprs[insn.i_format.rs] < 0) in kvm_compute_return_epc() 93 if ((long)arch->gprs[insn.i_format.rs] >= 0) in kvm_compute_return_epc() 102 arch->gprs[31] = epc + 8; in kvm_compute_return_epc() 103 if ((long)arch->gprs[insn.i_format.rs] < 0) in kvm_compute_return_epc() 112 arch->gprs[31] = epc + 8; in kvm_compute_return_epc() 113 if ((long)arch->gprs[insn.i_format.rs] >= 0) in kvm_compute_return_epc() 141 arch->gprs[31] = instpc + 8; in kvm_compute_return_epc() [all …]
|
| /linux/arch/s390/kvm/ |
| H A D | guestdbg.c | 62 u64 *cr9 = &vcpu->arch.sie_block->gcr[9]; in enable_all_hw_bp() 63 u64 *cr10 = &vcpu->arch.sie_block->gcr[10]; in enable_all_hw_bp() 64 u64 *cr11 = &vcpu->arch.sie_block->gcr[11]; in enable_all_hw_bp() 67 if (vcpu->arch.guestdbg.nr_hw_bp <= 0 || in enable_all_hw_bp() 68 vcpu->arch.guestdbg.hw_bp_info == NULL) in enable_all_hw_bp() 79 for (i = 0; i < vcpu->arch.guestdbg.nr_hw_bp; i++) { in enable_all_hw_bp() 80 start = vcpu->arch.guestdbg.hw_bp_info[i].addr; in enable_all_hw_bp() 81 len = vcpu->arch.guestdbg.hw_bp_info[i].len; in enable_all_hw_bp() 102 u64 *cr9 = &vcpu->arch.sie_block->gcr[9]; in enable_all_hw_wp() 103 u64 *cr10 = &vcpu->arch.sie_block->gcr[10]; in enable_all_hw_wp() [all …]
|
| H A D | kvm-s390.c | 311 kvm_clock_sync_scb(vcpu->arch.sie_block, *delta); in kvm_clock_sync() 313 kvm->arch.epoch = vcpu->arch.sie_block->epoch; in kvm_clock_sync() 314 kvm->arch.epdx = vcpu->arch.sie_block->epdx; in kvm_clock_sync() 316 if (vcpu->arch.cputm_enabled) in kvm_clock_sync() 317 vcpu->arch.cputm_start += *delta; in kvm_clock_sync() 318 if (vcpu->arch.vsie_block) in kvm_clock_sync() 319 kvm_clock_sync_scb(vcpu->arch.vsie_block, in kvm_clock_sync() 560 union asce asce = kvm->arch.gmap->asce; in kvm_s390_keyop() 726 gmap_sync_dirty_log(kvm->arch.gmap, memslot->base_gfn, last_gfn); in kvm_arch_sync_dirty_log() 787 kvm->arch.use_irqchip = 1; in kvm_vm_ioctl_enable_cap() [all …]
|
| /linux/arch/loongarch/kvm/ |
| H A D | vm.c | 33 kvm->arch.kvm_features |= BIT(KVM_LOONGARCH_VM_FEAT_LSX); in kvm_vm_init_features() 35 kvm->arch.kvm_features |= BIT(KVM_LOONGARCH_VM_FEAT_LASX); in kvm_vm_init_features() 37 kvm->arch.kvm_features |= BIT(KVM_LOONGARCH_VM_FEAT_X86BT); in kvm_vm_init_features() 39 kvm->arch.kvm_features |= BIT(KVM_LOONGARCH_VM_FEAT_ARMBT); in kvm_vm_init_features() 41 kvm->arch.kvm_features |= BIT(KVM_LOONGARCH_VM_FEAT_MIPSBT); in kvm_vm_init_features() 43 kvm->arch.kvm_features |= BIT(KVM_LOONGARCH_VM_FEAT_PTW); in kvm_vm_init_features() 45 kvm->arch.kvm_features |= BIT(KVM_LOONGARCH_VM_FEAT_MSGINT); in kvm_vm_init_features() 49 kvm->arch.kvm_features |= BIT(KVM_LOONGARCH_VM_FEAT_PMU); in kvm_vm_init_features() 52 kvm->arch.pv_features = BIT(KVM_FEATURE_IPI); in kvm_vm_init_features() 53 kvm->arch.kvm_features = BIT(KVM_LOONGARCH_VM_FEAT_PV_IPI); in kvm_vm_init_features() [all …]
|
| H A D | vcpu.c | 45 context = this_cpu_ptr(vcpu->kvm->arch.vmcs); in kvm_save_host_pmu() 60 context = this_cpu_ptr(vcpu->kvm->arch.vmcs); in kvm_restore_host_pmu() 74 struct loongarch_csrs *csr = vcpu->arch.csr; in kvm_save_guest_pmu() 88 struct loongarch_csrs *csr = vcpu->arch.csr; in kvm_restore_guest_pmu() 104 if (!kvm_guest_has_pmu(&vcpu->arch)) in kvm_own_pmu() 111 val |= (kvm_get_pmu_num(&vcpu->arch) + 1) << CSR_GCFG_GPERF_SHIFT; in kvm_own_pmu() 122 struct loongarch_csrs *csr = vcpu->arch.csr; in kvm_lose_pmu() 124 if (!(vcpu->arch.aux_inuse & KVM_LARCH_PMU)) in kvm_lose_pmu() 145 vcpu->arch.aux_inuse &= ~KVM_LARCH_PMU; in kvm_lose_pmu() 156 vcpu->arch in kvm_check_pmu() [all...] |
| /linux/scripts/ |
| H A D | head-object-list.txt | 14 arch/alpha/kernel/head.o 15 arch/arc/kernel/head.o 16 arch/arm/kernel/head-nommu.o 17 arch/arm/kernel/head.o 18 arch/csky/kernel/head.o 19 arch/hexagon/kernel/head.o 20 arch/loongarch/kernel/head.o 21 arch/m68k/68000/head.o 22 arch/m68k/coldfire/head.o 23 arch/m68k/kernel/head.o [all …]
|
| H A D | checkstack.pl | 42 my $arch = shift; 43 if ($arch eq "") { 44 $arch = `uname -m`; 45 chomp($arch); 56 if ($arch =~ '^(aarch|arm)64$') { 61 } elsif ($arch eq 'arm') { 65 } elsif ($arch =~ /^x86(_64)?$/ || $arch =~ /^i[3456]86$/) { 71 } elsif ($arch eq 'm68k') { 75 } elsif ($arch eq 'mips64') { 78 } elsif ($arch eq 'mips') { [all …]
|
| /linux/tools/perf/util/ |
| H A D | disasm.h | 25 struct arch { struct 44 bool (*ins_is_fused)(const struct arch *arch, const char *ins1, argument 46 const struct ins_ops *(*associate_instruction_ops)(struct arch *arch, const char *name); argument 93 int (*parse)(const struct arch *arch, struct ins_operands *ops, struct map_symbol *ms, 102 const struct arch *arch; member 111 const struct arch *arch__find(uint16_t e_machine, uint32_t e_flags, const char *cpuid); 112 bool arch__is_x86(const struct arch *arch); 113 bool arch__is_powerpc(const struct arch *arch); 123 int arch__associate_ins_ops(struct arch *arch, const char *name, const struct ins_ops *ops); 125 const struct arch *arch__new_arc(const struct e_machine_and_e_flags *id, const char *cpuid); [all …]
|
| /linux/arch/riscv/kvm/ |
| H A D | vcpu.c | 57 struct kvm_vcpu_csr *csr = &vcpu->arch.guest_csr; in kvm_riscv_vcpu_context_reset() 58 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_context_reset() 63 memset(&vcpu->arch.smstateen_csr, 0, sizeof(vcpu->arch.smstateen_csr)); in kvm_riscv_vcpu_context_reset() 93 vcpu->arch.last_exit_cpu = -1; in kvm_riscv_reset_vcpu() 105 bitmap_zero(vcpu->arch.irqs_pending, KVM_RISCV_VCPU_NR_IRQS); in kvm_riscv_reset_vcpu() 106 bitmap_zero(vcpu->arch.irqs_pending_mask, KVM_RISCV_VCPU_NR_IRQS); in kvm_riscv_reset_vcpu() 110 vcpu->arch.hfence_head = 0; in kvm_riscv_reset_vcpu() 111 vcpu->arch.hfence_tail = 0; in kvm_riscv_reset_vcpu() 112 memset(vcpu->arch.hfence_queue, 0, sizeof(vcpu->arch.hfence_queue)); in kvm_riscv_reset_vcpu() 131 spin_lock_init(&vcpu->arch.mp_state_lock); in kvm_arch_vcpu_create() [all …]
|
| /linux/arch/x86/kvm/ |
| H A D | xen.c | 40 struct gfn_to_pfn_cache *gpc = &kvm->arch.xen.shinfo_cache; in kvm_xen_shared_info_init() 67 BUILD_BUG_ON(offsetof(struct compat_shared_info, arch.wc_sec_hi) != 0x924); in kvm_xen_shared_info_init() 75 if (IS_ENABLED(CONFIG_64BIT) && kvm->arch.xen.long_mode) { in kvm_xen_shared_info_init() 85 wc_sec_hi = &shinfo->arch.wc_sec_hi; in kvm_xen_shared_info_init() 110 if (atomic_read(&vcpu->arch.xen.timer_pending) > 0) { in kvm_xen_inject_timer_irqs() 115 e.port = vcpu->arch.xen.timer_virq; in kvm_xen_inject_timer_irqs() 120 vcpu->arch.xen.timer_expires = 0; in kvm_xen_inject_timer_irqs() 121 atomic_set(&vcpu->arch.xen.timer_pending, 0); in kvm_xen_inject_timer_irqs() 128 arch.xen.timer); in xen_timer_callback() 132 if (atomic_read(&vcpu->arch in xen_timer_callback() [all...] |
| /linux/tools/perf/ |
| H A D | check-headers.sh | 187 check arch/x86/lib/memcpy_64.S '-I "^EXPORT_SYMBOL" -I "^#include <asm/export.h>" -I"^SYM_FU… 188 check arch/x86/lib/memset_64.S '-I "^EXPORT_SYMBOL" -I "^#include <asm/export.h>" -I"^SYM_FU… 189 check arch/x86/include/asm/amd/ibs.h '-I "^#include .*/msr-index.h"' 190 check arch/arm64/include/asm/cputype.h '-I "^#include [<\"]\(asm/\)*sysreg.h"' 199 check_2 tools/perf/arch/x86/entry/syscalls/syscall_32.tbl arch/x86/entry/syscalls/syscall_32.tbl 200 check_2 tools/perf/arch/x86/entry/syscalls/syscall_64.tbl arch/x86/entry/syscalls/syscall_64.tbl 201 check_2 tools/perf/arch/powerpc/entry/syscalls/syscall.tbl arch/powerpc/kernel/syscalls/syscall.tbl 202 check_2 tools/perf/arch/s390/entry/syscalls/syscall.tbl arch/s390/kernel/syscalls/syscall.tbl 203 check_2 tools/perf/arch/mips/entry/syscalls/syscall_n64.tbl arch/mips/kernel/syscalls/syscall_n64.t… 204 check_2 tools/perf/arch/arm/entry/syscalls/syscall.tbl arch/arm/tools/syscall.tbl [all …]
|