Home
last modified time | relevance | path

Searched refs:kvm (Results 1 – 25 of 272) sorted by relevance

1234567891011

/linux/arch/powerpc/include/asm/
H A Dkvm_book3s_uvmem.h9 int kvmppc_uvmem_slot_init(struct kvm *kvm, const struct kvm_memory_slot *slot);
10 void kvmppc_uvmem_slot_free(struct kvm *kvm,
12 unsigned long kvmppc_h_svm_page_in(struct kvm *kvm,
16 unsigned long kvmppc_h_svm_page_out(struct kvm *kvm,
20 unsigned long kvmppc_h_svm_init_start(struct kvm *kvm);
21 unsigned long kvmppc_h_svm_init_done(struct kvm *kvm);
22 int kvmppc_send_page_to_uv(struct kvm *kvm, unsigned long gfn);
23 unsigned long kvmppc_h_svm_init_abort(struct kvm *kvm);
25 struct kvm *kvm, bool skip_page_out);
26 int kvmppc_uvmem_memslot_create(struct kvm *kvm,
[all …]
H A Dkvm_ppc.h169 extern void kvmppc_set_hpt(struct kvm *kvm, struct kvm_hpt_info *info);
170 extern int kvmppc_alloc_reset_hpt(struct kvm *kvm, int order);
172 extern void kvmppc_rmap_reset(struct kvm *kvm);
176 extern long kvm_spapr_tce_attach_iommu_group(struct kvm *kvm, int tablefd,
178 extern void kvm_spapr_tce_release_iommu_group(struct kvm *kvm,
180 extern int kvmppc_switch_mmu_to_hpt(struct kvm *kvm);
181 extern int kvmppc_switch_mmu_to_radix(struct kvm *kvm);
182 extern void kvmppc_setup_partition_table(struct kvm *kvm);
184 extern int kvm_vm_ioctl_create_spapr_tce(struct kvm *kvm,
202 extern int kvmppc_core_init_vm(struct kvm *kvm);
[all …]
/linux/arch/x86/kvm/mmu/
H A Dpage_track.c23 static bool kvm_external_write_tracking_enabled(struct kvm *kvm) in kvm_external_write_tracking_enabled() argument
30 return smp_load_acquire(&kvm->arch.external_write_tracking_enabled); in kvm_external_write_tracking_enabled()
36 bool kvm_page_track_write_tracking_enabled(struct kvm *kvm) in kvm_page_track_write_tracking_enabled() argument
38 return kvm_external_write_tracking_enabled(kvm) || in kvm_page_track_write_tracking_enabled()
39 kvm_shadow_root_allocated(kvm) || !tdp_enabled; in kvm_page_track_write_tracking_enabled()
60 int kvm_page_track_create_memslot(struct kvm *kvm, in kvm_page_track_create_memslot() argument
64 if (!kvm_page_track_write_tracking_enabled(kvm)) in kvm_page_track_create_memslot()
90 void __kvm_write_track_add_gfn(struct kvm *kvm, struct kvm_memory_slot *slot, in __kvm_write_track_add_gfn() argument
93 lockdep_assert_held_write(&kvm->mmu_lock); in __kvm_write_track_add_gfn()
95 lockdep_assert_once(lockdep_is_held(&kvm->slots_lock) || in __kvm_write_track_add_gfn()
[all …]
H A Dtdp_mmu.c15 void kvm_mmu_init_tdp_mmu(struct kvm *kvm) in kvm_mmu_init_tdp_mmu() argument
17 INIT_LIST_HEAD(&kvm->arch.tdp_mmu_roots); in kvm_mmu_init_tdp_mmu()
18 spin_lock_init(&kvm->arch.tdp_mmu_pages_lock); in kvm_mmu_init_tdp_mmu()
22 static __always_inline bool kvm_lockdep_assert_mmu_lock_held(struct kvm *kvm, in kvm_lockdep_assert_mmu_lock_held() argument
26 lockdep_assert_held_read(&kvm->mmu_lock); in kvm_lockdep_assert_mmu_lock_held()
28 lockdep_assert_held_write(&kvm->mmu_lock); in kvm_lockdep_assert_mmu_lock_held()
33 void kvm_mmu_uninit_tdp_mmu(struct kvm *kvm) in kvm_mmu_uninit_tdp_mmu() argument
40 kvm_tdp_mmu_invalidate_roots(kvm, KVM_VALID_ROOTS); in kvm_mmu_uninit_tdp_mmu()
41 kvm_tdp_mmu_zap_invalidated_roots(kvm, false); in kvm_mmu_uninit_tdp_mmu()
43 WARN_ON(atomic64_read(&kvm->arch.tdp_mmu_pages)); in kvm_mmu_uninit_tdp_mmu()
[all …]
H A Dpage_track.h10 bool kvm_page_track_write_tracking_enabled(struct kvm *kvm);
14 int kvm_page_track_create_memslot(struct kvm *kvm,
18 void __kvm_write_track_add_gfn(struct kvm *kvm, struct kvm_memory_slot *slot,
20 void __kvm_write_track_remove_gfn(struct kvm *kvm,
23 bool kvm_gfn_is_write_tracked(struct kvm *kvm,
27 int kvm_page_track_init(struct kvm *kvm);
28 void kvm_page_track_cleanup(struct kvm *kvm);
30 void __kvm_page_track_write(struct kvm *kvm, gpa_t gpa, const u8 *new, int bytes);
31 void kvm_page_track_delete_slot(struct kvm *kvm, struct kvm_memory_slot *slot);
33 static inline bool kvm_page_track_has_external_user(struct kvm *kvm) in kvm_page_track_has_external_user() argument
[all …]
H A Dmmu.c283 static void kvm_flush_remote_tlbs_sptep(struct kvm *kvm, u64 *sptep) in kvm_flush_remote_tlbs_sptep() argument
288 kvm_flush_remote_tlbs_gfn(kvm, gfn, sp->role.level); in kvm_flush_remote_tlbs_sptep()
521 static u64 mmu_spte_clear_track_bits(struct kvm *kvm, u64 *sptep) in mmu_spte_clear_track_bits() argument
535 kvm_update_page_stats(kvm, level, -1); in mmu_spte_clear_track_bits()
602 if (kvm_has_mirrored_tdp(vcpu->kvm)) { in mmu_topup_memory_caches()
748 static void account_shadowed(struct kvm *kvm, struct kvm_mmu_page *sp) in account_shadowed() argument
754 kvm->arch.indirect_shadow_pages++; in account_shadowed()
765 slots = kvm_memslots_for_spte_role(kvm, sp->role); in account_shadowed()
770 return __kvm_write_track_add_gfn(kvm, slot, gfn); in account_shadowed()
774 if (kvm_mmu_slot_gfn_write_protect(kvm, slot, gfn, PG_LEVEL_4K)) in account_shadowed()
[all …]
/linux/include/linux/
H A Dkvm_host.h187 bool kvm_make_vcpus_request_mask(struct kvm *kvm, unsigned int req,
189 bool kvm_make_all_cpus_request(struct kvm *kvm, unsigned int req);
226 int kvm_io_bus_register_dev(struct kvm *kvm, enum kvm_bus bus_idx, gpa_t addr,
228 int kvm_io_bus_unregister_dev(struct kvm *kvm, enum kvm_bus bus_idx,
230 struct kvm_io_device *kvm_io_bus_get_dev(struct kvm *kvm, enum kvm_bus bus_idx,
271 bool kvm_unmap_gfn_range(struct kvm *kvm, struct kvm_gfn_range *range);
272 bool kvm_age_gfn(struct kvm *kvm, struct kvm_gfn_range *range);
273 bool kvm_test_age_gfn(struct kvm *kvm, struct kvm_gfn_range *range);
323 struct kvm *kvm; member
666 struct kvm *kvm, int irq_source_id, int level,
[all …]
/linux/virt/kvm/
H A Deventfd.c36 kvm_arch_irqfd_allowed(struct kvm *kvm, struct kvm_irqfd *args) in kvm_arch_irqfd_allowed() argument
46 struct kvm *kvm = irqfd->kvm; in irqfd_inject() local
49 kvm_set_irq(kvm, KVM_USERSPACE_IRQ_SOURCE_ID, irqfd->gsi, 1, in irqfd_inject()
51 kvm_set_irq(kvm, KVM_USERSPACE_IRQ_SOURCE_ID, irqfd->gsi, 0, in irqfd_inject()
54 kvm_set_irq(kvm, KVM_IRQFD_RESAMPLE_IRQ_SOURCE_ID, in irqfd_inject()
63 srcu_read_lock_held(&resampler->kvm->irq_srcu)) in irqfd_resampler_notify()
76 struct kvm *kvm; in irqfd_resampler_ack() local
81 kvm = resampler->kvm; in irqfd_resampler_ack()
83 kvm_set_irq(kvm, KVM_IRQFD_RESAMPLE_IRQ_SOURCE_ID, in irqfd_resampler_ack()
86 idx = srcu_read_lock(&kvm->irq_srcu); in irqfd_resampler_ack()
[all …]
H A Dkvm_main.c153 static void kvm_uevent_notify_change(unsigned int type, struct kvm *kvm);
159 __weak void kvm_arch_guest_memory_reclaimed(struct kvm *kvm) in kvm_arch_guest_memory_reclaimed() argument
246 bool kvm_make_vcpus_request_mask(struct kvm *kvm, unsigned int req, in kvm_make_vcpus_request_mask() argument
260 vcpu = kvm_get_vcpu(kvm, i); in kvm_make_vcpus_request_mask()
272 bool kvm_make_all_cpus_request(struct kvm *kvm, unsigned int req) in kvm_make_all_cpus_request() argument
285 kvm_for_each_vcpu(i, vcpu, kvm) in kvm_make_all_cpus_request()
295 void kvm_flush_remote_tlbs(struct kvm *kvm) in kvm_flush_remote_tlbs() argument
297 ++kvm->stat.generic.remote_tlb_flush_requests; in kvm_flush_remote_tlbs()
310 if (!kvm_arch_flush_remote_tlbs(kvm) in kvm_flush_remote_tlbs()
311 || kvm_make_all_cpus_request(kvm, KVM_REQ_TLB_FLUSH)) in kvm_flush_remote_tlbs()
[all …]
H A Dirqchip.c21 int kvm_irq_map_gsi(struct kvm *kvm, in kvm_irq_map_gsi() argument
28 irq_rt = srcu_dereference_check(kvm->irq_routing, &kvm->irq_srcu, in kvm_irq_map_gsi()
29 lockdep_is_held(&kvm->irq_lock)); in kvm_irq_map_gsi()
40 int kvm_irq_map_chip_pin(struct kvm *kvm, unsigned irqchip, unsigned pin) in kvm_irq_map_chip_pin() argument
44 irq_rt = srcu_dereference(kvm->irq_routing, &kvm->irq_srcu); in kvm_irq_map_chip_pin()
48 int kvm_send_userspace_msi(struct kvm *kvm, struct kvm_msi *msi) in kvm_send_userspace_msi() argument
52 if (!kvm_arch_irqchip_in_kernel(kvm) || (msi->flags & ~KVM_MSI_VALID_DEVID)) in kvm_send_userspace_msi()
61 return kvm_set_msi(&route, kvm, KVM_USERSPACE_IRQ_SOURCE_ID, 1, false); in kvm_send_userspace_msi()
70 int kvm_set_irq(struct kvm *kvm, int irq_source_id, u32 irq, int level, in kvm_set_irq() argument
82 idx = srcu_read_lock(&kvm->irq_srcu); in kvm_set_irq()
[all …]
H A Dkvm_mm.h14 #define KVM_MMU_LOCK_INIT(kvm) rwlock_init(&(kvm)->mmu_lock) argument
15 #define KVM_MMU_LOCK(kvm) write_lock(&(kvm)->mmu_lock) argument
16 #define KVM_MMU_UNLOCK(kvm) write_unlock(&(kvm)->mmu_lock) argument
18 #define KVM_MMU_LOCK_INIT(kvm) spin_lock_init(&(kvm)->mmu_lock) argument
19 #define KVM_MMU_LOCK(kvm) spin_lock(&(kvm) argument
20 KVM_MMU_UNLOCK(kvm) global() argument
31 gfn_to_pfn_cache_invalidate_start(struct kvm * kvm,unsigned long start,unsigned long end) gfn_to_pfn_cache_invalidate_start() argument
50 kvm_gmem_bind(struct kvm * kvm,struct kvm_memory_slot * slot,unsigned int fd,loff_t offset) kvm_gmem_bind() argument
[all...]
/linux/arch/x86/kvm/
H A Dirq_comm.c31 struct kvm *kvm, int irq_source_id, int level, in kvm_set_pic_irq() argument
34 struct kvm_pic *pic = kvm->arch.vpic; in kvm_set_pic_irq()
39 struct kvm *kvm, int irq_source_id, int level, in kvm_set_ioapic_irq() argument
42 struct kvm_ioapic *ioapic = kvm->arch.vioapic; in kvm_set_ioapic_irq()
47 int kvm_irq_delivery_to_apic(struct kvm *kvm, struct kvm_lapic *src, in kvm_irq_delivery_to_apic() argument
55 if (kvm_irq_delivery_to_apic_fast(kvm, src, irq, &r, dest_map)) in kvm_irq_delivery_to_apic()
66 kvm_for_each_vcpu(i, vcpu, kvm) { in kvm_irq_delivery_to_apic()
95 lowest = kvm_get_vcpu(kvm, idx); in kvm_irq_delivery_to_apic()
104 void kvm_set_msi_irq(struct kvm *kvm, struct kvm_kernel_irq_routing_entry *e, in kvm_set_msi_irq() argument
111 trace_kvm_msi_set_irq(msg.address_lo | (kvm->arch.x2apic_format ? in kvm_set_msi_irq()
[all …]
H A Dirq.h25 struct kvm;
53 struct kvm *kvm; member
62 int kvm_pic_init(struct kvm *kvm);
63 void kvm_pic_destroy(struct kvm *kvm);
64 int kvm_pic_read_irq(struct kvm *kvm);
67 static inline int irqchip_split(struct kvm *kvm) in irqchip_split() argument
69 int mode = kvm->arch.irqchip_mode; in irqchip_split()
76 static inline int irqchip_kernel(struct kvm *kvm) in irqchip_kernel() argument
78 int mode = kvm->arch.irqchip_mode; in irqchip_kernel()
85 static inline int pic_in_kernel(struct kvm *kvm) in pic_in_kernel() argument
[all …]
H A Dxen.h24 int kvm_xen_hvm_set_attr(struct kvm *kvm, struct kvm_xen_hvm_attr *data);
25 int kvm_xen_hvm_get_attr(struct kvm *kvm, struct kvm_xen_hvm_attr *data);
26 int kvm_xen_hvm_evtchn_send(struct kvm *kvm, struct kvm_irq_routing_xen_evtchn *evt);
28 int kvm_xen_hvm_config(struct kvm *kvm, struct kvm_xen_hvm_config *xhc);
29 void kvm_xen_init_vm(struct kvm *kvm);
30 void kvm_xen_destroy_vm(struct kvm *kvm);
34 struct kvm *kvm);
35 int kvm_xen_setup_evtchn(struct kvm *kvm,
53 static inline bool kvm_xen_msr_enabled(struct kvm *kvm) in kvm_xen_msr_enabled() argument
56 kvm->arch.xen_hvm_config.msr; in kvm_xen_msr_enabled()
[all …]
/linux/arch/s390/kvm/
H A Dpv.c22 bool kvm_s390_pv_is_protected(struct kvm *kvm) in kvm_s390_pv_is_protected() argument
24 lockdep_assert_held(&kvm->lock); in kvm_s390_pv_is_protected()
25 return !!kvm_s390_pv_get_handle(kvm); in kvm_s390_pv_is_protected()
59 static void kvm_s390_clear_pv_state(struct kvm *kvm) in kvm_s390_clear_pv_state() argument
61 kvm->arch.pv.handle = 0; in kvm_s390_clear_pv_state()
62 kvm->arch.pv.guest_len = 0; in kvm_s390_clear_pv_state()
63 kvm->arch.pv.stor_base = 0; in kvm_s390_clear_pv_state()
64 kvm->arch.pv.stor_var = NULL; in kvm_s390_clear_pv_state()
76 KVM_UV_EVENT(vcpu->kvm, 3, "PROTVIRT DESTROY VCPU %d: rc %x rrc %x", in kvm_s390_pv_destroy_cpu()
119 uvcb.guest_handle = kvm_s390_pv_get_handle(vcpu->kvm); in kvm_s390_pv_create_cpu()
[all …]
H A Dkvm-s390.c272 static int sca_switch_to_extended(struct kvm *kvm);
305 struct kvm *kvm; in kvm_clock_sync() local
310 list_for_each_entry(kvm, &vm_list, vm_list) { in kvm_clock_sync()
311 kvm_for_each_vcpu(i, vcpu, kvm) { in kvm_clock_sync()
314 kvm->arch.epoch = vcpu->arch.sie_block->epoch; in kvm_clock_sync()
315 kvm->arch.epdx = vcpu->arch.sie_block->epdx; in kvm_clock_sync()
573 int kvm_vm_ioctl_check_extension(struct kvm *kvm, long ext) in kvm_vm_ioctl_check_extension() argument
614 if (hpage && !(kvm && kvm_is_ucontrol(kvm))) in kvm_vm_ioctl_check_extension()
693 void kvm_arch_sync_dirty_log(struct kvm *kvm, struct kvm_memory_slot *memslot) in kvm_arch_sync_dirty_log() argument
698 struct gmap *gmap = kvm->arch.gmap; in kvm_arch_sync_dirty_log()
[all …]
H A Dkvm-s390.h75 debug_sprintf_event(d_vcpu->kvm->arch.dbf, d_loglevel, \
103 return test_bit(vcpu->vcpu_idx, vcpu->kvm->arch.idle_mask); in is_vcpu_idle()
106 static inline int kvm_is_ucontrol(struct kvm *kvm) in kvm_is_ucontrol() argument
109 if (kvm->arch.gmap) in kvm_is_ucontrol()
218 static inline int test_kvm_facility(struct kvm *kvm, unsigned long nr) in test_kvm_facility() argument
220 return __test_facility(nr, kvm->arch.model.fac_mask) && in test_kvm_facility()
221 __test_facility(nr, kvm->arch.model.fac_list); in test_kvm_facility()
235 static inline int test_kvm_cpu_feat(struct kvm *kvm, unsigned long nr) in test_kvm_cpu_feat() argument
238 return test_bit_inv(nr, kvm->arch.cpu_feat); in test_kvm_cpu_feat()
242 static inline int kvm_s390_user_cpu_state_ctrl(struct kvm *kvm) in kvm_s390_user_cpu_state_ctrl() argument
[all …]
/linux/arch/loongarch/kvm/
H A Dvm.c27 int kvm_arch_init_vm(struct kvm *kvm, unsigned long type) in kvm_arch_init_vm() argument
32 kvm->arch.pgd = kvm_pgd_alloc(); in kvm_arch_init_vm()
33 if (!kvm->arch.pgd) in kvm_arch_init_vm()
36 kvm->arch.phyid_map = kvzalloc(sizeof(struct kvm_phyid_map), GFP_KERNEL_ACCOUNT); in kvm_arch_init_vm()
37 if (!kvm->arch.phyid_map) { in kvm_arch_init_vm()
38 free_page((unsigned long)kvm->arch.pgd); in kvm_arch_init_vm()
39 kvm->arch.pgd = NULL; in kvm_arch_init_vm()
42 spin_lock_init(&kvm->arch.phyid_map_lock); in kvm_arch_init_vm()
44 kvm_init_vmcs(kvm); in kvm_arch_init_vm()
47 kvm->arch.pv_features = BIT(KVM_FEATURE_IPI); in kvm_arch_init_vm()
[all …]
/linux/arch/arm64/kvm/
H A Dnested.c30 void kvm_init_nested(struct kvm *kvm) in kvm_init_nested() argument
32 kvm->arch.nested_mmus = NULL; in kvm_init_nested()
33 kvm->arch.nested_mmus_size = 0; in kvm_init_nested()
36 static int init_nested_s2_mmu(struct kvm *kvm, struct kvm_s2_mmu *mmu) in init_nested_s2_mmu() argument
48 return kvm_init_stage2_mmu(kvm, mmu, kvm_get_pa_bits(kvm)); in init_nested_s2_mmu()
53 struct kvm *kvm = vcpu->kvm; in kvm_vcpu_init_nested() local
63 num_mmus = atomic_read(&kvm->online_vcpus) * S2_MMU_PER_VCPU; in kvm_vcpu_init_nested()
64 tmp = kvrealloc(kvm->arch.nested_mmus, in kvm_vcpu_init_nested()
65 size_mul(sizeof(*kvm->arch.nested_mmus), num_mmus), in kvm_vcpu_init_nested()
70 swap(kvm->arch.nested_mmus, tmp); in kvm_vcpu_init_nested()
[all …]
/linux/arch/riscv/kvm/
H A Daia_device.c15 static void unlock_vcpus(struct kvm *kvm, int vcpu_lock_idx) in unlock_vcpus() argument
20 tmp_vcpu = kvm_get_vcpu(kvm, vcpu_lock_idx); in unlock_vcpus()
25 static void unlock_all_vcpus(struct kvm *kvm) in unlock_all_vcpus() argument
27 unlock_vcpus(kvm, atomic_read(&kvm->online_vcpus) - 1); in unlock_all_vcpus()
30 static bool lock_all_vcpus(struct kvm *kvm) in lock_all_vcpus() argument
35 kvm_for_each_vcpu(c, tmp_vcpu, kvm) { in lock_all_vcpus()
37 unlock_vcpus(kvm, c - 1); in lock_all_vcpus()
49 struct kvm *kvm = dev->kvm; in aia_create() local
52 if (irqchip_in_kernel(kvm)) in aia_create()
56 if (!lock_all_vcpus(kvm)) in aia_create()
[all …]
H A Dvm.c30 int kvm_arch_init_vm(struct kvm *kvm, unsigned long type) in kvm_arch_init_vm() argument
34 r = kvm_riscv_gstage_alloc_pgd(kvm); in kvm_arch_init_vm()
38 r = kvm_riscv_gstage_vmid_init(kvm); in kvm_arch_init_vm()
40 kvm_riscv_gstage_free_pgd(kvm); in kvm_arch_init_vm()
44 kvm_riscv_aia_init_vm(kvm); in kvm_arch_init_vm()
46 kvm_riscv_guest_timer_init(kvm); in kvm_arch_init_vm()
51 void kvm_arch_destroy_vm(struct kvm *kvm) in kvm_arch_destroy_vm() argument
53 kvm_destroy_vcpus(kvm); in kvm_arch_destroy_vm()
55 kvm_riscv_aia_destroy_vm(kvm); in kvm_arch_destroy_vm()
58 int kvm_vm_ioctl_irq_line(struct kvm *kvm, struct kvm_irq_level *irql, in kvm_vm_ioctl_irq_line() argument
[all …]
/linux/arch/arm64/kvm/vgic/
H A Dvgic.h105 return vcpu->kvm->arch.vgic.implementation_rev; in vgic_get_implementation_rev()
136 static inline int vgic_write_guest_lock(struct kvm *kvm, gpa_t gpa, in vgic_write_guest_lock() argument
139 struct vgic_dist *dist = &kvm->arch.vgic; in vgic_write_guest_lock()
143 ret = kvm_write_guest_lock(kvm, gpa, data, len); in vgic_write_guest_lock()
182 struct vgic_irq *vgic_get_irq(struct kvm *kvm, u32 intid);
184 void vgic_put_irq(struct kvm *kvm, struct vgic_irq *irq);
188 bool vgic_queue_irq_unlock(struct kvm *kvm, struct vgic_irq *irq,
190 void vgic_kick_vcpus(struct kvm *kvm);
194 int vgic_check_iorange(struct kvm *kvm, phys_addr_t ioaddr,
211 int vgic_v2_map_resources(struct kvm *kvm);
[all …]
H A Dvgic-debug.c35 static void iter_next(struct kvm *kvm, struct vgic_state_iter *iter) in iter_next() argument
37 struct vgic_dist *dist = &kvm->arch.vgic; in iter_next()
64 static int iter_mark_lpis(struct kvm *kvm) in iter_mark_lpis() argument
66 struct vgic_dist *dist = &kvm->arch.vgic; in iter_mark_lpis()
82 static void iter_unmark_lpis(struct kvm *kvm) in iter_unmark_lpis() argument
84 struct vgic_dist *dist = &kvm->arch.vgic; in iter_unmark_lpis()
90 vgic_put_irq(kvm, irq); in iter_unmark_lpis()
94 static void iter_init(struct kvm *kvm, struct vgic_state_iter *iter, in iter_init() argument
97 int nr_cpus = atomic_read(&kvm->online_vcpus); in iter_init()
102 iter->nr_spis = kvm->arch.vgic.nr_spis; in iter_init()
[all …]
H A Dvgic-its.c31 static int update_lpi_config(struct kvm *kvm, struct vgic_irq *irq,
37 struct kvm *__k = (i)->dev->kvm; \
54 struct kvm *__k = (i)->dev->kvm; \
76 static struct vgic_irq *vgic_add_lpi(struct kvm *kvm, u32 intid, in vgic_add_lpi() argument
79 struct vgic_dist *dist = &kvm->arch.vgic; in vgic_add_lpi()
80 struct vgic_irq *irq = vgic_get_irq(kvm, intid), *oldirq; in vgic_add_lpi()
142 ret = update_lpi_config(kvm, irq, NULL, false); in vgic_add_lpi()
144 vgic_put_irq(kvm, irq); in vgic_add_lpi()
148 ret = vgic_v3_lpi_sync_pending_status(kvm, irq); in vgic_add_lpi()
150 vgic_put_irq(kvm, irq); in vgic_add_lpi()
[all …]
/linux/arch/powerpc/kvm/
H A DMakefile6 ccflags-y := -Ivirt/kvm -Iarch/powerpc/kvm
8 include $(srctree)/virt/kvm/Makefile.kvm
16 kvm-e500-objs := \
26 kvm-objs-$(CONFIG_KVM_E500V2) := $(kvm-e500-objs)
28 kvm-e500mc-objs := \
38 kvm-objs-$(CONFIG_KVM_E500MC) := $(kvm-e500mc-objs)
40 kvm-pr-y := \
53 kvm-book3s_64-builtin-objs-$(CONFIG_KVM_BOOK3S_64_HANDLER) += \
58 kvm-book3s_64-builtin-objs-$(CONFIG_KVM_BOOK3S_64_HANDLER) += \
62 kvm-hv-y += \
[all …]

1234567891011