xref: /linux/virt/kvm/kvm_mm.h (revision aec499c75cf8e0b599be4d559e6922b613085f8f)
1 // SPDX-License-Identifier: GPL-2.0-only
2 
3 #ifndef __KVM_MM_H__
4 #define __KVM_MM_H__ 1
5 
6 /*
7  * Architectures can choose whether to use an rwlock or spinlock
8  * for the mmu_lock.  These macros, for use in common code
9  * only, avoids using #ifdefs in places that must deal with
10  * multiple architectures.
11  */
12 
13 #ifdef KVM_HAVE_MMU_RWLOCK
14 #define KVM_MMU_LOCK_INIT(kvm)		rwlock_init(&(kvm)->mmu_lock)
15 #define KVM_MMU_LOCK(kvm)		write_lock(&(kvm)->mmu_lock)
16 #define KVM_MMU_UNLOCK(kvm)		write_unlock(&(kvm)->mmu_lock)
17 #define KVM_MMU_READ_LOCK(kvm)		read_lock(&(kvm)->mmu_lock)
18 #define KVM_MMU_READ_UNLOCK(kvm)	read_unlock(&(kvm)->mmu_lock)
19 #else
20 #define KVM_MMU_LOCK_INIT(kvm)		spin_lock_init(&(kvm)->mmu_lock)
21 #define KVM_MMU_LOCK(kvm)		spin_lock(&(kvm)->mmu_lock)
22 #define KVM_MMU_UNLOCK(kvm)		spin_unlock(&(kvm)->mmu_lock)
23 #define KVM_MMU_READ_LOCK(kvm)		spin_lock(&(kvm)->mmu_lock)
24 #define KVM_MMU_READ_UNLOCK(kvm)	spin_unlock(&(kvm)->mmu_lock)
25 #endif /* KVM_HAVE_MMU_RWLOCK */
26 
27 kvm_pfn_t hva_to_pfn(unsigned long addr, bool atomic, bool *async,
28 		     bool write_fault, bool *writable);
29 
30 #ifdef CONFIG_HAVE_KVM_PFNCACHE
31 void gfn_to_pfn_cache_invalidate_start(struct kvm *kvm,
32 				       unsigned long start,
33 				       unsigned long end,
34 				       bool may_block);
35 #else
36 static inline void gfn_to_pfn_cache_invalidate_start(struct kvm *kvm,
37 						     unsigned long start,
38 						     unsigned long end,
39 						     bool may_block)
40 {
41 }
42 #endif /* HAVE_KVM_PFNCACHE */
43 
44 #endif /* __KVM_MM_H__ */
45