| H A D | mmu.c | 30 #include "spte.h" 175 #define for_each_shadow_entry_lockless(_vcpu, _addr, _walker, spte) \ argument 178 ({ spte = mmu_spte_get_lockless(_walker.sptep); 1; }); \ 179 __shadow_walk_next(&(_walker), spte)) 184 static void mmu_spte_set(u64 *sptep, u64 spte); 295 u64 spte = make_mmio_spte(vcpu, gfn, access); in mark_mmio_spte() local 297 trace_mark_mmio_spte(sptep, gfn, spte); in mark_mmio_spte() 298 mmu_spte_set(sptep, spte); in mark_mmio_spte() 301 static gfn_t get_mmio_spte_gfn(u64 spte) in get_mmio_spte_gfn() argument 303 u64 gpa = spte in get_mmio_spte_gfn() 311 get_mmio_spte_access(u64 spte) get_mmio_spte_access() argument 316 check_mmio_spte(struct kvm_vcpu * vcpu,u64 spte) check_mmio_spte() argument 337 __set_spte(u64 * sptep,u64 spte) __set_spte() argument 343 __update_clear_spte_fast(u64 * sptep,u64 spte) __update_clear_spte_fast() argument 349 __update_clear_spte_slow(u64 * sptep,u64 spte) __update_clear_spte_slow() argument 365 u64 spte; global() member 368 count_spte_clear(u64 * sptep,u64 spte) count_spte_clear() argument 380 __set_spte(u64 * sptep,u64 spte) __set_spte() argument 399 __update_clear_spte_fast(u64 * sptep,u64 spte) __update_clear_spte_fast() argument 418 __update_clear_spte_slow(u64 * sptep,u64 spte) __update_clear_spte_slow() argument 455 union split_spte spte, *orig = (union split_spte *)sptep; __get_spte_lockless() local 1008 pte_list_add(struct kvm * kvm,struct kvm_mmu_memory_cache * cache,u64 * spte,struct kvm_rmap_head * rmap_head) pte_list_add() argument 1087 pte_list_remove(struct kvm * kvm,u64 * spte,struct kvm_rmap_head * rmap_head) pte_list_remove() argument 1184 rmap_remove(struct kvm * kvm,u64 * spte) rmap_remove() argument 1318 u64 spte = *sptep; spte_write_protect() local 1346 u64 spte = *sptep; spte_clear_dirty() local 1682 __rmap_add(struct kvm * kvm,struct kvm_mmu_memory_cache * cache,const struct kvm_memory_slot * slot,u64 * spte,gfn_t gfn,unsigned int access) __rmap_add() argument 1704 rmap_add(struct kvm_vcpu * vcpu,const struct kvm_memory_slot * slot,u64 * spte,gfn_t gfn,unsigned int access) rmap_add() argument 1722 u64 spte; kvm_rmap_age_gfn_range() local 1869 mark_unsync(u64 * spte) mark_unsync() argument 2244 clear_sp_write_flooding_count(u64 * spte) clear_sp_write_flooding_count() argument 2523 __shadow_walk_next(struct kvm_shadow_walk_iterator * iterator,u64 spte) __shadow_walk_next() argument 2543 u64 spte; __link_shadow_page() local 2604 mmu_page_zap_pte(struct kvm * kvm,struct kvm_mmu_page * sp,u64 * spte,struct list_head * invalid_list) mmu_page_zap_pte() argument 3040 u64 spte; mmu_set_spte() local 3153 u64 *spte, *start = NULL; __direct_pte_prefetch() local 3412 disallowed_hugepage_adjust(struct kvm_page_fault * fault,u64 spte,int cur_level) disallowed_hugepage_adjust() argument 3640 fast_pf_get_last_sptep(struct kvm_vcpu * vcpu,gpa_t gpa,u64 * spte) fast_pf_get_last_sptep() argument 3661 u64 spte; fast_page_fault() local 4367 u64 spte; get_walk() local 4442 u64 spte; handle_mmio_page_fault() local 4496 u64 spte; shadow_page_table_clear_flood() local 6213 u64 *spte; get_written_sptes() local 6247 u64 entry, gentry, *spte; kvm_mmu_track_write() local 7060 u64 *sptep, spte; shadow_mmu_split_huge_page() local 7104 u64 spte; shadow_mmu_try_split_huge_page() local [all...] |