/* SPDX-License-Identifier: GPL-2.0+ */ #pragma once /* * Contains declarations that are STUBBED, that is that are rendered no-ops, in * order to faciliate userland VMA testing. */ /* Forward declarations. */ struct mm_struct; struct vm_area_struct; struct vm_area_desc; struct pagetable_move_control; struct mmap_action; struct file; struct anon_vma; struct anon_vma_chain; struct address_space; struct unmap_desc; #define __bitwise #define __randomize_layout #define FIRST_USER_ADDRESS 0UL #define USER_PGTABLES_CEILING 0UL #define vma_policy(vma) NULL #define down_write_nest_lock(sem, nest_lock) #define data_race(expr) expr #define ASSERT_EXCLUSIVE_WRITER(x) struct vm_userfaultfd_ctx {}; struct mempolicy {}; struct mmu_gather {}; struct mutex {}; struct vm_fault {}; static inline void userfaultfd_unmap_complete(struct mm_struct *mm, struct list_head *uf) { } static inline unsigned long move_page_tables(struct pagetable_move_control *pmc) { return 0; } static inline void free_pgd_range(struct mmu_gather *tlb, unsigned long addr, unsigned long end, unsigned long floor, unsigned long ceiling) { } static inline int ksm_execve(struct mm_struct *mm) { return 0; } static inline void ksm_exit(struct mm_struct *mm) { } static inline void vma_numab_state_init(struct vm_area_struct *vma) { } static inline void vma_numab_state_free(struct vm_area_struct *vma) { } static inline void dup_anon_vma_name(struct vm_area_struct *orig_vma, struct vm_area_struct *new_vma) { } static inline void free_anon_vma_name(struct vm_area_struct *vma) { } static inline void mmap_action_prepare(struct mmap_action *action, struct vm_area_desc *desc) { } static inline int mmap_action_complete(struct mmap_action *action, struct vm_area_struct *vma) { return 0; } static inline void fixup_hugetlb_reservations(struct vm_area_struct *vma) { } static inline bool shmem_file(struct file *file) { return false; } static inline vm_flags_t ksm_vma_flags(const struct mm_struct *mm, const struct file *file, vm_flags_t vm_flags) { return vm_flags; } static inline void remap_pfn_range_prepare(struct vm_area_desc *desc, unsigned long pfn) { } static inline int remap_pfn_range_complete(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, unsigned long size, pgprot_t pgprot) { return 0; } static inline int do_munmap(struct mm_struct *, unsigned long, size_t, struct list_head *uf) { return 0; } /* Currently stubbed but we may later wish to un-stub. */ static inline void vm_acct_memory(long pages); static inline void mmap_assert_locked(struct mm_struct *mm) { } static inline void anon_vma_unlock_write(struct anon_vma *anon_vma) { } static inline void i_mmap_unlock_write(struct address_space *mapping) { } static inline int userfaultfd_unmap_prep(struct vm_area_struct *vma, unsigned long start, unsigned long end, struct list_head *unmaps) { return 0; } static inline void mmap_write_downgrade(struct mm_struct *mm) { } static inline void mmap_read_unlock(struct mm_struct *mm) { } static inline void mmap_write_unlock(struct mm_struct *mm) { } static inline int mmap_write_lock_killable(struct mm_struct *mm) { return 0; } static inline bool can_modify_mm(struct mm_struct *mm, unsigned long start, unsigned long end) { return true; } static inline void arch_unmap(struct mm_struct *mm, unsigned long start, unsigned long end) { } static inline bool mpol_equal(struct mempolicy *a, struct mempolicy *b) { return true; } static inline void khugepaged_enter_vma(struct vm_area_struct *vma, vm_flags_t vm_flags) { } static inline bool mapping_can_writeback(struct address_space *mapping) { return true; } static inline bool is_vm_hugetlb_page(struct vm_area_struct *vma) { return false; } static inline bool vma_soft_dirty_enabled(struct vm_area_struct *vma) { return false; } static inline bool userfaultfd_wp(struct vm_area_struct *vma) { return false; } static inline void mmap_assert_write_locked(struct mm_struct *mm) { } static inline void mutex_lock(struct mutex *lock) { } static inline void mutex_unlock(struct mutex *lock) { } static inline bool mutex_is_locked(struct mutex *lock) { return true; } static inline bool signal_pending(void *p) { return false; } static inline bool is_file_hugepages(struct file *file) { return false; } static inline int security_vm_enough_memory_mm(struct mm_struct *mm, long pages) { return 0; } static inline bool may_expand_vm(struct mm_struct *mm, vm_flags_t flags, unsigned long npages) { return true; } static inline int shmem_zero_setup(struct vm_area_struct *vma) { return 0; } static inline void vm_acct_memory(long pages) { } static inline void vma_interval_tree_insert(struct vm_area_struct *vma, struct rb_root_cached *rb) { } static inline void vma_interval_tree_remove(struct vm_area_struct *vma, struct rb_root_cached *rb) { } static inline void flush_dcache_mmap_unlock(struct address_space *mapping) { } static inline void anon_vma_interval_tree_insert(struct anon_vma_chain *avc, struct rb_root_cached *rb) { } static inline void anon_vma_interval_tree_remove(struct anon_vma_chain *avc, struct rb_root_cached *rb) { } static inline void uprobe_mmap(struct vm_area_struct *vma) { } static inline void uprobe_munmap(struct vm_area_struct *vma, unsigned long start, unsigned long end) { } static inline void i_mmap_lock_write(struct address_space *mapping) { } static inline void anon_vma_lock_write(struct anon_vma *anon_vma) { } static inline void vma_assert_write_locked(struct vm_area_struct *vma) { } static inline void ksm_add_vma(struct vm_area_struct *vma) { } static inline void perf_event_mmap(struct vm_area_struct *vma) { } static inline bool vma_is_dax(struct vm_area_struct *vma) { return false; } static inline struct vm_area_struct *get_gate_vma(struct mm_struct *mm) { return NULL; } static inline bool arch_validate_flags(vm_flags_t flags) { return true; } static inline void vma_close(struct vm_area_struct *vma) { } static inline int mmap_file(struct file *file, struct vm_area_struct *vma) { return 0; } static inline int is_hugepage_only_range(struct mm_struct *mm, unsigned long addr, unsigned long len) { return 0; } static inline bool capable(int cap) { return true; } static inline struct anon_vma_name *anon_vma_name(struct vm_area_struct *vma) { return NULL; } static inline bool is_mergeable_vm_userfaultfd_ctx(struct vm_area_struct *vma, struct vm_userfaultfd_ctx vm_ctx) { return true; } static inline bool anon_vma_name_eq(struct anon_vma_name *anon_name1, struct anon_vma_name *anon_name2) { return true; } static inline void might_sleep(void) { } static inline void fput(struct file *file) { } static inline void mpol_put(struct mempolicy *pol) { } static inline void lru_add_drain(void) { } static inline void tlb_gather_mmu(struct mmu_gather *tlb, struct mm_struct *mm) { } static inline void update_hiwater_rss(struct mm_struct *mm) { } static inline void update_hiwater_vm(struct mm_struct *mm) { } static inline void unmap_vmas(struct mmu_gather *tlb, struct unmap_desc *unmap) { } static inline void free_pgtables(struct mmu_gather *tlb, struct unmap_desc *unmap) { } static inline void mapping_unmap_writable(struct address_space *mapping) { } static inline void flush_dcache_mmap_lock(struct address_space *mapping) { } static inline void tlb_finish_mmu(struct mmu_gather *tlb) { } static inline struct file *get_file(struct file *f) { return f; } static inline int vma_dup_policy(struct vm_area_struct *src, struct vm_area_struct *dst) { return 0; } static inline void vma_adjust_trans_huge(struct vm_area_struct *vma, unsigned long start, unsigned long end, struct vm_area_struct *next) { } static inline void hugetlb_split(struct vm_area_struct *, unsigned long) {}