| /linux/include/linux/ |
| H A D | userfaultfd_k.h | 145 static inline bool is_mergeable_vm_userfaultfd_ctx(struct vm_area_struct *vma, in is_mergeable_vm_userfaultfd_ctx() 162 static inline bool uffd_disable_huge_pmd_share(struct vm_area_struct *vma) in uffd_disable_huge_pmd_share() 174 static inline bool uffd_disable_fault_around(struct vm_area_struct *vma) in uffd_disable_fault_around() 179 static inline bool userfaultfd_missing(struct vm_area_struct *vma) in userfaultfd_missing() 184 static inline bool userfaultfd_wp(struct vm_area_struct *vma) in userfaultfd_wp() 189 static inline bool userfaultfd_minor(struct vm_area_struct *vma) in userfaultfd_minor() 194 static inline bool userfaultfd_pte_wp(struct vm_area_struct *vma, in userfaultfd_pte_wp() 200 static inline bool userfaultfd_huge_pmd_wp(struct vm_area_struct *vma, in userfaultfd_huge_pmd_wp() 206 static inline bool userfaultfd_armed(struct vm_area_struct *vma) in userfaultfd_armed() 211 static inline bool vma_can_userfault(struct vm_area_struct *vma, in vma_can_userfault() [all …]
|
| H A D | mmap_lock.h | 113 vma_lock_init(struct vm_area_struct * vma,bool reset_refcnt) vma_lock_init() argument 136 vma_refcount_put(struct vm_area_struct * vma) vma_refcount_put() argument 156 vma_start_read_locked_nested(struct vm_area_struct * vma,int subclass) vma_start_read_locked_nested() argument 175 vma_start_read_locked(struct vm_area_struct * vma) vma_start_read_locked() argument 180 vma_end_read(struct vm_area_struct * vma) vma_end_read() argument 186 __is_vma_write_locked(struct vm_area_struct * vma,unsigned int * mm_lock_seq) __is_vma_write_locked() argument 206 vma_start_write(struct vm_area_struct * vma) vma_start_write() argument 230 vma_start_write_killable(struct vm_area_struct * vma) vma_start_write_killable() argument 239 vma_assert_write_locked(struct vm_area_struct * vma) vma_assert_write_locked() argument 246 vma_assert_locked(struct vm_area_struct * vma) vma_assert_locked() argument 259 vma_assert_attached(struct vm_area_struct * vma) vma_assert_attached() argument 264 vma_assert_detached(struct vm_area_struct * vma) vma_assert_detached() argument 269 vma_mark_attached(struct vm_area_struct * vma) vma_mark_attached() argument 307 vma_lock_init(struct vm_area_struct * vma,bool reset_refcnt) vma_lock_init() argument 308 vma_end_read(struct vm_area_struct * vma) vma_end_read() argument 309 vma_start_write(struct vm_area_struct * vma) vma_start_write() argument 311 vma_start_write_killable(struct vm_area_struct * vma) vma_start_write_killable() argument 312 vma_assert_write_locked(struct vm_area_struct * vma) vma_assert_write_locked() argument 314 vma_assert_attached(struct vm_area_struct * vma) vma_assert_attached() argument 315 vma_assert_detached(struct vm_area_struct * vma) vma_assert_detached() argument 316 vma_mark_attached(struct vm_area_struct * vma) vma_mark_attached() argument 317 vma_mark_detached(struct vm_area_struct * vma) vma_mark_detached() argument 325 vma_assert_locked(struct vm_area_struct * vma) vma_assert_locked() argument [all...] |
| H A D | rmap.h | 84 struct vm_area_struct *vma; member 160 anon_vma_prepare(struct vm_area_struct * vma) anon_vma_prepare() argument 168 anon_vma_merge(struct vm_area_struct * vma,struct vm_area_struct * next) anon_vma_merge() argument 230 folio_set_large_mapcount(struct folio * folio,int mapcount,struct vm_area_struct * vma) folio_set_large_mapcount() argument 244 folio_add_return_large_mapcount(struct folio * folio,int diff,struct vm_area_struct * vma) folio_add_return_large_mapcount() argument 294 folio_sub_return_large_mapcount(struct folio * folio,int diff,struct vm_area_struct * vma) folio_sub_return_large_mapcount() argument 345 folio_set_large_mapcount(struct folio * folio,int mapcount,struct vm_area_struct * vma) folio_set_large_mapcount() argument 352 folio_add_large_mapcount(struct folio * folio,int diff,struct vm_area_struct * vma) folio_add_large_mapcount() argument 358 folio_add_return_large_mapcount(struct folio * folio,int diff,struct vm_area_struct * vma) folio_add_return_large_mapcount() argument 364 folio_sub_large_mapcount(struct folio * folio,int diff,struct vm_area_struct * vma) folio_sub_large_mapcount() argument 370 folio_sub_return_large_mapcount(struct folio * folio,int diff,struct vm_area_struct * vma) folio_sub_return_large_mapcount() argument 376 folio_inc_large_mapcount(folio,vma) global() argument 378 folio_inc_return_large_mapcount(folio,vma) global() argument 380 folio_dec_large_mapcount(folio,vma) global() argument 382 folio_dec_return_large_mapcount(folio,vma) global() argument 472 folio_add_anon_rmap_pte(folio,page,vma,address,flags) global() argument 480 folio_add_file_rmap_pte(folio,page,vma) global() argument 488 folio_remove_rmap_pte(folio,page,vma) global() argument 502 hugetlb_try_dup_anon_rmap(struct folio * folio,struct vm_area_struct * vma) hugetlb_try_dup_anon_rmap() argument 934 struct vm_area_struct *vma; global() member 1045 anon_vma_prepare(vma) global() argument [all...] |
| H A D | hugetlb.h | 108 struct vm_area_struct *vma; member 254 hugetlb_zap_begin(struct vm_area_struct * vma,unsigned long * start,unsigned long * end) hugetlb_zap_begin() argument 261 hugetlb_zap_end(struct vm_area_struct * vma,struct zap_details * details) hugetlb_zap_end() argument 285 hugetlb_dup_vma_private(struct vm_area_struct * vma) hugetlb_dup_vma_private() argument 289 clear_vma_resv_huge_pages(struct vm_area_struct * vma) clear_vma_resv_huge_pages() argument 305 huge_pmd_unshare(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long addr,pte_t * ptep) huge_pmd_unshare() argument 311 huge_pmd_unshare_flush(struct mmu_gather * tlb,struct vm_area_struct * vma) huge_pmd_unshare_flush() argument 316 adjust_range_if_pmd_sharing_possible(struct vm_area_struct * vma,unsigned long * start,unsigned long * end) adjust_range_if_pmd_sharing_possible() argument 322 hugetlb_zap_begin(struct vm_area_struct * vma,unsigned long * start,unsigned long * end) hugetlb_zap_begin() argument 328 hugetlb_zap_end(struct vm_area_struct * vma,struct zap_details * details) hugetlb_zap_end() argument 342 move_hugetlb_page_tables(struct vm_area_struct * vma,struct vm_area_struct * new_vma,unsigned long old_addr,unsigned long new_addr,unsigned long len) move_hugetlb_page_tables() argument 365 hugetlb_vma_lock_read(struct vm_area_struct * vma) hugetlb_vma_lock_read() argument 369 hugetlb_vma_unlock_read(struct vm_area_struct * vma) hugetlb_vma_unlock_read() argument 373 hugetlb_vma_lock_write(struct vm_area_struct * vma) hugetlb_vma_lock_write() argument 377 hugetlb_vma_unlock_write(struct vm_area_struct * vma) hugetlb_vma_unlock_write() argument 381 hugetlb_vma_trylock_write(struct vm_area_struct * vma) hugetlb_vma_trylock_write() argument 386 hugetlb_vma_assert_locked(struct vm_area_struct * vma) hugetlb_vma_assert_locked() argument 441 hugetlb_change_protection(struct vm_area_struct * vma,unsigned long address,unsigned long end,pgprot_t newprot,unsigned long cp_flags) hugetlb_change_protection() argument 449 __unmap_hugepage_range(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long start,unsigned long end,struct folio * folio,zap_flags_t zap_flags) __unmap_hugepage_range() argument 457 hugetlb_fault(struct mm_struct * mm,struct vm_area_struct * vma,unsigned long address,unsigned int flags) hugetlb_fault() argument 464 hugetlb_unshare_all_pmds(struct vm_area_struct * vma) hugetlb_unshare_all_pmds() argument 466 fixup_hugetlb_reservations(struct vm_area_struct * vma) fixup_hugetlb_reservations() argument 470 hugetlb_split(struct vm_area_struct * vma,unsigned long addr) hugetlb_split() argument 472 hugetlb_vma_lock_alloc(struct vm_area_struct * vma) hugetlb_vma_lock_alloc() argument 770 hstate_vma(struct vm_area_struct * vma) hstate_vma() argument 1047 huge_ptep_modify_prot_start(struct vm_area_struct * vma,unsigned long addr,pte_t * ptep) huge_ptep_modify_prot_start() argument 1058 huge_ptep_modify_prot_commit(struct vm_area_struct * vma,unsigned long addr,pte_t * ptep,pte_t old_pte,pte_t pte) huge_ptep_modify_prot_commit() argument 1350 flush_hugetlb_tlb_range(vma,addr,end) global() argument 1353 __vma_shareable_lock(struct vm_area_struct * vma) __vma_shareable_lock() argument 1365 hugetlb_walk(struct vm_area_struct * vma,unsigned long addr,unsigned long sz) hugetlb_walk() argument [all...] |
| H A D | huge_mm.h | 104 #define thp_vma_allowable_order(vma, vm_flags, type, order) \ argument 220 static inline bool thp_vma_suitable_order(struct vm_area_struct *vma, in thp_vma_suitable_order() 245 static inline unsigned long thp_vma_suitable_orders(struct vm_area_struct *vma, in thp_vma_suitable_orders() 289 unsigned long thp_vma_allowable_orders(struct vm_area_struct *vma, in thp_vma_allowable_orders() 331 static inline bool vma_thp_disabled(struct vm_area_struct *vma, in vma_thp_disabled() 471 change_huge_pud(struct mmu_gather *tlb, struct vm_area_struct *vma, in change_huge_pud() 494 struct vm_area_struct *vma) in pmd_trans_huge_lock() 502 struct vm_area_struct *vma) in pud_trans_huge_lock() 578 static inline bool thp_vma_suitable_order(struct vm_area_struct *vma, in thp_vma_suitable_order() 584 static inline unsigned long thp_vma_suitable_orders(struct vm_area_struct *vma, in thp_vma_suitable_orders() [all …]
|
| /linux/mm/ |
| H A D | vma.c | 73 static bool vma_is_fork_child(struct vm_area_struct *vma) in vma_is_fork_child() argument 85 struct vm_area_struct *vma = merge_next ? vmg->next : vmg->prev; in is_mergeable_vma() local 140 init_multi_vma_prep(struct vma_prepare * vp,struct vm_area_struct * vma,struct vma_merge_struct * vmg) init_multi_vma_prep() argument 224 __vma_link_file(struct vm_area_struct * vma,struct address_space * mapping) __vma_link_file() argument 238 __remove_shared_vm_struct(struct vm_area_struct * vma,struct address_space * mapping) __remove_shared_vm_struct() argument 264 anon_vma_interval_tree_pre_update_vma(struct vm_area_struct * vma) anon_vma_interval_tree_pre_update_vma() argument 273 anon_vma_interval_tree_post_update_vma(struct vm_area_struct * vma) anon_vma_interval_tree_post_update_vma() argument 410 init_vma_prep(struct vma_prepare * vp,struct vm_area_struct * vma) init_vma_prep() argument 460 remove_vma(struct vm_area_struct * vma) remove_vma() argument 475 unmap_region(struct ma_state * mas,struct vm_area_struct * vma,struct vm_area_struct * prev,struct vm_area_struct * next) unmap_region() argument 497 __split_vma(struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long addr,int new_below) __split_vma() argument 590 split_vma(struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long addr,int new_below) split_vma() argument 646 struct vm_area_struct *vma; validate_mm() local 730 struct vm_area_struct *vma; commit_merge() local 771 can_merge_remove_vma(struct vm_area_struct * vma) can_merge_remove_vma() argument 1228 vma_shrink(struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long start,unsigned long end,pgoff_t pgoff) vma_shrink() argument 1285 struct vm_area_struct *vma; vms_clean_up_area() local 1308 struct vm_area_struct *vma; vms_complete_munmap_vmas() local 1354 struct vm_area_struct *vma; reattach_vmas() local 1522 init_vma_munmap(struct vma_munmap_struct * vms,struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long start,unsigned long end,struct list_head * uf,bool unlock) init_vma_munmap() argument 1558 do_vmi_align_munmap(struct vma_iterator * vmi,struct vm_area_struct * vma,struct mm_struct * mm,unsigned long start,unsigned long end,struct list_head * uf,bool unlock) do_vmi_align_munmap() argument 1610 struct vm_area_struct *vma; do_vmi_munmap() local 1645 struct vm_area_struct *vma = vmg->middle; vma_modify() local 1683 vma_modify_flags(struct vma_iterator * vmi,struct vm_area_struct * prev,struct vm_area_struct * vma,unsigned long start,unsigned long end,vm_flags_t * vm_flags_ptr) vma_modify_flags() argument 1708 vma_modify_name(struct vma_iterator * vmi,struct vm_area_struct * prev,struct vm_area_struct * vma,unsigned long start,unsigned long end,struct anon_vma_name * new_name) vma_modify_name() argument 1720 vma_modify_policy(struct vma_iterator * vmi,struct vm_area_struct * prev,struct vm_area_struct * vma,unsigned long start,unsigned long end,struct mempolicy * new_pol) vma_modify_policy() argument 1732 vma_modify_flags_uffd(struct vma_iterator * vmi,struct vm_area_struct * prev,struct vm_area_struct * vma,unsigned long start,unsigned long end,vm_flags_t vm_flags,struct vm_userfaultfd_ctx new_ctx,bool give_up_on_oom) vma_modify_flags_uffd() argument 1751 vma_merge_extend(struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long delta) vma_merge_extend() argument 1785 unlink_file_vma_batch_add(struct unlink_vma_file_batch * vb,struct vm_area_struct * vma) unlink_file_vma_batch_add() argument 1804 vma_link_file(struct vm_area_struct * vma,bool hold_rmap_lock) vma_link_file() argument 1818 vma_link(struct mm_struct * mm,struct vm_area_struct * vma) vma_link() argument 1842 struct vm_area_struct *vma = *vmap; copy_vma() local 1997 find_mergeable_anon_vma(struct vm_area_struct * vma) find_mergeable_anon_vma() argument 2036 vma_is_shared_writable(struct vm_area_struct * vma) vma_is_shared_writable() argument 2042 vma_fs_can_writeback(struct vm_area_struct * vma) vma_fs_can_writeback() argument 2056 vma_needs_dirty_tracking(struct vm_area_struct * vma) vma_needs_dirty_tracking() argument 2079 vma_wants_writenotify(struct vm_area_struct * vma,pgprot_t vm_page_prot) vma_wants_writenotify() argument 2193 struct vm_area_struct *vma; mm_take_all_locks() local 2289 struct vm_area_struct *vma; mm_drop_all_locks() local 2450 __mmap_new_file_vma(struct mmap_state * map,struct vm_area_struct * vma) __mmap_new_file_vma() argument 2503 struct vm_area_struct *vma; __mmap_new_vma() local 2573 __mmap_complete(struct mmap_state * map,struct vm_area_struct * vma) __mmap_complete() argument 2658 set_vma_user_defined_fields(struct vm_area_struct * vma,struct mmap_state * map) set_vma_user_defined_fields() argument 2697 call_action_complete(struct mmap_state * map,struct vm_area_desc * desc,struct vm_area_struct * vma) call_action_complete() argument 2718 struct vm_area_struct *vma = NULL; __mmap_region() local 2852 do_brk_flags(struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long addr,unsigned long len,vm_flags_t vm_flags) do_brk_flags() argument 3037 acct_stack_growth(struct vm_area_struct * vma,unsigned long size,unsigned long grow) acct_stack_growth() argument 3076 expand_upwards(struct vm_area_struct * vma,unsigned long address) expand_upwards() argument 3162 expand_downwards(struct vm_area_struct * vma,unsigned long address) expand_downwards() argument 3259 insert_vm_struct(struct mm_struct * mm,struct vm_area_struct * vma) insert_vm_struct() argument [all...] |
| H A D | rmap.c | 185 int __anon_vma_prepare(struct vm_area_struct *vma) in __anon_vma_prepare() argument 149 anon_vma_chain_link(struct vm_area_struct * vma,struct anon_vma_chain * avc,struct anon_vma * anon_vma) anon_vma_chain_link() argument 333 anon_vma_fork(struct vm_area_struct * vma,struct vm_area_struct * pvma) anon_vma_fork() argument 395 unlink_anon_vmas(struct vm_area_struct * vma) unlink_anon_vmas() argument 764 page_address_in_vma(const struct folio * folio,const struct page * page,const struct vm_area_struct * vma) page_address_in_vma() argument 825 folio_referenced_one(struct folio * folio,struct vm_area_struct * vma,unsigned long address,void * arg) folio_referenced_one() argument 908 invalid_folio_referenced_vma(struct vm_area_struct * vma,void * arg) invalid_folio_referenced_vma() argument 985 struct vm_area_struct *vma = pvmw->vma; page_vma_mkclean_one() local 1059 page_mkclean_one(struct folio * folio,struct vm_area_struct * vma,unsigned long address,void * arg) page_mkclean_one() argument 1070 invalid_mkclean_vma(struct vm_area_struct * vma,void * arg) invalid_mkclean_vma() argument 1111 mapping_wrprotect_range_one(struct folio * folio,struct vm_area_struct * vma,unsigned long address,void * arg) mapping_wrprotect_range_one() argument 1190 pfn_mkclean_range(unsigned long pfn,unsigned long nr_pages,pgoff_t pgoff,struct vm_area_struct * vma) pfn_mkclean_range() argument 1232 __folio_add_rmap(struct folio * folio,struct page * page,int nr_pages,struct vm_area_struct * vma,enum pgtable_level level) __folio_add_rmap() argument 1319 folio_move_anon_rmap(struct folio * folio,struct vm_area_struct * vma) folio_move_anon_rmap() argument 1342 __folio_set_anon(struct folio * folio,struct vm_area_struct * vma,unsigned long address,bool exclusive) __folio_set_anon() argument 1375 __page_check_anon_rmap(const struct folio * folio,const struct page * page,struct vm_area_struct * vma,unsigned long address) __page_check_anon_rmap() argument 1396 __folio_add_anon_rmap(struct folio * folio,struct page * page,int nr_pages,struct vm_area_struct * vma,unsigned long address,rmap_t flags,enum pgtable_level level) __folio_add_anon_rmap() argument 1475 folio_add_anon_rmap_ptes(struct folio * folio,struct page * page,int nr_pages,struct vm_area_struct * vma,unsigned long address,rmap_t flags) folio_add_anon_rmap_ptes() argument 1496 folio_add_anon_rmap_pmd(struct folio * folio,struct page * page,struct vm_area_struct * vma,unsigned long address,rmap_t flags) folio_add_anon_rmap_pmd() argument 1521 folio_add_new_anon_rmap(struct folio * folio,struct vm_area_struct * vma,unsigned long address,rmap_t flags) folio_add_new_anon_rmap() argument 1580 __folio_add_file_rmap(struct folio * folio,struct page * page,int nr_pages,struct vm_area_struct * vma,enum pgtable_level level) __folio_add_file_rmap() argument 1609 folio_add_file_rmap_ptes(struct folio * folio,struct page * page,int nr_pages,struct vm_area_struct * vma) folio_add_file_rmap_ptes() argument 1625 folio_add_file_rmap_pmd(struct folio * folio,struct page * page,struct vm_area_struct * vma) folio_add_file_rmap_pmd() argument 1645 folio_add_file_rmap_pud(struct folio * folio,struct page * page,struct vm_area_struct * vma) folio_add_file_rmap_pud() argument 1656 __folio_remove_rmap(struct folio * folio,struct page * page,int nr_pages,struct vm_area_struct * vma,enum pgtable_level level) __folio_remove_rmap() argument 1777 folio_remove_rmap_ptes(struct folio * folio,struct page * page,int nr_pages,struct vm_area_struct * vma) folio_remove_rmap_ptes() argument 1793 folio_remove_rmap_pmd(struct folio * folio,struct page * page,struct vm_area_struct * vma) folio_remove_rmap_pmd() argument 1813 folio_remove_rmap_pud(struct folio * folio,struct page * page,struct vm_area_struct * vma) folio_remove_rmap_pud() argument 1828 struct vm_area_struct *vma = pvmw->vma; folio_unmap_pte_batch() local 1852 try_to_unmap_one(struct folio * folio,struct vm_area_struct * vma,unsigned long address,void * arg) try_to_unmap_one() argument 2239 invalid_migration_vma(struct vm_area_struct * vma,void * arg) invalid_migration_vma() argument 2281 try_to_migrate_one(struct folio * folio,struct vm_area_struct * vma,unsigned long address,void * arg) try_to_migrate_one() argument 2681 struct vm_area_struct *vma; make_device_exclusive() local 2851 struct vm_area_struct *vma = avc->vma; rmap_walk_anon() local 2896 struct vm_area_struct *vma; __rmap_walk_file() local 2988 hugetlb_add_anon_rmap(struct folio * folio,struct vm_area_struct * vma,unsigned long address,rmap_t flags) hugetlb_add_anon_rmap() argument 3003 hugetlb_add_new_anon_rmap(struct folio * folio,struct vm_area_struct * vma,unsigned long address) hugetlb_add_new_anon_rmap() argument [all...] |
| H A D | memory.c | 374 struct vm_area_struct *vma, unsigned long floor, in free_pgtables() argument 586 print_bad_page_map(struct vm_area_struct * vma,unsigned long addr,unsigned long long entry,struct page * page,enum pgtable_level level) print_bad_page_map() argument 615 print_bad_pte(vma,addr,pte,page) global() argument 683 __vm_normal_page(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,bool special,unsigned long long entry,enum pgtable_level level) __vm_normal_page() argument 751 vm_normal_page(struct vm_area_struct * vma,unsigned long addr,pte_t pte) vm_normal_page() argument 770 vm_normal_folio(struct vm_area_struct * vma,unsigned long addr,pte_t pte) vm_normal_folio() argument 793 vm_normal_page_pmd(struct vm_area_struct * vma,unsigned long addr,pmd_t pmd) vm_normal_page_pmd() argument 812 vm_normal_folio_pmd(struct vm_area_struct * vma,unsigned long addr,pmd_t pmd) vm_normal_folio_pmd() argument 834 vm_normal_page_pud(struct vm_area_struct * vma,unsigned long addr,pud_t pud) vm_normal_page_pud() argument 868 restore_exclusive_pte(struct vm_area_struct * vma,struct folio * folio,struct page * page,unsigned long address,pte_t * ptep,pte_t orig_pte) restore_exclusive_pte() argument 902 try_restore_exclusive_pte(struct vm_area_struct * vma,unsigned long addr,pte_t * ptep,pte_t orig_pte) try_restore_exclusive_pte() argument 1186 folio_prealloc(struct mm_struct * src_mm,struct vm_area_struct * vma,unsigned long addr,bool need_zero) folio_prealloc() argument 1591 zap_install_uffd_wp_if_needed(struct vm_area_struct * vma,unsigned long addr,pte_t * pte,int nr,struct zap_details * details,pte_t pteval) zap_install_uffd_wp_if_needed() argument 1621 zap_present_folio_ptes(struct mmu_gather * tlb,struct vm_area_struct * vma,struct folio * folio,struct page * page,pte_t * pte,pte_t ptent,unsigned int nr,unsigned long addr,struct zap_details * details,int * rss,bool * force_flush,bool * force_break,bool * any_skipped) zap_present_folio_ptes() argument 1672 zap_present_ptes(struct mmu_gather * tlb,struct vm_area_struct * vma,pte_t * pte,pte_t ptent,unsigned int max_nr,unsigned long addr,struct zap_details * details,int * rss,bool * force_flush,bool * force_break,bool * any_skipped) zap_present_ptes() argument 1718 zap_nonpresent_ptes(struct mmu_gather * tlb,struct vm_area_struct * vma,pte_t * pte,pte_t ptent,unsigned int max_nr,unsigned long addr,struct zap_details * details,int * rss,bool * any_skipped) zap_nonpresent_ptes() argument 1789 do_zap_pte_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pte_t * pte,unsigned long addr,unsigned long end,struct zap_details * details,int * rss,bool * force_flush,bool * force_break,bool * any_skipped) do_zap_pte_range() argument 1825 zap_pte_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pmd_t * pmd,unsigned long addr,unsigned long end,struct zap_details * details) zap_pte_range() argument 1917 zap_pmd_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pud_t * pud,unsigned long addr,unsigned long end,struct zap_details * details) zap_pmd_range() argument 1959 zap_pud_range(struct mmu_gather * tlb,struct vm_area_struct * vma,p4d_t * p4d,unsigned long addr,unsigned long end,struct zap_details * details) zap_pud_range() argument 1987 zap_p4d_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pgd_t * pgd,unsigned long addr,unsigned long end,struct zap_details * details) zap_p4d_range() argument 2006 unmap_page_range(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long addr,unsigned long end,struct zap_details * details) unmap_page_range() argument 2027 unmap_single_vma(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long start_addr,unsigned long end_addr,struct zap_details * details) unmap_single_vma() argument 2087 unmap_vmas(struct mmu_gather * tlb,struct ma_state * mas,struct vm_area_struct * vma,unsigned long start_addr,unsigned long end_addr,unsigned long tree_end) unmap_vmas() argument 2123 zap_page_range_single_batched(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long address,unsigned long size,struct zap_details * details) zap_page_range_single_batched() argument 2162 zap_page_range_single(struct vm_area_struct * vma,unsigned long address,unsigned long size,struct zap_details * details) zap_page_range_single() argument 2183 zap_vma_ptes(struct vm_area_struct * vma,unsigned long address,unsigned long size) zap_vma_ptes() argument 2226 vm_mixed_zeropage_allowed(struct vm_area_struct * vma) vm_mixed_zeropage_allowed() argument 2256 validate_page_before_insert(struct vm_area_struct * vma,struct page * page) validate_page_before_insert() argument 2274 insert_page_into_pte_locked(struct vm_area_struct * vma,pte_t * pte,unsigned long addr,struct page * page,pgprot_t prot,bool mkwrite) insert_page_into_pte_locked() argument 2315 insert_page(struct vm_area_struct * vma,unsigned long addr,struct page * page,pgprot_t prot,bool mkwrite) insert_page() argument 2336 insert_page_in_batch_locked(struct vm_area_struct * vma,pte_t * pte,unsigned long addr,struct page * page,pgprot_t prot) insert_page_in_batch_locked() argument 2350 insert_pages(struct vm_area_struct * vma,unsigned long addr,struct page ** pages,unsigned long * num,pgprot_t prot) insert_pages() argument 2423 vm_insert_pages(struct vm_area_struct * vma,unsigned long addr,struct page ** pages,unsigned long * num) vm_insert_pages() argument 2470 vm_insert_page(struct vm_area_struct * vma,unsigned long addr,struct page * page) vm_insert_page() argument 2497 __vm_map_pages(struct vm_area_struct * vma,struct page ** pages,unsigned long num,unsigned long offset) __vm_map_pages() argument 2540 vm_map_pages(struct vm_area_struct * vma,struct page ** pages,unsigned long num) vm_map_pages() argument 2560 vm_map_pages_zero(struct vm_area_struct * vma,struct page ** pages,unsigned long num) vm_map_pages_zero() argument 2567 insert_pfn(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,pgprot_t prot,bool mkwrite) insert_pfn() argument 2651 vmf_insert_pfn_prot(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,pgprot_t pgprot) vmf_insert_pfn_prot() argument 2698 vmf_insert_pfn(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn) vmf_insert_pfn() argument 2705 vm_mixed_ok(struct vm_area_struct * vma,unsigned long pfn,bool mkwrite) vm_mixed_ok() argument 2719 __vm_insert_mixed(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,bool mkwrite) __vm_insert_mixed() argument 2785 vmf_insert_mixed(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn) vmf_insert_mixed() argument 2797 vmf_insert_mixed_mkwrite(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn) vmf_insert_mixed_mkwrite() argument 2920 remap_pfn_range_internal(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t prot) remap_pfn_range_internal() argument 2953 remap_pfn_range_notrack(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t prot) remap_pfn_range_notrack() argument 2999 remap_pfn_range_track(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t prot) remap_pfn_range_track() argument 3036 do_remap_pfn_range(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t prot) do_remap_pfn_range() argument 3042 do_remap_pfn_range(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t prot) do_remap_pfn_range() argument 3061 remap_pfn_range_prepare_vma(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size) remap_pfn_range_prepare_vma() argument 3089 remap_pfn_range(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t prot) remap_pfn_range() argument 3102 remap_pfn_range_complete(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t prot) remap_pfn_range_complete() argument 3123 vm_iomap_memory(struct vm_area_struct * vma,phys_addr_t start,unsigned long len) vm_iomap_memory() argument 3408 struct vm_area_struct *vma = vmf->vma; __wp_page_copy_user() local 3500 __get_fault_gfp_mask(struct vm_area_struct * vma) __get_fault_gfp_mask() argument 3555 struct vm_area_struct *vma = vmf->vma; fault_dirty_shared_page() local 3609 struct vm_area_struct *vma = vmf->vma; wp_page_reuse() local 3642 struct vm_area_struct *vma = vmf->vma; vmf_can_call_fault() local 3667 struct vm_area_struct *vma = vmf->vma; __vmf_anon_prepare() local 3703 struct vm_area_struct *vma = vmf->vma; wp_page_copy() local 3894 struct vm_area_struct *vma = vmf->vma; wp_pfn_shared() local 3917 struct vm_area_struct *vma = vmf->vma; wp_page_shared() local 3956 __wp_can_reuse_large_anon_folio(struct folio * folio,struct vm_area_struct * vma) __wp_can_reuse_large_anon_folio() argument 4022 __wp_can_reuse_large_anon_folio(struct folio * folio,struct vm_area_struct * vma) __wp_can_reuse_large_anon_folio() argument 4029 wp_can_reuse_anon_folio(struct folio * folio,struct vm_area_struct * vma) wp_can_reuse_anon_folio() argument 4095 struct vm_area_struct *vma = vmf->vma; do_wp_page() local 4186 unmap_mapping_range_vma(struct vm_area_struct * vma,unsigned long start_addr,unsigned long end_addr,struct zap_details * details) unmap_mapping_range_vma() argument 4198 struct vm_area_struct *vma; unmap_mapping_range_tree() local 4320 struct vm_area_struct *vma = vmf->vma; remove_device_exclusive_entry() local 4361 should_try_to_free_swap(struct folio * folio,struct vm_area_struct * vma,unsigned int fault_flags) should_try_to_free_swap() argument 4452 struct vm_area_struct *vma = vmf->vma; __alloc_swap_folio() local 4530 struct vm_area_struct *vma = vmf->vma; alloc_swap_folio() local 4626 struct vm_area_struct *vma = vmf->vma; do_swap_page() local 5088 struct vm_area_struct *vma = vmf->vma; alloc_anon_folio() local 5178 struct vm_area_struct *vma = vmf->vma; do_anonymous_page() local 5298 struct vm_area_struct *vma = vmf->vma; __do_fault() local 5355 struct vm_area_struct *vma = vmf->vma; deposit_prealloc_pte() local 5368 struct vm_area_struct *vma = vmf->vma; do_set_pmd() local 5459 struct vm_area_struct *vma = vmf->vma; set_pte_range() local 5517 struct vm_area_struct *vma = vmf->vma; finish_fault() local 5772 struct vm_area_struct *vma = vmf->vma; do_cow_fault() local 5814 struct vm_area_struct *vma = vmf->vma; do_shared_fault() local 5864 struct vm_area_struct *vma = vmf->vma; do_fault() local 5910 struct vm_area_struct *vma = vmf->vma; numa_migrate_check() local 5953 numa_rebuild_single_mapping(struct vm_fault * vmf,struct vm_area_struct * vma,unsigned long fault_addr,pte_t * fault_pte,bool writable) numa_rebuild_single_mapping() argument 5968 numa_rebuild_large_mapping(struct vm_fault * vmf,struct vm_area_struct * vma,struct folio * folio,pte_t fault_pte,bool ignore_writable,bool pte_write_upgrade) numa_rebuild_large_mapping() argument 6009 struct vm_area_struct *vma = vmf->vma; do_numa_page() local 6100 struct vm_area_struct *vma = vmf->vma; create_huge_pmd() local 6111 struct vm_area_struct *vma = vmf->vma; wp_huge_pmd() local 6144 struct vm_area_struct *vma = vmf->vma; create_huge_pud() local 6158 struct vm_area_struct *vma = vmf->vma; wp_huge_pud() local 6314 __handle_mm_fault(struct vm_area_struct * vma,unsigned long address,unsigned int flags) __handle_mm_fault() argument 6486 lru_gen_enter_fault(struct vm_area_struct * vma) lru_gen_enter_fault() argument 6497 lru_gen_enter_fault(struct vm_area_struct * vma) lru_gen_enter_fault() argument 6506 sanitize_fault_flags(struct vm_area_struct * vma,unsigned int * flags) sanitize_fault_flags() argument 6548 handle_mm_fault(struct vm_area_struct * vma,unsigned long address,unsigned int flags,struct pt_regs * regs) handle_mm_fault() argument 6701 pfnmap_lockdep_assert(struct vm_area_struct * vma) pfnmap_lockdep_assert() argument 6748 struct vm_area_struct *vma = args->vma; follow_pfnmap_start() local 6851 generic_access_phys(struct vm_area_struct * vma,unsigned long addr,void * buf,int len,int write) generic_access_phys() argument 6926 struct vm_area_struct *vma = NULL; __access_remote_vm() local 7053 struct vm_area_struct *vma = NULL; __copy_remote_vm_str() local 7150 struct vm_area_struct *vma; print_vma_addr() local 7281 copy_user_gigantic_page(struct folio * dst,struct folio * src,unsigned long addr_hint,struct vm_area_struct * vma,unsigned int nr_pages) copy_user_gigantic_page() argument 7304 struct vm_area_struct *vma; global() member 7319 copy_user_large_folio(struct folio * dst,struct folio * src,unsigned long addr_hint,struct vm_area_struct * vma) copy_user_large_folio() argument 7394 vma_pgtable_walk_begin(struct vm_area_struct * vma) vma_pgtable_walk_begin() argument 7400 vma_pgtable_walk_end(struct vm_area_struct * vma) vma_pgtable_walk_end() argument [all...] |
| H A D | mmap_lock.c | 53 __vma_enter_locked(struct vm_area_struct * vma,bool detaching,int state) __vma_enter_locked() argument 93 __vma_exit_locked(struct vm_area_struct * vma,bool * detached) __vma_exit_locked() argument 99 __vma_start_write(struct vm_area_struct * vma,unsigned int mm_lock_seq,int state) __vma_start_write() argument 127 vma_mark_detached(struct vm_area_struct * vma) vma_mark_detached() argument 163 vma_start_read(struct mm_struct * mm,struct vm_area_struct * vma) vma_start_read() argument 249 struct vm_area_struct *vma; lock_vma_under_rcu() local 297 struct vm_area_struct *vma; lock_next_vma_under_mmap_lock() local 322 struct vm_area_struct *vma; lock_next_vma() local 448 struct vm_area_struct *vma; lock_mm_and_find_vma() local 510 struct vm_area_struct *vma; lock_mm_and_find_vma() local [all...] |
| H A D | mprotect.c | 41 static bool maybe_change_pte_writable(struct vm_area_struct *vma, pte_t pte) in maybe_change_pte_writable() argument 61 static bool can_change_private_pte_writable(struct vm_area_struct *vma, in can_change_private_pte_writable() argument 79 static bool can_change_shared_pte_writable(struct vm_area_struct *vma, in can_change_shared_pte_writable() argument 97 bool can_change_pte_writable(struct vm_area_struct *vma, unsigne argument 120 prot_commit_flush_ptes(struct vm_area_struct * vma,unsigned long addr,pte_t * ptep,pte_t oldpte,pte_t ptent,int nr_ptes,int idx,bool set_write,struct mmu_gather * tlb) prot_commit_flush_ptes() argument 172 commit_anon_folio_batch(struct vm_area_struct * vma,struct folio * folio,struct page * first_page,unsigned long addr,pte_t * ptep,pte_t oldpte,pte_t ptent,int nr_ptes,struct mmu_gather * tlb) commit_anon_folio_batch() argument 191 set_write_prot_commit_flush_ptes(struct vm_area_struct * vma,struct folio * folio,struct page * page,unsigned long addr,pte_t * ptep,pte_t oldpte,pte_t ptent,int nr_ptes,struct mmu_gather * tlb) set_write_prot_commit_flush_ptes() argument 215 change_pte_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pmd_t * pmd,unsigned long addr,unsigned long end,pgprot_t newprot,unsigned long cp_flags) change_pte_range() argument 393 pgtable_split_needed(struct vm_area_struct * vma,unsigned long cp_flags) pgtable_split_needed() argument 409 pgtable_populate_needed(struct vm_area_struct * vma,unsigned long cp_flags) pgtable_populate_needed() argument 425 change_pmd_prepare(vma,pmd,cp_flags) global() argument 440 change_prepare(vma,high,low,addr,cp_flags) global() argument 452 change_pmd_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pud_t * pud,unsigned long addr,unsigned long end,pgprot_t newprot,unsigned long cp_flags) change_pmd_range() argument 522 change_pud_range(struct mmu_gather * tlb,struct vm_area_struct * vma,p4d_t * p4d,unsigned long addr,unsigned long end,pgprot_t newprot,unsigned long cp_flags) change_pud_range() argument 581 change_p4d_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pgd_t * pgd,unsigned long addr,unsigned long end,pgprot_t newprot,unsigned long cp_flags) change_p4d_range() argument 604 change_protection_range(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long addr,unsigned long end,pgprot_t newprot,unsigned long cp_flags) change_protection_range() argument 634 change_protection(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long start,unsigned long end,unsigned long cp_flags) change_protection() argument 696 mprotect_fixup(struct vma_iterator * vmi,struct mmu_gather * tlb,struct vm_area_struct * vma,struct vm_area_struct ** pprev,unsigned long start,unsigned long end,vm_flags_t newflags) mprotect_fixup() argument 805 struct vm_area_struct *vma, *prev; do_mprotect_pkey() local [all...] |
| H A D | madvise.c | 80 struct vm_area_struct *vma; member 110 struct anon_vma_name *anon_vma_name(struct vm_area_struct *vma) in anon_vma_name() argument 119 replace_anon_vma_name(struct vm_area_struct * vma,struct anon_vma_name * anon_name) replace_anon_vma_name() argument 139 replace_anon_vma_name(struct vm_area_struct * vma,struct anon_vma_name * anon_name) replace_anon_vma_name() argument 155 struct vm_area_struct *vma = madv_behavior->vma; madvise_update_vma() local 190 struct vm_area_struct *vma = walk->private; swapin_walk_pmd_entry() local 234 shmem_swapin_range(struct vm_area_struct * vma,unsigned long start,unsigned long end,struct address_space * mapping) shmem_swapin_range() argument 283 struct vm_area_struct *vma = madv_behavior->vma; madvise_willneed() local 329 can_do_file_pageout(struct vm_area_struct * vma) can_do_file_pageout() argument 362 struct vm_area_struct *vma = walk->vma; madvise_cold_or_pageout_pte_range() local 578 struct vm_area_struct *vma = madv_behavior->vma; madvise_cold_page_range() local 591 can_madv_lru_vma(struct vm_area_struct * vma) can_madv_lru_vma() argument 598 struct vm_area_struct *vma = madv_behavior->vma; madvise_cold() local 613 madvise_pageout_page_range(struct mmu_gather * tlb,struct vm_area_struct * vma,struct madvise_behavior_range * range) madvise_pageout_page_range() argument 630 struct vm_area_struct *vma = madv_behavior->vma; madvise_pageout() local 660 struct vm_area_struct *vma = walk->vma; madvise_free_pte_range() local 803 struct vm_area_struct *vma = madv_behavior->vma; madvise_free_single_vma() local 875 struct vm_area_struct *vma = madv_behavior->vma; madvise_dontneed_free_valid_vma() local 917 struct vm_area_struct *vma; madvise_dontneed_free() local 1014 struct vm_area_struct *vma = madv_behavior->vma; madvise_remove() local 1054 is_valid_guard_vma(struct vm_area_struct * vma,bool allow_locked) is_valid_guard_vma() argument 1125 struct vm_area_struct *vma = madv_behavior->vma; madvise_guard_install() local 1254 struct vm_area_struct *vma = madv_behavior->vma; madvise_guard_remove() local 1301 struct vm_area_struct *vma = madv_behavior->vma; can_madvise_modify() local 1350 struct vm_area_struct *vma = madv_behavior->vma; madvise_vma_behavior() local 1599 is_vma_lock_sufficient(struct vm_area_struct * vma,struct madvise_behavior * madv_behavior) is_vma_lock_sufficient() argument 1641 struct vm_area_struct *vma; try_vma_read_lock() local 1679 struct vm_area_struct *prev, *vma; madvise_walk_vmas() local [all...] |
| H A D | nommu.c | 92 struct vm_area_struct *vma; in kobjsize() local 148 struct vm_area_struct *vma; in __vmalloc_user_flags() local 345 int vm_insert_page(struct vm_area_struct *vma, unsigned long addr, in vm_insert_page() 352 int vm_insert_pages(struct vm_area_struct *vma, unsigned long addr, in vm_insert_pages() 359 int vm_map_pages(struct vm_area_struct *vma, struct page **pages, in vm_map_pages() 366 int vm_map_pages_zero(struct vm_area_struct *vma, struct page **pages, in vm_map_pages_zero() 562 static void setup_vma_to_mm(struct vm_area_struct *vma, struct mm_struct *mm) in setup_vma_to_mm() 578 static void cleanup_vma_from_mm(struct vm_area_struct *vma) in cleanup_vma_from_mm() 597 static int delete_vma_from_mm(struct vm_area_struct *vma) in delete_vma_from_mm() 616 static void delete_vma(struct mm_struct *mm, struct vm_area_struct *vma) in delete_vma() [all …]
|
| H A D | hugetlb.c | 284 if (__vma_shareable_lock(vma)) { in subpool_vma() argument 292 hugetlb_vma_lock_read(struct vm_area_struct * vma) hugetlb_vma_lock_read() argument 305 hugetlb_vma_unlock_read(struct vm_area_struct * vma) hugetlb_vma_unlock_read() argument 318 hugetlb_vma_lock_write(struct vm_area_struct * vma) hugetlb_vma_lock_write() argument 331 hugetlb_vma_unlock_write(struct vm_area_struct * vma) hugetlb_vma_unlock_write() argument 344 hugetlb_vma_trylock_write(struct vm_area_struct * vma) hugetlb_vma_trylock_write() argument 360 hugetlb_vma_assert_locked(struct vm_area_struct * vma) hugetlb_vma_assert_locked() argument 383 struct vm_area_struct *vma = vma_lock->vma; __hugetlb_vma_unlock_write_put() local 396 __hugetlb_vma_unlock_write_free(struct vm_area_struct * vma) __hugetlb_vma_unlock_write_free() argument 410 hugetlb_vma_lock_free(struct vm_area_struct * vma) hugetlb_vma_lock_free() argument 430 hugetlb_vma_lock_alloc(struct vm_area_struct * vma) hugetlb_vma_lock_alloc() argument 1023 vma_hugecache_offset(struct hstate * h,struct vm_area_struct * vma,unsigned long address) vma_hugecache_offset() argument 1038 vma_kernel_pagesize(struct vm_area_struct * vma) vma_kernel_pagesize() argument 1052 vma_mmu_pagesize(struct vm_area_struct * vma) vma_mmu_pagesize() argument 1085 get_vma_private_data(struct vm_area_struct * vma) get_vma_private_data() argument 1090 set_vma_private_data(struct vm_area_struct * vma,unsigned long value) set_vma_private_data() argument 1180 vma_resv_map(struct vm_area_struct * vma) vma_resv_map() argument 1195 set_vma_resv_flags(struct vm_area_struct * vma,unsigned long flags) set_vma_resv_flags() argument 1219 is_vma_resv_set(struct vm_area_struct * vma,unsigned long flag) is_vma_resv_set() argument 1233 __vma_private_lock(struct vm_area_struct * vma) __vma_private_lock() argument 1240 hugetlb_dup_vma_private(struct vm_area_struct * vma) hugetlb_dup_vma_private() argument 1271 clear_vma_resv_huge_pages(struct vm_area_struct * vma) clear_vma_resv_huge_pages() argument 1382 dequeue_hugetlb_folio_vma(struct hstate * h,struct vm_area_struct * vma,unsigned long address,long gbl_chg) dequeue_hugetlb_folio_vma() argument 2236 alloc_buddy_hugetlb_folio_with_mpol(struct hstate * h,struct vm_area_struct * vma,unsigned long addr) alloc_buddy_hugetlb_folio_with_mpol() argument 2511 __vma_reservation_common(struct hstate * h,struct vm_area_struct * vma,unsigned long addr,enum vma_resv_mode mode) __vma_reservation_common() argument 2591 vma_needs_reservation(struct hstate * h,struct vm_area_struct * vma,unsigned long addr) vma_needs_reservation() argument 2597 vma_commit_reservation(struct hstate * h,struct vm_area_struct * vma,unsigned long addr) vma_commit_reservation() argument 2603 vma_end_reservation(struct hstate * h,struct vm_area_struct * vma,unsigned long addr) vma_end_reservation() argument 2609 vma_add_reservation(struct hstate * h,struct vm_area_struct * vma,unsigned long addr) vma_add_reservation() argument 2615 vma_del_reservation(struct hstate * h,struct vm_area_struct * vma,unsigned long addr) vma_del_reservation() argument 2640 restore_reserve_on_error(struct hstate * h,struct vm_area_struct * vma,unsigned long address,struct folio * folio) restore_reserve_on_error() argument 2895 alloc_hugetlb_folio(struct vm_area_struct * vma,unsigned long addr,bool cow_from_owner) alloc_hugetlb_folio() argument 4690 hugetlb_vm_op_open(struct vm_area_struct * vma) hugetlb_vm_op_open() argument 4728 hugetlb_vm_op_close(struct vm_area_struct * vma) hugetlb_vm_op_close() argument 4759 hugetlb_vm_op_split(struct vm_area_struct * vma,unsigned long addr) hugetlb_vm_op_split() argument 4766 hugetlb_split(struct vm_area_struct * vma,unsigned long addr) hugetlb_split() argument 4799 hugetlb_vm_op_pagesize(struct vm_area_struct * vma) hugetlb_vm_op_pagesize() argument 4831 make_huge_pte(struct vm_area_struct * vma,struct folio * folio,bool try_mkwrite) make_huge_pte() argument 4848 set_huge_ptep_writable(struct vm_area_struct * vma,unsigned long address,pte_t * ptep) set_huge_ptep_writable() argument 4858 set_huge_ptep_maybe_writable(struct vm_area_struct * vma,unsigned long address,pte_t * ptep) set_huge_ptep_maybe_writable() argument 4866 hugetlb_install_folio(struct vm_area_struct * vma,pte_t * ptep,unsigned long addr,struct folio * new_folio,pte_t old,unsigned long sz) hugetlb_install_folio() argument 5063 move_huge_pte(struct vm_area_struct * vma,unsigned long old_addr,unsigned long new_addr,pte_t * src_pte,pte_t * dst_pte,unsigned long sz) move_huge_pte() argument 5102 move_hugetlb_page_tables(struct vm_area_struct * vma,struct vm_area_struct * new_vma,unsigned long old_addr,unsigned long new_addr,unsigned long len) move_hugetlb_page_tables() argument 5167 __unmap_hugepage_range(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long start,unsigned long end,struct folio * folio,zap_flags_t zap_flags) __unmap_hugepage_range() argument 5323 __hugetlb_zap_begin(struct vm_area_struct * vma,unsigned long * start,unsigned long * end) __hugetlb_zap_begin() argument 5335 __hugetlb_zap_end(struct vm_area_struct * vma,struct zap_details * details) __hugetlb_zap_end() argument 5362 unmap_hugepage_range(struct vm_area_struct * vma,unsigned long start,unsigned long end,struct folio * folio,zap_flags_t zap_flags) unmap_hugepage_range() argument 5388 unmap_ref_private(struct mm_struct * mm,struct vm_area_struct * vma,struct folio * folio,unsigned long address) unmap_ref_private() argument 5447 struct vm_area_struct *vma = vmf->vma; hugetlb_wp() local 5641 hugetlbfs_pagecache_present(struct hstate * h,struct vm_area_struct * vma,unsigned long address) hugetlbfs_pagecache_present() argument 5722 struct vm_area_struct *vma = vmf->vma; hugetlb_no_page() local 5967 hugetlb_fault(struct mm_struct * mm,struct vm_area_struct * vma,unsigned long address,unsigned int flags) hugetlb_fault() argument 6163 alloc_hugetlb_folio_vma(struct hstate * h,struct vm_area_struct * vma,unsigned long address) alloc_hugetlb_folio_vma() argument 6409 hugetlb_change_protection(struct vm_area_struct * vma,unsigned long address,unsigned long end,pgprot_t newprot,unsigned long cp_flags) hugetlb_change_protection() argument 6780 page_table_shareable(struct vm_area_struct * svma,struct vm_area_struct * vma,unsigned long addr,pgoff_t idx) page_table_shareable() argument 6807 want_pmd_share(struct vm_area_struct * vma,unsigned long addr) want_pmd_share() argument 6833 adjust_range_if_pmd_sharing_possible(struct vm_area_struct * vma,unsigned long * start,unsigned long * end) adjust_range_if_pmd_sharing_possible() argument 6864 huge_pmd_share(struct mm_struct * mm,struct vm_area_struct * vma,unsigned long addr,pud_t * pud) huge_pmd_share() argument 6925 huge_pmd_unshare(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long addr,pte_t * ptep) huge_pmd_unshare() argument 6959 huge_pmd_unshare_flush(struct mmu_gather * tlb,struct vm_area_struct * vma) huge_pmd_unshare_flush() argument 6973 huge_pmd_share(struct mm_struct * mm,struct vm_area_struct * vma,unsigned long addr,pud_t * pud) huge_pmd_share() argument 6979 huge_pmd_unshare(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long addr,pte_t * ptep) huge_pmd_unshare() argument 6985 huge_pmd_unshare_flush(struct mmu_gather * tlb,struct vm_area_struct * vma) huge_pmd_unshare_flush() argument 6989 adjust_range_if_pmd_sharing_possible(struct vm_area_struct * vma,unsigned long * start,unsigned long * end) adjust_range_if_pmd_sharing_possible() argument 6994 want_pmd_share(struct vm_area_struct * vma,unsigned long addr) want_pmd_share() argument 7001 huge_pte_alloc(struct mm_struct * mm,struct vm_area_struct * vma,unsigned long addr,unsigned long sz) huge_pte_alloc() argument 7246 hugetlb_unshare_pmds(struct vm_area_struct * vma,unsigned long start,unsigned long end,bool take_locks) hugetlb_unshare_pmds() argument 7307 hugetlb_unshare_all_pmds(struct vm_area_struct * vma) hugetlb_unshare_all_pmds() argument 7322 fixup_hugetlb_reservations(struct vm_area_struct * vma) fixup_hugetlb_reservations() argument [all...] |
| H A D | mmap.c | 81 void vma_set_page_prot(struct vm_area_struct *vma) in vma_set_page_prot() 693 struct vm_area_struct *vma, *prev; in generic_get_unmapped_area() local 741 struct vm_area_struct *vma, *prev; in generic_get_unmapped_area_topdown() local 927 struct vm_area_struct *vma; in find_vma_prev() local 954 int expand_stack_locked(struct vm_area_struct *vma, unsigned long address) in expand_stack_locked() 961 struct vm_area_struct *vma, *prev; in find_extend_vma_locked() local 976 int expand_stack_locked(struct vm_area_struct *vma, unsigned long address) in expand_stack_locked() 983 struct vm_area_struct *vma; in find_extend_vma_locked() local 1003 #define vma_expand_up(vma,addr) expand_upwards(vma, addr) argument 1004 #define vma_expand_down(vma, addr) (-EFAULT) argument [all …]
|
| H A D | pgtable-generic.c | 70 int ptep_set_access_flags(struct vm_area_struct *vma, in ptep_set_access_flags() argument 84 int ptep_clear_flush_young(struct vm_area_struct *vma, in ptep_clear_flush_young() argument 96 pte_t ptep_clear_flush(struct vm_area_struct *vma, unsigned long address, in ptep_clear_flush() argument 111 int pmdp_set_access_flags(struct vm_area_struct *vma, in pmdp_set_access_flags() argument 126 pmdp_clear_flush_young(struct vm_area_struct * vma,unsigned long address,pmd_t * pmdp) pmdp_clear_flush_young() argument 139 pmdp_huge_clear_flush(struct vm_area_struct * vma,unsigned long address,pmd_t * pmdp) pmdp_huge_clear_flush() argument 151 pudp_huge_clear_flush(struct vm_area_struct * vma,unsigned long address,pud_t * pudp) pudp_huge_clear_flush() argument 199 pmdp_invalidate(struct vm_area_struct * vma,unsigned long address,pmd_t * pmdp) pmdp_invalidate() argument 210 pmdp_invalidate_ad(struct vm_area_struct * vma,unsigned long address,pmd_t * pmdp) pmdp_invalidate_ad() argument 219 pmdp_collapse_flush(struct vm_area_struct * vma,unsigned long address,pmd_t * pmdp) pmdp_collapse_flush() argument [all...] |
| H A D | huge_memory.c | 85 static inline bool file_thp_enabled(struct vm_area_struct *vma) in file_thp_enabled() argument 100 unsigned long __thp_vma_allowable_orders(struct vm_area_struct *vma, in __thp_vma_allowable_orders() argument 1073 maybe_pmd_mkwrite(pmd_t pmd,struct vm_area_struct * vma) maybe_pmd_mkwrite() argument 1253 vma_alloc_anon_folio_pmd(struct vm_area_struct * vma,unsigned long addr) vma_alloc_anon_folio_pmd() argument 1297 map_anon_folio_pmd_nopf(struct folio * folio,pmd_t * pmd,struct vm_area_struct * vma,unsigned long haddr) map_anon_folio_pmd_nopf() argument 1311 map_anon_folio_pmd_pf(struct folio * folio,pmd_t * pmd,struct vm_area_struct * vma,unsigned long haddr) map_anon_folio_pmd_pf() argument 1323 struct vm_area_struct *vma = vmf->vma; __do_huge_pmd_anonymous_page() local 1374 struct vm_area_struct *vma = vmf->vma; do_huge_pmd_device_private() local 1419 vma_thp_gfp_mask(struct vm_area_struct * vma) vma_thp_gfp_mask() argument 1447 set_huge_zero_folio(pgtable_t pgtable,struct mm_struct * mm,struct vm_area_struct * vma,unsigned long haddr,pmd_t * pmd,struct folio * zero_folio) set_huge_zero_folio() argument 1460 struct vm_area_struct *vma = vmf->vma; do_huge_pmd_anonymous_page() local 1523 insert_pmd(struct vm_area_struct * vma,unsigned long addr,pmd_t * pmd,struct folio_or_pfn fop,pgprot_t prot,bool write) insert_pmd() argument 1608 struct vm_area_struct *vma = vmf->vma; vmf_insert_pfn_pmd() local 1633 struct vm_area_struct *vma = vmf->vma; vmf_insert_folio_pmd() local 1648 maybe_pud_mkwrite(pud_t pud,struct vm_area_struct * vma) maybe_pud_mkwrite() argument 1655 insert_pud(struct vm_area_struct * vma,unsigned long addr,pud_t * pud,struct folio_or_pfn fop,pgprot_t prot,bool write) insert_pud() argument 1716 struct vm_area_struct *vma = vmf->vma; vmf_insert_pfn_pud() local 1749 struct vm_area_struct *vma = vmf->vma; vmf_insert_folio_pud() local 1773 touch_pmd(struct vm_area_struct * vma,unsigned long addr,pmd_t * pmd,bool write) touch_pmd() argument 1953 touch_pud(struct vm_area_struct * vma,unsigned long addr,pud_t * pud,bool write) touch_pud() argument 1968 copy_huge_pud(struct mm_struct * dst_mm,struct mm_struct * src_mm,pud_t * dst_pud,pud_t * src_pud,unsigned long addr,struct vm_area_struct * vma) copy_huge_pud() argument 2028 struct vm_area_struct *vma = vmf->vma; do_huge_zero_wp_pmd() local 2060 struct vm_area_struct *vma = vmf->vma; do_huge_pmd_wp_page() local 2151 can_change_pmd_writable(struct vm_area_struct * vma,unsigned long addr,pmd_t pmd) can_change_pmd_writable() argument 2184 struct vm_area_struct *vma = vmf->vma; do_huge_pmd_numa_page() local 2262 madvise_free_huge_pmd(struct mmu_gather * tlb,struct vm_area_struct * vma,pmd_t * pmd,unsigned long addr,unsigned long next) madvise_free_huge_pmd() argument 2341 zap_huge_pmd(struct mmu_gather * tlb,struct vm_area_struct * vma,pmd_t * pmd,unsigned long addr) zap_huge_pmd() argument 2425 pmd_move_must_withdraw(spinlock_t * new_pmd_ptl,spinlock_t * old_pmd_ptl,struct vm_area_struct * vma) pmd_move_must_withdraw() argument 2461 move_huge_pmd(struct vm_area_struct * vma,unsigned long old_addr,unsigned long new_addr,pmd_t * old_pmd,pmd_t * new_pmd) move_huge_pmd() argument 2555 change_huge_pmd(struct mmu_gather * tlb,struct vm_area_struct * vma,pmd_t * pmd,unsigned long addr,pgprot_t newprot,unsigned long cp_flags) change_huge_pmd() argument 2656 change_huge_pud(struct mmu_gather * tlb,struct vm_area_struct * vma,pud_t * pudp,unsigned long addr,pgprot_t newprot,unsigned long cp_flags) change_huge_pud() argument 2821 __pmd_trans_huge_lock(pmd_t * pmd,struct vm_area_struct * vma) __pmd_trans_huge_lock() argument 2838 __pud_trans_huge_lock(pud_t * pud,struct vm_area_struct * vma) __pud_trans_huge_lock() argument 2850 zap_huge_pud(struct mmu_gather * tlb,struct vm_area_struct * vma,pud_t * pud,unsigned long addr) zap_huge_pud() argument 2885 __split_huge_pud_locked(struct vm_area_struct * vma,pud_t * pud,unsigned long haddr) __split_huge_pud_locked() argument 2917 __split_huge_pud(struct vm_area_struct * vma,pud_t * pud,unsigned long address) __split_huge_pud() argument 2937 __split_huge_pud(struct vm_area_struct * vma,pud_t * pud,unsigned long address) __split_huge_pud() argument 2943 __split_huge_zero_page_pmd(struct vm_area_struct * vma,unsigned long haddr,pmd_t * pmd) __split_huge_zero_page_pmd() argument 2984 __split_huge_pmd_locked(struct vm_area_struct * vma,pmd_t * pmd,unsigned long haddr,bool freeze) __split_huge_pmd_locked() argument 3265 split_huge_pmd_locked(struct vm_area_struct * vma,unsigned long address,pmd_t * pmd,bool freeze) split_huge_pmd_locked() argument 3273 __split_huge_pmd(struct vm_area_struct * vma,pmd_t * pmd,unsigned long address,bool freeze) __split_huge_pmd() argument 3289 split_huge_pmd_address(struct vm_area_struct * vma,unsigned long address,bool freeze) split_huge_pmd_address() argument 3300 split_huge_pmd_if_needed(struct vm_area_struct * vma,unsigned long address) split_huge_pmd_if_needed() argument 3312 vma_adjust_trans_huge(struct vm_area_struct * vma,unsigned long start,unsigned long end,struct vm_area_struct * next) vma_adjust_trans_huge() argument 3351 __discard_anon_folio_pmd_locked(struct vm_area_struct * vma,unsigned long addr,pmd_t * pmdp,struct folio * folio) __discard_anon_folio_pmd_locked() argument 3414 unmap_huge_pmd_locked(struct vm_area_struct * vma,unsigned long addr,pmd_t * pmdp,struct folio * folio) unmap_huge_pmd_locked() argument 4568 vma_not_suitable_for_thp_split(struct vm_area_struct * vma) vma_not_suitable_for_thp_split() argument 4611 struct vm_area_struct *vma = vma_lookup(mm, addr); split_huge_pages_pid() local 4866 struct vm_area_struct *vma = pvmw->vma; set_pmd_migration_entry() local 4918 struct vm_area_struct *vma = pvmw->vma; remove_migration_pmd() local [all...] |
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_vm.h | 112 * DOC: Provide accessors for vma members to facilitate easy change of in xe_vma_start() argument 117 return vma->gpuva.va.addr; in xe_vma_size() argument 122 return vma->gpuva.va.range; in xe_vma_end() argument 127 return xe_vma_start(vma) + xe_vma_size(vma); in xe_vma_bo_offset() argument 132 xe_vma_bo(struct xe_vma * vma) xe_vma_bo() argument 138 xe_vma_vm(struct xe_vma * vma) xe_vma_vm() argument 143 xe_vma_read_only(struct xe_vma * vma) xe_vma_read_only() argument 148 xe_vma_userptr(struct xe_vma * vma) xe_vma_userptr() argument 153 xe_vma_is_null(struct xe_vma * vma) xe_vma_is_null() argument 158 xe_vma_is_cpu_addr_mirror(struct xe_vma * vma) xe_vma_is_cpu_addr_mirror() argument 163 xe_vma_has_no_bo(struct xe_vma * vma) xe_vma_has_no_bo() argument 168 xe_vma_is_userptr(struct xe_vma * vma) xe_vma_is_userptr() argument 188 to_userptr_vma(struct xe_vma * vma) to_userptr_vma() argument [all...] |
| H A D | xe_vm.c | 603 struct xe_vma *vma; xe_vma_svm_prefetch_op_fini() local 651 xe_vm_populate_rebind(struct xe_vma_op * op,struct xe_vma * vma,u8 tile_mask) xe_vm_populate_rebind() argument 666 xe_vm_ops_add_rebind(struct xe_vma_ops * vops,struct xe_vma * vma,u8 tile_mask) xe_vm_ops_add_rebind() argument 691 struct xe_vma *vma, *next; xe_vm_rebind() local 743 xe_vma_rebind(struct xe_vm * vm,struct xe_vma * vma,u8 tile_mask) xe_vma_rebind() argument 787 xe_vm_populate_range_rebind(struct xe_vma_op * op,struct xe_vma * vma,struct xe_svm_range * range,u8 tile_mask) xe_vm_populate_range_rebind() argument 801 xe_vm_ops_add_range_rebind(struct xe_vma_ops * vops,struct xe_vma * vma,struct xe_svm_range * range,u8 tile_mask) xe_vm_ops_add_range_rebind() argument 831 xe_vm_range_rebind(struct xe_vm * vm,struct xe_vma * vma,struct xe_svm_range * range,u8 tile_mask) xe_vm_range_rebind() argument 960 xe_vma_free(struct xe_vma * vma) xe_vma_free() argument 975 struct xe_vma *vma; xe_vma_create() local 1057 xe_vma_destroy_late(struct xe_vma * vma) xe_vma_destroy_late() argument 1082 struct xe_vma *vma = vma_destroy_work_func() local 1091 struct xe_vma *vma = container_of(cb, struct xe_vma, destroy_cb); vma_destroy_cb() local 1097 xe_vma_destroy(struct xe_vma * vma,struct dma_fence * fence) xe_vma_destroy() argument 1137 xe_vm_lock_vma(struct drm_exec * exec,struct xe_vma * vma) xe_vm_lock_vma() argument 1152 xe_vma_destroy_unlocked(struct xe_vma * vma) xe_vma_destroy_unlocked() argument 1186 xe_vm_insert_vma(struct xe_vm * vm,struct xe_vma * vma) xe_vm_insert_vma() argument 1201 xe_vm_remove_vma(struct xe_vm * vm,struct xe_vma * vma) xe_vm_remove_vma() argument 1337 xelp_pte_encode_vma(u64 pte,struct xe_vma * vma,u16 pat_index,u32 pt_level) xelp_pte_encode_vma() argument 1716 struct xe_vma *vma, *next_vma; xe_vm_close_and_put() local 2021 struct xe_vma *vma = gpuva_to_vma(gpuva); get_mem_attrs() local 2104 vma_matches(struct xe_vma * vma,u64 page_addr) vma_matches() argument 2121 struct xe_vma *vma = NULL; xe_vm_find_vma_by_addr() local 2139 prep_vma_destroy(struct xe_vm * vm,struct xe_vma * vma,bool post_commit) prep_vma_destroy() argument 2155 struct xe_vma *vma; print_op() local 2312 struct xe_vma *vma = gpuva_to_vma(op->base.prefetch.va); vm_bind_ioctl_ops_create() local 2411 struct xe_vma *vma; new_vma() local 2466 xe_vma_max_pte_size(struct xe_vma * vma) xe_vma_max_pte_size() argument 2480 xe_vma_set_pte_size(struct xe_vma * vma,u64 size) xe_vma_set_pte_size() argument 2576 xe_vma_has_default_mem_attrs(struct xe_vma * vma) xe_vma_has_default_mem_attrs() argument 2600 struct xe_vma *vma; vm_bind_ioctl_ops_parse() local 2779 struct xe_vma *vma = gpuva_to_vma(op->base.unmap.va); xe_vma_op_unwind() local 2792 struct xe_vma *vma = gpuva_to_vma(op->base.remap.unmap->va); xe_vma_op_unwind() local 2843 vma_lock_and_validate(struct drm_exec * exec,struct xe_vma * vma,bool res_evict,bool validate) vma_lock_and_validate() argument 2862 check_ufence(struct xe_vma * vma) check_ufence() argument 2880 struct xe_vma *vma = gpuva_to_vma(op->base.prefetch.va); prefetch_ranges() local 2972 struct xe_vma *vma = gpuva_to_vma(op->base.prefetch.va); op_lock_and_prep() local 3205 vma_add_ufence(struct xe_vma * vma,struct xe_user_fence * ufence) vma_add_ufence() argument 3957 xe_vm_invalidate_vma(struct xe_vma * vma) xe_vm_invalidate_vma() argument 4021 struct xe_vma *vma = gpuva_to_vma(gpuva); xe_vm_validate_protected() local 4075 struct xe_vma *vma = gpuva_to_vma(gpuva); xe_vm_snapshot_capture() local 4220 xe_vma_need_vram_for_atomic(struct xe_device * xe,struct xe_vma * vma,bool is_atomic) xe_vma_need_vram_for_atomic() argument 4277 struct xe_vma *vma = NULL; xe_vm_alloc_vma() local 4334 struct xe_vma *vma; xe_vm_alloc_vma() local [all...] |
| /linux/tools/testing/vma/ |
| H A D | vma.c | 18 #define vma_iter_prealloc(vmi, vma) \ argument 72 struct vm_area_struct *vma = vm_area_alloc(mm); in alloc_vma() local 87 static int attach_vma(struct mm_struct *mm, struct vm_area_struct *vma) in attach_vma() 97 static void detach_free_vma(struct vm_area_struct *vma) in detach_free_vma() 110 struct vm_area_struct *vma = alloc_vma(mm, start, end, pgoff, vm_flags); in alloc_and_link_vma() local 133 struct vm_area_struct *vma; in merge_new() local 155 struct vm_area_struct *vma; in merge_existing() local 253 struct vm_area_struct *vma; in cleanup_mm() local 271 static bool vma_write_started(struct vm_area_struct *vma) in vma_write_started() 287 static void __vma_set_dummy_anon_vma(struct vm_area_struct *vma, in __vma_set_dummy_anon_vma() [all …]
|
| H A D | vma_internal.h | 385 #define vma_policy(vma) NULL argument 883 vma_is_shared_maywrite(struct vm_area_struct * vma) vma_is_shared_maywrite() argument 902 vma_assert_attached(struct vm_area_struct * vma) vma_assert_attached() argument 907 vma_assert_detached(struct vm_area_struct * vma) vma_assert_detached() argument 913 vma_mark_attached(struct vm_area_struct * vma) vma_mark_attached() argument 920 vma_mark_detached(struct vm_area_struct * vma) vma_mark_detached() argument 937 vma_init(struct vm_area_struct * vma,struct mm_struct * mm) vma_init() argument 990 vma_set_range(struct vm_area_struct * vma,unsigned long start,unsigned long end,pgoff_t pgoff) vma_set_range() argument 1043 vma_is_anonymous(struct vm_area_struct * vma) vma_is_anonymous() argument 1056 struct vm_area_struct *vma; find_vma_prev() local 1076 anon_vma_name(struct vm_area_struct * vma) anon_vma_name() argument 1081 is_mergeable_vm_userfaultfd_ctx(struct vm_area_struct * vma,struct vm_userfaultfd_ctx vm_ctx) is_mergeable_vm_userfaultfd_ctx() argument 1097 vma_pages(struct vm_area_struct * vma) vma_pages() argument 1127 unmap_vmas(struct mmu_gather * tlb,struct ma_state * mas,struct vm_area_struct * vma,unsigned long start_addr,unsigned long end_addr,unsigned long tree_end) unmap_vmas() argument 1133 free_pgtables(struct mmu_gather * tlb,struct ma_state * mas,struct vm_area_struct * vma,unsigned long floor,unsigned long ceiling,bool mm_wr_locked) free_pgtables() argument 1171 vma_start_write(struct vm_area_struct * vma) vma_start_write() argument 1178 vma_start_write_killable(struct vm_area_struct * vma) vma_start_write_killable() argument 1185 vma_adjust_trans_huge(struct vm_area_struct * vma,unsigned long start,unsigned long end,struct vm_area_struct * next) vma_adjust_trans_huge() argument 1209 vma_interval_tree_insert(struct vm_area_struct * vma,struct rb_root_cached * rb) vma_interval_tree_insert() argument 1214 vma_interval_tree_remove(struct vm_area_struct * vma,struct rb_root_cached * rb) vma_interval_tree_remove() argument 1233 uprobe_mmap(struct vm_area_struct * vma) uprobe_mmap() argument 1237 uprobe_munmap(struct vm_area_struct * vma,unsigned long start,unsigned long end) uprobe_munmap() argument 1250 vma_assert_write_locked(struct vm_area_struct * vma) vma_assert_write_locked() argument 1254 unlink_anon_vmas(struct vm_area_struct * vma) unlink_anon_vmas() argument 1268 anon_vma_merge(struct vm_area_struct * vma,struct vm_area_struct * next) anon_vma_merge() argument 1273 userfaultfd_unmap_prep(struct vm_area_struct * vma,unsigned long start,unsigned long end,struct list_head * unmaps) userfaultfd_unmap_prep() argument 1320 khugepaged_enter_vma(struct vm_area_struct * vma,vm_flags_t vm_flags) khugepaged_enter_vma() argument 1330 is_vm_hugetlb_page(struct vm_area_struct * vma) is_vm_hugetlb_page() argument 1335 vma_soft_dirty_enabled(struct vm_area_struct * vma) vma_soft_dirty_enabled() argument 1340 userfaultfd_wp(struct vm_area_struct * vma) userfaultfd_wp() argument 1383 shmem_zero_setup(struct vm_area_struct * vma) shmem_zero_setup() argument 1388 vma_set_anonymous(struct vm_area_struct * vma) vma_set_anonymous() argument 1393 ksm_add_vma(struct vm_area_struct * vma) ksm_add_vma() argument 1397 perf_event_mmap(struct vm_area_struct * vma) perf_event_mmap() argument 1401 vma_is_dax(struct vm_area_struct * vma) vma_is_dax() argument 1414 vma_set_page_prot(struct vm_area_struct * vma) vma_set_page_prot() argument 1436 vma_close(struct vm_area_struct * vma) vma_close() argument 1440 mmap_file(struct file * file,struct vm_area_struct * vma) mmap_file() argument 1445 stack_guard_start_gap(struct vm_area_struct * vma) stack_guard_start_gap() argument 1457 vm_start_gap(struct vm_area_struct * vma) vm_start_gap() argument 1468 vm_end_gap(struct vm_area_struct * vma) vm_end_gap() argument 1486 vma_is_accessible(struct vm_area_struct * vma) vma_is_accessible() argument 1513 __anon_vma_prepare(struct vm_area_struct * vma) __anon_vma_prepare() argument 1526 anon_vma_prepare(struct vm_area_struct * vma) anon_vma_prepare() argument 1606 vm_flags_init(struct vm_area_struct * vma,vm_flags_t flags) vm_flags_init() argument 1618 vm_flags_reset(struct vm_area_struct * vma,vm_flags_t flags) vm_flags_reset() argument 1625 vm_flags_reset_once(struct vm_area_struct * vma,vm_flags_t flags) vm_flags_reset_once() argument 1637 vm_flags_set(struct vm_area_struct * vma,vm_flags_t flags) vm_flags_set() argument 1644 vm_flags_clear(struct vm_area_struct * vma,vm_flags_t flags) vm_flags_clear() argument 1725 vma_lock_init(struct vm_area_struct * vma,bool reset_refcnt) vma_lock_init() argument 1731 vma_numab_state_init(struct vm_area_struct * vma) vma_numab_state_init() argument 1735 vma_numab_state_free(struct vm_area_struct * vma) vma_numab_state_free() argument 1744 free_anon_vma_name(struct vm_area_struct * vma) free_anon_vma_name() argument 1758 mmap_action_complete(struct mmap_action * action,struct vm_area_struct * vma) mmap_action_complete() argument 1764 __compat_vma_mmap(const struct file_operations * f_op,struct file * file,struct vm_area_struct * vma) __compat_vma_mmap() argument 1791 compat_vma_mmap(struct file * file,struct vm_area_struct * vma) compat_vma_mmap() argument 1811 vfs_mmap(struct file * file,struct vm_area_struct * vma) vfs_mmap() argument 1824 fixup_hugetlb_reservations(struct vm_area_struct * vma) fixup_hugetlb_reservations() argument 1828 vma_set_file(struct vm_area_struct * vma,struct file * file) vma_set_file() argument 1851 remap_pfn_range_complete(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t pgprot) remap_pfn_range_complete() argument [all...] |
| /linux/arch/powerpc/include/asm/book3s/64/ |
| H A D | tlbflush.h | 47 static inline void flush_pmd_tlb_range(struct vm_area_struct *vma, in flush_pmd_tlb_range() 55 static inline void flush_pud_tlb_range(struct vm_area_struct *vma, in flush_pud_tlb_range() 63 static inline void flush_hugetlb_tlb_range(struct vm_area_struct *vma, in flush_hugetlb_tlb_range() 71 static inline void flush_tlb_range(struct vm_area_struct *vma, in flush_tlb_range() 91 static inline void local_flush_tlb_page(struct vm_area_struct *vma, in local_flush_tlb_page() 120 static inline void flush_tlb_page(struct vm_area_struct *vma, in flush_tlb_page() 128 #define flush_tlb_page(vma, addr) local_flush_tlb_page(vma, addr) argument 132 static inline void flush_tlb_fix_spurious_fault(struct vm_area_struct *vma, in flush_tlb_fix_spurious_fault()
|
| /linux/arch/parisc/kernel/ |
| H A D | cache.c | 331 __flush_cache_page(struct vm_area_struct *vma, unsigned long vmaddr, in __flush_cache_page() 393 void flush_icache_pages(struct vm_area_struct *vma, struct page *page, in flush_icache_pages() 476 struct vm_area_struct *vma; in flush_dcache_folio() local 623 static void flush_cache_page_if_present(struct vm_area_struct *vma, in flush_cache_page_if_present() 648 unsigned long vaddr, struct vm_area_struct *vma) in copy_user_highpage() 660 void copy_to_user_page(struct vm_area_struct *vma, struct page *page, in copy_to_user_page() 668 void copy_from_user_page(struct vm_area_struct *vma, struct page *page, in copy_from_user_page() 705 static void flush_cache_pages(struct vm_area_struct *vma, unsigned long start, unsigned long end) in flush_cache_pages() 715 struct vm_area_struct *vma; in mm_total_size() local 729 struct vm_area_struct *vma; in flush_cache_mm() local [all …]
|
| /linux/drivers/gpu/drm/i915/ |
| H A D | i915_gem_evict.c | 67 static bool grab_vma(struct i915_vma *vma, struct i915_gem_ww_ctx *ww) in grab_vma() 86 static void ungrab_vma(struct i915_vma *vma) in ungrab_vma() 98 struct i915_vma *vma, in mark_free() 112 static bool defer_evict(struct i915_vma *vma) in defer_evict() 157 struct i915_vma *vma, *next; in i915_gem_evict_something() local 337 struct i915_vma *vma, *next; in i915_gem_evict_for_node() local 478 struct i915_vma *vma, *vn; in i915_gem_evict_vm() local
|
| /linux/arch/x86/mm/ |
| H A D | pgtable.c | 391 int ptep_set_access_flags(struct vm_area_struct *vma, in ptep_set_access_flags() 404 int pmdp_set_access_flags(struct vm_area_struct *vma, in pmdp_set_access_flags() 425 int pudp_set_access_flags(struct vm_area_struct *vma, unsigned long address, in pudp_set_access_flags() 446 int ptep_test_and_clear_young(struct vm_area_struct *vma, in ptep_test_and_clear_young() 459 int pmdp_test_and_clear_young(struct vm_area_struct *vma, in pmdp_test_and_clear_young() 473 int pudp_test_and_clear_young(struct vm_area_struct *vma, in pudp_test_and_clear_young() 486 int ptep_clear_flush_young(struct vm_area_struct *vma, in ptep_clear_flush_young() 506 int pmdp_clear_flush_young(struct vm_area_struct *vma, in pmdp_clear_flush_young() 520 pmd_t pmdp_invalidate_ad(struct vm_area_struct *vma, unsigned long address, in pmdp_invalidate_ad() 535 pud_t pudp_invalidate(struct vm_area_struct *vma, unsigned long address, in pudp_invalidate() [all …]
|
| /linux/drivers/gpu/drm/nouveau/ |
| H A D | nouveau_vmm.c | 29 nouveau_vma_unmap(struct nouveau_vma *vma) in nouveau_vma_unmap() 38 nouveau_vma_map(struct nouveau_vma *vma, struct nouveau_mem *mem) in nouveau_vma_map() 51 struct nouveau_vma *vma; in nouveau_vma_find() local 64 struct nouveau_vma *vma = *pvma; in nouveau_vma_del() local 81 struct nouveau_vma *vma; in nouveau_vma_new() local
|