Home
last modified time | relevance | path

Searched defs:vma (Results 1 – 25 of 601) sorted by relevance

12345678910>>...25

/linux/include/linux/
H A Duserfaultfd_k.h145 static inline bool is_mergeable_vm_userfaultfd_ctx(struct vm_area_struct *vma, in is_mergeable_vm_userfaultfd_ctx()
162 static inline bool uffd_disable_huge_pmd_share(struct vm_area_struct *vma) in uffd_disable_huge_pmd_share()
174 static inline bool uffd_disable_fault_around(struct vm_area_struct *vma) in uffd_disable_fault_around()
179 static inline bool userfaultfd_missing(struct vm_area_struct *vma) in userfaultfd_missing()
184 static inline bool userfaultfd_wp(struct vm_area_struct *vma) in userfaultfd_wp()
189 static inline bool userfaultfd_minor(struct vm_area_struct *vma) in userfaultfd_minor()
194 static inline bool userfaultfd_pte_wp(struct vm_area_struct *vma, in userfaultfd_pte_wp()
200 static inline bool userfaultfd_huge_pmd_wp(struct vm_area_struct *vma, in userfaultfd_huge_pmd_wp()
206 static inline bool userfaultfd_armed(struct vm_area_struct *vma) in userfaultfd_armed()
211 static inline bool vma_can_userfault(struct vm_area_struct *vma, in vma_can_userfault()
[all …]
H A Dmmap_lock.h113 vma_lock_init(struct vm_area_struct * vma,bool reset_refcnt) vma_lock_init() argument
136 vma_refcount_put(struct vm_area_struct * vma) vma_refcount_put() argument
156 vma_start_read_locked_nested(struct vm_area_struct * vma,int subclass) vma_start_read_locked_nested() argument
175 vma_start_read_locked(struct vm_area_struct * vma) vma_start_read_locked() argument
180 vma_end_read(struct vm_area_struct * vma) vma_end_read() argument
186 __is_vma_write_locked(struct vm_area_struct * vma,unsigned int * mm_lock_seq) __is_vma_write_locked() argument
206 vma_start_write(struct vm_area_struct * vma) vma_start_write() argument
230 vma_start_write_killable(struct vm_area_struct * vma) vma_start_write_killable() argument
239 vma_assert_write_locked(struct vm_area_struct * vma) vma_assert_write_locked() argument
246 vma_assert_locked(struct vm_area_struct * vma) vma_assert_locked() argument
259 vma_assert_attached(struct vm_area_struct * vma) vma_assert_attached() argument
264 vma_assert_detached(struct vm_area_struct * vma) vma_assert_detached() argument
269 vma_mark_attached(struct vm_area_struct * vma) vma_mark_attached() argument
307 vma_lock_init(struct vm_area_struct * vma,bool reset_refcnt) vma_lock_init() argument
308 vma_end_read(struct vm_area_struct * vma) vma_end_read() argument
309 vma_start_write(struct vm_area_struct * vma) vma_start_write() argument
311 vma_start_write_killable(struct vm_area_struct * vma) vma_start_write_killable() argument
312 vma_assert_write_locked(struct vm_area_struct * vma) vma_assert_write_locked() argument
314 vma_assert_attached(struct vm_area_struct * vma) vma_assert_attached() argument
315 vma_assert_detached(struct vm_area_struct * vma) vma_assert_detached() argument
316 vma_mark_attached(struct vm_area_struct * vma) vma_mark_attached() argument
317 vma_mark_detached(struct vm_area_struct * vma) vma_mark_detached() argument
325 vma_assert_locked(struct vm_area_struct * vma) vma_assert_locked() argument
[all...]
H A Drmap.h84 struct vm_area_struct *vma; member
160 anon_vma_prepare(struct vm_area_struct * vma) anon_vma_prepare() argument
168 anon_vma_merge(struct vm_area_struct * vma,struct vm_area_struct * next) anon_vma_merge() argument
230 folio_set_large_mapcount(struct folio * folio,int mapcount,struct vm_area_struct * vma) folio_set_large_mapcount() argument
244 folio_add_return_large_mapcount(struct folio * folio,int diff,struct vm_area_struct * vma) folio_add_return_large_mapcount() argument
294 folio_sub_return_large_mapcount(struct folio * folio,int diff,struct vm_area_struct * vma) folio_sub_return_large_mapcount() argument
345 folio_set_large_mapcount(struct folio * folio,int mapcount,struct vm_area_struct * vma) folio_set_large_mapcount() argument
352 folio_add_large_mapcount(struct folio * folio,int diff,struct vm_area_struct * vma) folio_add_large_mapcount() argument
358 folio_add_return_large_mapcount(struct folio * folio,int diff,struct vm_area_struct * vma) folio_add_return_large_mapcount() argument
364 folio_sub_large_mapcount(struct folio * folio,int diff,struct vm_area_struct * vma) folio_sub_large_mapcount() argument
370 folio_sub_return_large_mapcount(struct folio * folio,int diff,struct vm_area_struct * vma) folio_sub_return_large_mapcount() argument
376 folio_inc_large_mapcount(folio,vma) global() argument
378 folio_inc_return_large_mapcount(folio,vma) global() argument
380 folio_dec_large_mapcount(folio,vma) global() argument
382 folio_dec_return_large_mapcount(folio,vma) global() argument
472 folio_add_anon_rmap_pte(folio,page,vma,address,flags) global() argument
480 folio_add_file_rmap_pte(folio,page,vma) global() argument
488 folio_remove_rmap_pte(folio,page,vma) global() argument
502 hugetlb_try_dup_anon_rmap(struct folio * folio,struct vm_area_struct * vma) hugetlb_try_dup_anon_rmap() argument
934 struct vm_area_struct *vma; global() member
1045 anon_vma_prepare(vma) global() argument
[all...]
H A Dhugetlb.h108 struct vm_area_struct *vma; member
254 hugetlb_zap_begin(struct vm_area_struct * vma,unsigned long * start,unsigned long * end) hugetlb_zap_begin() argument
261 hugetlb_zap_end(struct vm_area_struct * vma,struct zap_details * details) hugetlb_zap_end() argument
285 hugetlb_dup_vma_private(struct vm_area_struct * vma) hugetlb_dup_vma_private() argument
289 clear_vma_resv_huge_pages(struct vm_area_struct * vma) clear_vma_resv_huge_pages() argument
305 huge_pmd_unshare(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long addr,pte_t * ptep) huge_pmd_unshare() argument
311 huge_pmd_unshare_flush(struct mmu_gather * tlb,struct vm_area_struct * vma) huge_pmd_unshare_flush() argument
316 adjust_range_if_pmd_sharing_possible(struct vm_area_struct * vma,unsigned long * start,unsigned long * end) adjust_range_if_pmd_sharing_possible() argument
322 hugetlb_zap_begin(struct vm_area_struct * vma,unsigned long * start,unsigned long * end) hugetlb_zap_begin() argument
328 hugetlb_zap_end(struct vm_area_struct * vma,struct zap_details * details) hugetlb_zap_end() argument
342 move_hugetlb_page_tables(struct vm_area_struct * vma,struct vm_area_struct * new_vma,unsigned long old_addr,unsigned long new_addr,unsigned long len) move_hugetlb_page_tables() argument
365 hugetlb_vma_lock_read(struct vm_area_struct * vma) hugetlb_vma_lock_read() argument
369 hugetlb_vma_unlock_read(struct vm_area_struct * vma) hugetlb_vma_unlock_read() argument
373 hugetlb_vma_lock_write(struct vm_area_struct * vma) hugetlb_vma_lock_write() argument
377 hugetlb_vma_unlock_write(struct vm_area_struct * vma) hugetlb_vma_unlock_write() argument
381 hugetlb_vma_trylock_write(struct vm_area_struct * vma) hugetlb_vma_trylock_write() argument
386 hugetlb_vma_assert_locked(struct vm_area_struct * vma) hugetlb_vma_assert_locked() argument
441 hugetlb_change_protection(struct vm_area_struct * vma,unsigned long address,unsigned long end,pgprot_t newprot,unsigned long cp_flags) hugetlb_change_protection() argument
449 __unmap_hugepage_range(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long start,unsigned long end,struct folio * folio,zap_flags_t zap_flags) __unmap_hugepage_range() argument
457 hugetlb_fault(struct mm_struct * mm,struct vm_area_struct * vma,unsigned long address,unsigned int flags) hugetlb_fault() argument
464 hugetlb_unshare_all_pmds(struct vm_area_struct * vma) hugetlb_unshare_all_pmds() argument
466 fixup_hugetlb_reservations(struct vm_area_struct * vma) fixup_hugetlb_reservations() argument
470 hugetlb_split(struct vm_area_struct * vma,unsigned long addr) hugetlb_split() argument
472 hugetlb_vma_lock_alloc(struct vm_area_struct * vma) hugetlb_vma_lock_alloc() argument
770 hstate_vma(struct vm_area_struct * vma) hstate_vma() argument
1047 huge_ptep_modify_prot_start(struct vm_area_struct * vma,unsigned long addr,pte_t * ptep) huge_ptep_modify_prot_start() argument
1058 huge_ptep_modify_prot_commit(struct vm_area_struct * vma,unsigned long addr,pte_t * ptep,pte_t old_pte,pte_t pte) huge_ptep_modify_prot_commit() argument
1350 flush_hugetlb_tlb_range(vma,addr,end) global() argument
1353 __vma_shareable_lock(struct vm_area_struct * vma) __vma_shareable_lock() argument
1365 hugetlb_walk(struct vm_area_struct * vma,unsigned long addr,unsigned long sz) hugetlb_walk() argument
[all...]
H A Dhuge_mm.h104 #define thp_vma_allowable_order(vma, vm_flags, type, order) \ argument
220 static inline bool thp_vma_suitable_order(struct vm_area_struct *vma, in thp_vma_suitable_order()
245 static inline unsigned long thp_vma_suitable_orders(struct vm_area_struct *vma, in thp_vma_suitable_orders()
289 unsigned long thp_vma_allowable_orders(struct vm_area_struct *vma, in thp_vma_allowable_orders()
331 static inline bool vma_thp_disabled(struct vm_area_struct *vma, in vma_thp_disabled()
471 change_huge_pud(struct mmu_gather *tlb, struct vm_area_struct *vma, in change_huge_pud()
494 struct vm_area_struct *vma) in pmd_trans_huge_lock()
502 struct vm_area_struct *vma) in pud_trans_huge_lock()
578 static inline bool thp_vma_suitable_order(struct vm_area_struct *vma, in thp_vma_suitable_order()
584 static inline unsigned long thp_vma_suitable_orders(struct vm_area_struct *vma, in thp_vma_suitable_orders()
[all …]
/linux/mm/
H A Dvma.c73 static bool vma_is_fork_child(struct vm_area_struct *vma) in vma_is_fork_child() argument
85 struct vm_area_struct *vma = merge_next ? vmg->next : vmg->prev; in is_mergeable_vma() local
140 init_multi_vma_prep(struct vma_prepare * vp,struct vm_area_struct * vma,struct vma_merge_struct * vmg) init_multi_vma_prep() argument
224 __vma_link_file(struct vm_area_struct * vma,struct address_space * mapping) __vma_link_file() argument
238 __remove_shared_vm_struct(struct vm_area_struct * vma,struct address_space * mapping) __remove_shared_vm_struct() argument
264 anon_vma_interval_tree_pre_update_vma(struct vm_area_struct * vma) anon_vma_interval_tree_pre_update_vma() argument
273 anon_vma_interval_tree_post_update_vma(struct vm_area_struct * vma) anon_vma_interval_tree_post_update_vma() argument
410 init_vma_prep(struct vma_prepare * vp,struct vm_area_struct * vma) init_vma_prep() argument
460 remove_vma(struct vm_area_struct * vma) remove_vma() argument
475 unmap_region(struct ma_state * mas,struct vm_area_struct * vma,struct vm_area_struct * prev,struct vm_area_struct * next) unmap_region() argument
497 __split_vma(struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long addr,int new_below) __split_vma() argument
590 split_vma(struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long addr,int new_below) split_vma() argument
646 struct vm_area_struct *vma; validate_mm() local
730 struct vm_area_struct *vma; commit_merge() local
771 can_merge_remove_vma(struct vm_area_struct * vma) can_merge_remove_vma() argument
1228 vma_shrink(struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long start,unsigned long end,pgoff_t pgoff) vma_shrink() argument
1285 struct vm_area_struct *vma; vms_clean_up_area() local
1308 struct vm_area_struct *vma; vms_complete_munmap_vmas() local
1354 struct vm_area_struct *vma; reattach_vmas() local
1522 init_vma_munmap(struct vma_munmap_struct * vms,struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long start,unsigned long end,struct list_head * uf,bool unlock) init_vma_munmap() argument
1558 do_vmi_align_munmap(struct vma_iterator * vmi,struct vm_area_struct * vma,struct mm_struct * mm,unsigned long start,unsigned long end,struct list_head * uf,bool unlock) do_vmi_align_munmap() argument
1610 struct vm_area_struct *vma; do_vmi_munmap() local
1645 struct vm_area_struct *vma = vmg->middle; vma_modify() local
1683 vma_modify_flags(struct vma_iterator * vmi,struct vm_area_struct * prev,struct vm_area_struct * vma,unsigned long start,unsigned long end,vm_flags_t * vm_flags_ptr) vma_modify_flags() argument
1708 vma_modify_name(struct vma_iterator * vmi,struct vm_area_struct * prev,struct vm_area_struct * vma,unsigned long start,unsigned long end,struct anon_vma_name * new_name) vma_modify_name() argument
1720 vma_modify_policy(struct vma_iterator * vmi,struct vm_area_struct * prev,struct vm_area_struct * vma,unsigned long start,unsigned long end,struct mempolicy * new_pol) vma_modify_policy() argument
1732 vma_modify_flags_uffd(struct vma_iterator * vmi,struct vm_area_struct * prev,struct vm_area_struct * vma,unsigned long start,unsigned long end,vm_flags_t vm_flags,struct vm_userfaultfd_ctx new_ctx,bool give_up_on_oom) vma_modify_flags_uffd() argument
1751 vma_merge_extend(struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long delta) vma_merge_extend() argument
1785 unlink_file_vma_batch_add(struct unlink_vma_file_batch * vb,struct vm_area_struct * vma) unlink_file_vma_batch_add() argument
1804 vma_link_file(struct vm_area_struct * vma,bool hold_rmap_lock) vma_link_file() argument
1818 vma_link(struct mm_struct * mm,struct vm_area_struct * vma) vma_link() argument
1842 struct vm_area_struct *vma = *vmap; copy_vma() local
1997 find_mergeable_anon_vma(struct vm_area_struct * vma) find_mergeable_anon_vma() argument
2036 vma_is_shared_writable(struct vm_area_struct * vma) vma_is_shared_writable() argument
2042 vma_fs_can_writeback(struct vm_area_struct * vma) vma_fs_can_writeback() argument
2056 vma_needs_dirty_tracking(struct vm_area_struct * vma) vma_needs_dirty_tracking() argument
2079 vma_wants_writenotify(struct vm_area_struct * vma,pgprot_t vm_page_prot) vma_wants_writenotify() argument
2193 struct vm_area_struct *vma; mm_take_all_locks() local
2289 struct vm_area_struct *vma; mm_drop_all_locks() local
2450 __mmap_new_file_vma(struct mmap_state * map,struct vm_area_struct * vma) __mmap_new_file_vma() argument
2503 struct vm_area_struct *vma; __mmap_new_vma() local
2573 __mmap_complete(struct mmap_state * map,struct vm_area_struct * vma) __mmap_complete() argument
2658 set_vma_user_defined_fields(struct vm_area_struct * vma,struct mmap_state * map) set_vma_user_defined_fields() argument
2697 call_action_complete(struct mmap_state * map,struct vm_area_desc * desc,struct vm_area_struct * vma) call_action_complete() argument
2718 struct vm_area_struct *vma = NULL; __mmap_region() local
2852 do_brk_flags(struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long addr,unsigned long len,vm_flags_t vm_flags) do_brk_flags() argument
3037 acct_stack_growth(struct vm_area_struct * vma,unsigned long size,unsigned long grow) acct_stack_growth() argument
3076 expand_upwards(struct vm_area_struct * vma,unsigned long address) expand_upwards() argument
3162 expand_downwards(struct vm_area_struct * vma,unsigned long address) expand_downwards() argument
3259 insert_vm_struct(struct mm_struct * mm,struct vm_area_struct * vma) insert_vm_struct() argument
[all...]
H A Drmap.c185 int __anon_vma_prepare(struct vm_area_struct *vma) in __anon_vma_prepare() argument
149 anon_vma_chain_link(struct vm_area_struct * vma,struct anon_vma_chain * avc,struct anon_vma * anon_vma) anon_vma_chain_link() argument
333 anon_vma_fork(struct vm_area_struct * vma,struct vm_area_struct * pvma) anon_vma_fork() argument
395 unlink_anon_vmas(struct vm_area_struct * vma) unlink_anon_vmas() argument
764 page_address_in_vma(const struct folio * folio,const struct page * page,const struct vm_area_struct * vma) page_address_in_vma() argument
825 folio_referenced_one(struct folio * folio,struct vm_area_struct * vma,unsigned long address,void * arg) folio_referenced_one() argument
908 invalid_folio_referenced_vma(struct vm_area_struct * vma,void * arg) invalid_folio_referenced_vma() argument
985 struct vm_area_struct *vma = pvmw->vma; page_vma_mkclean_one() local
1059 page_mkclean_one(struct folio * folio,struct vm_area_struct * vma,unsigned long address,void * arg) page_mkclean_one() argument
1070 invalid_mkclean_vma(struct vm_area_struct * vma,void * arg) invalid_mkclean_vma() argument
1111 mapping_wrprotect_range_one(struct folio * folio,struct vm_area_struct * vma,unsigned long address,void * arg) mapping_wrprotect_range_one() argument
1190 pfn_mkclean_range(unsigned long pfn,unsigned long nr_pages,pgoff_t pgoff,struct vm_area_struct * vma) pfn_mkclean_range() argument
1232 __folio_add_rmap(struct folio * folio,struct page * page,int nr_pages,struct vm_area_struct * vma,enum pgtable_level level) __folio_add_rmap() argument
1319 folio_move_anon_rmap(struct folio * folio,struct vm_area_struct * vma) folio_move_anon_rmap() argument
1342 __folio_set_anon(struct folio * folio,struct vm_area_struct * vma,unsigned long address,bool exclusive) __folio_set_anon() argument
1375 __page_check_anon_rmap(const struct folio * folio,const struct page * page,struct vm_area_struct * vma,unsigned long address) __page_check_anon_rmap() argument
1396 __folio_add_anon_rmap(struct folio * folio,struct page * page,int nr_pages,struct vm_area_struct * vma,unsigned long address,rmap_t flags,enum pgtable_level level) __folio_add_anon_rmap() argument
1475 folio_add_anon_rmap_ptes(struct folio * folio,struct page * page,int nr_pages,struct vm_area_struct * vma,unsigned long address,rmap_t flags) folio_add_anon_rmap_ptes() argument
1496 folio_add_anon_rmap_pmd(struct folio * folio,struct page * page,struct vm_area_struct * vma,unsigned long address,rmap_t flags) folio_add_anon_rmap_pmd() argument
1521 folio_add_new_anon_rmap(struct folio * folio,struct vm_area_struct * vma,unsigned long address,rmap_t flags) folio_add_new_anon_rmap() argument
1580 __folio_add_file_rmap(struct folio * folio,struct page * page,int nr_pages,struct vm_area_struct * vma,enum pgtable_level level) __folio_add_file_rmap() argument
1609 folio_add_file_rmap_ptes(struct folio * folio,struct page * page,int nr_pages,struct vm_area_struct * vma) folio_add_file_rmap_ptes() argument
1625 folio_add_file_rmap_pmd(struct folio * folio,struct page * page,struct vm_area_struct * vma) folio_add_file_rmap_pmd() argument
1645 folio_add_file_rmap_pud(struct folio * folio,struct page * page,struct vm_area_struct * vma) folio_add_file_rmap_pud() argument
1656 __folio_remove_rmap(struct folio * folio,struct page * page,int nr_pages,struct vm_area_struct * vma,enum pgtable_level level) __folio_remove_rmap() argument
1777 folio_remove_rmap_ptes(struct folio * folio,struct page * page,int nr_pages,struct vm_area_struct * vma) folio_remove_rmap_ptes() argument
1793 folio_remove_rmap_pmd(struct folio * folio,struct page * page,struct vm_area_struct * vma) folio_remove_rmap_pmd() argument
1813 folio_remove_rmap_pud(struct folio * folio,struct page * page,struct vm_area_struct * vma) folio_remove_rmap_pud() argument
1828 struct vm_area_struct *vma = pvmw->vma; folio_unmap_pte_batch() local
1852 try_to_unmap_one(struct folio * folio,struct vm_area_struct * vma,unsigned long address,void * arg) try_to_unmap_one() argument
2239 invalid_migration_vma(struct vm_area_struct * vma,void * arg) invalid_migration_vma() argument
2281 try_to_migrate_one(struct folio * folio,struct vm_area_struct * vma,unsigned long address,void * arg) try_to_migrate_one() argument
2681 struct vm_area_struct *vma; make_device_exclusive() local
2851 struct vm_area_struct *vma = avc->vma; rmap_walk_anon() local
2896 struct vm_area_struct *vma; __rmap_walk_file() local
2988 hugetlb_add_anon_rmap(struct folio * folio,struct vm_area_struct * vma,unsigned long address,rmap_t flags) hugetlb_add_anon_rmap() argument
3003 hugetlb_add_new_anon_rmap(struct folio * folio,struct vm_area_struct * vma,unsigned long address) hugetlb_add_new_anon_rmap() argument
[all...]
H A Dmemory.c374 struct vm_area_struct *vma, unsigned long floor, in free_pgtables() argument
586 print_bad_page_map(struct vm_area_struct * vma,unsigned long addr,unsigned long long entry,struct page * page,enum pgtable_level level) print_bad_page_map() argument
615 print_bad_pte(vma,addr,pte,page) global() argument
683 __vm_normal_page(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,bool special,unsigned long long entry,enum pgtable_level level) __vm_normal_page() argument
751 vm_normal_page(struct vm_area_struct * vma,unsigned long addr,pte_t pte) vm_normal_page() argument
770 vm_normal_folio(struct vm_area_struct * vma,unsigned long addr,pte_t pte) vm_normal_folio() argument
793 vm_normal_page_pmd(struct vm_area_struct * vma,unsigned long addr,pmd_t pmd) vm_normal_page_pmd() argument
812 vm_normal_folio_pmd(struct vm_area_struct * vma,unsigned long addr,pmd_t pmd) vm_normal_folio_pmd() argument
834 vm_normal_page_pud(struct vm_area_struct * vma,unsigned long addr,pud_t pud) vm_normal_page_pud() argument
868 restore_exclusive_pte(struct vm_area_struct * vma,struct folio * folio,struct page * page,unsigned long address,pte_t * ptep,pte_t orig_pte) restore_exclusive_pte() argument
902 try_restore_exclusive_pte(struct vm_area_struct * vma,unsigned long addr,pte_t * ptep,pte_t orig_pte) try_restore_exclusive_pte() argument
1186 folio_prealloc(struct mm_struct * src_mm,struct vm_area_struct * vma,unsigned long addr,bool need_zero) folio_prealloc() argument
1591 zap_install_uffd_wp_if_needed(struct vm_area_struct * vma,unsigned long addr,pte_t * pte,int nr,struct zap_details * details,pte_t pteval) zap_install_uffd_wp_if_needed() argument
1621 zap_present_folio_ptes(struct mmu_gather * tlb,struct vm_area_struct * vma,struct folio * folio,struct page * page,pte_t * pte,pte_t ptent,unsigned int nr,unsigned long addr,struct zap_details * details,int * rss,bool * force_flush,bool * force_break,bool * any_skipped) zap_present_folio_ptes() argument
1672 zap_present_ptes(struct mmu_gather * tlb,struct vm_area_struct * vma,pte_t * pte,pte_t ptent,unsigned int max_nr,unsigned long addr,struct zap_details * details,int * rss,bool * force_flush,bool * force_break,bool * any_skipped) zap_present_ptes() argument
1718 zap_nonpresent_ptes(struct mmu_gather * tlb,struct vm_area_struct * vma,pte_t * pte,pte_t ptent,unsigned int max_nr,unsigned long addr,struct zap_details * details,int * rss,bool * any_skipped) zap_nonpresent_ptes() argument
1789 do_zap_pte_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pte_t * pte,unsigned long addr,unsigned long end,struct zap_details * details,int * rss,bool * force_flush,bool * force_break,bool * any_skipped) do_zap_pte_range() argument
1825 zap_pte_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pmd_t * pmd,unsigned long addr,unsigned long end,struct zap_details * details) zap_pte_range() argument
1917 zap_pmd_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pud_t * pud,unsigned long addr,unsigned long end,struct zap_details * details) zap_pmd_range() argument
1959 zap_pud_range(struct mmu_gather * tlb,struct vm_area_struct * vma,p4d_t * p4d,unsigned long addr,unsigned long end,struct zap_details * details) zap_pud_range() argument
1987 zap_p4d_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pgd_t * pgd,unsigned long addr,unsigned long end,struct zap_details * details) zap_p4d_range() argument
2006 unmap_page_range(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long addr,unsigned long end,struct zap_details * details) unmap_page_range() argument
2027 unmap_single_vma(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long start_addr,unsigned long end_addr,struct zap_details * details) unmap_single_vma() argument
2087 unmap_vmas(struct mmu_gather * tlb,struct ma_state * mas,struct vm_area_struct * vma,unsigned long start_addr,unsigned long end_addr,unsigned long tree_end) unmap_vmas() argument
2123 zap_page_range_single_batched(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long address,unsigned long size,struct zap_details * details) zap_page_range_single_batched() argument
2162 zap_page_range_single(struct vm_area_struct * vma,unsigned long address,unsigned long size,struct zap_details * details) zap_page_range_single() argument
2183 zap_vma_ptes(struct vm_area_struct * vma,unsigned long address,unsigned long size) zap_vma_ptes() argument
2226 vm_mixed_zeropage_allowed(struct vm_area_struct * vma) vm_mixed_zeropage_allowed() argument
2256 validate_page_before_insert(struct vm_area_struct * vma,struct page * page) validate_page_before_insert() argument
2274 insert_page_into_pte_locked(struct vm_area_struct * vma,pte_t * pte,unsigned long addr,struct page * page,pgprot_t prot,bool mkwrite) insert_page_into_pte_locked() argument
2315 insert_page(struct vm_area_struct * vma,unsigned long addr,struct page * page,pgprot_t prot,bool mkwrite) insert_page() argument
2336 insert_page_in_batch_locked(struct vm_area_struct * vma,pte_t * pte,unsigned long addr,struct page * page,pgprot_t prot) insert_page_in_batch_locked() argument
2350 insert_pages(struct vm_area_struct * vma,unsigned long addr,struct page ** pages,unsigned long * num,pgprot_t prot) insert_pages() argument
2423 vm_insert_pages(struct vm_area_struct * vma,unsigned long addr,struct page ** pages,unsigned long * num) vm_insert_pages() argument
2470 vm_insert_page(struct vm_area_struct * vma,unsigned long addr,struct page * page) vm_insert_page() argument
2497 __vm_map_pages(struct vm_area_struct * vma,struct page ** pages,unsigned long num,unsigned long offset) __vm_map_pages() argument
2540 vm_map_pages(struct vm_area_struct * vma,struct page ** pages,unsigned long num) vm_map_pages() argument
2560 vm_map_pages_zero(struct vm_area_struct * vma,struct page ** pages,unsigned long num) vm_map_pages_zero() argument
2567 insert_pfn(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,pgprot_t prot,bool mkwrite) insert_pfn() argument
2651 vmf_insert_pfn_prot(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,pgprot_t pgprot) vmf_insert_pfn_prot() argument
2698 vmf_insert_pfn(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn) vmf_insert_pfn() argument
2705 vm_mixed_ok(struct vm_area_struct * vma,unsigned long pfn,bool mkwrite) vm_mixed_ok() argument
2719 __vm_insert_mixed(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,bool mkwrite) __vm_insert_mixed() argument
2785 vmf_insert_mixed(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn) vmf_insert_mixed() argument
2797 vmf_insert_mixed_mkwrite(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn) vmf_insert_mixed_mkwrite() argument
2920 remap_pfn_range_internal(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t prot) remap_pfn_range_internal() argument
2953 remap_pfn_range_notrack(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t prot) remap_pfn_range_notrack() argument
2999 remap_pfn_range_track(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t prot) remap_pfn_range_track() argument
3036 do_remap_pfn_range(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t prot) do_remap_pfn_range() argument
3042 do_remap_pfn_range(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t prot) do_remap_pfn_range() argument
3061 remap_pfn_range_prepare_vma(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size) remap_pfn_range_prepare_vma() argument
3089 remap_pfn_range(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t prot) remap_pfn_range() argument
3102 remap_pfn_range_complete(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t prot) remap_pfn_range_complete() argument
3123 vm_iomap_memory(struct vm_area_struct * vma,phys_addr_t start,unsigned long len) vm_iomap_memory() argument
3408 struct vm_area_struct *vma = vmf->vma; __wp_page_copy_user() local
3500 __get_fault_gfp_mask(struct vm_area_struct * vma) __get_fault_gfp_mask() argument
3555 struct vm_area_struct *vma = vmf->vma; fault_dirty_shared_page() local
3609 struct vm_area_struct *vma = vmf->vma; wp_page_reuse() local
3642 struct vm_area_struct *vma = vmf->vma; vmf_can_call_fault() local
3667 struct vm_area_struct *vma = vmf->vma; __vmf_anon_prepare() local
3703 struct vm_area_struct *vma = vmf->vma; wp_page_copy() local
3894 struct vm_area_struct *vma = vmf->vma; wp_pfn_shared() local
3917 struct vm_area_struct *vma = vmf->vma; wp_page_shared() local
3956 __wp_can_reuse_large_anon_folio(struct folio * folio,struct vm_area_struct * vma) __wp_can_reuse_large_anon_folio() argument
4022 __wp_can_reuse_large_anon_folio(struct folio * folio,struct vm_area_struct * vma) __wp_can_reuse_large_anon_folio() argument
4029 wp_can_reuse_anon_folio(struct folio * folio,struct vm_area_struct * vma) wp_can_reuse_anon_folio() argument
4095 struct vm_area_struct *vma = vmf->vma; do_wp_page() local
4186 unmap_mapping_range_vma(struct vm_area_struct * vma,unsigned long start_addr,unsigned long end_addr,struct zap_details * details) unmap_mapping_range_vma() argument
4198 struct vm_area_struct *vma; unmap_mapping_range_tree() local
4320 struct vm_area_struct *vma = vmf->vma; remove_device_exclusive_entry() local
4361 should_try_to_free_swap(struct folio * folio,struct vm_area_struct * vma,unsigned int fault_flags) should_try_to_free_swap() argument
4452 struct vm_area_struct *vma = vmf->vma; __alloc_swap_folio() local
4530 struct vm_area_struct *vma = vmf->vma; alloc_swap_folio() local
4626 struct vm_area_struct *vma = vmf->vma; do_swap_page() local
5088 struct vm_area_struct *vma = vmf->vma; alloc_anon_folio() local
5178 struct vm_area_struct *vma = vmf->vma; do_anonymous_page() local
5298 struct vm_area_struct *vma = vmf->vma; __do_fault() local
5355 struct vm_area_struct *vma = vmf->vma; deposit_prealloc_pte() local
5368 struct vm_area_struct *vma = vmf->vma; do_set_pmd() local
5459 struct vm_area_struct *vma = vmf->vma; set_pte_range() local
5517 struct vm_area_struct *vma = vmf->vma; finish_fault() local
5772 struct vm_area_struct *vma = vmf->vma; do_cow_fault() local
5814 struct vm_area_struct *vma = vmf->vma; do_shared_fault() local
5864 struct vm_area_struct *vma = vmf->vma; do_fault() local
5910 struct vm_area_struct *vma = vmf->vma; numa_migrate_check() local
5953 numa_rebuild_single_mapping(struct vm_fault * vmf,struct vm_area_struct * vma,unsigned long fault_addr,pte_t * fault_pte,bool writable) numa_rebuild_single_mapping() argument
5968 numa_rebuild_large_mapping(struct vm_fault * vmf,struct vm_area_struct * vma,struct folio * folio,pte_t fault_pte,bool ignore_writable,bool pte_write_upgrade) numa_rebuild_large_mapping() argument
6009 struct vm_area_struct *vma = vmf->vma; do_numa_page() local
6100 struct vm_area_struct *vma = vmf->vma; create_huge_pmd() local
6111 struct vm_area_struct *vma = vmf->vma; wp_huge_pmd() local
6144 struct vm_area_struct *vma = vmf->vma; create_huge_pud() local
6158 struct vm_area_struct *vma = vmf->vma; wp_huge_pud() local
6314 __handle_mm_fault(struct vm_area_struct * vma,unsigned long address,unsigned int flags) __handle_mm_fault() argument
6486 lru_gen_enter_fault(struct vm_area_struct * vma) lru_gen_enter_fault() argument
6497 lru_gen_enter_fault(struct vm_area_struct * vma) lru_gen_enter_fault() argument
6506 sanitize_fault_flags(struct vm_area_struct * vma,unsigned int * flags) sanitize_fault_flags() argument
6548 handle_mm_fault(struct vm_area_struct * vma,unsigned long address,unsigned int flags,struct pt_regs * regs) handle_mm_fault() argument
6701 pfnmap_lockdep_assert(struct vm_area_struct * vma) pfnmap_lockdep_assert() argument
6748 struct vm_area_struct *vma = args->vma; follow_pfnmap_start() local
6851 generic_access_phys(struct vm_area_struct * vma,unsigned long addr,void * buf,int len,int write) generic_access_phys() argument
6926 struct vm_area_struct *vma = NULL; __access_remote_vm() local
7053 struct vm_area_struct *vma = NULL; __copy_remote_vm_str() local
7150 struct vm_area_struct *vma; print_vma_addr() local
7281 copy_user_gigantic_page(struct folio * dst,struct folio * src,unsigned long addr_hint,struct vm_area_struct * vma,unsigned int nr_pages) copy_user_gigantic_page() argument
7304 struct vm_area_struct *vma; global() member
7319 copy_user_large_folio(struct folio * dst,struct folio * src,unsigned long addr_hint,struct vm_area_struct * vma) copy_user_large_folio() argument
7394 vma_pgtable_walk_begin(struct vm_area_struct * vma) vma_pgtable_walk_begin() argument
7400 vma_pgtable_walk_end(struct vm_area_struct * vma) vma_pgtable_walk_end() argument
[all...]
H A Dmmap_lock.c53 __vma_enter_locked(struct vm_area_struct * vma,bool detaching,int state) __vma_enter_locked() argument
93 __vma_exit_locked(struct vm_area_struct * vma,bool * detached) __vma_exit_locked() argument
99 __vma_start_write(struct vm_area_struct * vma,unsigned int mm_lock_seq,int state) __vma_start_write() argument
127 vma_mark_detached(struct vm_area_struct * vma) vma_mark_detached() argument
163 vma_start_read(struct mm_struct * mm,struct vm_area_struct * vma) vma_start_read() argument
249 struct vm_area_struct *vma; lock_vma_under_rcu() local
297 struct vm_area_struct *vma; lock_next_vma_under_mmap_lock() local
322 struct vm_area_struct *vma; lock_next_vma() local
448 struct vm_area_struct *vma; lock_mm_and_find_vma() local
510 struct vm_area_struct *vma; lock_mm_and_find_vma() local
[all...]
H A Dmprotect.c41 static bool maybe_change_pte_writable(struct vm_area_struct *vma, pte_t pte) in maybe_change_pte_writable() argument
61 static bool can_change_private_pte_writable(struct vm_area_struct *vma, in can_change_private_pte_writable() argument
79 static bool can_change_shared_pte_writable(struct vm_area_struct *vma, in can_change_shared_pte_writable() argument
97 bool can_change_pte_writable(struct vm_area_struct *vma, unsigne argument
120 prot_commit_flush_ptes(struct vm_area_struct * vma,unsigned long addr,pte_t * ptep,pte_t oldpte,pte_t ptent,int nr_ptes,int idx,bool set_write,struct mmu_gather * tlb) prot_commit_flush_ptes() argument
172 commit_anon_folio_batch(struct vm_area_struct * vma,struct folio * folio,struct page * first_page,unsigned long addr,pte_t * ptep,pte_t oldpte,pte_t ptent,int nr_ptes,struct mmu_gather * tlb) commit_anon_folio_batch() argument
191 set_write_prot_commit_flush_ptes(struct vm_area_struct * vma,struct folio * folio,struct page * page,unsigned long addr,pte_t * ptep,pte_t oldpte,pte_t ptent,int nr_ptes,struct mmu_gather * tlb) set_write_prot_commit_flush_ptes() argument
215 change_pte_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pmd_t * pmd,unsigned long addr,unsigned long end,pgprot_t newprot,unsigned long cp_flags) change_pte_range() argument
393 pgtable_split_needed(struct vm_area_struct * vma,unsigned long cp_flags) pgtable_split_needed() argument
409 pgtable_populate_needed(struct vm_area_struct * vma,unsigned long cp_flags) pgtable_populate_needed() argument
425 change_pmd_prepare(vma,pmd,cp_flags) global() argument
440 change_prepare(vma,high,low,addr,cp_flags) global() argument
452 change_pmd_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pud_t * pud,unsigned long addr,unsigned long end,pgprot_t newprot,unsigned long cp_flags) change_pmd_range() argument
522 change_pud_range(struct mmu_gather * tlb,struct vm_area_struct * vma,p4d_t * p4d,unsigned long addr,unsigned long end,pgprot_t newprot,unsigned long cp_flags) change_pud_range() argument
581 change_p4d_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pgd_t * pgd,unsigned long addr,unsigned long end,pgprot_t newprot,unsigned long cp_flags) change_p4d_range() argument
604 change_protection_range(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long addr,unsigned long end,pgprot_t newprot,unsigned long cp_flags) change_protection_range() argument
634 change_protection(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long start,unsigned long end,unsigned long cp_flags) change_protection() argument
696 mprotect_fixup(struct vma_iterator * vmi,struct mmu_gather * tlb,struct vm_area_struct * vma,struct vm_area_struct ** pprev,unsigned long start,unsigned long end,vm_flags_t newflags) mprotect_fixup() argument
805 struct vm_area_struct *vma, *prev; do_mprotect_pkey() local
[all...]
H A Dmadvise.c80 struct vm_area_struct *vma; member
110 struct anon_vma_name *anon_vma_name(struct vm_area_struct *vma) in anon_vma_name() argument
119 replace_anon_vma_name(struct vm_area_struct * vma,struct anon_vma_name * anon_name) replace_anon_vma_name() argument
139 replace_anon_vma_name(struct vm_area_struct * vma,struct anon_vma_name * anon_name) replace_anon_vma_name() argument
155 struct vm_area_struct *vma = madv_behavior->vma; madvise_update_vma() local
190 struct vm_area_struct *vma = walk->private; swapin_walk_pmd_entry() local
234 shmem_swapin_range(struct vm_area_struct * vma,unsigned long start,unsigned long end,struct address_space * mapping) shmem_swapin_range() argument
283 struct vm_area_struct *vma = madv_behavior->vma; madvise_willneed() local
329 can_do_file_pageout(struct vm_area_struct * vma) can_do_file_pageout() argument
362 struct vm_area_struct *vma = walk->vma; madvise_cold_or_pageout_pte_range() local
578 struct vm_area_struct *vma = madv_behavior->vma; madvise_cold_page_range() local
591 can_madv_lru_vma(struct vm_area_struct * vma) can_madv_lru_vma() argument
598 struct vm_area_struct *vma = madv_behavior->vma; madvise_cold() local
613 madvise_pageout_page_range(struct mmu_gather * tlb,struct vm_area_struct * vma,struct madvise_behavior_range * range) madvise_pageout_page_range() argument
630 struct vm_area_struct *vma = madv_behavior->vma; madvise_pageout() local
660 struct vm_area_struct *vma = walk->vma; madvise_free_pte_range() local
803 struct vm_area_struct *vma = madv_behavior->vma; madvise_free_single_vma() local
875 struct vm_area_struct *vma = madv_behavior->vma; madvise_dontneed_free_valid_vma() local
917 struct vm_area_struct *vma; madvise_dontneed_free() local
1014 struct vm_area_struct *vma = madv_behavior->vma; madvise_remove() local
1054 is_valid_guard_vma(struct vm_area_struct * vma,bool allow_locked) is_valid_guard_vma() argument
1125 struct vm_area_struct *vma = madv_behavior->vma; madvise_guard_install() local
1254 struct vm_area_struct *vma = madv_behavior->vma; madvise_guard_remove() local
1301 struct vm_area_struct *vma = madv_behavior->vma; can_madvise_modify() local
1350 struct vm_area_struct *vma = madv_behavior->vma; madvise_vma_behavior() local
1599 is_vma_lock_sufficient(struct vm_area_struct * vma,struct madvise_behavior * madv_behavior) is_vma_lock_sufficient() argument
1641 struct vm_area_struct *vma; try_vma_read_lock() local
1679 struct vm_area_struct *prev, *vma; madvise_walk_vmas() local
[all...]
H A Dnommu.c92 struct vm_area_struct *vma; in kobjsize() local
148 struct vm_area_struct *vma; in __vmalloc_user_flags() local
345 int vm_insert_page(struct vm_area_struct *vma, unsigned long addr, in vm_insert_page()
352 int vm_insert_pages(struct vm_area_struct *vma, unsigned long addr, in vm_insert_pages()
359 int vm_map_pages(struct vm_area_struct *vma, struct page **pages, in vm_map_pages()
366 int vm_map_pages_zero(struct vm_area_struct *vma, struct page **pages, in vm_map_pages_zero()
562 static void setup_vma_to_mm(struct vm_area_struct *vma, struct mm_struct *mm) in setup_vma_to_mm()
578 static void cleanup_vma_from_mm(struct vm_area_struct *vma) in cleanup_vma_from_mm()
597 static int delete_vma_from_mm(struct vm_area_struct *vma) in delete_vma_from_mm()
616 static void delete_vma(struct mm_struct *mm, struct vm_area_struct *vma) in delete_vma()
[all …]
H A Dhugetlb.c284 if (__vma_shareable_lock(vma)) { in subpool_vma() argument
292 hugetlb_vma_lock_read(struct vm_area_struct * vma) hugetlb_vma_lock_read() argument
305 hugetlb_vma_unlock_read(struct vm_area_struct * vma) hugetlb_vma_unlock_read() argument
318 hugetlb_vma_lock_write(struct vm_area_struct * vma) hugetlb_vma_lock_write() argument
331 hugetlb_vma_unlock_write(struct vm_area_struct * vma) hugetlb_vma_unlock_write() argument
344 hugetlb_vma_trylock_write(struct vm_area_struct * vma) hugetlb_vma_trylock_write() argument
360 hugetlb_vma_assert_locked(struct vm_area_struct * vma) hugetlb_vma_assert_locked() argument
383 struct vm_area_struct *vma = vma_lock->vma; __hugetlb_vma_unlock_write_put() local
396 __hugetlb_vma_unlock_write_free(struct vm_area_struct * vma) __hugetlb_vma_unlock_write_free() argument
410 hugetlb_vma_lock_free(struct vm_area_struct * vma) hugetlb_vma_lock_free() argument
430 hugetlb_vma_lock_alloc(struct vm_area_struct * vma) hugetlb_vma_lock_alloc() argument
1023 vma_hugecache_offset(struct hstate * h,struct vm_area_struct * vma,unsigned long address) vma_hugecache_offset() argument
1038 vma_kernel_pagesize(struct vm_area_struct * vma) vma_kernel_pagesize() argument
1052 vma_mmu_pagesize(struct vm_area_struct * vma) vma_mmu_pagesize() argument
1085 get_vma_private_data(struct vm_area_struct * vma) get_vma_private_data() argument
1090 set_vma_private_data(struct vm_area_struct * vma,unsigned long value) set_vma_private_data() argument
1180 vma_resv_map(struct vm_area_struct * vma) vma_resv_map() argument
1195 set_vma_resv_flags(struct vm_area_struct * vma,unsigned long flags) set_vma_resv_flags() argument
1219 is_vma_resv_set(struct vm_area_struct * vma,unsigned long flag) is_vma_resv_set() argument
1233 __vma_private_lock(struct vm_area_struct * vma) __vma_private_lock() argument
1240 hugetlb_dup_vma_private(struct vm_area_struct * vma) hugetlb_dup_vma_private() argument
1271 clear_vma_resv_huge_pages(struct vm_area_struct * vma) clear_vma_resv_huge_pages() argument
1382 dequeue_hugetlb_folio_vma(struct hstate * h,struct vm_area_struct * vma,unsigned long address,long gbl_chg) dequeue_hugetlb_folio_vma() argument
2236 alloc_buddy_hugetlb_folio_with_mpol(struct hstate * h,struct vm_area_struct * vma,unsigned long addr) alloc_buddy_hugetlb_folio_with_mpol() argument
2511 __vma_reservation_common(struct hstate * h,struct vm_area_struct * vma,unsigned long addr,enum vma_resv_mode mode) __vma_reservation_common() argument
2591 vma_needs_reservation(struct hstate * h,struct vm_area_struct * vma,unsigned long addr) vma_needs_reservation() argument
2597 vma_commit_reservation(struct hstate * h,struct vm_area_struct * vma,unsigned long addr) vma_commit_reservation() argument
2603 vma_end_reservation(struct hstate * h,struct vm_area_struct * vma,unsigned long addr) vma_end_reservation() argument
2609 vma_add_reservation(struct hstate * h,struct vm_area_struct * vma,unsigned long addr) vma_add_reservation() argument
2615 vma_del_reservation(struct hstate * h,struct vm_area_struct * vma,unsigned long addr) vma_del_reservation() argument
2640 restore_reserve_on_error(struct hstate * h,struct vm_area_struct * vma,unsigned long address,struct folio * folio) restore_reserve_on_error() argument
2895 alloc_hugetlb_folio(struct vm_area_struct * vma,unsigned long addr,bool cow_from_owner) alloc_hugetlb_folio() argument
4690 hugetlb_vm_op_open(struct vm_area_struct * vma) hugetlb_vm_op_open() argument
4728 hugetlb_vm_op_close(struct vm_area_struct * vma) hugetlb_vm_op_close() argument
4759 hugetlb_vm_op_split(struct vm_area_struct * vma,unsigned long addr) hugetlb_vm_op_split() argument
4766 hugetlb_split(struct vm_area_struct * vma,unsigned long addr) hugetlb_split() argument
4799 hugetlb_vm_op_pagesize(struct vm_area_struct * vma) hugetlb_vm_op_pagesize() argument
4831 make_huge_pte(struct vm_area_struct * vma,struct folio * folio,bool try_mkwrite) make_huge_pte() argument
4848 set_huge_ptep_writable(struct vm_area_struct * vma,unsigned long address,pte_t * ptep) set_huge_ptep_writable() argument
4858 set_huge_ptep_maybe_writable(struct vm_area_struct * vma,unsigned long address,pte_t * ptep) set_huge_ptep_maybe_writable() argument
4866 hugetlb_install_folio(struct vm_area_struct * vma,pte_t * ptep,unsigned long addr,struct folio * new_folio,pte_t old,unsigned long sz) hugetlb_install_folio() argument
5063 move_huge_pte(struct vm_area_struct * vma,unsigned long old_addr,unsigned long new_addr,pte_t * src_pte,pte_t * dst_pte,unsigned long sz) move_huge_pte() argument
5102 move_hugetlb_page_tables(struct vm_area_struct * vma,struct vm_area_struct * new_vma,unsigned long old_addr,unsigned long new_addr,unsigned long len) move_hugetlb_page_tables() argument
5167 __unmap_hugepage_range(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long start,unsigned long end,struct folio * folio,zap_flags_t zap_flags) __unmap_hugepage_range() argument
5323 __hugetlb_zap_begin(struct vm_area_struct * vma,unsigned long * start,unsigned long * end) __hugetlb_zap_begin() argument
5335 __hugetlb_zap_end(struct vm_area_struct * vma,struct zap_details * details) __hugetlb_zap_end() argument
5362 unmap_hugepage_range(struct vm_area_struct * vma,unsigned long start,unsigned long end,struct folio * folio,zap_flags_t zap_flags) unmap_hugepage_range() argument
5388 unmap_ref_private(struct mm_struct * mm,struct vm_area_struct * vma,struct folio * folio,unsigned long address) unmap_ref_private() argument
5447 struct vm_area_struct *vma = vmf->vma; hugetlb_wp() local
5641 hugetlbfs_pagecache_present(struct hstate * h,struct vm_area_struct * vma,unsigned long address) hugetlbfs_pagecache_present() argument
5722 struct vm_area_struct *vma = vmf->vma; hugetlb_no_page() local
5967 hugetlb_fault(struct mm_struct * mm,struct vm_area_struct * vma,unsigned long address,unsigned int flags) hugetlb_fault() argument
6163 alloc_hugetlb_folio_vma(struct hstate * h,struct vm_area_struct * vma,unsigned long address) alloc_hugetlb_folio_vma() argument
6409 hugetlb_change_protection(struct vm_area_struct * vma,unsigned long address,unsigned long end,pgprot_t newprot,unsigned long cp_flags) hugetlb_change_protection() argument
6780 page_table_shareable(struct vm_area_struct * svma,struct vm_area_struct * vma,unsigned long addr,pgoff_t idx) page_table_shareable() argument
6807 want_pmd_share(struct vm_area_struct * vma,unsigned long addr) want_pmd_share() argument
6833 adjust_range_if_pmd_sharing_possible(struct vm_area_struct * vma,unsigned long * start,unsigned long * end) adjust_range_if_pmd_sharing_possible() argument
6864 huge_pmd_share(struct mm_struct * mm,struct vm_area_struct * vma,unsigned long addr,pud_t * pud) huge_pmd_share() argument
6925 huge_pmd_unshare(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long addr,pte_t * ptep) huge_pmd_unshare() argument
6959 huge_pmd_unshare_flush(struct mmu_gather * tlb,struct vm_area_struct * vma) huge_pmd_unshare_flush() argument
6973 huge_pmd_share(struct mm_struct * mm,struct vm_area_struct * vma,unsigned long addr,pud_t * pud) huge_pmd_share() argument
6979 huge_pmd_unshare(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long addr,pte_t * ptep) huge_pmd_unshare() argument
6985 huge_pmd_unshare_flush(struct mmu_gather * tlb,struct vm_area_struct * vma) huge_pmd_unshare_flush() argument
6989 adjust_range_if_pmd_sharing_possible(struct vm_area_struct * vma,unsigned long * start,unsigned long * end) adjust_range_if_pmd_sharing_possible() argument
6994 want_pmd_share(struct vm_area_struct * vma,unsigned long addr) want_pmd_share() argument
7001 huge_pte_alloc(struct mm_struct * mm,struct vm_area_struct * vma,unsigned long addr,unsigned long sz) huge_pte_alloc() argument
7246 hugetlb_unshare_pmds(struct vm_area_struct * vma,unsigned long start,unsigned long end,bool take_locks) hugetlb_unshare_pmds() argument
7307 hugetlb_unshare_all_pmds(struct vm_area_struct * vma) hugetlb_unshare_all_pmds() argument
7322 fixup_hugetlb_reservations(struct vm_area_struct * vma) fixup_hugetlb_reservations() argument
[all...]
H A Dmmap.c81 void vma_set_page_prot(struct vm_area_struct *vma) in vma_set_page_prot()
693 struct vm_area_struct *vma, *prev; in generic_get_unmapped_area() local
741 struct vm_area_struct *vma, *prev; in generic_get_unmapped_area_topdown() local
927 struct vm_area_struct *vma; in find_vma_prev() local
954 int expand_stack_locked(struct vm_area_struct *vma, unsigned long address) in expand_stack_locked()
961 struct vm_area_struct *vma, *prev; in find_extend_vma_locked() local
976 int expand_stack_locked(struct vm_area_struct *vma, unsigned long address) in expand_stack_locked()
983 struct vm_area_struct *vma; in find_extend_vma_locked() local
1003 #define vma_expand_up(vma,addr) expand_upwards(vma, addr) argument
1004 #define vma_expand_down(vma, addr) (-EFAULT) argument
[all …]
H A Dpgtable-generic.c70 int ptep_set_access_flags(struct vm_area_struct *vma, in ptep_set_access_flags() argument
84 int ptep_clear_flush_young(struct vm_area_struct *vma, in ptep_clear_flush_young() argument
96 pte_t ptep_clear_flush(struct vm_area_struct *vma, unsigned long address, in ptep_clear_flush() argument
111 int pmdp_set_access_flags(struct vm_area_struct *vma, in pmdp_set_access_flags() argument
126 pmdp_clear_flush_young(struct vm_area_struct * vma,unsigned long address,pmd_t * pmdp) pmdp_clear_flush_young() argument
139 pmdp_huge_clear_flush(struct vm_area_struct * vma,unsigned long address,pmd_t * pmdp) pmdp_huge_clear_flush() argument
151 pudp_huge_clear_flush(struct vm_area_struct * vma,unsigned long address,pud_t * pudp) pudp_huge_clear_flush() argument
199 pmdp_invalidate(struct vm_area_struct * vma,unsigned long address,pmd_t * pmdp) pmdp_invalidate() argument
210 pmdp_invalidate_ad(struct vm_area_struct * vma,unsigned long address,pmd_t * pmdp) pmdp_invalidate_ad() argument
219 pmdp_collapse_flush(struct vm_area_struct * vma,unsigned long address,pmd_t * pmdp) pmdp_collapse_flush() argument
[all...]
H A Dhuge_memory.c85 static inline bool file_thp_enabled(struct vm_area_struct *vma) in file_thp_enabled() argument
100 unsigned long __thp_vma_allowable_orders(struct vm_area_struct *vma, in __thp_vma_allowable_orders() argument
1073 maybe_pmd_mkwrite(pmd_t pmd,struct vm_area_struct * vma) maybe_pmd_mkwrite() argument
1253 vma_alloc_anon_folio_pmd(struct vm_area_struct * vma,unsigned long addr) vma_alloc_anon_folio_pmd() argument
1297 map_anon_folio_pmd_nopf(struct folio * folio,pmd_t * pmd,struct vm_area_struct * vma,unsigned long haddr) map_anon_folio_pmd_nopf() argument
1311 map_anon_folio_pmd_pf(struct folio * folio,pmd_t * pmd,struct vm_area_struct * vma,unsigned long haddr) map_anon_folio_pmd_pf() argument
1323 struct vm_area_struct *vma = vmf->vma; __do_huge_pmd_anonymous_page() local
1374 struct vm_area_struct *vma = vmf->vma; do_huge_pmd_device_private() local
1419 vma_thp_gfp_mask(struct vm_area_struct * vma) vma_thp_gfp_mask() argument
1447 set_huge_zero_folio(pgtable_t pgtable,struct mm_struct * mm,struct vm_area_struct * vma,unsigned long haddr,pmd_t * pmd,struct folio * zero_folio) set_huge_zero_folio() argument
1460 struct vm_area_struct *vma = vmf->vma; do_huge_pmd_anonymous_page() local
1523 insert_pmd(struct vm_area_struct * vma,unsigned long addr,pmd_t * pmd,struct folio_or_pfn fop,pgprot_t prot,bool write) insert_pmd() argument
1608 struct vm_area_struct *vma = vmf->vma; vmf_insert_pfn_pmd() local
1633 struct vm_area_struct *vma = vmf->vma; vmf_insert_folio_pmd() local
1648 maybe_pud_mkwrite(pud_t pud,struct vm_area_struct * vma) maybe_pud_mkwrite() argument
1655 insert_pud(struct vm_area_struct * vma,unsigned long addr,pud_t * pud,struct folio_or_pfn fop,pgprot_t prot,bool write) insert_pud() argument
1716 struct vm_area_struct *vma = vmf->vma; vmf_insert_pfn_pud() local
1749 struct vm_area_struct *vma = vmf->vma; vmf_insert_folio_pud() local
1773 touch_pmd(struct vm_area_struct * vma,unsigned long addr,pmd_t * pmd,bool write) touch_pmd() argument
1953 touch_pud(struct vm_area_struct * vma,unsigned long addr,pud_t * pud,bool write) touch_pud() argument
1968 copy_huge_pud(struct mm_struct * dst_mm,struct mm_struct * src_mm,pud_t * dst_pud,pud_t * src_pud,unsigned long addr,struct vm_area_struct * vma) copy_huge_pud() argument
2028 struct vm_area_struct *vma = vmf->vma; do_huge_zero_wp_pmd() local
2060 struct vm_area_struct *vma = vmf->vma; do_huge_pmd_wp_page() local
2151 can_change_pmd_writable(struct vm_area_struct * vma,unsigned long addr,pmd_t pmd) can_change_pmd_writable() argument
2184 struct vm_area_struct *vma = vmf->vma; do_huge_pmd_numa_page() local
2262 madvise_free_huge_pmd(struct mmu_gather * tlb,struct vm_area_struct * vma,pmd_t * pmd,unsigned long addr,unsigned long next) madvise_free_huge_pmd() argument
2341 zap_huge_pmd(struct mmu_gather * tlb,struct vm_area_struct * vma,pmd_t * pmd,unsigned long addr) zap_huge_pmd() argument
2425 pmd_move_must_withdraw(spinlock_t * new_pmd_ptl,spinlock_t * old_pmd_ptl,struct vm_area_struct * vma) pmd_move_must_withdraw() argument
2461 move_huge_pmd(struct vm_area_struct * vma,unsigned long old_addr,unsigned long new_addr,pmd_t * old_pmd,pmd_t * new_pmd) move_huge_pmd() argument
2555 change_huge_pmd(struct mmu_gather * tlb,struct vm_area_struct * vma,pmd_t * pmd,unsigned long addr,pgprot_t newprot,unsigned long cp_flags) change_huge_pmd() argument
2656 change_huge_pud(struct mmu_gather * tlb,struct vm_area_struct * vma,pud_t * pudp,unsigned long addr,pgprot_t newprot,unsigned long cp_flags) change_huge_pud() argument
2821 __pmd_trans_huge_lock(pmd_t * pmd,struct vm_area_struct * vma) __pmd_trans_huge_lock() argument
2838 __pud_trans_huge_lock(pud_t * pud,struct vm_area_struct * vma) __pud_trans_huge_lock() argument
2850 zap_huge_pud(struct mmu_gather * tlb,struct vm_area_struct * vma,pud_t * pud,unsigned long addr) zap_huge_pud() argument
2885 __split_huge_pud_locked(struct vm_area_struct * vma,pud_t * pud,unsigned long haddr) __split_huge_pud_locked() argument
2917 __split_huge_pud(struct vm_area_struct * vma,pud_t * pud,unsigned long address) __split_huge_pud() argument
2937 __split_huge_pud(struct vm_area_struct * vma,pud_t * pud,unsigned long address) __split_huge_pud() argument
2943 __split_huge_zero_page_pmd(struct vm_area_struct * vma,unsigned long haddr,pmd_t * pmd) __split_huge_zero_page_pmd() argument
2984 __split_huge_pmd_locked(struct vm_area_struct * vma,pmd_t * pmd,unsigned long haddr,bool freeze) __split_huge_pmd_locked() argument
3265 split_huge_pmd_locked(struct vm_area_struct * vma,unsigned long address,pmd_t * pmd,bool freeze) split_huge_pmd_locked() argument
3273 __split_huge_pmd(struct vm_area_struct * vma,pmd_t * pmd,unsigned long address,bool freeze) __split_huge_pmd() argument
3289 split_huge_pmd_address(struct vm_area_struct * vma,unsigned long address,bool freeze) split_huge_pmd_address() argument
3300 split_huge_pmd_if_needed(struct vm_area_struct * vma,unsigned long address) split_huge_pmd_if_needed() argument
3312 vma_adjust_trans_huge(struct vm_area_struct * vma,unsigned long start,unsigned long end,struct vm_area_struct * next) vma_adjust_trans_huge() argument
3351 __discard_anon_folio_pmd_locked(struct vm_area_struct * vma,unsigned long addr,pmd_t * pmdp,struct folio * folio) __discard_anon_folio_pmd_locked() argument
3414 unmap_huge_pmd_locked(struct vm_area_struct * vma,unsigned long addr,pmd_t * pmdp,struct folio * folio) unmap_huge_pmd_locked() argument
4568 vma_not_suitable_for_thp_split(struct vm_area_struct * vma) vma_not_suitable_for_thp_split() argument
4611 struct vm_area_struct *vma = vma_lookup(mm, addr); split_huge_pages_pid() local
4866 struct vm_area_struct *vma = pvmw->vma; set_pmd_migration_entry() local
4918 struct vm_area_struct *vma = pvmw->vma; remove_migration_pmd() local
[all...]
/linux/drivers/gpu/drm/xe/
H A Dxe_vm.h112 * DOC: Provide accessors for vma members to facilitate easy change of in xe_vma_start() argument
117 return vma->gpuva.va.addr; in xe_vma_size() argument
122 return vma->gpuva.va.range; in xe_vma_end() argument
127 return xe_vma_start(vma) + xe_vma_size(vma); in xe_vma_bo_offset() argument
132 xe_vma_bo(struct xe_vma * vma) xe_vma_bo() argument
138 xe_vma_vm(struct xe_vma * vma) xe_vma_vm() argument
143 xe_vma_read_only(struct xe_vma * vma) xe_vma_read_only() argument
148 xe_vma_userptr(struct xe_vma * vma) xe_vma_userptr() argument
153 xe_vma_is_null(struct xe_vma * vma) xe_vma_is_null() argument
158 xe_vma_is_cpu_addr_mirror(struct xe_vma * vma) xe_vma_is_cpu_addr_mirror() argument
163 xe_vma_has_no_bo(struct xe_vma * vma) xe_vma_has_no_bo() argument
168 xe_vma_is_userptr(struct xe_vma * vma) xe_vma_is_userptr() argument
188 to_userptr_vma(struct xe_vma * vma) to_userptr_vma() argument
[all...]
H A Dxe_vm.c603 struct xe_vma *vma; xe_vma_svm_prefetch_op_fini() local
651 xe_vm_populate_rebind(struct xe_vma_op * op,struct xe_vma * vma,u8 tile_mask) xe_vm_populate_rebind() argument
666 xe_vm_ops_add_rebind(struct xe_vma_ops * vops,struct xe_vma * vma,u8 tile_mask) xe_vm_ops_add_rebind() argument
691 struct xe_vma *vma, *next; xe_vm_rebind() local
743 xe_vma_rebind(struct xe_vm * vm,struct xe_vma * vma,u8 tile_mask) xe_vma_rebind() argument
787 xe_vm_populate_range_rebind(struct xe_vma_op * op,struct xe_vma * vma,struct xe_svm_range * range,u8 tile_mask) xe_vm_populate_range_rebind() argument
801 xe_vm_ops_add_range_rebind(struct xe_vma_ops * vops,struct xe_vma * vma,struct xe_svm_range * range,u8 tile_mask) xe_vm_ops_add_range_rebind() argument
831 xe_vm_range_rebind(struct xe_vm * vm,struct xe_vma * vma,struct xe_svm_range * range,u8 tile_mask) xe_vm_range_rebind() argument
960 xe_vma_free(struct xe_vma * vma) xe_vma_free() argument
975 struct xe_vma *vma; xe_vma_create() local
1057 xe_vma_destroy_late(struct xe_vma * vma) xe_vma_destroy_late() argument
1082 struct xe_vma *vma = vma_destroy_work_func() local
1091 struct xe_vma *vma = container_of(cb, struct xe_vma, destroy_cb); vma_destroy_cb() local
1097 xe_vma_destroy(struct xe_vma * vma,struct dma_fence * fence) xe_vma_destroy() argument
1137 xe_vm_lock_vma(struct drm_exec * exec,struct xe_vma * vma) xe_vm_lock_vma() argument
1152 xe_vma_destroy_unlocked(struct xe_vma * vma) xe_vma_destroy_unlocked() argument
1186 xe_vm_insert_vma(struct xe_vm * vm,struct xe_vma * vma) xe_vm_insert_vma() argument
1201 xe_vm_remove_vma(struct xe_vm * vm,struct xe_vma * vma) xe_vm_remove_vma() argument
1337 xelp_pte_encode_vma(u64 pte,struct xe_vma * vma,u16 pat_index,u32 pt_level) xelp_pte_encode_vma() argument
1716 struct xe_vma *vma, *next_vma; xe_vm_close_and_put() local
2021 struct xe_vma *vma = gpuva_to_vma(gpuva); get_mem_attrs() local
2104 vma_matches(struct xe_vma * vma,u64 page_addr) vma_matches() argument
2121 struct xe_vma *vma = NULL; xe_vm_find_vma_by_addr() local
2139 prep_vma_destroy(struct xe_vm * vm,struct xe_vma * vma,bool post_commit) prep_vma_destroy() argument
2155 struct xe_vma *vma; print_op() local
2312 struct xe_vma *vma = gpuva_to_vma(op->base.prefetch.va); vm_bind_ioctl_ops_create() local
2411 struct xe_vma *vma; new_vma() local
2466 xe_vma_max_pte_size(struct xe_vma * vma) xe_vma_max_pte_size() argument
2480 xe_vma_set_pte_size(struct xe_vma * vma,u64 size) xe_vma_set_pte_size() argument
2576 xe_vma_has_default_mem_attrs(struct xe_vma * vma) xe_vma_has_default_mem_attrs() argument
2600 struct xe_vma *vma; vm_bind_ioctl_ops_parse() local
2779 struct xe_vma *vma = gpuva_to_vma(op->base.unmap.va); xe_vma_op_unwind() local
2792 struct xe_vma *vma = gpuva_to_vma(op->base.remap.unmap->va); xe_vma_op_unwind() local
2843 vma_lock_and_validate(struct drm_exec * exec,struct xe_vma * vma,bool res_evict,bool validate) vma_lock_and_validate() argument
2862 check_ufence(struct xe_vma * vma) check_ufence() argument
2880 struct xe_vma *vma = gpuva_to_vma(op->base.prefetch.va); prefetch_ranges() local
2972 struct xe_vma *vma = gpuva_to_vma(op->base.prefetch.va); op_lock_and_prep() local
3205 vma_add_ufence(struct xe_vma * vma,struct xe_user_fence * ufence) vma_add_ufence() argument
3957 xe_vm_invalidate_vma(struct xe_vma * vma) xe_vm_invalidate_vma() argument
4021 struct xe_vma *vma = gpuva_to_vma(gpuva); xe_vm_validate_protected() local
4075 struct xe_vma *vma = gpuva_to_vma(gpuva); xe_vm_snapshot_capture() local
4220 xe_vma_need_vram_for_atomic(struct xe_device * xe,struct xe_vma * vma,bool is_atomic) xe_vma_need_vram_for_atomic() argument
4277 struct xe_vma *vma = NULL; xe_vm_alloc_vma() local
4334 struct xe_vma *vma; xe_vm_alloc_vma() local
[all...]
/linux/tools/testing/vma/
H A Dvma.c18 #define vma_iter_prealloc(vmi, vma) \ argument
72 struct vm_area_struct *vma = vm_area_alloc(mm); in alloc_vma() local
87 static int attach_vma(struct mm_struct *mm, struct vm_area_struct *vma) in attach_vma()
97 static void detach_free_vma(struct vm_area_struct *vma) in detach_free_vma()
110 struct vm_area_struct *vma = alloc_vma(mm, start, end, pgoff, vm_flags); in alloc_and_link_vma() local
133 struct vm_area_struct *vma; in merge_new() local
155 struct vm_area_struct *vma; in merge_existing() local
253 struct vm_area_struct *vma; in cleanup_mm() local
271 static bool vma_write_started(struct vm_area_struct *vma) in vma_write_started()
287 static void __vma_set_dummy_anon_vma(struct vm_area_struct *vma, in __vma_set_dummy_anon_vma()
[all …]
H A Dvma_internal.h385 #define vma_policy(vma) NULL argument
883 vma_is_shared_maywrite(struct vm_area_struct * vma) vma_is_shared_maywrite() argument
902 vma_assert_attached(struct vm_area_struct * vma) vma_assert_attached() argument
907 vma_assert_detached(struct vm_area_struct * vma) vma_assert_detached() argument
913 vma_mark_attached(struct vm_area_struct * vma) vma_mark_attached() argument
920 vma_mark_detached(struct vm_area_struct * vma) vma_mark_detached() argument
937 vma_init(struct vm_area_struct * vma,struct mm_struct * mm) vma_init() argument
990 vma_set_range(struct vm_area_struct * vma,unsigned long start,unsigned long end,pgoff_t pgoff) vma_set_range() argument
1043 vma_is_anonymous(struct vm_area_struct * vma) vma_is_anonymous() argument
1056 struct vm_area_struct *vma; find_vma_prev() local
1076 anon_vma_name(struct vm_area_struct * vma) anon_vma_name() argument
1081 is_mergeable_vm_userfaultfd_ctx(struct vm_area_struct * vma,struct vm_userfaultfd_ctx vm_ctx) is_mergeable_vm_userfaultfd_ctx() argument
1097 vma_pages(struct vm_area_struct * vma) vma_pages() argument
1127 unmap_vmas(struct mmu_gather * tlb,struct ma_state * mas,struct vm_area_struct * vma,unsigned long start_addr,unsigned long end_addr,unsigned long tree_end) unmap_vmas() argument
1133 free_pgtables(struct mmu_gather * tlb,struct ma_state * mas,struct vm_area_struct * vma,unsigned long floor,unsigned long ceiling,bool mm_wr_locked) free_pgtables() argument
1171 vma_start_write(struct vm_area_struct * vma) vma_start_write() argument
1178 vma_start_write_killable(struct vm_area_struct * vma) vma_start_write_killable() argument
1185 vma_adjust_trans_huge(struct vm_area_struct * vma,unsigned long start,unsigned long end,struct vm_area_struct * next) vma_adjust_trans_huge() argument
1209 vma_interval_tree_insert(struct vm_area_struct * vma,struct rb_root_cached * rb) vma_interval_tree_insert() argument
1214 vma_interval_tree_remove(struct vm_area_struct * vma,struct rb_root_cached * rb) vma_interval_tree_remove() argument
1233 uprobe_mmap(struct vm_area_struct * vma) uprobe_mmap() argument
1237 uprobe_munmap(struct vm_area_struct * vma,unsigned long start,unsigned long end) uprobe_munmap() argument
1250 vma_assert_write_locked(struct vm_area_struct * vma) vma_assert_write_locked() argument
1254 unlink_anon_vmas(struct vm_area_struct * vma) unlink_anon_vmas() argument
1268 anon_vma_merge(struct vm_area_struct * vma,struct vm_area_struct * next) anon_vma_merge() argument
1273 userfaultfd_unmap_prep(struct vm_area_struct * vma,unsigned long start,unsigned long end,struct list_head * unmaps) userfaultfd_unmap_prep() argument
1320 khugepaged_enter_vma(struct vm_area_struct * vma,vm_flags_t vm_flags) khugepaged_enter_vma() argument
1330 is_vm_hugetlb_page(struct vm_area_struct * vma) is_vm_hugetlb_page() argument
1335 vma_soft_dirty_enabled(struct vm_area_struct * vma) vma_soft_dirty_enabled() argument
1340 userfaultfd_wp(struct vm_area_struct * vma) userfaultfd_wp() argument
1383 shmem_zero_setup(struct vm_area_struct * vma) shmem_zero_setup() argument
1388 vma_set_anonymous(struct vm_area_struct * vma) vma_set_anonymous() argument
1393 ksm_add_vma(struct vm_area_struct * vma) ksm_add_vma() argument
1397 perf_event_mmap(struct vm_area_struct * vma) perf_event_mmap() argument
1401 vma_is_dax(struct vm_area_struct * vma) vma_is_dax() argument
1414 vma_set_page_prot(struct vm_area_struct * vma) vma_set_page_prot() argument
1436 vma_close(struct vm_area_struct * vma) vma_close() argument
1440 mmap_file(struct file * file,struct vm_area_struct * vma) mmap_file() argument
1445 stack_guard_start_gap(struct vm_area_struct * vma) stack_guard_start_gap() argument
1457 vm_start_gap(struct vm_area_struct * vma) vm_start_gap() argument
1468 vm_end_gap(struct vm_area_struct * vma) vm_end_gap() argument
1486 vma_is_accessible(struct vm_area_struct * vma) vma_is_accessible() argument
1513 __anon_vma_prepare(struct vm_area_struct * vma) __anon_vma_prepare() argument
1526 anon_vma_prepare(struct vm_area_struct * vma) anon_vma_prepare() argument
1606 vm_flags_init(struct vm_area_struct * vma,vm_flags_t flags) vm_flags_init() argument
1618 vm_flags_reset(struct vm_area_struct * vma,vm_flags_t flags) vm_flags_reset() argument
1625 vm_flags_reset_once(struct vm_area_struct * vma,vm_flags_t flags) vm_flags_reset_once() argument
1637 vm_flags_set(struct vm_area_struct * vma,vm_flags_t flags) vm_flags_set() argument
1644 vm_flags_clear(struct vm_area_struct * vma,vm_flags_t flags) vm_flags_clear() argument
1725 vma_lock_init(struct vm_area_struct * vma,bool reset_refcnt) vma_lock_init() argument
1731 vma_numab_state_init(struct vm_area_struct * vma) vma_numab_state_init() argument
1735 vma_numab_state_free(struct vm_area_struct * vma) vma_numab_state_free() argument
1744 free_anon_vma_name(struct vm_area_struct * vma) free_anon_vma_name() argument
1758 mmap_action_complete(struct mmap_action * action,struct vm_area_struct * vma) mmap_action_complete() argument
1764 __compat_vma_mmap(const struct file_operations * f_op,struct file * file,struct vm_area_struct * vma) __compat_vma_mmap() argument
1791 compat_vma_mmap(struct file * file,struct vm_area_struct * vma) compat_vma_mmap() argument
1811 vfs_mmap(struct file * file,struct vm_area_struct * vma) vfs_mmap() argument
1824 fixup_hugetlb_reservations(struct vm_area_struct * vma) fixup_hugetlb_reservations() argument
1828 vma_set_file(struct vm_area_struct * vma,struct file * file) vma_set_file() argument
1851 remap_pfn_range_complete(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t pgprot) remap_pfn_range_complete() argument
[all...]
/linux/arch/powerpc/include/asm/book3s/64/
H A Dtlbflush.h47 static inline void flush_pmd_tlb_range(struct vm_area_struct *vma, in flush_pmd_tlb_range()
55 static inline void flush_pud_tlb_range(struct vm_area_struct *vma, in flush_pud_tlb_range()
63 static inline void flush_hugetlb_tlb_range(struct vm_area_struct *vma, in flush_hugetlb_tlb_range()
71 static inline void flush_tlb_range(struct vm_area_struct *vma, in flush_tlb_range()
91 static inline void local_flush_tlb_page(struct vm_area_struct *vma, in local_flush_tlb_page()
120 static inline void flush_tlb_page(struct vm_area_struct *vma, in flush_tlb_page()
128 #define flush_tlb_page(vma, addr) local_flush_tlb_page(vma, addr) argument
132 static inline void flush_tlb_fix_spurious_fault(struct vm_area_struct *vma, in flush_tlb_fix_spurious_fault()
/linux/arch/parisc/kernel/
H A Dcache.c331 __flush_cache_page(struct vm_area_struct *vma, unsigned long vmaddr, in __flush_cache_page()
393 void flush_icache_pages(struct vm_area_struct *vma, struct page *page, in flush_icache_pages()
476 struct vm_area_struct *vma; in flush_dcache_folio() local
623 static void flush_cache_page_if_present(struct vm_area_struct *vma, in flush_cache_page_if_present()
648 unsigned long vaddr, struct vm_area_struct *vma) in copy_user_highpage()
660 void copy_to_user_page(struct vm_area_struct *vma, struct page *page, in copy_to_user_page()
668 void copy_from_user_page(struct vm_area_struct *vma, struct page *page, in copy_from_user_page()
705 static void flush_cache_pages(struct vm_area_struct *vma, unsigned long start, unsigned long end) in flush_cache_pages()
715 struct vm_area_struct *vma; in mm_total_size() local
729 struct vm_area_struct *vma; in flush_cache_mm() local
[all …]
/linux/drivers/gpu/drm/i915/
H A Di915_gem_evict.c67 static bool grab_vma(struct i915_vma *vma, struct i915_gem_ww_ctx *ww) in grab_vma()
86 static void ungrab_vma(struct i915_vma *vma) in ungrab_vma()
98 struct i915_vma *vma, in mark_free()
112 static bool defer_evict(struct i915_vma *vma) in defer_evict()
157 struct i915_vma *vma, *next; in i915_gem_evict_something() local
337 struct i915_vma *vma, *next; in i915_gem_evict_for_node() local
478 struct i915_vma *vma, *vn; in i915_gem_evict_vm() local
/linux/arch/x86/mm/
H A Dpgtable.c391 int ptep_set_access_flags(struct vm_area_struct *vma, in ptep_set_access_flags()
404 int pmdp_set_access_flags(struct vm_area_struct *vma, in pmdp_set_access_flags()
425 int pudp_set_access_flags(struct vm_area_struct *vma, unsigned long address, in pudp_set_access_flags()
446 int ptep_test_and_clear_young(struct vm_area_struct *vma, in ptep_test_and_clear_young()
459 int pmdp_test_and_clear_young(struct vm_area_struct *vma, in pmdp_test_and_clear_young()
473 int pudp_test_and_clear_young(struct vm_area_struct *vma, in pudp_test_and_clear_young()
486 int ptep_clear_flush_young(struct vm_area_struct *vma, in ptep_clear_flush_young()
506 int pmdp_clear_flush_young(struct vm_area_struct *vma, in pmdp_clear_flush_young()
520 pmd_t pmdp_invalidate_ad(struct vm_area_struct *vma, unsigned long address, in pmdp_invalidate_ad()
535 pud_t pudp_invalidate(struct vm_area_struct *vma, unsigned long address, in pudp_invalidate()
[all …]
/linux/drivers/gpu/drm/nouveau/
H A Dnouveau_vmm.c29 nouveau_vma_unmap(struct nouveau_vma *vma) in nouveau_vma_unmap()
38 nouveau_vma_map(struct nouveau_vma *vma, struct nouveau_mem *mem) in nouveau_vma_map()
51 struct nouveau_vma *vma; in nouveau_vma_find() local
64 struct nouveau_vma *vma = *pvma; in nouveau_vma_del() local
81 struct nouveau_vma *vma; in nouveau_vma_new() local

12345678910>>...25