Home
last modified time | relevance | path

Searched refs:vm_flags (Results 1 – 25 of 271) sorted by relevance

1234567891011

/linux/include/trace/events/
H A Dfs_dax.h18 __field(unsigned long, vm_flags)
31 __entry->vm_flags = vmf->vma->vm_flags;
43 __entry->vm_flags & VM_SHARED ? "shared" : "private",
70 __field(unsigned long, vm_flags)
79 __entry->vm_flags = vmf->vma->vm_flags;
89 __entry->vm_flags & VM_SHARED ? "shared" : "private",
111 __field(unsigned long, vm_flags)
122 __entry->vm_flags
[all...]
/linux/mm/
H A Dmmap.c84 static pgprot_t vm_pgprot_modify(pgprot_t oldprot, unsigned long vm_flags) in vm_pgprot_modify() argument
86 return pgprot_modify(oldprot, vm_get_page_prot(vm_flags)); in vm_pgprot_modify()
92 unsigned long vm_flags = vma->vm_flags; in vma_set_page_prot() local
95 vm_page_prot = vm_pgprot_modify(vma->vm_page_prot, vm_flags); in vma_set_page_prot()
97 vm_flags &= ~VM_SHARED; in vma_set_page_prot()
98 vm_page_prot = vm_pgprot_modify(vm_page_prot, vm_flags); in vma_set_page_prot()
720 struct file *file, unsigned long vm_flags, in is_mergeable_vma() argument
732 if ((vma->vm_flags ^ vm_flags) & ~VM_SOFTDIRTY) in is_mergeable_vma()
772 can_vma_merge_before(struct vm_area_struct *vma, unsigned long vm_flags, in can_vma_merge_before() argument
777 if (is_mergeable_vma(vma, file, vm_flags, vm_userfaultfd_ctx, anon_name, true) && in can_vma_merge_before()
[all …]
H A Dnommu.c131 pgprot_t prot, unsigned long vm_flags, int node, in follow_pfn()
514 if (region->vm_flags & VM_MAPPED_COPY)
839 unsigned long vm_flags; in validate_mmap_request()
841 vm_flags = calc_vm_prot_bits(prot, 0) | calc_vm_flag_bits(flags); in validate_mmap_request()
848 vm_flags |= VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC; in validate_mmap_request()
852 vm_flags |= (capabilities & NOMMU_VMFLAGS);
854 vm_flags |= VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC;
865 vm_flags |= VM_MAYOVERLAY; in determine_vm_flags()
868 vm_flags |= VM_SHARED | VM_MAYSHARE | in determine_vm_flags()
872 return vm_flags; in determine_vm_flags()
152 __vmalloc_node_range(unsigned long size,unsigned long align,unsigned long start,unsigned long end,gfp_t gfp_mask,pgprot_t prot,unsigned long vm_flags,int node,const void * caller) __vmalloc_node_range() argument
860 unsigned long vm_flags; determine_vm_flags() local
1024 do_mmap(struct file * file,unsigned long addr,unsigned long len,unsigned long prot,unsigned long flags,vm_flags_t vm_flags,unsigned long pgoff,unsigned long * populate,struct list_head * uf) do_mmap() argument
[all...]
H A Dmremap.c659 unsigned long vm_flags = vma->vm_flags; in move_vma() local
696 MADV_UNMERGEABLE, &vm_flags); in move_vma()
700 if (vm_flags & VM_ACCOUNT) { in move_vma()
710 if (vm_flags & VM_ACCOUNT) in move_vma()
744 if (vm_flags & VM_ACCOUNT && !(flags & MREMAP_DONTUNMAP)) { in move_vma()
762 vm_stat_account(mm, vma->vm_flags, new_len >> PAGE_SHIFT); in move_vma()
765 if (unlikely(vma->vm_flags & VM_PFNMAP)) in move_vma()
787 if (vm_flags & VM_ACCOUNT && !(flags & MREMAP_DONTUNMAP)) in move_vma()
792 if (vm_flags & VM_LOCKED) { in move_vma()
832 if (!old_len && !(vma->vm_flags & (VM_SHARED | VM_MAYSHARE))) { in vma_to_resize()
[all …]
H A Dmseal.c21 return (vma->vm_flags & VM_SEALED); in vma_is_sealed()
51 if (vma->vm_file || vma->vm_flags & VM_SHARED) in is_ro_anon()
58 if (!(vma->vm_flags & VM_WRITE) || in is_ro_anon()
115 vm_flags_t oldflags = vma->vm_flags; in mseal_fixup()
186 newflags = vma->vm_flags | VM_SEALED; in apply_mm_seal()
H A Dexecmem.c20 unsigned long vm_flags = VM_FLUSH_RESET_PERMS; in __execmem_alloc() local
29 vm_flags |= VM_DEFER_KMEMLEAK; in __execmem_alloc()
32 pgprot, vm_flags, NUMA_NO_NODE, in __execmem_alloc()
38 pgprot, vm_flags, NUMA_NO_NODE, in __execmem_alloc()
H A Dmlock.c345 if (!(vma->vm_flags & VM_LOCKED)) in allow_mlock_munlock()
382 if (vma->vm_flags & VM_LOCKED) in mlock_pte_range()
407 if (vma->vm_flags & VM_LOCKED) in mlock_pte_range()
484 vm_flags_t oldflags = vma->vm_flags; in mlock_fixup()
509 * vm_flags is protected by the mmap_lock held in write mode. in mlock_fixup()
556 newflags = vma->vm_flags & ~VM_LOCKED_MASK; in apply_vma_lock_flags()
597 if (vma->vm_flags & VM_LOCKED) { in count_mm_mlocked_page_nr()
677 vm_flags_t vm_flags = VM_LOCKED; in SYSCALL_DEFINE3() local
683 vm_flags |= VM_LOCKONFAULT; in SYSCALL_DEFINE3()
685 return do_mlock(start, len, vm_flags); in SYSCALL_DEFINE3()
[all...]
H A Dmprotect.c48 if (WARN_ON_ONCE(!(vma->vm_flags & VM_WRITE))) in can_change_pte_writable()
63 if (!(vma->vm_flags & VM_SHARED)) { in can_change_pte_writable()
102 if (prot_numa && !(vma->vm_flags & VM_SHARED) && in change_pte_range()
132 if (is_cow_mapping(vma->vm_flags) && in change_pte_range()
584 unsigned long oldflags = vma->vm_flags; in mprotect_fixup()
601 (vma->vm_flags & (VM_PFNMAP|VM_MIXEDMAP)) && in mprotect_fixup()
735 if (!(vma->vm_flags & VM_GROWSDOWN)) in do_mprotect_pkey()
743 if (!(vma->vm_flags & VM_GROWSUP)) in do_mprotect_pkey()
775 if (rier && (vma->vm_flags & VM_MAYEXEC)) in do_mprotect_pkey()
787 newflags |= (vma->vm_flags & ~mask_off_old_flags); in do_mprotect_pkey()
H A Dmemory.c531 pr_alert("addr:%px vm_flags:%08lx anon_vma:%px mapping:%px index:%lx\n", in print_bad_pte()
532 (void *)addr, vma->vm_flags, vma->anon_vma, mapping, index); in print_bad_pte()
594 if (vma->vm_flags & (VM_PFNMAP | VM_MIXEDMAP)) in vm_normal_page()
615 if (unlikely(vma->vm_flags & (VM_PFNMAP|VM_MIXEDMAP))) { in vm_normal_page()
616 if (vma->vm_flags & VM_MIXEDMAP) { in vm_normal_page()
625 if (!is_cow_mapping(vma->vm_flags)) in vm_normal_page()
668 if (unlikely(vma->vm_flags & (VM_PFNMAP|VM_MIXEDMAP))) { in vm_normal_page_pmd()
669 if (vma->vm_flags & VM_MIXEDMAP) { in vm_normal_page_pmd()
678 if (!is_cow_mapping(vma->vm_flags)) in vm_normal_page_pmd()
785 unsigned long vm_flags in copy_nonpresent_pte()
783 unsigned long vm_flags = dst_vma->vm_flags; copy_nonpresent_pte() local
5357 unsigned long vm_flags = vma->vm_flags; __handle_mm_fault() local
[all...]
/linux/arch/powerpc/include/asm/book3s/64/
H A Dhash-pkey.h8 static inline u64 hash__vmflag_to_pte_pkey_bits(u64 vm_flags) in hash__vmflag_to_pte_pkey_bits() argument
10 return (((vm_flags & VM_PKEY_BIT0) ? H_PTE_PKEY_BIT0 : 0x0UL) | in hash__vmflag_to_pte_pkey_bits()
11 ((vm_flags & VM_PKEY_BIT1) ? H_PTE_PKEY_BIT1 : 0x0UL) | in hash__vmflag_to_pte_pkey_bits()
12 ((vm_flags & VM_PKEY_BIT2) ? H_PTE_PKEY_BIT2 : 0x0UL) | in hash__vmflag_to_pte_pkey_bits()
13 ((vm_flags & VM_PKEY_BIT3) ? H_PTE_PKEY_BIT3 : 0x0UL) | in hash__vmflag_to_pte_pkey_bits()
14 ((vm_flags & VM_PKEY_BIT4) ? H_PTE_PKEY_BIT4 : 0x0UL)); in hash__vmflag_to_pte_pkey_bits()
H A Dpkeys.h8 static inline u64 vmflag_to_pte_pkey_bits(u64 vm_flags) in vmflag_to_pte_pkey_bits() argument
15 return hash__vmflag_to_pte_pkey_bits(vm_flags); in vmflag_to_pte_pkey_bits()
/linux/arch/sparc/include/asm/
H A Dmman.h57 #define arch_validate_flags(vm_flags) arch_validate_flags(vm_flags) argument
61 static inline bool arch_validate_flags(unsigned long vm_flags) in arch_validate_flags() argument
67 if (vm_flags & VM_SPARC_ADI) { in arch_validate_flags()
72 if (vm_flags & (VM_PFNMAP | VM_MIXEDMAP)) in arch_validate_flags()
83 if (vm_flags & VM_MERGEABLE) in arch_validate_flags()
/linux/arch/x86/mm/
H A Dpgprot.c35 pgprot_t vm_get_page_prot(unsigned long vm_flags) in vm_get_page_prot() argument
37 unsigned long val = pgprot_val(protection_map[vm_flags & in vm_get_page_prot()
48 if (vm_flags & VM_PKEY_BIT0) in vm_get_page_prot()
50 if (vm_flags & VM_PKEY_BIT1) in vm_get_page_prot()
52 if (vm_flags & VM_PKEY_BIT2) in vm_get_page_prot()
54 if (vm_flags & VM_PKEY_BIT3) in vm_get_page_prot()
/linux/include/linux/
H A Duserfaultfd_k.h168 return vma->vm_flags & (VM_UFFD_WP | VM_UFFD_MINOR); in uffd_disable_huge_pmd_share()
180 return vma->vm_flags & (VM_UFFD_WP | VM_UFFD_MINOR); in uffd_disable_fault_around()
185 return vma->vm_flags & VM_UFFD_MISSING; in userfaultfd_missing()
190 return vma->vm_flags & VM_UFFD_WP; in userfaultfd_wp()
195 return vma->vm_flags & VM_UFFD_MINOR; in userfaultfd_minor()
212 return vma->vm_flags & __VM_UFFD_FLAGS; in userfaultfd_armed()
216 unsigned long vm_flags, in vma_can_userfault() argument
219 vm_flags &= __VM_UFFD_FLAGS; in vma_can_userfault()
221 if ((vm_flags & VM_UFFD_MINOR) && in vma_can_userfault()
229 if (wp_async && (vm_flags == VM_UFFD_WP)) in vma_can_userfault()
[all …]
H A Dmman.h192 static inline bool map_deny_write_exec(struct vm_area_struct *vma, unsigned long vm_flags) in map_deny_write_exec() argument
197 if ((vm_flags & VM_EXEC) && (vm_flags & VM_WRITE)) in map_deny_write_exec()
200 if (!(vma->vm_flags & VM_EXEC) && (vm_flags & VM_EXEC)) in map_deny_write_exec()
H A Dhuge_mm.h88 #define thp_vma_allowable_order(vma, vm_flags, tva_flags, order) \
89 (!!thp_vma_allowable_orders(vma, vm_flags, tva_flags, BIT(order)))
225 unsigned long vm_flags,
232 * @vm_flags: use these vm_flags instead of vma->vm_flags
246 unsigned long vm_flags, in thp_vma_allowable_orders()
254 if (vm_flags & VM_HUGEPAGE) in thp_vma_allowable_orders()
257 ((vm_flags & VM_HUGEPAGE) && hugepage_global_enabled())) in thp_vma_allowable_orders()
265 return __thp_vma_allowable_orders(vma, vm_flags, tva_flag
87 thp_vma_allowable_order(vma,vm_flags,smaps,in_pf,enforce_sysfs,order) global() argument
236 thp_vma_allowable_orders(struct vm_area_struct * vma,unsigned long vm_flags,bool smaps,bool in_pf,bool enforce_sysfs,unsigned long orders) thp_vma_allowable_orders() argument
407 thp_vma_allowable_orders(struct vm_area_struct * vma,unsigned long vm_flags,bool smaps,bool in_pf,bool enforce_sysfs,unsigned long orders) thp_vma_allowable_orders() argument
448 hugepage_madvise(struct vm_area_struct * vma,unsigned long * vm_flags,int advice) hugepage_madvise() argument
[all...]
/linux/arch/arm64/include/asm/
H A Dmman.h53 static inline bool arch_validate_flags(unsigned long vm_flags) in arch_validate_flags() argument
59 return !(vm_flags & VM_MTE) || (vm_flags & VM_MTE_ALLOWED); in arch_validate_flags()
61 #define arch_validate_flags(vm_flags) arch_validate_flags(vm_flags) argument
/linux/tools/testing/selftests/bpf/progs/
H A Dbpf_iter_task_vmas.c43 perm_str[0] = (vma->vm_flags & VM_READ) ? 'r' : '-'; in proc_maps()
44 perm_str[1] = (vma->vm_flags & VM_WRITE) ? 'w' : '-'; in proc_maps()
45 perm_str[2] = (vma->vm_flags & VM_EXEC) ? 'x' : '-'; in proc_maps()
46 perm_str[3] = (vma->vm_flags & VM_MAYSHARE) ? 's' : 'p'; in proc_maps()
/linux/arch/arm64/mm/
H A Dmmap.c84 pgprot_t vm_get_page_prot(unsigned long vm_flags) in vm_get_page_prot() argument
86 pteval_t prot = pgprot_val(protection_map[vm_flags & in vm_get_page_prot()
89 if (vm_flags & VM_ARM64_BTI) in vm_get_page_prot()
102 if (vm_flags & VM_MTE) in vm_get_page_prot()
H A Dfault.c509 unsigned long vm_flags; in is_el0_instruction_abort()
529 * vm_flags tells us what bits we must have in vma->vm_flags in do_page_fault()
531 * vma->vm_flags & vm_flags and returns an error if the in do_page_fault()
536 vm_flags = VM_EXEC; in do_page_fault()
540 vm_flags = VM_WRITE; in do_page_fault()
544 vm_flags = VM_READ; in do_page_fault()
546 vm_flags |= VM_WRITE; in do_page_fault()
549 vm_flags | in do_page_fault()
494 __do_page_fault(struct mm_struct * mm,struct vm_area_struct * vma,unsigned long addr,unsigned int mm_flags,unsigned long vm_flags,struct pt_regs * regs) __do_page_fault() argument
528 unsigned long vm_flags; do_page_fault() local
[all...]
/linux/arch/nios2/mm/
H A Dcacheflush.c90 if (!(vma->vm_flags & VM_MAYSHARE)) in flush_aliases()
138 if (vma == NULL || (vma->vm_flags & VM_EXEC)) in flush_cache_range()
159 if (vma->vm_flags & VM_EXEC) in flush_cache_page()
236 if (vma->vm_flags & VM_EXEC) in update_mmu_cache_range()
268 if (vma->vm_flags & VM_EXEC) in copy_from_user_page()
279 if (vma->vm_flags & VM_EXEC) in copy_to_user_page()
/linux/fs/
H A Duserfaultfd.c110 const bool uffd_wp_changed = (vma->vm_flags ^ flags) & VM_UFFD_WP; in userfaultfd_set_vm_flags()
118 if ((vma->vm_flags & VM_SHARED) && uffd_wp_changed) in userfaultfd_set_vm_flags()
629 vma->vm_flags & ~__VM_UFFD_FLAGS); in userfaultfd_event_wait_completion()
667 userfaultfd_set_vm_flags(vma, vma->vm_flags & ~__VM_UFFD_FLAGS); in dup_userfaultfd()
754 userfaultfd_set_vm_flags(vma, vma->vm_flags & ~__VM_UFFD_FLAGS);
897 !!(vma->vm_flags & __VM_UFFD_FLAGS)); in userfaultfd_release()
906 new_flags = vma->vm_flags & ~__VM_UFFD_FLAGS; in userfaultfd_release()
1300 unsigned long vm_flags, new_flags; in userfaultfd_register()
1319 vm_flags = 0; in userfaultfd_register()
1321 vm_flags | in userfaultfd_register()
1297 unsigned long vm_flags, new_flags; userfaultfd_register() local
[all...]
/linux/arch/hexagon/mm/
H A Dvm_fault.c70 if (!(vma->vm_flags & VM_EXEC)) in do_page_fault()
74 if (!(vma->vm_flags & VM_READ)) in do_page_fault()
78 if (!(vma->vm_flags & VM_WRITE)) in do_page_fault()
/linux/drivers/sbus/char/
H A Dflash.c44 if ((vma->vm_flags & VM_READ) && in flash_mmap()
45 (vma->vm_flags & VM_WRITE)) { in flash_mmap()
49 if (vma->vm_flags & VM_READ) { in flash_mmap()
52 } else if (vma->vm_flags & VM_WRITE) { in flash_mmap()
/linux/arch/arm/mm/
H A Dfault.c271 unsigned long vm_flags = VM_ACCESS_FLAGS; in do_page_fault()
293 vm_flags = VM_WRITE; in do_page_fault()
297 vm_flags = VM_EXEC; in do_page_fault()
321 if (!(vma->vm_flags & vm_flags)) { in do_page_fault()
360 if (!(vma->vm_flags & vm_flags)) { in do_page_fault()
274 unsigned long vm_flags = VM_ACCESS_FLAGS; do_page_fault() local

1234567891011