/linux/include/trace/events/ |
H A D | fs_dax.h | 18 __field(unsigned long, vm_flags) 31 __entry->vm_flags = vmf->vma->vm_flags; 43 __entry->vm_flags & VM_SHARED ? "shared" : "private", 70 __field(unsigned long, vm_flags) 79 __entry->vm_flags = vmf->vma->vm_flags; 89 __entry->vm_flags & VM_SHARED ? "shared" : "private", 111 __field(unsigned long, vm_flags) 122 __entry->vm_flags [all...] |
/linux/mm/ |
H A D | mmap.c | 84 static pgprot_t vm_pgprot_modify(pgprot_t oldprot, unsigned long vm_flags) in vm_pgprot_modify() argument 86 return pgprot_modify(oldprot, vm_get_page_prot(vm_flags)); in vm_pgprot_modify() 92 unsigned long vm_flags = vma->vm_flags; in vma_set_page_prot() local 95 vm_page_prot = vm_pgprot_modify(vma->vm_page_prot, vm_flags); in vma_set_page_prot() 97 vm_flags &= ~VM_SHARED; in vma_set_page_prot() 98 vm_page_prot = vm_pgprot_modify(vm_page_prot, vm_flags); in vma_set_page_prot() 720 struct file *file, unsigned long vm_flags, in is_mergeable_vma() argument 732 if ((vma->vm_flags ^ vm_flags) & ~VM_SOFTDIRTY) in is_mergeable_vma() 772 can_vma_merge_before(struct vm_area_struct *vma, unsigned long vm_flags, in can_vma_merge_before() argument 777 if (is_mergeable_vma(vma, file, vm_flags, vm_userfaultfd_ctx, anon_name, true) && in can_vma_merge_before() [all …]
|
H A D | nommu.c | 131 pgprot_t prot, unsigned long vm_flags, int node, in follow_pfn() 514 if (region->vm_flags & VM_MAPPED_COPY) 839 unsigned long vm_flags; in validate_mmap_request() 841 vm_flags = calc_vm_prot_bits(prot, 0) | calc_vm_flag_bits(flags); in validate_mmap_request() 848 vm_flags |= VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC; in validate_mmap_request() 852 vm_flags |= (capabilities & NOMMU_VMFLAGS); 854 vm_flags |= VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC; 865 vm_flags |= VM_MAYOVERLAY; in determine_vm_flags() 868 vm_flags |= VM_SHARED | VM_MAYSHARE | in determine_vm_flags() 872 return vm_flags; in determine_vm_flags() 152 __vmalloc_node_range(unsigned long size,unsigned long align,unsigned long start,unsigned long end,gfp_t gfp_mask,pgprot_t prot,unsigned long vm_flags,int node,const void * caller) __vmalloc_node_range() argument 860 unsigned long vm_flags; determine_vm_flags() local 1024 do_mmap(struct file * file,unsigned long addr,unsigned long len,unsigned long prot,unsigned long flags,vm_flags_t vm_flags,unsigned long pgoff,unsigned long * populate,struct list_head * uf) do_mmap() argument [all...] |
H A D | mremap.c | 659 unsigned long vm_flags = vma->vm_flags; in move_vma() local 696 MADV_UNMERGEABLE, &vm_flags); in move_vma() 700 if (vm_flags & VM_ACCOUNT) { in move_vma() 710 if (vm_flags & VM_ACCOUNT) in move_vma() 744 if (vm_flags & VM_ACCOUNT && !(flags & MREMAP_DONTUNMAP)) { in move_vma() 762 vm_stat_account(mm, vma->vm_flags, new_len >> PAGE_SHIFT); in move_vma() 765 if (unlikely(vma->vm_flags & VM_PFNMAP)) in move_vma() 787 if (vm_flags & VM_ACCOUNT && !(flags & MREMAP_DONTUNMAP)) in move_vma() 792 if (vm_flags & VM_LOCKED) { in move_vma() 832 if (!old_len && !(vma->vm_flags & (VM_SHARED | VM_MAYSHARE))) { in vma_to_resize() [all …]
|
H A D | mseal.c | 21 return (vma->vm_flags & VM_SEALED); in vma_is_sealed() 51 if (vma->vm_file || vma->vm_flags & VM_SHARED) in is_ro_anon() 58 if (!(vma->vm_flags & VM_WRITE) || in is_ro_anon() 115 vm_flags_t oldflags = vma->vm_flags; in mseal_fixup() 186 newflags = vma->vm_flags | VM_SEALED; in apply_mm_seal()
|
H A D | execmem.c | 20 unsigned long vm_flags = VM_FLUSH_RESET_PERMS; in __execmem_alloc() local 29 vm_flags |= VM_DEFER_KMEMLEAK; in __execmem_alloc() 32 pgprot, vm_flags, NUMA_NO_NODE, in __execmem_alloc() 38 pgprot, vm_flags, NUMA_NO_NODE, in __execmem_alloc()
|
H A D | mlock.c | 345 if (!(vma->vm_flags & VM_LOCKED)) in allow_mlock_munlock() 382 if (vma->vm_flags & VM_LOCKED) in mlock_pte_range() 407 if (vma->vm_flags & VM_LOCKED) in mlock_pte_range() 484 vm_flags_t oldflags = vma->vm_flags; in mlock_fixup() 509 * vm_flags is protected by the mmap_lock held in write mode. in mlock_fixup() 556 newflags = vma->vm_flags & ~VM_LOCKED_MASK; in apply_vma_lock_flags() 597 if (vma->vm_flags & VM_LOCKED) { in count_mm_mlocked_page_nr() 677 vm_flags_t vm_flags = VM_LOCKED; in SYSCALL_DEFINE3() local 683 vm_flags |= VM_LOCKONFAULT; in SYSCALL_DEFINE3() 685 return do_mlock(start, len, vm_flags); in SYSCALL_DEFINE3() [all...] |
H A D | mprotect.c | 48 if (WARN_ON_ONCE(!(vma->vm_flags & VM_WRITE))) in can_change_pte_writable() 63 if (!(vma->vm_flags & VM_SHARED)) { in can_change_pte_writable() 102 if (prot_numa && !(vma->vm_flags & VM_SHARED) && in change_pte_range() 132 if (is_cow_mapping(vma->vm_flags) && in change_pte_range() 584 unsigned long oldflags = vma->vm_flags; in mprotect_fixup() 601 (vma->vm_flags & (VM_PFNMAP|VM_MIXEDMAP)) && in mprotect_fixup() 735 if (!(vma->vm_flags & VM_GROWSDOWN)) in do_mprotect_pkey() 743 if (!(vma->vm_flags & VM_GROWSUP)) in do_mprotect_pkey() 775 if (rier && (vma->vm_flags & VM_MAYEXEC)) in do_mprotect_pkey() 787 newflags |= (vma->vm_flags & ~mask_off_old_flags); in do_mprotect_pkey()
|
H A D | memory.c | 531 pr_alert("addr:%px vm_flags:%08lx anon_vma:%px mapping:%px index:%lx\n", in print_bad_pte() 532 (void *)addr, vma->vm_flags, vma->anon_vma, mapping, index); in print_bad_pte() 594 if (vma->vm_flags & (VM_PFNMAP | VM_MIXEDMAP)) in vm_normal_page() 615 if (unlikely(vma->vm_flags & (VM_PFNMAP|VM_MIXEDMAP))) { in vm_normal_page() 616 if (vma->vm_flags & VM_MIXEDMAP) { in vm_normal_page() 625 if (!is_cow_mapping(vma->vm_flags)) in vm_normal_page() 668 if (unlikely(vma->vm_flags & (VM_PFNMAP|VM_MIXEDMAP))) { in vm_normal_page_pmd() 669 if (vma->vm_flags & VM_MIXEDMAP) { in vm_normal_page_pmd() 678 if (!is_cow_mapping(vma->vm_flags)) in vm_normal_page_pmd() 785 unsigned long vm_flags in copy_nonpresent_pte() 783 unsigned long vm_flags = dst_vma->vm_flags; copy_nonpresent_pte() local 5357 unsigned long vm_flags = vma->vm_flags; __handle_mm_fault() local [all...] |
/linux/arch/powerpc/include/asm/book3s/64/ |
H A D | hash-pkey.h | 8 static inline u64 hash__vmflag_to_pte_pkey_bits(u64 vm_flags) in hash__vmflag_to_pte_pkey_bits() argument 10 return (((vm_flags & VM_PKEY_BIT0) ? H_PTE_PKEY_BIT0 : 0x0UL) | in hash__vmflag_to_pte_pkey_bits() 11 ((vm_flags & VM_PKEY_BIT1) ? H_PTE_PKEY_BIT1 : 0x0UL) | in hash__vmflag_to_pte_pkey_bits() 12 ((vm_flags & VM_PKEY_BIT2) ? H_PTE_PKEY_BIT2 : 0x0UL) | in hash__vmflag_to_pte_pkey_bits() 13 ((vm_flags & VM_PKEY_BIT3) ? H_PTE_PKEY_BIT3 : 0x0UL) | in hash__vmflag_to_pte_pkey_bits() 14 ((vm_flags & VM_PKEY_BIT4) ? H_PTE_PKEY_BIT4 : 0x0UL)); in hash__vmflag_to_pte_pkey_bits()
|
H A D | pkeys.h | 8 static inline u64 vmflag_to_pte_pkey_bits(u64 vm_flags) in vmflag_to_pte_pkey_bits() argument 15 return hash__vmflag_to_pte_pkey_bits(vm_flags); in vmflag_to_pte_pkey_bits()
|
/linux/arch/sparc/include/asm/ |
H A D | mman.h | 57 #define arch_validate_flags(vm_flags) arch_validate_flags(vm_flags) argument 61 static inline bool arch_validate_flags(unsigned long vm_flags) in arch_validate_flags() argument 67 if (vm_flags & VM_SPARC_ADI) { in arch_validate_flags() 72 if (vm_flags & (VM_PFNMAP | VM_MIXEDMAP)) in arch_validate_flags() 83 if (vm_flags & VM_MERGEABLE) in arch_validate_flags()
|
/linux/arch/x86/mm/ |
H A D | pgprot.c | 35 pgprot_t vm_get_page_prot(unsigned long vm_flags) in vm_get_page_prot() argument 37 unsigned long val = pgprot_val(protection_map[vm_flags & in vm_get_page_prot() 48 if (vm_flags & VM_PKEY_BIT0) in vm_get_page_prot() 50 if (vm_flags & VM_PKEY_BIT1) in vm_get_page_prot() 52 if (vm_flags & VM_PKEY_BIT2) in vm_get_page_prot() 54 if (vm_flags & VM_PKEY_BIT3) in vm_get_page_prot()
|
/linux/include/linux/ |
H A D | userfaultfd_k.h | 168 return vma->vm_flags & (VM_UFFD_WP | VM_UFFD_MINOR); in uffd_disable_huge_pmd_share() 180 return vma->vm_flags & (VM_UFFD_WP | VM_UFFD_MINOR); in uffd_disable_fault_around() 185 return vma->vm_flags & VM_UFFD_MISSING; in userfaultfd_missing() 190 return vma->vm_flags & VM_UFFD_WP; in userfaultfd_wp() 195 return vma->vm_flags & VM_UFFD_MINOR; in userfaultfd_minor() 212 return vma->vm_flags & __VM_UFFD_FLAGS; in userfaultfd_armed() 216 unsigned long vm_flags, in vma_can_userfault() argument 219 vm_flags &= __VM_UFFD_FLAGS; in vma_can_userfault() 221 if ((vm_flags & VM_UFFD_MINOR) && in vma_can_userfault() 229 if (wp_async && (vm_flags == VM_UFFD_WP)) in vma_can_userfault() [all …]
|
H A D | mman.h | 192 static inline bool map_deny_write_exec(struct vm_area_struct *vma, unsigned long vm_flags) in map_deny_write_exec() argument 197 if ((vm_flags & VM_EXEC) && (vm_flags & VM_WRITE)) in map_deny_write_exec() 200 if (!(vma->vm_flags & VM_EXEC) && (vm_flags & VM_EXEC)) in map_deny_write_exec()
|
H A D | huge_mm.h | 88 #define thp_vma_allowable_order(vma, vm_flags, tva_flags, order) \ 89 (!!thp_vma_allowable_orders(vma, vm_flags, tva_flags, BIT(order))) 225 unsigned long vm_flags, 232 * @vm_flags: use these vm_flags instead of vma->vm_flags 246 unsigned long vm_flags, in thp_vma_allowable_orders() 254 if (vm_flags & VM_HUGEPAGE) in thp_vma_allowable_orders() 257 ((vm_flags & VM_HUGEPAGE) && hugepage_global_enabled())) in thp_vma_allowable_orders() 265 return __thp_vma_allowable_orders(vma, vm_flags, tva_flag 87 thp_vma_allowable_order(vma,vm_flags,smaps,in_pf,enforce_sysfs,order) global() argument 236 thp_vma_allowable_orders(struct vm_area_struct * vma,unsigned long vm_flags,bool smaps,bool in_pf,bool enforce_sysfs,unsigned long orders) thp_vma_allowable_orders() argument 407 thp_vma_allowable_orders(struct vm_area_struct * vma,unsigned long vm_flags,bool smaps,bool in_pf,bool enforce_sysfs,unsigned long orders) thp_vma_allowable_orders() argument 448 hugepage_madvise(struct vm_area_struct * vma,unsigned long * vm_flags,int advice) hugepage_madvise() argument [all...] |
/linux/arch/arm64/include/asm/ |
H A D | mman.h | 53 static inline bool arch_validate_flags(unsigned long vm_flags) in arch_validate_flags() argument 59 return !(vm_flags & VM_MTE) || (vm_flags & VM_MTE_ALLOWED); in arch_validate_flags() 61 #define arch_validate_flags(vm_flags) arch_validate_flags(vm_flags) argument
|
/linux/tools/testing/selftests/bpf/progs/ |
H A D | bpf_iter_task_vmas.c | 43 perm_str[0] = (vma->vm_flags & VM_READ) ? 'r' : '-'; in proc_maps() 44 perm_str[1] = (vma->vm_flags & VM_WRITE) ? 'w' : '-'; in proc_maps() 45 perm_str[2] = (vma->vm_flags & VM_EXEC) ? 'x' : '-'; in proc_maps() 46 perm_str[3] = (vma->vm_flags & VM_MAYSHARE) ? 's' : 'p'; in proc_maps()
|
/linux/arch/arm64/mm/ |
H A D | mmap.c | 84 pgprot_t vm_get_page_prot(unsigned long vm_flags) in vm_get_page_prot() argument 86 pteval_t prot = pgprot_val(protection_map[vm_flags & in vm_get_page_prot() 89 if (vm_flags & VM_ARM64_BTI) in vm_get_page_prot() 102 if (vm_flags & VM_MTE) in vm_get_page_prot()
|
H A D | fault.c | 509 unsigned long vm_flags; in is_el0_instruction_abort() 529 * vm_flags tells us what bits we must have in vma->vm_flags in do_page_fault() 531 * vma->vm_flags & vm_flags and returns an error if the in do_page_fault() 536 vm_flags = VM_EXEC; in do_page_fault() 540 vm_flags = VM_WRITE; in do_page_fault() 544 vm_flags = VM_READ; in do_page_fault() 546 vm_flags |= VM_WRITE; in do_page_fault() 549 vm_flags | in do_page_fault() 494 __do_page_fault(struct mm_struct * mm,struct vm_area_struct * vma,unsigned long addr,unsigned int mm_flags,unsigned long vm_flags,struct pt_regs * regs) __do_page_fault() argument 528 unsigned long vm_flags; do_page_fault() local [all...] |
/linux/arch/nios2/mm/ |
H A D | cacheflush.c | 90 if (!(vma->vm_flags & VM_MAYSHARE)) in flush_aliases() 138 if (vma == NULL || (vma->vm_flags & VM_EXEC)) in flush_cache_range() 159 if (vma->vm_flags & VM_EXEC) in flush_cache_page() 236 if (vma->vm_flags & VM_EXEC) in update_mmu_cache_range() 268 if (vma->vm_flags & VM_EXEC) in copy_from_user_page() 279 if (vma->vm_flags & VM_EXEC) in copy_to_user_page()
|
/linux/fs/ |
H A D | userfaultfd.c | 110 const bool uffd_wp_changed = (vma->vm_flags ^ flags) & VM_UFFD_WP; in userfaultfd_set_vm_flags() 118 if ((vma->vm_flags & VM_SHARED) && uffd_wp_changed) in userfaultfd_set_vm_flags() 629 vma->vm_flags & ~__VM_UFFD_FLAGS); in userfaultfd_event_wait_completion() 667 userfaultfd_set_vm_flags(vma, vma->vm_flags & ~__VM_UFFD_FLAGS); in dup_userfaultfd() 754 userfaultfd_set_vm_flags(vma, vma->vm_flags & ~__VM_UFFD_FLAGS); 897 !!(vma->vm_flags & __VM_UFFD_FLAGS)); in userfaultfd_release() 906 new_flags = vma->vm_flags & ~__VM_UFFD_FLAGS; in userfaultfd_release() 1300 unsigned long vm_flags, new_flags; in userfaultfd_register() 1319 vm_flags = 0; in userfaultfd_register() 1321 vm_flags | in userfaultfd_register() 1297 unsigned long vm_flags, new_flags; userfaultfd_register() local [all...] |
/linux/arch/hexagon/mm/ |
H A D | vm_fault.c | 70 if (!(vma->vm_flags & VM_EXEC)) in do_page_fault() 74 if (!(vma->vm_flags & VM_READ)) in do_page_fault() 78 if (!(vma->vm_flags & VM_WRITE)) in do_page_fault()
|
/linux/drivers/sbus/char/ |
H A D | flash.c | 44 if ((vma->vm_flags & VM_READ) && in flash_mmap() 45 (vma->vm_flags & VM_WRITE)) { in flash_mmap() 49 if (vma->vm_flags & VM_READ) { in flash_mmap() 52 } else if (vma->vm_flags & VM_WRITE) { in flash_mmap()
|
/linux/arch/arm/mm/ |
H A D | fault.c | 271 unsigned long vm_flags = VM_ACCESS_FLAGS; in do_page_fault() 293 vm_flags = VM_WRITE; in do_page_fault() 297 vm_flags = VM_EXEC; in do_page_fault() 321 if (!(vma->vm_flags & vm_flags)) { in do_page_fault() 360 if (!(vma->vm_flags & vm_flags)) { in do_page_fault() 274 unsigned long vm_flags = VM_ACCESS_FLAGS; do_page_fault() local
|