/linux/arch/powerpc/include/asm/book3s/64/ |
H A D | tlbflush.h | 85 static inline void local_flush_tlb_mm(struct mm_struct *mm) in local_flush_tlb_mm() argument 88 radix__local_flush_tlb_mm(mm); in local_flush_tlb_mm() 98 static inline void local_flush_tlb_page_psize(struct mm_struct *mm, in local_flush_tlb_page_psize() argument 102 radix__local_flush_tlb_page_psize(mm, vmaddr, psize); in local_flush_tlb_page_psize() 114 static inline void flush_tlb_mm(struct mm_struct *mm) in flush_tlb_mm() argument 117 radix__flush_tlb_mm(mm); in flush_tlb_mm() 127 #define flush_tlb_mm(mm) local_flush_tlb_mm(mm) argument
|
H A D | radix.h | 172 static inline unsigned long radix__pte_update(struct mm_struct *mm, in radix__pte_update() argument 182 assert_pte_locked(mm, addr); in radix__pte_update() 187 static inline pte_t radix__ptep_get_and_clear_full(struct mm_struct *mm, in radix__ptep_get_and_clear_full() argument 197 old_pte = radix__pte_update(mm, addr, ptep, ~0ul, 0, 0); in radix__ptep_get_and_clear_full() 212 static inline void radix__set_pte_at(struct mm_struct *mm, unsigned long addr, in radix__set_pte_at() argument 285 extern unsigned long radix__pmd_hugepage_update(struct mm_struct *mm, unsigned long addr, 288 extern unsigned long radix__pud_hugepage_update(struct mm_struct *mm, unsigned long addr, 293 extern void radix__pgtable_trans_huge_deposit(struct mm_struct *mm, pmd_t *pmdp, 295 extern pgtable_t radix__pgtable_trans_huge_withdraw(struct mm_struct *mm, pmd_t *pmdp); 296 extern pmd_t radix__pmdp_huge_get_and_clear(struct mm_struct *mm, [all …]
|
/linux/mm/ |
H A D | memory.c | 180 void mm_trace_rss_stat(struct mm_struct *mm, int member) in mm_trace_rss_stat() argument 182 trace_rss_stat(mm, member); in mm_trace_rss_stat() 195 mm_dec_nr_ptes(tlb->mm); in free_pte_range() 229 mm_dec_nr_pmds(tlb->mm); in free_pmd_range() 263 mm_dec_nr_puds(tlb->mm); in free_pud_range() 355 pgd = pgd_offset(tlb->mm, addr); in free_pgd_range() 420 void pmd_install(struct mm_struct *mm, pmd_t *pmd, pgtable_t *pte) in pmd_install() argument 422 spinlock_t *ptl = pmd_lock(mm, pmd); in pmd_install() 425 mm_inc_nr_ptes(mm); in pmd_install() 440 pmd_populate(mm, pmd, *pte); in pmd_install() [all …]
|
H A D | pgalloc-track.h | 6 static inline p4d_t *p4d_alloc_track(struct mm_struct *mm, pgd_t *pgd, in p4d_alloc_track() argument 11 if (__p4d_alloc(mm, pgd, address)) in p4d_alloc_track() 19 static inline pud_t *pud_alloc_track(struct mm_struct *mm, p4d_t *p4d, in pud_alloc_track() argument 24 if (__pud_alloc(mm, p4d, address)) in pud_alloc_track() 32 static inline pmd_t *pmd_alloc_track(struct mm_struct *mm, pud_t *pud, in pmd_alloc_track() argument 37 if (__pmd_alloc(mm, pud, address)) in pmd_alloc_track()
|
/linux/arch/arm/kernel/ |
H A D | smp_tlb.c | 33 struct mm_struct *mm = (struct mm_struct *)arg; in ipi_flush_tlb_mm() local 35 local_flush_tlb_mm(mm); in ipi_flush_tlb_mm() 164 static void broadcast_tlb_mm_a15_erratum(struct mm_struct *mm) in broadcast_tlb_mm_a15_erratum() argument 173 a15_erratum_get_cpumask(this_cpu, mm, &mask); in broadcast_tlb_mm_a15_erratum() 187 void flush_tlb_mm(struct mm_struct *mm) in flush_tlb_mm() argument 190 on_each_cpu_mask(mm_cpumask(mm), ipi_flush_tlb_mm, mm, 1); in flush_tlb_mm() 192 __flush_tlb_mm(mm); in flush_tlb_mm() 193 broadcast_tlb_mm_a15_erratum(mm); in flush_tlb_mm()
|
H A D | process.c | 322 struct vm_area_struct *get_gate_vma(struct mm_struct *mm) in get_gate_vma() argument 327 int in_gate_area(struct mm_struct *mm, unsigned long addr) in in_gate_area() argument 349 static unsigned long sigpage_addr(const struct mm_struct *mm, in sigpage_addr() argument 358 first = PAGE_ALIGN(mm->start_stack); in sigpage_addr() 385 current->mm->context.sigpage = new_vma->vm_start; in sigpage_mremap() 397 struct mm_struct *mm = current->mm; in arch_setup_additional_pages() local 412 if (mmap_write_lock_killable(mm)) in arch_setup_additional_pages() 414 hint = sigpage_addr(mm, npages); in arch_setup_additional_pages() 421 vma = _install_special_mapping(mm, addr, PAGE_SIZE, in arch_setup_additional_pages() 430 mm->context.sigpage = addr; in arch_setup_additional_pages() [all …]
|
/linux/drivers/infiniband/hw/mlx5/ |
H A D | doorbell.c | 45 struct mm_struct *mm; member 57 if ((current->mm == page->mm) && in mlx5_ib_db_map_user() 76 mmgrab(current->mm); in mlx5_ib_db_map_user() 77 page->mm = current->mm; in mlx5_ib_db_map_user() 99 mmdrop(db->u.user_page->mm); in mlx5_ib_db_unmap_user()
|
/linux/lib/kunit/ |
H A D | user_alloc.c | 28 struct mm_struct *mm; in kunit_attach_mm() local 30 if (current->mm) in kunit_attach_mm() 37 mm = mm_alloc(); in kunit_attach_mm() 38 if (!mm) in kunit_attach_mm() 42 mm->task_size = TASK_SIZE; in kunit_attach_mm() 45 arch_pick_mmap_layout(mm, ¤t->signal->rlim[RLIMIT_STACK]); in kunit_attach_mm() 48 kthread_use_mm(mm); in kunit_attach_mm()
|
/linux/arch/csky/mm/ |
H A D | fault.c | 100 static inline void bad_area_nosemaphore(struct pt_regs *regs, struct mm_struct *mm, int code, unsig… in bad_area_nosemaphore() argument 191 struct mm_struct *mm; in do_page_fault() local 198 mm = tsk->mm; in do_page_fault() 227 if (unlikely(faulthandler_disabled() || !mm)) { in do_page_fault() 240 vma = lock_mm_and_find_vma(mm, addr, regs); in do_page_fault() 242 bad_area_nosemaphore(regs, mm, code, addr); in do_page_fault() 253 mmap_read_unlock(mm); in do_page_fault() 254 bad_area_nosemaphore(regs, mm, code, addr); in do_page_fault() 291 mmap_read_unlock(mm); in do_page_fault()
|
/linux/arch/sparc/include/asm/ |
H A D | tlb_64.h | 16 void smp_flush_tlb_mm(struct mm_struct *mm); 17 #define do_flush_tlb_mm(mm) smp_flush_tlb_mm(mm) argument 19 #define do_flush_tlb_mm(mm) __flush_tlb_mm(CTX_HWBITS(mm->context), SECONDARY_CONTEXT) argument
|
H A D | mmu_context_32.h | 13 int init_new_context(struct task_struct *tsk, struct mm_struct *mm); 21 void destroy_context(struct mm_struct *mm); 24 void switch_mm(struct mm_struct *old_mm, struct mm_struct *mm, 28 #define activate_mm(active_mm, mm) switch_mm((active_mm), (mm), NULL) argument
|
/linux/arch/s390/include/asm/ |
H A D | hugetlb.h | 18 void set_huge_pte_at(struct mm_struct *mm, unsigned long addr, 20 void __set_huge_pte_at(struct mm_struct *mm, unsigned long addr, 22 pte_t huge_ptep_get(struct mm_struct *mm, unsigned long addr, pte_t *ptep); 23 pte_t huge_ptep_get_and_clear(struct mm_struct *mm, 48 static inline void huge_pte_clear(struct mm_struct *mm, unsigned long addr, in huge_pte_clear() argument 75 static inline void huge_ptep_set_wrprotect(struct mm_struct *mm, in huge_ptep_set_wrprotect() argument 78 pte_t pte = huge_ptep_get_and_clear(mm, addr, ptep); in huge_ptep_set_wrprotect() 79 __set_huge_pte_at(mm, addr, ptep, pte_wrprotect(pte)); in huge_ptep_set_wrprotect()
|
/linux/fs/ |
H A D | binfmt_flat.c | 115 sp = (unsigned long __user *)current->mm->start_stack; in create_flat_tables() 123 current->mm->start_stack = (unsigned long)sp & -FLAT_STACK_ALIGN; in create_flat_tables() 124 sp = (unsigned long __user *)current->mm->start_stack; in create_flat_tables() 136 current->mm->arg_start = (unsigned long)p; in create_flat_tables() 147 current->mm->arg_end = (unsigned long)p; in create_flat_tables() 149 current->mm->env_start = (unsigned long) p; in create_flat_tables() 160 current->mm->env_end = (unsigned long)p; in create_flat_tables() 350 ptr = (unsigned long __user *)(current->mm->start_code + r.reloc.offset); in old_reloc() 352 ptr = (unsigned long __user *)(current->mm->start_data + r.reloc.offset); in old_reloc() 362 val += current->mm->start_code; in old_reloc() [all …]
|
/linux/arch/powerpc/include/asm/ |
H A D | hugetlb.h | 20 int slice_is_hugepage_only_range(struct mm_struct *mm, unsigned long addr, 23 static inline int is_hugepage_only_range(struct mm_struct *mm, in is_hugepage_only_range() argument 28 return slice_is_hugepage_only_range(mm, addr, len); in is_hugepage_only_range() 34 void set_huge_pte_at(struct mm_struct *mm, unsigned long addr, pte_t *ptep, 38 static inline pte_t huge_ptep_get_and_clear(struct mm_struct *mm, in huge_ptep_get_and_clear() argument 41 return __pte(pte_update(mm, addr, ptep, ~0UL, 0, 1)); in huge_ptep_get_and_clear()
|
/linux/arch/powerpc/include/asm/book3s/32/ |
H A D | pgalloc.h | 8 static inline pgd_t *pgd_alloc(struct mm_struct *mm) in pgd_alloc() argument 11 pgtable_gfp_flags(mm, GFP_KERNEL)); in pgd_alloc() 14 static inline void pgd_free(struct mm_struct *mm, pgd_t *pgd) in pgd_free() argument 24 #define pmd_free(mm, x) do { } while (0) argument 28 static inline void pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmdp, in pmd_populate_kernel() argument 34 static inline void pmd_populate(struct mm_struct *mm, pmd_t *pmdp, in pmd_populate() argument
|
/linux/drivers/misc/ocxl/ |
H A D | link.c | 42 struct mm_struct *mm; member 152 rc = copro_handle_mm_fault(fault->pe_data.mm, fault->dar, fault->dsisr, in xsl_fault_handler_bh() 179 hash_page_mm(fault->pe_data.mm, fault->dar, access, 0x300, in xsl_fault_handler_bh() 185 mmput(fault->pe_data.mm); in xsl_fault_handler_bh() 237 if (!pe_data->mm) { in xsl_fault_handler() 247 WARN_ON(pe_data->mm->context.id != pid); in xsl_fault_handler() 249 if (mmget_not_zero(pe_data->mm)) { in xsl_fault_handler() 495 struct mm_struct *mm, in arch_invalidate_secondary_tlbs() argument 502 pid = mm->context.id; in arch_invalidate_secondary_tlbs() 539 u64 amr, u16 bdf, struct mm_struct *mm, in ocxl_link_add_pe() argument [all …]
|
/linux/drivers/gpu/drm/ttm/ |
H A D | ttm_range_manager.c | 48 struct drm_mm mm; member 65 struct drm_mm *mm = &rman->mm; in ttm_range_man_alloc() local 85 ret = drm_mm_insert_node_in_range(mm, &node->mm_nodes[0], in ttm_range_man_alloc() 153 drm_mm_print(&rman->mm, printer); in ttm_range_man_debug() 196 drm_mm_init(&rman->mm, 0, p_size); in ttm_range_man_init_nocheck() 219 struct drm_mm *mm = &rman->mm; in ttm_range_man_fini_nocheck() local 232 drm_mm_takedown(mm); in ttm_range_man_fini_nocheck()
|
/linux/lib/ |
H A D | test_hmm.c | 10 #include <linux/mm.h> 27 #include <linux/sched/mm.h> 175 ret = mmu_interval_notifier_insert(&dmirror->notifier, current->mm, in dmirror_fops_open() 290 struct mm_struct *mm = dmirror->notifier.mm; in dmirror_range_fault() local 302 mmap_read_lock(mm); in dmirror_range_fault() 304 mmap_read_unlock(mm); in dmirror_range_fault() 330 struct mm_struct *mm = dmirror->notifier.mm; in dmirror_fault() local 343 /* Since the mm i in dmirror_fault() 782 struct mm_struct *mm = dmirror->notifier.mm; dmirror_exclusive() local 905 struct mm_struct *mm = dmirror->notifier.mm; dmirror_migrate_to_system() local 965 struct mm_struct *mm = dmirror->notifier.mm; dmirror_migrate_to_device() local 1111 struct mm_struct *mm = dmirror->notifier.mm; dmirror_range_snapshot() local 1167 struct mm_struct *mm = dmirror->notifier.mm; dmirror_snapshot() local [all...] |
/linux/arch/microblaze/include/asm/ |
H A D | pgalloc.h | 29 #define pgd_alloc(mm) get_pgd() argument 31 extern pte_t *pte_alloc_one_kernel(struct mm_struct *mm); 33 #define __pte_free_tlb(tlb, pte, addr) pte_free((tlb)->mm, (pte)) 35 #define pmd_populate(mm, pmd, pte) \ argument 38 #define pmd_populate_kernel(mm, pmd, pte) \ argument
|
/linux/arch/m68k/mm/ |
H A D | mcfmmu.c | 81 current->mm = NULL; in paging_init() 89 struct mm_struct *mm; in cf_tlb_miss() local 103 mm = (!user_mode(regs) && KMAPAREA(mmuar)) ? &init_mm : current->mm; in cf_tlb_miss() 104 if (!mm) in cf_tlb_miss() 107 pgd = pgd_offset(mm, mmuar); in cf_tlb_miss() 135 asid = mm->context & 0xff; in cf_tlb_miss() 215 struct mm_struct *mm; in steal_context() local 222 mm = context_mm[next_mmu_context]; in steal_context() 223 flush_tlb_mm(mm); in steal_context() 224 destroy_context(mm); in steal_context()
|
/linux/arch/csky/include/asm/ |
H A D | mmu_context.h | 17 #define cpu_asid(mm) (atomic64_read(&mm->context.asid) & ASID_MASK) argument 19 #define init_new_context(tsk,mm) ({ atomic64_set(&(mm)->context.asid, 0); 0; }) argument 21 void check_and_switch_context(struct mm_struct *mm, unsigned int cpu);
|
/linux/arch/hexagon/mm/ |
H A D | vm_fault.c | 40 struct mm_struct *mm = current->mm; in do_page_fault() local 51 if (unlikely(in_interrupt() || !mm)) in do_page_fault() 61 vma = lock_mm_and_find_vma(mm, address, regs); in do_page_fault() 103 mmap_read_unlock(mm); in do_page_fault() 107 mmap_read_unlock(mm); in do_page_fault() 134 mmap_read_unlock(mm); in do_page_fault()
|
/linux/arch/powerpc/mm/book3s64/ |
H A D | hash_pgtable.c | 190 unsigned long hash__pmd_hugepage_update(struct mm_struct *mm, unsigned long addr, in hash__pmd_hugepage_update() argument 199 assert_spin_locked(pmd_lockptr(mm, pmdp)); in hash__pmd_hugepage_update() 219 hpte_do_hugepage_flush(mm, addr, pmdp, old); in hash__pmd_hugepage_update() 267 void hash__pgtable_trans_huge_deposit(struct mm_struct *mm, pmd_t *pmdp, in hash__pgtable_trans_huge_deposit() argument 272 assert_spin_locked(pmd_lockptr(mm, pmdp)); in hash__pgtable_trans_huge_deposit() 287 pgtable_t hash__pgtable_trans_huge_withdraw(struct mm_struct *mm, pmd_t *pmdp) in hash__pgtable_trans_huge_withdraw() argument 292 assert_spin_locked(pmd_lockptr(mm, pmdp)); in hash__pgtable_trans_huge_withdraw() 312 void hpte_do_hugepage_flush(struct mm_struct *mm, unsigned long addr, in hpte_do_hugepage_flush() argument 322 psize = get_slice_psize(mm, addr); in hpte_do_hugepage_flush() 332 vsid = get_user_vsid(&mm->context, addr, ssize); in hpte_do_hugepage_flush() [all …]
|
/linux/arch/alpha/mm/ |
H A D | fault.c | 89 struct mm_struct *mm = current->mm; in do_page_fault() local 111 if (!mm || faulthandler_disabled()) in do_page_fault() 122 vma = lock_mm_and_find_vma(mm, address, regs); in do_page_fault() 178 mmap_read_unlock(mm); in do_page_fault() 185 mmap_read_unlock(mm); in do_page_fault() 210 mmap_read_unlock(mm); in do_page_fault() 217 mmap_read_unlock(mm); in do_page_fault()
|
/linux/arch/arm64/include/asm/ |
H A D | efi.h | 29 int efi_create_mapping(struct mm_struct *mm, efi_memory_desc_t *md); 30 int efi_set_mapping_permissions(struct mm_struct *mm, efi_memory_desc_t *md, 121 static inline void efi_set_pgd(struct mm_struct *mm) in efi_set_pgd() argument 123 __switch_mm(mm); in efi_set_pgd() 126 if (mm != current->active_mm) { in efi_set_pgd() 134 update_saved_ttbr0(current, mm); in efi_set_pgd()
|