Home
last modified time | relevance | path

Searched +full:mm +full:- +full:0 (Results 1 – 25 of 1049) sorted by relevance

12345678910>>...42

/linux/drivers/gpu/drm/tests/
H A Ddrm_buddy_test.c1 // SPDX-License-Identifier: MIT
31 struct drm_buddy mm; in drm_test_buddy_alloc_range_bias() local
37 mm_size = (SZ_8M-1) & ~(ps-1); /* Multiple roots */ in drm_test_buddy_alloc_range_bias()
41 KUNIT_ASSERT_FALSE_MSG(test, drm_buddy_init(&mm, mm_size, ps), in drm_test_buddy_alloc_range_bias()
55 for (i = 0; i < count; i++) { in drm_test_buddy_alloc_range_bias()
65 drm_buddy_alloc_blocks(&mm, bias_start, in drm_test_buddy_alloc_range_bias()
69 "buddy_alloc failed with bias(%x-%x), size=%u, ps=%u\n", in drm_test_buddy_alloc_range_bias()
74 drm_buddy_alloc_blocks(&mm, bias_start, in drm_test_buddy_alloc_range_bias()
78 "buddy_alloc didn't fail with bias(%x-%x), size=%u, ps=%u\n", in drm_test_buddy_alloc_range_bias()
83 drm_buddy_alloc_blocks(&mm, bias_start + ps, in drm_test_buddy_alloc_range_bias()
[all …]
H A Ddrm_mm_test.c1 // SPDX-License-Identifier: GPL-2.0-only
32 [BOTTOMUP] = { "bottom-up", DRM_MM_INSERT_LOW },
33 [TOPDOWN] = { "top-down", DRM_MM_INSERT_HIGH },
38 static bool assert_no_holes(struct kunit *test, const struct drm_mm *mm) in assert_no_holes() argument
44 count = 0; in assert_no_holes()
45 drm_mm_for_each_hole(hole, mm, hole_start, hole_end) in assert_no_holes()
53 drm_mm_for_each_node(hole, mm) { in assert_no_holes()
63 static bool assert_one_hole(struct kunit *test, const struct drm_mm *mm, u64 start, u64 end) in assert_one_hole() argument
73 count = 0; in assert_one_hole()
74 drm_mm_for_each_hole(hole, mm, hole_start, hole_end) { in assert_one_hole()
[all …]
/linux/drivers/gpu/drm/
H A Ddrm_buddy.c1 // SPDX-License-Identifier: MIT
14 static struct drm_buddy_block *drm_block_alloc(struct drm_buddy *mm, in drm_block_alloc() argument
27 block->header = offset; in drm_block_alloc()
28 block->header |= order; in drm_block_alloc()
29 block->parent = parent; in drm_block_alloc()
31 BUG_ON(block->header & DRM_BUDDY_HEADER_UNUSED); in drm_block_alloc()
35 static void drm_block_free(struct drm_buddy *mm, in drm_block_free() argument
41 static void list_insert_sorted(struct drm_buddy *mm, in list_insert_sorted() argument
47 head = &mm->free_list[drm_buddy_block_order(block)]; in list_insert_sorted()
49 list_add(&block->link, head); in list_insert_sorted()
[all …]
/linux/arch/m68k/include/asm/
H A Dmmu_context.h1 /* SPDX-License-Identifier: GPL-2.0 */
5 #include <asm-generic/mm_hooks.h>
28 static inline void get_mmu_context(struct mm_struct *mm) in get_mmu_context() argument
32 if (mm->context != NO_CONTEXT) in get_mmu_context()
42 ctx = 0; in get_mmu_context()
45 mm->context = ctx; in get_mmu_context()
46 context_mm[ctx] = mm; in get_mmu_context()
52 #define init_new_context(tsk, mm) (((mm)->context = NO_CONTEXT), 0) argument
58 static inline void destroy_context(struct mm_struct *mm) in destroy_context() argument
60 if (mm->context != NO_CONTEXT) { in destroy_context()
[all …]
/linux/arch/s390/include/asm/
H A Dmmu_context.h1 /* SPDX-License-Identifier: GPL-2.0 */
5 * Derived from "include/asm-i386/mmu_context.h"
16 #include <asm-generic/mm_hooks.h>
20 struct mm_struct *mm) in init_new_context() argument
24 spin_lock_init(&mm->context.lock); in init_new_context()
25 INIT_LIST_HEAD(&mm->context.gmap_list); in init_new_context()
26 cpumask_clear(&mm->context.cpu_attach_mask); in init_new_context()
27 atomic_set(&mm->context.flush_count, 0); in init_new_context()
28 atomic_set(&mm->context.protected_count, 0); in init_new_context()
29 mm->context.gmap_asce = 0; in init_new_context()
[all …]
/linux/drivers/net/ethernet/mscc/
H A Docelot_mm.c1 // SPDX-License-Identifier: (GPL-2.0 OR MIT)
3 * Hardware library for MAC Merge Layer and Frame Preemption on TSN-capable
6 * Copyright 2022-2023 NXP
37 case 0: in ocelot_mm_verify_status()
54 struct ocelot_port *ocelot_port = ocelot->ports[port]; in ocelot_port_update_active_preemptible_tcs()
55 struct ocelot_mm_state *mm = &ocelot->mm[port]; in ocelot_port_update_active_preemptible_tcs() local
56 u32 val = 0; in ocelot_port_update_active_preemptible_tcs()
58 lockdep_assert_held(&ocelot->fwd_domain_lock); in ocelot_port_update_active_preemptible_tcs()
65 if ((ocelot_port->phy_mode != PHY_INTERFACE_MODE_QSGMII || in ocelot_port_update_active_preemptible_tcs()
66 ocelot_port->speed == SPEED_1000) && mm->tx_active) in ocelot_port_update_active_preemptible_tcs()
[all …]
/linux/mm/
H A Dmmu_notifier.c1 // SPDX-License-Identifier: GPL-2.0-only
3 * linux/mm/mmu_notifier.c
13 #include <linux/mm.h>
19 #include <linux/sched/mm.h>
35 * mm->notifier_subscriptions inside the mm_take_all_locks() protected
40 /* all mmu notifiers registered in this mm are queued in this list */
53 * This is a collision-retry read-side/write-side 'lock', a lot like a
54 * seqcount, however this allows multiple write-sides to hold it at
56 * this mm, such that PTES cannot be read into SPTEs (shadow PTEs) while any
59 * Note that the core mm creates nested invalidate_range_start()/end() regions
[all …]
H A Ddebug.c1 // SPDX-License-Identifier: GPL-2.0
3 * mm/debug.c
5 * mm/ specific debug routines.
10 #include <linux/mm.h>
36 {0, NULL}
41 {0, NULL}
46 {0, NULL}
49 #define DEF_PAGETYPE_NAME(_name) [PGTY_##_name - 0xf0] = __stringify(_name)
63 unsigned i = (page_type >> 24) - 0xf0; in page_type_name()
74 int mapcount = atomic_read(&page->_mapcount); in __dump_folio()
[all …]
H A Dmmap.c1 // SPDX-License-Identifier: GPL-2.0-only
3 * mm/mmap.c
14 #include <linux/backing-dev.h>
15 #include <linux/mm.h>
48 #include <linux/sched/mm.h>
63 #define arch_mmap_check(addr, len, flags) (0)
80 /* Update vma->vm_page_prot to reflect vma->vm_flags. */
83 unsigned long vm_flags = vma->vm_flags; in vma_set_page_prot()
86 vm_page_prot = vm_pgprot_modify(vma->vm_page_prot, vm_flags); in vma_set_page_prot()
91 /* remove_protection_ptes reads vma->vm_page_prot without mmap_lock */ in vma_set_page_prot()
[all …]
/linux/kernel/
H A Dfork.c1 // SPDX-License-Identifier: GPL-2.0-only
9 * 'fork.c' contains the help-routines for the 'fork' system call
12 * management can be a bitch. See 'mm/memory.c': 'copy_page_range()'
18 #include <linux/sched/mm.h>
45 #include <linux/mm.h>
84 #include <linux/posix-timers.h>
85 #include <linux/user-return-notifier.h>
149 DEFINE_PER_CPU(unsigned long, process_counts) = 0;
164 int total = 0; in nr_processes()
211 for (i = 0; i < NR_CACHED_STACKS; i++) { in try_release_thread_stack_to_cache()
[all …]
/linux/arch/powerpc/mm/book3s64/
H A Dmmu_context.c1 // SPDX-License-Identifier: GPL-2.0-or-later
3 * MMU context allocation for 64-bit kernels.
13 #include <linux/mm.h>
62 * id 0 (aka. ctx->id) is special, we always allocate a new one, even if in realloc_context_ids()
67 * the array, so that we can test if they're non-zer in realloc_context_ids()
95 hash__init_new_context(struct mm_struct * mm) hash__init_new_context() argument
157 hash__init_new_context(struct mm_struct * mm) hash__init_new_context() argument
164 radix__init_new_context(struct mm_struct * mm) radix__init_new_context() argument
195 init_new_context(struct task_struct * tsk,struct mm_struct * mm) init_new_context() argument
261 destroy_pagetable_cache(struct mm_struct * mm) destroy_pagetable_cache() argument
275 destroy_context(struct mm_struct * mm) destroy_context() argument
301 arch_exit_mmap(struct mm_struct * mm) arch_exit_mmap() argument
[all...]
H A Dradix_tlb.c1 // SPDX-License-Identifier: GPL-2.0-or-later
5 * Copyright 2015-2016, Aneesh Kumar K.V, IBM Corporation.
8 #include <linux/mm.h>
12 #include <linux/sched/mm.h>
15 #include <asm/ppc-opcode.h>
38 asm volatile(PPC_TLBIEL(%0, %1, %2, %3, 1) in tlbiel_radix_set_isa300()
57 tlbiel_radix_set_isa300(0, is, 0, RIC_FLUSH_ALL, 0); in tlbiel_all_isa300()
61 tlbiel_radix_set_isa300(set, is, 0, in tlbiel_all_isa300()
62 RIC_FLUSH_TLB, 0); in tlbiel_all_isa300()
67 tlbiel_radix_set_isa300(0, is, 0, RIC_FLUSH_ALL, 1); in tlbiel_all_isa300()
[all …]
H A Dsubpage_prot.c1 // SPDX-License-Identifier: GPL-2.0-or-later
3 * Copyright 2007-2008 Paul Mackerras, IBM Corp.
22 void subpage_prot_free(struct mm_struct *mm) in subpage_prot_free() argument
24 struct subpage_prot_table *spt = mm_ctx_subpage_prot(&mm->context); in subpage_prot_free()
31 for (i = 0; i < 4; ++i) { in subpage_prot_free()
32 if (spt->low_prot[i]) { in subpage_prot_free()
33 free_page((unsigned long)spt->low_prot[i]); in subpage_prot_free()
34 spt->low_prot[i] = NULL; in subpage_prot_free()
37 addr = 0; in subpage_prot_free()
38 for (i = 0; i < (TASK_SIZE_USER64 >> 43); ++i) { in subpage_prot_free()
[all …]
/linux/fs/proc/
H A Dtask_nommu.c1 // SPDX-License-Identifier: GPL-2.0
3 #include <linux/mm.h>
11 #include <linux/sched/mm.h>
17 * "non-shared". Shared memory may get counted more than once, for
18 * each process that owns it. Non-shared memory is counted
21 void task_mem(struct seq_file *m, struct mm_struct *mm) in task_mem() argument
23 VMA_ITERATOR(vmi, mm, 0); in task_mem()
26 unsigned long bytes = 0, sbytes = 0, slack = 0, size; in task_mem()
28 mmap_read_lock(mm); in task_mem()
32 region = vma->vm_region; in task_mem()
[all …]
/linux/arch/powerpc/include/asm/
H A Dmmu_context.h1 /* SPDX-License-Identifier: GPL-2.0 */
7 #include <linux/mm.h>
18 extern int init_new_context(struct task_struct *tsk, struct mm_struct *mm);
20 extern void destroy_context(struct mm_struct *mm);
24 extern bool mm_iommu_preregistered(struct mm_struct *mm);
25 extern long mm_iommu_new(struct mm_struct *mm,
28 extern long mm_iommu_newdev(struct mm_struct *mm, unsigned long ua,
31 extern long mm_iommu_put(struct mm_struct *mm,
33 extern void mm_iommu_init(struct mm_struct *mm);
34 extern struct mm_iommu_table_group_mem_t *mm_iommu_lookup(struct mm_struct *mm,
[all …]
/linux/arch/sparc/mm/
H A Dtsb.c1 // SPDX-License-Identifier: GPL-2.0
2 /* arch/sparc64/mm/tsb.c
25 return vaddr & (nentries - 1); in tsb_hash()
37 for (idx = 0; idx < KERNEL_TSB_NENTRIES; idx++) { in flush_tsb_kernel_range_scan()
41 match |= (ent->tag << 22); in flush_tsb_kernel_range_scan()
43 ent->tag = (1UL << TSB_TAG_INVALID_BIT); in flush_tsb_kernel_range_scan()
56 if ((end - start) >> PAGE_SHIFT >= 2 * KERNEL_TSB_NENTRIES) in flush_tsb_kernel_range()
64 if (tag_compare(ent->tag, v)) in flush_tsb_kernel_range()
65 ent->tag = (1UL << TSB_TAG_INVALID_BIT); in flush_tsb_kernel_range()
75 v &= ~0x1UL; in __flush_tsb_one_entry()
[all …]
/linux/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_mqd_manager_v9.c1 // SPDX-License-Identifier: GPL-2.0 OR MIT
3 * Copyright 2016-2022 Advanced Micro Devices, Inc.
37 static void update_mqd(struct mqd_manager *mm, void *mqd,
41 static uint64_t mqd_stride_v9(struct mqd_manager *mm, in mqd_stride_v9() argument
44 if (mm->dev->kfd->cwsr_enabled && in mqd_stride_v9()
45 q->type == KFD_QUEUE_TYPE_COMPUTE) in mqd_stride_v9()
46 return ALIGN(q->ctl_stack_size, PAGE_SIZE) + in mqd_stride_v9()
49 return mm->mqd_size; in mqd_stride_v9()
62 static void update_cu_mask(struct mqd_manager *mm, void *mqd, in update_cu_mask() argument
66 uint32_t se_mask[KFD_MAX_NUM_SE] = {0}; in update_cu_mask()
[all …]
H A Dkfd_mqd_manager.c1 // SPDX-License-Identifier: GPL-2.0 OR MIT
3 * Copyright 2014-2022 Advanced Micro Devices, Inc.
57 mqd_mem_obj->gtt_mem = dev->dqm->hiq_sdma_mqd.gtt_mem; in allocate_hiq_mqd()
58 mqd_mem_obj->gpu_addr = dev->dqm->hiq_sdma_mqd.gpu_addr; in allocate_hiq_mqd()
59 mqd_mem_obj->cpu_ptr = dev->dqm->hiq_sdma_mqd.cpu_ptr; in allocate_hiq_mqd()
74 offset = (q->sdma_engine_id * in allocate_sdma_mqd()
75 dev->kfd->device_info.num_sdma_queues_per_engine + in allocate_sdma_mqd()
76 q->sdma_queue_id) * in allocate_sdma_mqd()
77 dev->dqm->mqd_mgrs[KFD_MQD_TYPE_SDMA]->mqd_size; in allocate_sdma_mqd()
79 offset += dev->dqm->mqd_mgrs[KFD_MQD_TYPE_HIQ]->mqd_size * in allocate_sdma_mqd()
[all …]
/linux/arch/x86/kernel/
H A Dldt.c1 // SPDX-License-Identifier: GPL-2.0
19 #include <linux/mm.h>
42 void load_mm_ldt(struct mm_struct *mm) in load_mm_ldt() argument
47 ldt = READ_ONCE(mm->context.ldt); in load_mm_ldt()
50 * Any change to mm->context.ldt is followed by an IPI to all in load_mm_ldt()
51 * CPUs with the mm active. The LDT will not be freed until in load_mm_ldt()
65 if (WARN_ON_ONCE((unsigned long)ldt->slot > 1)) { in load_mm_ldt()
67 * Whoops -- either the new LDT isn't mapped in load_mm_ldt()
68 * (if slot == -1) or is mapped into a bogus in load_mm_ldt()
76 * If page table isolation is enabled, ldt->entries in load_mm_ldt()
[all …]
/linux/arch/arm/mm/
H A Dpgd.c1 // SPDX-License-Identifier: GPL-2.0-only
3 * linux/arch/arm/mm/pgd.c
5 * Copyright (C) 1998-2005 Russell King
7 #include <linux/mm.h>
17 #include "mm.h"
20 #define _pgd_alloc(mm) kmalloc_array(PTRS_PER_PGD, sizeof(pgd_t), GFP_KERNEL | __GFP_ZERO)
21 #define _pgd_free(mm, pg
30 pgd_alloc(struct mm_struct * mm) pgd_alloc() argument
142 pgd_free(struct mm_struct * mm,pgd_t * pgd_base) pgd_free() argument
[all...]
/linux/drivers/misc/cxl/
H A Dfault.c1 // SPDX-License-Identifier: GPL-2.0-or-later
8 #include <linux/sched/mm.h>
10 #include <linux/mm.h>
24 return ((sste->vsid_data == cpu_to_be64(slb->vsid)) && in sste_matches()
25 (sste->esid_data == cpu_to_be64(slb->esid))); in sste_matches()
36 unsigned int mask = (ctx->sst_size >> 7) - 1; /* SSTP0[SegTableSize] */ in find_free_sste()
40 if (slb->vsid & SLB_VSID_B_1T) in find_free_sste()
41 hash = (slb->esid >> SID_SHIFT_1T) & mask; in find_free_sste()
43 hash = (slb->esid >> SID_SHIFT) & mask; in find_free_sste()
45 primary = ctx->sstp + (hash << 3); in find_free_sste()
[all …]
/linux/kernel/sched/
H A Dmembarrier.c1 // SPDX-License-Identifier: GPL-2.0-or-later
3 * Copyright (C) 2010-2017 Mathieu Desnoyers <mathieu.desnoyers@efficios.com>
17 * int x = 0, y = 0;
22 * CPU1 after the IPI-induced memory barrier:
29 * b: send IPI IPI-induced mb
36 * BUG_ON(r1 == 0 && r2 == 0)
41 * can be reordered after (a) (although not after (c)), so we get r1 == 0
42 * and r2 == 0. This violates the guarantee that membarrier() is
46 * before the IPI-induced memory barrier on CPU1.
53 * int x = 0, y = 0;
[all …]
/linux/arch/sparc/include/asm/
H A Dmmu_context_64.h1 /* SPDX-License-Identifier: GPL-2.0 */
16 #include <asm-generic/mm_hooks.h>
24 void get_new_mmu_context(struct mm_struct *mm);
27 int init_new_context(struct task_struct *tsk, struct mm_struct *mm);
29 void destroy_context(struct mm_struct *mm);
37 static inline void tsb_context_switch_ctx(struct mm_struct *mm, in tsb_context_switch_ctx() argument
40 __tsb_context_switch(__pa(mm->pgd), in tsb_context_switch_ctx()
41 &mm->context.tsb_block[MM_TSB_BASE], in tsb_context_switch_ctx()
43 (mm->context.tsb_block[MM_TSB_HUGE].tsb ? in tsb_context_switch_ctx()
44 &mm->context.tsb_block[MM_TSB_HUGE] : in tsb_context_switch_ctx()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/core/
H A Dmm.c24 #include <core/mm.h>
26 #define node(root, dir) ((root)->nl_entry.dir == &mm->nodes) ? NULL : \
27 list_entry((root)->nl_entry.dir, struct nvkm_mm_node, nl_entry)
30 nvkm_mm_dump(struct nvkm_mm *mm, const char *header) in nvkm_mm_dump() argument
36 list_for_each_entry(node, &mm->nodes, nl_entry) { in nvkm_mm_dump()
38 node->offset, node->length, node->type); in nvkm_mm_dump()
41 list_for_each_entry(node, &mm->free, fl_entry) { in nvkm_mm_dump()
43 node->offset, node->length, node->type); in nvkm_mm_dump()
48 nvkm_mm_free(struct nvkm_mm *mm, struct nvkm_mm_node **pthis) in nvkm_mm_free() argument
56 if (prev && prev->type == NVKM_MM_TYPE_NONE) { in nvkm_mm_free()
[all …]
/linux/arch/riscv/mm/
H A Dhugetlbpage.c1 // SPDX-License-Identifier: GPL-2.0
6 pte_t huge_ptep_get(struct mm_struct *mm, unsigned long addr, pte_t *ptep) in huge_ptep_get() argument
17 for (i = 0; i < pte_num; i++, ptep++) { in huge_ptep_get()
30 pte_t *huge_pte_alloc(struct mm_struct *mm, in huge_pte_alloc() argument
42 pgd = pgd_offset(mm, addr); in huge_pte_alloc()
43 p4d = p4d_alloc(mm, pgd, addr); in huge_pte_alloc()
47 pud = pud_alloc(mm, p4d, addr); in huge_pte_alloc()
58 pte = huge_pmd_share(mm, vma, addr, pud); in huge_pte_alloc()
60 pte = (pte_t *)pmd_alloc(mm, pud, addr); in huge_pte_alloc()
64 pmd = pmd_alloc(mm, pud, addr); in huge_pte_alloc()
[all …]

12345678910>>...42