Home
last modified time | relevance | path

Searched refs:mem (Results 1 – 25 of 1284) sorted by relevance

12345678910>>...52

/linux/arch/powerpc/mm/book3s64/
H A Diommu_api.c60 struct mm_iommu_table_group_mem_t *mem, *mem2; in mm_iommu_do_alloc() local
73 mem = kzalloc_obj(*mem); in mm_iommu_do_alloc()
74 if (!mem) { in mm_iommu_do_alloc()
80 mem->pageshift = __ffs(dev_hpa | (entries << PAGE_SHIFT)); in mm_iommu_do_alloc()
81 mem->dev_hpa = dev_hpa; in mm_iommu_do_alloc()
84 mem->dev_hpa = MM_IOMMU_TABLE_INVALID_HPA; in mm_iommu_do_alloc()
91 mem->pageshift = __ffs(ua | (entries << PAGE_SHIFT)); in mm_iommu_do_alloc()
92 mem->hpas = vzalloc(array_size(entries, sizeof(mem->hpas[0]))); in mm_iommu_do_alloc()
93 if (!mem->hpas) { in mm_iommu_do_alloc()
94 kfree(mem); in mm_iommu_do_alloc()
[all …]
/linux/drivers/gpu/drm/i915/
H A Dintel_memory_region.c37 static int __iopagetest(struct intel_memory_region *mem, in __iopagetest() argument
52 dev_err(mem->i915->drm.dev, in __iopagetest()
54 &mem->region, &mem->io.start, &offset, caller, in __iopagetest()
62 static int iopagetest(struct intel_memory_region *mem, in iopagetest() argument
71 va = ioremap_wc(mem->io.start + offset, PAGE_SIZE); in iopagetest()
73 dev_err(mem->i915->drm.dev, in iopagetest()
75 &mem->io.start, &offset, caller); in iopagetest()
80 err = __iopagetest(mem, va, PAGE_SIZE, val[i], offset, caller); in iopagetest()
84 err = __iopagetest(mem, va, PAGE_SIZE, ~val[i], offset, caller); in iopagetest()
99 static int iomemtest(struct intel_memory_region *mem, in iomemtest() argument
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/
H A Dmem.c35 struct page **mem; member
57 struct nvkm_mem *mem = nvkm_mem(memory); in nvkm_mem_addr() local
58 if (mem->pages == 1 && mem->mem) in nvkm_mem_addr()
59 return mem->dma[0]; in nvkm_mem_addr()
73 struct nvkm_mem *mem = nvkm_mem(memory); in nvkm_mem_map_dma() local
75 .memory = &mem->memory, in nvkm_mem_map_dma()
77 .dma = mem->dma, in nvkm_mem_map_dma()
85 struct nvkm_mem *mem = nvkm_mem(memory); in nvkm_mem_dtor() local
86 if (mem->mem) { in nvkm_mem_dtor()
87 while (mem->pages--) { in nvkm_mem_dtor()
[all …]
/linux/drivers/gpu/drm/nouveau/
H A Dnouveau_mem.c37 nouveau_mem_map(struct nouveau_mem *mem, in nouveau_mem_map() argument
53 args.nv50.kind = mem->kind; in nouveau_mem_map()
54 args.nv50.comp = mem->comp; in nouveau_mem_map()
61 if (mem->mem.type & NVIF_MEM_VRAM) in nouveau_mem_map()
67 args.gf100.kind = mem->kind; in nouveau_mem_map()
75 return nvif_vmm_map(vmm, vma->addr, mem->mem.size, &args, argc, &mem->mem, 0); in nouveau_mem_map()
79 nouveau_mem_fini(struct nouveau_mem *mem) in nouveau_mem_fini() argument
81 nvif_vmm_put(&mem->drm->client.vmm.vmm, &mem->vma[1]); in nouveau_mem_fini()
82 nvif_vmm_put(&mem->drm->client.vmm.vmm, &mem->vma[0]); in nouveau_mem_fini()
83 mutex_lock(&mem->drm->client_mutex); in nouveau_mem_fini()
[all …]
H A Dnouveau_vmm.c31 if (vma->mem) { in nouveau_vma_unmap()
33 vma->mem = NULL; in nouveau_vma_unmap()
38 nouveau_vma_map(struct nouveau_vma *vma, struct nouveau_mem *mem) in nouveau_vma_map() argument
41 int ret = nouveau_mem_map(mem, &vma->vmm->vmm, &tmp); in nouveau_vma_map()
44 vma->mem = mem; in nouveau_vma_map()
80 struct nouveau_mem *mem = nouveau_mem(nvbo->bo.resource); in nouveau_vma_new() local
95 vma->mem = NULL; in nouveau_vma_new()
100 mem->mem.page == nvbo->page) { in nouveau_vma_new()
101 ret = nvif_vmm_get(&vmm->vmm, LAZY, false, mem->mem.page, 0, in nouveau_vma_new()
102 mem->mem.size, &tmp); in nouveau_vma_new()
[all …]
/linux/drivers/infiniband/sw/siw/
H A Dsiw_mem.c29 struct siw_mem *mem; in siw_mem_id2obj() local
32 mem = xa_load(&sdev->mem_xa, stag_index); in siw_mem_id2obj()
33 if (likely(mem && kref_get_unless_zero(&mem->ref))) { in siw_mem_id2obj()
35 return mem; in siw_mem_id2obj()
61 struct siw_mem *mem = kzalloc_obj(*mem); in siw_mr_add_mem() local
65 if (!mem) in siw_mr_add_mem()
68 mem->mem_obj = mem_obj; in siw_mr_add_mem()
69 mem->stag_valid = 0; in siw_mr_add_mem()
70 mem->sdev = sdev; in siw_mr_add_mem()
71 mem->va = start; in siw_mr_add_mem()
[all …]
/linux/drivers/pci/endpoint/
H A Dpci-epc-mem.c23 static int pci_epc_mem_get_order(struct pci_epc_mem *mem, size_t size) in pci_epc_mem_get_order() argument
26 unsigned int page_shift = ilog2(mem->window.page_size); in pci_epc_mem_get_order()
51 struct pci_epc_mem *mem = NULL; in pci_epc_multi_mem_init() local
77 mem = kzalloc_obj(*mem); in pci_epc_multi_mem_init()
78 if (!mem) { in pci_epc_multi_mem_init()
87 kfree(mem); in pci_epc_multi_mem_init()
92 mem->window.phys_base = windows[i].phys_base; in pci_epc_multi_mem_init()
93 mem->window.size = windows[i].size; in pci_epc_multi_mem_init()
94 mem->window.page_size = page_size; in pci_epc_multi_mem_init()
95 mem->bitmap = bitmap; in pci_epc_multi_mem_init()
[all …]
/linux/kernel/dma/
H A Dcoherent.c31 struct dma_coherent_mem * mem) in dma_get_device_base() argument
33 if (mem->use_dev_dma_pfn_offset) in dma_get_device_base()
34 return phys_to_dma(dev, PFN_PHYS(mem->pfn_base)); in dma_get_device_base()
35 return mem->device_base; in dma_get_device_base()
77 static void _dma_release_coherent_memory(struct dma_coherent_mem *mem) in _dma_release_coherent_memory() argument
79 if (!mem) in _dma_release_coherent_memory()
82 memunmap(mem->virt_base); in _dma_release_coherent_memory()
83 bitmap_free(mem->bitmap); in _dma_release_coherent_memory()
84 kfree(mem); in _dma_release_coherent_memory()
88 struct dma_coherent_mem *mem) in dma_assign_coherent_memory() argument
[all …]
H A Dswiotlb.c229 struct io_tlb_pool *mem = &io_tlb_default_mem.defpool; in swiotlb_print_info() local
231 if (!mem->nslabs) { in swiotlb_print_info()
236 pr_info("mapped [mem %pa-%pa] (%luMB)\n", &mem->start, &mem->end, in swiotlb_print_info()
237 (mem->nslabs << IO_TLB_SHIFT) >> 20); in swiotlb_print_info()
258 struct io_tlb_pool *mem = &io_tlb_default_mem.defpool; in swiotlb_update_mem_attributes() local
261 if (!mem->nslabs || mem->late_alloc) in swiotlb_update_mem_attributes()
263 bytes = PAGE_ALIGN(mem->nslabs << IO_TLB_SHIFT); in swiotlb_update_mem_attributes()
264 set_memory_decrypted((unsigned long)mem->vaddr, bytes >> PAGE_SHIFT); in swiotlb_update_mem_attributes()
267 static void swiotlb_init_io_tlb_pool(struct io_tlb_pool *mem, phys_addr_t start, in swiotlb_init_io_tlb_pool() argument
273 mem->nslabs = nslabs; in swiotlb_init_io_tlb_pool()
[all …]
/linux/drivers/base/
H A Dmemory.c94 struct memory_block *mem = to_memory_block(dev); in memory_block_release() local
96 WARN_ON(mem->altmap); in memory_block_release()
97 kfree(mem); in memory_block_release()
161 struct memory_block *mem = to_memory_block(dev); in phys_index_show() local
163 return sysfs_emit(buf, "%08lx\n", memory_block_id(mem->start_section_nr)); in phys_index_show()
182 struct memory_block *mem = to_memory_block(dev); in state_show() local
189 switch (mem->state) { in state_show()
201 return sysfs_emit(buf, "ERROR-UNKNOWN-%d\n", mem->state); in state_show()
213 static unsigned long memblk_nr_poison(struct memory_block *mem);
215 static inline unsigned long memblk_nr_poison(struct memory_block *mem) in memblk_nr_poison() argument
[all …]
/linux/tools/perf/
H A Dbuiltin-mem.c52 struct perf_mem *mem = (struct perf_mem *)opt->value; in parse_record_events() local
68 mem->operation = 0; in parse_record_events()
72 static int __cmd_record(int argc, const char **argv, struct perf_mem *mem, in __cmd_record() argument
105 if (mem->cpu_list) in __cmd_record()
121 (mem->operation & MEM_OPERATION_LOAD) && in __cmd_record()
122 (mem->operation & MEM_OPERATION_STORE)) { in __cmd_record()
126 if (mem->operation & MEM_OPERATION_LOAD) in __cmd_record()
129 if (mem->operation & MEM_OPERATION_STORE) in __cmd_record()
138 if (mem->phys_addr) in __cmd_record()
141 if (mem->data_page_size) in __cmd_record()
[all …]
/linux/tools/testing/selftests/mm/
H A Dmkdirty.c41 static void do_test_write_sigsegv(char *mem) in do_test_write_sigsegv() argument
43 char orig = *mem; in do_test_write_sigsegv()
53 *mem = orig + 1; in do_test_write_sigsegv()
58 ksft_test_result(ret == 1 && *mem == orig, in do_test_write_sigsegv()
65 char *mem, *mmap_mem; in mmap_thp_range() local
73 mem = (char *)(((uintptr_t)mmap_mem + thpsize) & ~(thpsize - 1)); in mmap_thp_range()
75 if (madvise(mem, thpsize, MADV_HUGEPAGE)) { in mmap_thp_range()
83 return mem; in mmap_thp_range()
89 char *mem; in test_ptrace_write() local
94 mem = mmap(NULL, pagesize, PROT_READ, MAP_PRIVATE|MAP_ANON, -1, 0); in test_ptrace_write()
[all …]
H A Dpagemap_ioctl.c164 char *mem; in gethugetlb_mem() local
171 mem = shmat(*shmid, 0, 0); in gethugetlb_mem()
172 if (mem == (char *)-1) { in gethugetlb_mem()
177 mem = mmap(NULL, size, PROT_READ | PROT_WRITE, in gethugetlb_mem()
179 if (mem == MAP_FAILED) in gethugetlb_mem()
183 return mem; in gethugetlb_mem()
189 char *mem, *vec; in userfaultfd_tests() local
192 mem = mmap(NULL, mem_size, PROT_NONE, MAP_PRIVATE | MAP_ANON, -1, 0); in userfaultfd_tests()
193 if (mem == MAP_FAILED) in userfaultfd_tests()
196 wp_init(mem, mem_siz in userfaultfd_tests()
244 char *mem, *m[2]; sanity_tests_sd() local
680 base_tests(char * prefix,char * mem,unsigned long long mem_size,int skip) base_tests() argument
1057 char *mem, *fmem; sanity_tests() local
1228 char *mem, *mem2; mprotect_tests() local
1315 get_dirty_pages_reset(char * mem,unsigned int count,int reset,int page_size) get_dirty_pages_reset() argument
1351 thread_proc(void * mem) thread_proc() argument
1396 char *mem; transact_test() local
1489 char *mmap_mem, *mem; zeropfn_tests() local
1548 char *mem, *map, *fmem; main() local
[all...]
H A Dcow.c119 static int child_memcmp_fn(char *mem, size_t size, in child_memcmp_fn()
126 memcpy(old, mem, size); in child_vmsplice_memcmp_fn()
134 return memcmp(old, mem, size); in child_vmsplice_memcmp_fn()
137 static int child_vmsplice_memcmp_fn(char *mem, size_t size, in child_vmsplice_memcmp_fn()
141 .iov_base = mem, in child_vmsplice_memcmp_fn()
153 memcpy(old, mem, size); in child_vmsplice_memcmp_fn()
166 if (munmap(mem, size) < 0) in child_vmsplice_memcmp_fn()
184 typedef int (*child_fn)(char *mem, size_t size, struct comm_pipes *comm_pipes); in do_test_cow_in_parent()
186 static void do_test_cow_in_parent(char *mem, size_t size, bool do_mprotect, in do_test_cow_in_parent()
205 exit(fn(mem, siz in do_test_cow_in_parent()
107 child_memcmp_fn(char * mem,size_t size,struct comm_pipes * comm_pipes) child_memcmp_fn() argument
125 child_vmsplice_memcmp_fn(char * mem,size_t size,struct comm_pipes * comm_pipes) child_vmsplice_memcmp_fn() argument
174 do_test_cow_in_parent(char * mem,size_t size,bool do_mprotect,child_fn fn,bool xfail) do_test_cow_in_parent() argument
251 test_cow_in_parent(char * mem,size_t size,bool is_hugetlb) test_cow_in_parent() argument
256 test_cow_in_parent_mprotect(char * mem,size_t size,bool is_hugetlb) test_cow_in_parent_mprotect() argument
261 test_vmsplice_in_child(char * mem,size_t size,bool is_hugetlb) test_vmsplice_in_child() argument
267 test_vmsplice_in_child_mprotect(char * mem,size_t size,bool is_hugetlb) test_vmsplice_in_child_mprotect() argument
274 do_test_vmsplice_in_parent(char * mem,size_t size,bool before_fork,bool xfail) do_test_vmsplice_in_parent() argument
388 test_vmsplice_before_fork(char * mem,size_t size,bool is_hugetlb) test_vmsplice_before_fork() argument
393 test_vmsplice_after_fork(char * mem,size_t size,bool is_hugetlb) test_vmsplice_after_fork() argument
399 do_test_iouring(char * mem,size_t size,bool use_fork) do_test_iouring() argument
571 test_iouring_ro(char * mem,size_t size,bool is_hugetlb) test_iouring_ro() argument
576 test_iouring_fork(char * mem,size_t size,bool is_hugetlb) test_iouring_fork() argument
590 do_test_ro_pin(char * mem,size_t size,enum ro_pin_test test,bool fast) do_test_ro_pin() argument
732 test_ro_pin_on_shared(char * mem,size_t size,bool is_hugetlb) test_ro_pin_on_shared() argument
737 test_ro_fast_pin_on_shared(char * mem,size_t size,bool is_hugetlb) test_ro_fast_pin_on_shared() argument
742 test_ro_pin_on_ro_previously_shared(char * mem,size_t size,bool is_hugetlb) test_ro_pin_on_ro_previously_shared() argument
748 test_ro_fast_pin_on_ro_previously_shared(char * mem,size_t size,bool is_hugetlb) test_ro_fast_pin_on_ro_previously_shared() argument
754 test_ro_pin_on_ro_exclusive(char * mem,size_t size,bool is_hugetlb) test_ro_pin_on_ro_exclusive() argument
760 test_ro_fast_pin_on_ro_exclusive(char * mem,size_t size,bool is_hugetlb) test_ro_fast_pin_on_ro_exclusive() argument
770 char *mem; do_run_with_base_page() local
831 char *mem, *mmap_mem, *tmp, *mremap_mem = MAP_FAILED; do_run_with_thp() local
1043 char *mem, *dummy; run_with_hugetlb() local
1264 do_test_anon_thp_collapse(char * mem,size_t size,enum anon_thp_collapse_test test) do_test_anon_thp_collapse() argument
1408 test_anon_thp_collapse_unshared(char * mem,size_t size,bool is_hugetlb) test_anon_thp_collapse_unshared() argument
1415 test_anon_thp_collapse_fully_shared(char * mem,size_t size,bool is_hugetlb) test_anon_thp_collapse_fully_shared() argument
1422 test_anon_thp_collapse_lower_shared(char * mem,size_t size,bool is_hugetlb) test_anon_thp_collapse_lower_shared() argument
1429 test_anon_thp_collapse_upper_shared(char * mem,size_t size,bool is_hugetlb) test_anon_thp_collapse_upper_shared() argument
1500 test_cow(char * mem,const char * smem,size_t size) test_cow() argument
1520 test_ro_pin(char * mem,const char * smem,size_t size) test_ro_pin() argument
1525 test_ro_fast_pin(char * mem,const char * smem,size_t size) test_ro_fast_pin() argument
1532 char *mem, *smem; run_with_zeropage() local
1564 char *mem, *smem, *mmap_mem, *mmap_smem; run_with_huge_zeropage() local
1632 char *mem, *smem; run_with_memfd() local
1680 char *mem, *smem; run_with_tmpfile() local
1738 char *mem, *smem; run_with_memfd_hugetlb() local
[all...]
H A Duffd-wp-mremap.c50 char *mmap_mem, *mem; in mmap_aligned() local
56 mem = (char *)(((uintptr_t)mmap_mem + size - 1) & ~(size - 1)); in mmap_aligned()
57 munmap(mmap_mem, mem - mmap_mem); in mmap_aligned()
58 munmap(mem + size, mmap_mem + mmap_size - mem - size); in mmap_aligned()
60 return mem; in mmap_aligned()
68 char *mem, *addr; in alloc_one_folio() local
91 mem = mmap_aligned(size, prot, flags); in alloc_one_folio()
98 mem = mmap(NULL, size, prot, flags, -1, 0); in alloc_one_folio()
101 if (mem == MAP_FAILED) { in alloc_one_folio()
102 mem = NULL; in alloc_one_folio()
[all …]
H A Dmemfd_secret.c63 char *mem; in test_mlock_limit() local
69 mem = mmap(NULL, len, prot, mode, fd, 0); in test_mlock_limit()
70 if (mem == MAP_FAILED) { in test_mlock_limit()
74 munmap(mem, len); in test_mlock_limit()
77 mem = mmap(NULL, len, prot, mode, fd, 0); in test_mlock_limit()
78 if (mem != MAP_FAILED) { in test_mlock_limit()
80 munmap(mem, len); in test_mlock_limit()
92 char *mem; in test_vmsplice() local
99 mem = mmap(NULL, page_size, prot, mode, fd, 0); in test_vmsplice()
100 if (mem == MAP_FAILED) { in test_vmsplice()
[all …]
/linux/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_amdkfd_gpuvm.c78 struct kgd_mem *mem) in kfd_mem_is_attached() argument
82 list_for_each_entry(entry, &mem->attachments, list) in kfd_mem_is_attached()
115 uint64_t mem; in amdgpu_amdkfd_gpuvm_init_mem_limits() local
121 mem = si.totalram - si.totalhigh; in amdgpu_amdkfd_gpuvm_init_mem_limits()
122 mem *= si.mem_unit; in amdgpu_amdkfd_gpuvm_init_mem_limits()
125 kfd_mem_limit.max_system_mem_limit = mem - (mem >> 6); in amdgpu_amdkfd_gpuvm_init_mem_limits()
332 struct kgd_mem *mem, struct amdgpu_bo **bo_out) in create_dmamap_sg_bo() argument
338 ret = amdgpu_bo_reserve(mem->bo, false); in create_dmamap_sg_bo()
342 if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_USERPTR) in create_dmamap_sg_bo()
343 flags |= mem->bo->flags & (AMDGPU_GEM_CREATE_COHERENT | in create_dmamap_sg_bo()
[all …]
/linux/drivers/spi/
H A Dspi-mem.c107 static int spi_check_buswidth_req(struct spi_mem *mem, u8 buswidth, bool tx) in spi_check_buswidth_req() argument
109 u32 mode = mem->spi->mode; in spi_check_buswidth_req()
145 static bool spi_mem_check_buswidth(struct spi_mem *mem, in spi_mem_check_buswidth() argument
148 if (spi_check_buswidth_req(mem, op->cmd.buswidth, true)) in spi_mem_check_buswidth()
152 spi_check_buswidth_req(mem, op->addr.buswidth, true)) in spi_mem_check_buswidth()
156 spi_check_buswidth_req(mem, op->dummy.buswidth, true)) in spi_mem_check_buswidth()
160 spi_check_buswidth_req(mem, op->data.buswidth, in spi_mem_check_buswidth()
167 bool spi_mem_default_supports_op(struct spi_mem *mem, in spi_mem_default_supports_op() argument
170 struct spi_controller *ctlr = mem->spi->controller; in spi_mem_default_supports_op()
204 if (op->max_freq && mem->spi->controller->min_speed_hz && in spi_mem_default_supports_op()
[all …]
/linux/drivers/gpu/drm/ttm/
H A Dttm_agp_backend.c47 struct agp_memory *mem; member
55 struct agp_memory *mem; in ttm_agp_bind() local
59 if (agp_be->mem) in ttm_agp_bind()
62 mem = agp_allocate_memory(agp_be->bridge, ttm->num_pages, AGP_USER_MEMORY); in ttm_agp_bind()
63 if (unlikely(mem == NULL)) in ttm_agp_bind()
66 mem->page_count = 0; in ttm_agp_bind()
73 mem->pages[mem->page_count++] = page; in ttm_agp_bind()
75 agp_be->mem = mem; in ttm_agp_bind()
77 mem->is_flushed = 1; in ttm_agp_bind()
78 mem->type = (cached) ? AGP_USER_CACHED_MEMORY : AGP_USER_MEMORY; in ttm_agp_bind()
[all …]
/linux/drivers/char/hw_random/
H A Dintel-rng.c155 static inline u8 hwstatus_get(void __iomem *mem) in hwstatus_get() argument
157 return readb(mem + INTEL_RNG_HW_STATUS); in hwstatus_get()
160 static inline u8 hwstatus_set(void __iomem *mem, in hwstatus_set() argument
163 writeb(hw_status, mem + INTEL_RNG_HW_STATUS); in hwstatus_set()
164 return hwstatus_get(mem); in hwstatus_set()
169 void __iomem *mem = (void __iomem *)rng->priv; in intel_rng_data_present() local
173 data = !!(readb(mem + INTEL_RNG_STATUS) & in intel_rng_data_present()
184 void __iomem *mem = (void __iomem *)rng->priv; in intel_rng_data_read() local
186 *data = readb(mem + INTEL_RNG_DATA); in intel_rng_data_read()
193 void __iomem *mem = (void __iomem *)rng->priv; in intel_rng_init() local
[all …]
/linux/drivers/firmware/efi/
H A Dcper.c301 int cper_mem_err_location(struct cper_mem_err_compact *mem, char *msg) in cper_mem_err_location() argument
310 if (mem->validation_bits & CPER_MEM_VALID_NODE) in cper_mem_err_location()
311 n += scnprintf(msg + n, len - n, "node:%d ", mem->node); in cper_mem_err_location()
312 if (mem->validation_bits & CPER_MEM_VALID_CARD) in cper_mem_err_location()
313 n += scnprintf(msg + n, len - n, "card:%d ", mem->card); in cper_mem_err_location()
314 if (mem->validation_bits & CPER_MEM_VALID_MODULE) in cper_mem_err_location()
315 n += scnprintf(msg + n, len - n, "module:%d ", mem->module); in cper_mem_err_location()
316 if (mem->validation_bits & CPER_MEM_VALID_RANK_NUMBER) in cper_mem_err_location()
317 n += scnprintf(msg + n, len - n, "rank:%d ", mem->rank); in cper_mem_err_location()
318 if (mem->validation_bits & CPER_MEM_VALID_BANK) in cper_mem_err_location()
[all …]
/linux/tools/testing/selftests/kvm/
H A Dguest_memfd_test.c44 void *mem; in test_mmap_cow() local
46 mem = mmap(NULL, size, PROT_READ | PROT_WRITE, MAP_PRIVATE, fd, 0); in test_mmap_cow()
47 TEST_ASSERT(mem == MAP_FAILED, "Copy-on-write not allowed by guest_memfd."); in test_mmap_cow()
53 char *mem; in test_mmap_supported() local
57 mem = kvm_mmap(total_size, PROT_READ | PROT_WRITE, MAP_SHARED, fd); in test_mmap_supported()
59 memset(mem, val, total_size); in test_mmap_supported()
61 TEST_ASSERT_EQ(READ_ONCE(mem[i]), val); in test_mmap_supported()
68 TEST_ASSERT_EQ(READ_ONCE(mem[i]), 0x00); in test_mmap_supported()
70 TEST_ASSERT_EQ(READ_ONCE(mem[i]), val); in test_mmap_supported()
72 memset(mem, val, page_size); in test_mmap_supported()
[all …]
/linux/mm/
H A Dcma_debug.c69 static void cma_add_to_cma_mem_list(struct cma *cma, struct cma_mem *mem) in cma_add_to_cma_mem_list() argument
72 hlist_add_head(&mem->node, &cma->mem_head); in cma_add_to_cma_mem_list()
78 struct cma_mem *mem = NULL; in cma_get_entry_from_list() local
82 mem = hlist_entry(cma->mem_head.first, struct cma_mem, node); in cma_get_entry_from_list()
83 hlist_del_init(&mem->node); in cma_get_entry_from_list()
87 return mem; in cma_get_entry_from_list()
92 struct cma_mem *mem = NULL; in cma_free_mem() local
95 mem = cma_get_entry_from_list(cma); in cma_free_mem()
96 if (mem == NULL) in cma_free_mem()
99 if (mem->n <= count) { in cma_free_mem()
[all …]
/linux/drivers/net/ipa/
H A Dipa_qmi.c285 const struct ipa_mem *mem; in init_modem_driver_req() local
298 mem = ipa_mem_find(ipa, IPA_MEM_MODEM_HEADER); in init_modem_driver_req()
299 if (mem->size) { in init_modem_driver_req()
301 req.hdr_tbl_info.start = ipa->mem_offset + mem->offset; in init_modem_driver_req()
302 req.hdr_tbl_info.end = req.hdr_tbl_info.start + mem->size - 1; in init_modem_driver_req()
305 mem = ipa_mem_find(ipa, IPA_MEM_V4_ROUTE); in init_modem_driver_req()
307 req.v4_route_tbl_info.start = ipa->mem_offset + mem->offset; in init_modem_driver_req()
310 mem = ipa_mem_find(ipa, IPA_MEM_V6_ROUTE); in init_modem_driver_req()
312 req.v6_route_tbl_info.start = ipa->mem_offset + mem->offset; in init_modem_driver_req()
315 mem = ipa_mem_find(ipa, IPA_MEM_V4_FILTER); in init_modem_driver_req()
[all …]
/linux/drivers/media/platform/mediatek/vcodec/common/
H A Dmtk_vcodec_util.c48 int mtk_vcodec_mem_alloc(void *priv, struct mtk_vcodec_mem *mem) in mtk_vcodec_mem_alloc() argument
66 mem->va = dma_alloc_attrs(&plat_dev->dev, mem->size, &mem->dma_addr, in mtk_vcodec_mem_alloc()
68 if (!mem->va) { in mtk_vcodec_mem_alloc()
70 __func__, mem->size); in mtk_vcodec_mem_alloc()
74 mtk_v4l2_debug(plat_dev, 3, "[%d] - va = %p dma = 0x%lx size = 0x%zx", id, mem->va, in mtk_vcodec_mem_alloc()
75 (unsigned long)mem->dma_addr, mem->size); in mtk_vcodec_mem_alloc()
81 void mtk_vcodec_mem_free(void *priv, struct mtk_vcodec_mem *mem) in mtk_vcodec_mem_free() argument
99 if (!mem->va) { in mtk_vcodec_mem_free()
101 if (mem->size) in mtk_vcodec_mem_free()
102 mtk_v4l2_err(plat_dev, "Failed to free %zu bytes", mem->size); in mtk_vcodec_mem_free()
[all …]

12345678910>>...52