Home
last modified time | relevance | path

Searched refs:mem (Results 1 – 25 of 1286) sorted by relevance

12345678910>>...52

/linux/arch/powerpc/mm/book3s64/
H A Diommu_api.c60 struct mm_iommu_table_group_mem_t *mem, *mem2; in mm_iommu_do_alloc() local
73 mem = kzalloc(sizeof(*mem), GFP_KERNEL); in mm_iommu_do_alloc()
74 if (!mem) { in mm_iommu_do_alloc()
80 mem->pageshift = __ffs(dev_hpa | (entries << PAGE_SHIFT)); in mm_iommu_do_alloc()
81 mem->dev_hpa = dev_hpa; in mm_iommu_do_alloc()
84 mem->dev_hpa = MM_IOMMU_TABLE_INVALID_HPA; in mm_iommu_do_alloc()
91 mem->pageshift = __ffs(ua | (entries << PAGE_SHIFT)); in mm_iommu_do_alloc()
92 mem->hpas = vzalloc(array_size(entries, sizeof(mem->hpas[0]))); in mm_iommu_do_alloc()
93 if (!mem->hpas) { in mm_iommu_do_alloc()
94 kfree(mem); in mm_iommu_do_alloc()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/
H A Dmem.c35 struct page **mem; member
57 struct nvkm_mem *mem = nvkm_mem(memory); in nvkm_mem_addr() local
58 if (mem->pages == 1 && mem->mem) in nvkm_mem_addr()
59 return mem->dma[0]; in nvkm_mem_addr()
73 struct nvkm_mem *mem = nvkm_mem(memory); in nvkm_mem_map_dma() local
75 .memory = &mem->memory, in nvkm_mem_map_dma()
77 .dma = mem->dma, in nvkm_mem_map_dma()
85 struct nvkm_mem *mem = nvkm_mem(memory); in nvkm_mem_dtor() local
86 if (mem->mem) { in nvkm_mem_dtor()
87 while (mem->pages--) { in nvkm_mem_dtor()
[all …]
/linux/drivers/gpu/drm/nouveau/
H A Dnouveau_mem.c37 nouveau_mem_map(struct nouveau_mem *mem, in nouveau_mem_map() argument
53 args.nv50.kind = mem->kind; in nouveau_mem_map()
54 args.nv50.comp = mem->comp; in nouveau_mem_map()
61 if (mem->mem.type & NVIF_MEM_VRAM) in nouveau_mem_map()
67 args.gf100.kind = mem->kind; in nouveau_mem_map()
75 return nvif_vmm_map(vmm, vma->addr, mem->mem.size, &args, argc, &mem->mem, 0); in nouveau_mem_map()
79 nouveau_mem_fini(struct nouveau_mem *mem) in nouveau_mem_fini() argument
81 nvif_vmm_put(&mem->drm->client.vmm.vmm, &mem->vma[1]); in nouveau_mem_fini()
82 nvif_vmm_put(&mem->drm->client.vmm.vmm, &mem->vma[0]); in nouveau_mem_fini()
83 mutex_lock(&mem->drm->client_mutex); in nouveau_mem_fini()
[all …]
H A Dnouveau_vmm.c31 if (vma->mem) { in nouveau_vma_unmap()
33 vma->mem = NULL; in nouveau_vma_unmap()
38 nouveau_vma_map(struct nouveau_vma *vma, struct nouveau_mem *mem) in nouveau_vma_map() argument
41 int ret = nouveau_mem_map(mem, &vma->vmm->vmm, &tmp); in nouveau_vma_map()
44 vma->mem = mem; in nouveau_vma_map()
80 struct nouveau_mem *mem = nouveau_mem(nvbo->bo.resource); in nouveau_vma_new() local
95 vma->mem = NULL; in nouveau_vma_new()
100 mem->mem.page == nvbo->page) { in nouveau_vma_new()
101 ret = nvif_vmm_get(&vmm->vmm, LAZY, false, mem->mem.page, 0, in nouveau_vma_new()
102 mem->mem.size, &tmp); in nouveau_vma_new()
[all …]
/linux/drivers/infiniband/sw/siw/
H A Dsiw_mem.c29 struct siw_mem *mem; in siw_mem_id2obj() local
32 mem = xa_load(&sdev->mem_xa, stag_index); in siw_mem_id2obj()
33 if (likely(mem && kref_get_unless_zero(&mem->ref))) { in siw_mem_id2obj()
35 return mem; in siw_mem_id2obj()
61 struct siw_mem *mem = kzalloc(sizeof(*mem), GFP_KERNEL); in siw_mr_add_mem() local
65 if (!mem) in siw_mr_add_mem()
68 mem->mem_obj = mem_obj; in siw_mr_add_mem()
69 mem->stag_valid = 0; in siw_mr_add_mem()
70 mem->sdev = sdev; in siw_mr_add_mem()
71 mem->va = start; in siw_mr_add_mem()
[all …]
/linux/kernel/dma/
H A Dcoherent.c31 struct dma_coherent_mem * mem) in dma_get_device_base() argument
33 if (mem->use_dev_dma_pfn_offset) in dma_get_device_base()
34 return phys_to_dma(dev, PFN_PHYS(mem->pfn_base)); in dma_get_device_base()
35 return mem->device_base; in dma_get_device_base()
77 static void _dma_release_coherent_memory(struct dma_coherent_mem *mem) in _dma_release_coherent_memory() argument
79 if (!mem) in _dma_release_coherent_memory()
82 memunmap(mem->virt_base); in _dma_release_coherent_memory()
83 bitmap_free(mem->bitmap); in _dma_release_coherent_memory()
84 kfree(mem); in _dma_release_coherent_memory()
88 struct dma_coherent_mem *mem) in dma_assign_coherent_memory() argument
[all …]
H A Dswiotlb.c229 struct io_tlb_pool *mem = &io_tlb_default_mem.defpool; in swiotlb_print_info() local
231 if (!mem->nslabs) { in swiotlb_print_info()
236 pr_info("mapped [mem %pa-%pa] (%luMB)\n", &mem->start, &mem->end, in swiotlb_print_info()
237 (mem->nslabs << IO_TLB_SHIFT) >> 20); in swiotlb_print_info()
258 struct io_tlb_pool *mem = &io_tlb_default_mem.defpool; in swiotlb_update_mem_attributes() local
261 if (!mem->nslabs || mem->late_alloc) in swiotlb_update_mem_attributes()
263 bytes = PAGE_ALIGN(mem->nslabs << IO_TLB_SHIFT); in swiotlb_update_mem_attributes()
264 set_memory_decrypted((unsigned long)mem->vaddr, bytes >> PAGE_SHIFT); in swiotlb_update_mem_attributes()
267 static void swiotlb_init_io_tlb_pool(struct io_tlb_pool *mem, phys_addr_t start, in swiotlb_init_io_tlb_pool() argument
273 mem->nslabs = nslabs; in swiotlb_init_io_tlb_pool()
[all …]
/linux/drivers/base/
H A Dmemory.c94 struct memory_block *mem = to_memory_block(dev); in memory_block_release() local
96 WARN_ON(mem->altmap); in memory_block_release()
97 kfree(mem); in memory_block_release()
161 struct memory_block *mem = to_memory_block(dev); in phys_index_show() local
163 return sysfs_emit(buf, "%08lx\n", memory_block_id(mem->start_section_nr)); in phys_index_show()
182 struct memory_block *mem = to_memory_block(dev); in state_show() local
189 switch (mem->state) { in state_show()
201 return sysfs_emit(buf, "ERROR-UNKNOWN-%d\n", mem->state); in state_show()
213 static unsigned long memblk_nr_poison(struct memory_block *mem);
215 static inline unsigned long memblk_nr_poison(struct memory_block *mem) in memblk_nr_poison() argument
[all …]
/linux/tools/perf/
H A Dbuiltin-mem.c52 struct perf_mem *mem = (struct perf_mem *)opt->value; in parse_record_events() local
68 mem->operation = 0; in parse_record_events()
72 static int __cmd_record(int argc, const char **argv, struct perf_mem *mem, in __cmd_record() argument
105 if (mem->cpu_list) in __cmd_record()
121 (mem->operation & MEM_OPERATION_LOAD) && in __cmd_record()
122 (mem->operation & MEM_OPERATION_STORE)) { in __cmd_record()
126 if (mem->operation & MEM_OPERATION_LOAD) in __cmd_record()
129 if (mem->operation & MEM_OPERATION_STORE) in __cmd_record()
138 if (mem->phys_addr) in __cmd_record()
141 if (mem->data_page_size) in __cmd_record()
[all …]
/linux/tools/testing/selftests/mm/
H A Dmkdirty.c41 static void do_test_write_sigsegv(char *mem) in do_test_write_sigsegv() argument
43 char orig = *mem; in do_test_write_sigsegv()
53 *mem = orig + 1; in do_test_write_sigsegv()
58 ksft_test_result(ret == 1 && *mem == orig, in do_test_write_sigsegv()
65 char *mem, *mmap_mem; in mmap_thp_range() local
73 mem = (char *)(((uintptr_t)mmap_mem + thpsize) & ~(thpsize - 1)); in mmap_thp_range()
75 if (madvise(mem, thpsize, MADV_HUGEPAGE)) { in mmap_thp_range()
83 return mem; in mmap_thp_range()
89 char *mem; in test_ptrace_write() local
94 mem = mmap(NULL, pagesize, PROT_READ, MAP_PRIVATE|MAP_ANON, -1, 0); in test_ptrace_write()
[all …]
H A Dpagemap_ioctl.c164 char *mem; in gethugetlb_mem() local
171 mem = shmat(*shmid, 0, 0); in gethugetlb_mem()
172 if (mem == (char *)-1) { in gethugetlb_mem()
177 mem = mmap(NULL, size, PROT_READ | PROT_WRITE, in gethugetlb_mem()
179 if (mem == MAP_FAILED) in gethugetlb_mem()
183 return mem; in gethugetlb_mem()
189 char *mem, *vec; in userfaultfd_tests() local
192 mem = mmap(NULL, mem_size, PROT_NONE, MAP_PRIVATE | MAP_ANON, -1, 0); in userfaultfd_tests()
193 if (mem == MAP_FAILED) in userfaultfd_tests()
196 wp_init(mem, mem_size); in userfaultfd_tests()
[all …]
H A Dcow.c107 static int child_memcmp_fn(char *mem, size_t size, in child_memcmp_fn() argument
114 memcpy(old, mem, size); in child_memcmp_fn()
122 return memcmp(old, mem, size); in child_memcmp_fn()
125 static int child_vmsplice_memcmp_fn(char *mem, size_t size, in child_vmsplice_memcmp_fn() argument
129 .iov_base = mem, in child_vmsplice_memcmp_fn()
141 memcpy(old, mem, size); in child_vmsplice_memcmp_fn()
154 if (munmap(mem, size) < 0) in child_vmsplice_memcmp_fn()
172 typedef int (*child_fn)(char *mem, size_t size, struct comm_pipes *comm_pipes);
174 static void do_test_cow_in_parent(char *mem, size_t size, bool do_mprotect, in do_test_cow_in_parent() argument
193 exit(fn(mem, size, &comm_pipes)); in do_test_cow_in_parent()
[all …]
H A Duffd-wp-mremap.c50 char *mmap_mem, *mem; in mmap_aligned() local
56 mem = (char *)(((uintptr_t)mmap_mem + size - 1) & ~(size - 1)); in mmap_aligned()
57 munmap(mmap_mem, mem - mmap_mem); in mmap_aligned()
58 munmap(mem + size, mmap_mem + mmap_size - mem - size); in mmap_aligned()
60 return mem; in mmap_aligned()
68 char *mem, *addr; in alloc_one_folio() local
91 mem = mmap_aligned(size, prot, flags); in alloc_one_folio()
98 mem = mmap(NULL, size, prot, flags, -1, 0); in alloc_one_folio()
101 if (mem == MAP_FAILED) { in alloc_one_folio()
102 mem = NULL; in alloc_one_folio()
[all …]
H A Dmemfd_secret.c63 char *mem; in test_mlock_limit() local
69 mem = mmap(NULL, len, prot, mode, fd, 0); in test_mlock_limit()
70 if (mem == MAP_FAILED) { in test_mlock_limit()
74 munmap(mem, len); in test_mlock_limit()
77 mem = mmap(NULL, len, prot, mode, fd, 0); in test_mlock_limit()
78 if (mem != MAP_FAILED) { in test_mlock_limit()
80 munmap(mem, len); in test_mlock_limit()
92 char *mem; in test_vmsplice() local
99 mem = mmap(NULL, page_size, prot, mode, fd, 0); in test_vmsplice()
100 if (mem == MAP_FAILED) { in test_vmsplice()
[all …]
/linux/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_amdkfd_gpuvm.c78 struct kgd_mem *mem) in kfd_mem_is_attached() argument
82 list_for_each_entry(entry, &mem->attachments, list) in kfd_mem_is_attached()
115 uint64_t mem; in amdgpu_amdkfd_gpuvm_init_mem_limits() local
121 mem = si.totalram - si.totalhigh; in amdgpu_amdkfd_gpuvm_init_mem_limits()
122 mem *= si.mem_unit; in amdgpu_amdkfd_gpuvm_init_mem_limits()
125 kfd_mem_limit.max_system_mem_limit = mem - (mem >> 6); in amdgpu_amdkfd_gpuvm_init_mem_limits()
332 struct kgd_mem *mem, struct amdgpu_bo **bo_out) in create_dmamap_sg_bo() argument
338 ret = amdgpu_bo_reserve(mem->bo, false); in create_dmamap_sg_bo()
342 if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_USERPTR) in create_dmamap_sg_bo()
343 flags |= mem->bo->flags & (AMDGPU_GEM_CREATE_COHERENT | in create_dmamap_sg_bo()
[all …]
/linux/drivers/char/hw_random/
H A Dintel-rng.c155 static inline u8 hwstatus_get(void __iomem *mem) in hwstatus_get() argument
157 return readb(mem + INTEL_RNG_HW_STATUS); in hwstatus_get()
160 static inline u8 hwstatus_set(void __iomem *mem, in hwstatus_set() argument
163 writeb(hw_status, mem + INTEL_RNG_HW_STATUS); in hwstatus_set()
164 return hwstatus_get(mem); in hwstatus_set()
169 void __iomem *mem = (void __iomem *)rng->priv; in intel_rng_data_present() local
173 data = !!(readb(mem + INTEL_RNG_STATUS) & in intel_rng_data_present()
184 void __iomem *mem = (void __iomem *)rng->priv; in intel_rng_data_read() local
186 *data = readb(mem + INTEL_RNG_DATA); in intel_rng_data_read()
193 void __iomem *mem = (void __iomem *)rng->priv; in intel_rng_init() local
[all …]
H A Dxiphera-trng.c31 void __iomem *mem; member
42 if (readl(trng->mem + STATUS_REG) == TRNG_NEW_RAND_AVAILABLE) { in xiphera_trng_read()
43 *(u32 *)buf = readl(trng->mem + RAND_REG); in xiphera_trng_read()
48 writel(HOST_TO_TRNG_READ, trng->mem + CONTROL_REG); in xiphera_trng_read()
49 writel(HOST_TO_TRNG_ENABLE, trng->mem + CONTROL_REG); in xiphera_trng_read()
70 trng->mem = devm_platform_ioremap_resource(pdev, 0); in xiphera_trng_probe()
71 if (IS_ERR(trng->mem)) in xiphera_trng_probe()
72 return PTR_ERR(trng->mem); in xiphera_trng_probe()
78 writel(HOST_TO_TRNG_RESET, trng->mem + CONTROL_REG); in xiphera_trng_probe()
81 if (readl(trng->mem + STATUS_REG) != TRNG_ACK_RESET) { in xiphera_trng_probe()
[all …]
H A Dmxc-rnga.c57 void __iomem *mem; member
68 int level = (__raw_readl(mxc_rng->mem + RNGA_STATUS) & in mxc_rnga_data_present()
84 *data = __raw_readl(mxc_rng->mem + RNGA_OUTPUT_FIFO); in mxc_rnga_data_read()
87 err = __raw_readl(mxc_rng->mem + RNGA_STATUS) & RNGA_STATUS_ERROR_INT; in mxc_rnga_data_read()
92 ctrl = __raw_readl(mxc_rng->mem + RNGA_CONTROL); in mxc_rnga_data_read()
94 mxc_rng->mem + RNGA_CONTROL); in mxc_rnga_data_read()
106 ctrl = __raw_readl(mxc_rng->mem + RNGA_CONTROL); in mxc_rnga_init()
107 __raw_writel(ctrl & ~RNGA_CONTROL_SLEEP, mxc_rng->mem + RNGA_CONTROL); in mxc_rnga_init()
110 osc = __raw_readl(mxc_rng->mem + RNGA_STATUS); in mxc_rnga_init()
117 ctrl = __raw_readl(mxc_rng->mem + RNGA_CONTROL); in mxc_rnga_init()
[all …]
/linux/tools/testing/selftests/kvm/
H A Dguest_memfd_test.c44 void *mem; in test_mmap_cow() local
46 mem = mmap(NULL, size, PROT_READ | PROT_WRITE, MAP_PRIVATE, fd, 0); in test_mmap_cow()
47 TEST_ASSERT(mem == MAP_FAILED, "Copy-on-write not allowed by guest_memfd."); in test_mmap_cow()
53 char *mem; in test_mmap_supported() local
57 mem = kvm_mmap(total_size, PROT_READ | PROT_WRITE, MAP_SHARED, fd); in test_mmap_supported()
59 memset(mem, val, total_size); in test_mmap_supported()
61 TEST_ASSERT_EQ(READ_ONCE(mem[i]), val); in test_mmap_supported()
68 TEST_ASSERT_EQ(READ_ONCE(mem[i]), 0x00); in test_mmap_supported()
70 TEST_ASSERT_EQ(READ_ONCE(mem[i]), val); in test_mmap_supported()
72 memset(mem, val, page_size); in test_mmap_supported()
[all …]
/linux/drivers/net/ipa/
H A Dipa_qmi.c285 const struct ipa_mem *mem; in init_modem_driver_req() local
298 mem = ipa_mem_find(ipa, IPA_MEM_MODEM_HEADER); in init_modem_driver_req()
299 if (mem->size) { in init_modem_driver_req()
301 req.hdr_tbl_info.start = ipa->mem_offset + mem->offset; in init_modem_driver_req()
302 req.hdr_tbl_info.end = req.hdr_tbl_info.start + mem->size - 1; in init_modem_driver_req()
305 mem = ipa_mem_find(ipa, IPA_MEM_V4_ROUTE); in init_modem_driver_req()
307 req.v4_route_tbl_info.start = ipa->mem_offset + mem->offset; in init_modem_driver_req()
310 mem = ipa_mem_find(ipa, IPA_MEM_V6_ROUTE); in init_modem_driver_req()
312 req.v6_route_tbl_info.start = ipa->mem_offset + mem->offset; in init_modem_driver_req()
315 mem = ipa_mem_find(ipa, IPA_MEM_V4_FILTER); in init_modem_driver_req()
[all …]
/linux/drivers/media/platform/mediatek/vcodec/common/
H A Dmtk_vcodec_util.c48 int mtk_vcodec_mem_alloc(void *priv, struct mtk_vcodec_mem *mem) in mtk_vcodec_mem_alloc() argument
66 mem->va = dma_alloc_attrs(&plat_dev->dev, mem->size, &mem->dma_addr, in mtk_vcodec_mem_alloc()
68 if (!mem->va) { in mtk_vcodec_mem_alloc()
70 __func__, mem->size); in mtk_vcodec_mem_alloc()
74 mtk_v4l2_debug(plat_dev, 3, "[%d] - va = %p dma = 0x%lx size = 0x%zx", id, mem->va, in mtk_vcodec_mem_alloc()
75 (unsigned long)mem->dma_addr, mem->size); in mtk_vcodec_mem_alloc()
81 void mtk_vcodec_mem_free(void *priv, struct mtk_vcodec_mem *mem) in mtk_vcodec_mem_free() argument
99 if (!mem->va) { in mtk_vcodec_mem_free()
101 if (mem->size) in mtk_vcodec_mem_free()
102 mtk_v4l2_err(plat_dev, "Failed to free %zu bytes", mem->size); in mtk_vcodec_mem_free()
[all …]
/linux/drivers/gpu/drm/nouveau/nvif/
H A Dmem.c29 struct nvif_mem *mem) in nvif_mem_ctor_map() argument
31 int ret = nvif_mem_ctor(mmu, name, mmu->mem, NVIF_MEM_MAPPABLE | type, in nvif_mem_ctor_map()
32 0, size, NULL, 0, mem); in nvif_mem_ctor_map()
34 ret = nvif_object_map(&mem->object, NULL, 0); in nvif_mem_ctor_map()
36 nvif_mem_dtor(mem); in nvif_mem_ctor_map()
42 nvif_mem_dtor(struct nvif_mem *mem) in nvif_mem_dtor() argument
44 nvif_object_dtor(&mem->object); in nvif_mem_dtor()
50 struct nvif_mem *mem) in nvif_mem_ctor_type() argument
56 mem->object.client = NULL; in nvif_mem_ctor_type()
73 args, sizeof(*args) + argc, &mem->object); in nvif_mem_ctor_type()
[all …]
/linux/drivers/uio/
H A Duio_mf624.c35 void __iomem *INTCSR_reg = info->mem[0].internal_addr + INTCSR; in mf624_disable_interrupt()
63 void __iomem *INTCSR_reg = info->mem[0].internal_addr + INTCSR; in mf624_enable_interrupt()
90 void __iomem *INTCSR_reg = info->mem[0].internal_addr + INTCSR; in mf624_irq_handler()
117 static int mf624_setup_mem(struct pci_dev *dev, int bar, struct uio_mem *mem, const char *name) in mf624_setup_mem() argument
122 mem->name = name; in mf624_setup_mem()
123 mem->addr = start & PAGE_MASK; in mf624_setup_mem()
124 mem->offs = start & ~PAGE_MASK; in mf624_setup_mem()
125 if (!mem->addr) in mf624_setup_mem()
127 mem->size = ((start & ~PAGE_MASK) + len + PAGE_SIZE - 1) & PAGE_MASK; in mf624_setup_mem()
128 mem->memtype = UIO_MEM_PHYS; in mf624_setup_mem()
[all …]
/linux/drivers/leds/
H A Dleds-bcm6328.c70 void __iomem *mem; member
117 mode = led->mem + BCM6328_REG_MODE_HI; in bcm6328_led_mode()
119 mode = led->mem + BCM6328_REG_MODE_LO; in bcm6328_led_mode()
212 val = bcm6328_led_read(led->mem + BCM6328_REG_INIT); in bcm6328_blink_set()
215 bcm6328_led_write(led->mem + BCM6328_REG_INIT, val); in bcm6328_blink_set()
236 val = bcm6328_led_read(led->mem + BCM6328_REG_INIT); in bcm6328_blink_set()
239 bcm6328_led_write(led->mem + BCM6328_REG_INIT, val); in bcm6328_blink_set()
256 void __iomem *mem, spinlock_t *lock) in bcm6328_hwled() argument
262 val = bcm6328_led_read(mem + BCM6328_REG_HWDIS); in bcm6328_hwled()
264 bcm6328_led_write(mem + BCM6328_REG_HWDIS, val); in bcm6328_hwled()
[all …]
/linux/kernel/
H A Dcrash_core.c167 int crash_prepare_elf64_headers(struct crash_mem *mem, int need_kernel_map, in crash_prepare_elf64_headers() argument
180 nr_phdr += mem->nr_ranges; in crash_prepare_elf64_headers()
242 for (i = 0; i < mem->nr_ranges; i++) { in crash_prepare_elf64_headers()
243 mstart = mem->ranges[i].start; in crash_prepare_elf64_headers()
244 mend = mem->ranges[i].end; in crash_prepare_elf64_headers()
282 int crash_exclude_mem_range(struct crash_mem *mem, in crash_exclude_mem_range() argument
288 for (i = 0; i < mem->nr_ranges; i++) { in crash_exclude_mem_range()
289 start = mem->ranges[i].start; in crash_exclude_mem_range()
290 end = mem->ranges[i].end; in crash_exclude_mem_range()
315 memmove(&mem->ranges[i], &mem->ranges[i + 1], in crash_exclude_mem_range()
[all …]

12345678910>>...52