Home
last modified time | relevance | path

Searched refs:vmap (Results 1 – 25 of 151) sorted by relevance

1234567

/linux/drivers/gpu/drm/nouveau/nvkm/subdev/bios/
H A Dvmap.c32 u32 vmap = 0; in nvbios_vmap_table() local
36 vmap = nvbios_rd32(bios, bit_P.offset + 0x20); in nvbios_vmap_table()
37 if (vmap) { in nvbios_vmap_table()
38 *ver = nvbios_rd08(bios, vmap + 0); in nvbios_vmap_table()
42 *hdr = nvbios_rd08(bios, vmap + 1); in nvbios_vmap_table()
43 *cnt = nvbios_rd08(bios, vmap + 3); in nvbios_vmap_table()
44 *len = nvbios_rd08(bios, vmap + 2); in nvbios_vmap_table()
45 return vmap; in nvbios_vmap_table()
60 u32 vmap = nvbios_vmap_table(bios, ver, hdr, cnt, len); in nvbios_vmap_parse() local
62 switch (!!vmap * *ver) { in nvbios_vmap_parse()
[all …]
/linux/drivers/gpu/drm/xe/display/
H A Dxe_panic.c16 struct iosys_map vmap; member
23 if (!panic->vmap.is_iomem && iosys_map_is_set(&panic->vmap)) { in xe_panic_kunmap()
24 drm_clflush_virt_range(panic->vmap.vaddr, PAGE_SIZE); in xe_panic_kunmap()
25 kunmap_local(panic->vmap.vaddr); in xe_panic_kunmap()
27 iosys_map_clear(&panic->vmap); in xe_panic_kunmap()
63 iosys_map_set_vaddr_iomem(&panic->vmap, in xe_panic_page_set_pixel()
67 iosys_map_set_vaddr(&panic->vmap, in xe_panic_page_set_pixel()
74 if (iosys_map_is_set(&panic->vmap)) in xe_panic_page_set_pixel()
75 iosys_map_wr(&panic->vmap, offset, u32, color); in xe_panic_page_set_pixel()
H A Dxe_hdcp_gsc.c81 xe_map_memset(xe, &bo->vmap, 0, 0, xe_bo_size(bo)); in intel_hdcp_gsc_initialize_message()
131 struct iosys_map *map = &gsc_context->hdcp_bo->vmap; in xe_gsc_send_sync()
171 addr_in_wr_off = xe_gsc_emit_header(xe, &gsc_context->hdcp_bo->vmap, in intel_hdcp_gsc_msg_send()
174 xe_map_memcpy_to(xe, &gsc_context->hdcp_bo->vmap, addr_in_wr_off, in intel_hdcp_gsc_msg_send()
197 xe_map_memcpy_from(xe, msg_out, &gsc_context->hdcp_bo->vmap, in intel_hdcp_gsc_msg_send()
H A Dxe_dsb_buffer.c24 iosys_map_wr(&dsb_buf->bo->vmap, idx * 4, u32, val); in intel_dsb_buffer_write()
29 return iosys_map_rd(&dsb_buf->bo->vmap, idx * 4, u32); in intel_dsb_buffer_read()
36 iosys_map_memset(&dsb_buf->bo->vmap, idx * 4, val, size); in intel_dsb_buffer_memset()
/linux/drivers/net/
H A Dvrf.c110 struct vrf_map vmap; member
136 return &nn_vrf->vmap; in netns_vrf_map()
182 static struct vrf_map_elem *vrf_map_lookup_elem(struct vrf_map *vmap, in vrf_map_lookup_elem() argument
189 hash_for_each_possible(vmap->ht, me, hnode, key) { in vrf_map_lookup_elem()
197 static void vrf_map_add_elem(struct vrf_map *vmap, struct vrf_map_elem *me) in vrf_map_add_elem() argument
203 hash_add(vmap->ht, &me->hnode, key); in vrf_map_add_elem()
211 static void vrf_map_lock(struct vrf_map *vmap) __acquires(&vmap->vmap_lock) in vrf_map_lock() argument
213 spin_lock(&vmap->vmap_lock); in vrf_map_lock()
216 static void vrf_map_unlock(struct vrf_map *vmap) __releases(&vmap->vmap_lock) in vrf_map_unlock() argument
218 spin_unlock(&vmap->vmap_lock); in vrf_map_unlock()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/volt/
H A Dbase.c87 u32 vmap; in nvkm_volt_map_min() local
89 vmap = nvbios_vmap_entry_parse(bios, id, &ver, &len, &info); in nvkm_volt_map_min()
90 if (vmap) { in nvkm_volt_map_min()
109 u32 vmap; in nvkm_volt_map() local
111 vmap = nvbios_vmap_entry_parse(bios, id, &ver, &len, &info); in nvkm_volt_map()
112 if (vmap) { in nvkm_volt_map()
295 struct nvbios_vmap vmap; in nvkm_volt_ctor() local
301 if (nvbios_vmap_parse(bios, &ver, &hdr, &cnt, &len, &vmap)) { in nvkm_volt_ctor()
302 volt->max0_id = vmap.max0; in nvkm_volt_ctor()
303 volt->max1_id = vmap.max1; in nvkm_volt_ctor()
[all …]
/linux/mm/
H A Dvma_exec.c107 int create_init_stack_vma(struct mm_struct *mm, struct vm_area_struct **vmap, in create_init_stack_vma() argument
152 *vmap = vma; in create_init_stack_vma()
161 *vmap = NULL; in create_init_stack_vma()
/linux/drivers/gpu/drm/xe/
H A Dxe_sriov_vf_ccs.c164 xe_map_memset(xe, &sa_manager->bo->vmap, offset, MI_NOOP, in alloc_bb_pool()
166 xe_map_memset(xe, &sa_manager->shadow->vmap, offset, MI_NOOP, in alloc_bb_pool()
170 xe_map_wr(xe, &sa_manager->bo->vmap, offset, u32, MI_BATCH_BUFFER_END); in alloc_bb_pool()
171 xe_map_wr(xe, &sa_manager->shadow->vmap, offset, u32, MI_BATCH_BUFFER_END); in alloc_bb_pool()
396 xe_map_wr(xe, &lrc->bo->vmap, XE_SRIOV_VF_CCS_RW_BB_ADDR_OFFSET, u32, addr); in xe_sriov_vf_ccs_rw_update_bb_addr()
H A Dxe_lmtt.c87 xe_map_memset(lmtt_to_xe(lmtt), &bo->vmap, 0, 0, xe_bo_size(bo)); in lmtt_pt_alloc()
310 lmtt_assert(lmtt, !pte || !iosys_map_rd(&pt->bo->vmap, idx * sizeof(u32), u32)); in lmtt_write_pte()
312 xe_map_wr(lmtt_to_xe(lmtt), &pt->bo->vmap, idx * sizeof(u32), u32, pte); in lmtt_write_pte()
315 lmtt_assert(lmtt, !pte || !iosys_map_rd(&pt->bo->vmap, idx * sizeof(u64), u64)); in lmtt_write_pte()
317 xe_map_wr(lmtt_to_xe(lmtt), &pt->bo->vmap, idx * sizeof(u64), u64, pte); in lmtt_write_pte()
H A Dxe_bo.c1296 if (iosys_map_is_null(&bo->vmap)) { in xe_bo_evict_pinned_copy()
1303 xe_map_memcpy_from(xe, backup->vmap.vaddr, &bo->vmap, 0, in xe_bo_evict_pinned_copy()
1446 if (iosys_map_is_null(&bo->vmap)) { in xe_bo_restore_pinned()
1453 xe_map_memcpy_to(xe, &bo->vmap, 0, backup->vmap.vaddr, in xe_bo_restore_pinned()
1640 struct iosys_map vmap; in xe_ttm_access_memory() local
1669 iosys_map_set_vaddr_iomem(&vmap, (u8 __iomem *)vram->mapping + in xe_ttm_access_memory()
1672 xe_map_memcpy_to(xe, &vmap, page_offset, buf, byte_count); in xe_ttm_access_memory()
1674 xe_map_memcpy_from(xe, buf, &vmap, page_offset, byte_count); in xe_ttm_access_memory()
2700 xe_map_memcpy_to(xe, &bo->vmap, 0, data, size); in xe_managed_bo_create_from_data()
2727 xe_assert(xe, !(*src)->vmap.is_iomem); in xe_managed_bo_reinit_in_vram()
[all …]
/linux/kernel/dma/
H A Dremap.c29 vaddr = vmap(pages, PAGE_ALIGN(size) >> PAGE_SHIFT, in dma_common_pages_remap()
53 vaddr = vmap(pages, count, VM_DMA_COHERENT, prot); in dma_common_contiguous_remap()
/linux/tools/testing/selftests/net/
H A Dtest_vxlan_vnifiltering.sh221 for vmap in $(echo $vattrs | cut -d "," -f1- --output-delimiter=' ')
223 local vid=$(echo $vmap | awk -F'-' '{print ($1)}')
224 local family=$(echo $vmap | awk -F'-' '{print ($2)}')
225 local localip=$(echo $vmap | awk -F'-' '{print ($3)}')
226 local group=$(echo $vmap | awk -F'-' '{print ($4)}')
227 local vtype=$(echo $vmap | awk -F'-' '{print ($5)}')
228 local port=$(echo $vmap | awk -F'-' '{print ($6)}')
/linux/Documentation/translations/zh_CN/mm/
H A Dvmalloced-kernel-stacks.rst48 - vmalloc空间的堆栈需要可靠地工作。例如,如果vmap页表是按需创建的,当堆栈指向
91 工作。架构特定的vmap堆栈分配器照顾到了这个细节。
H A Dhighmem.rst114 * vmap()。这可以用来将多个物理页长期映射到一个连续的虚拟空间。它需要全局同步来解除
/linux/drivers/gpu/drm/tiny/
H A Dsharp-memory.c129 const struct iosys_map *vmap, in sharp_memory_set_tx_buffer_data() argument
143 drm_fb_xrgb8888_to_mono(&dst, &pitch, vmap, fb, &clip, fmtcnv_state); in sharp_memory_set_tx_buffer_data()
150 const struct iosys_map *vmap, in sharp_memory_update_display() argument
166 sharp_memory_set_tx_buffer_data(&tx_buffer[2], fb, vmap, clip, pitch, fmtcnv_state); in sharp_memory_update_display()
209 static void sharp_memory_fb_dirty(struct drm_framebuffer *fb, const struct iosys_map *vmap, in sharp_memory_fb_dirty() argument
222 sharp_memory_update_display(smd, fb, vmap, clip, fmtconv_state); in sharp_memory_fb_dirty()
/linux/drivers/gpu/drm/etnaviv/
H A Detnaviv_gem.c337 etnaviv_obj->vaddr = etnaviv_obj->ops->vmap(etnaviv_obj); in etnaviv_gem_vmap()
366 return vmap(pages, obj->base.size >> PAGE_SHIFT, VM_MAP, prot); in etnaviv_gem_vmap_impl()
497 .vmap = etnaviv_gem_vmap_impl,
566 .vmap = etnaviv_gem_prime_vmap,
730 .vmap = etnaviv_gem_vmap_impl,
/linux/drivers/gpu/drm/sitronix/
H A Dst7586.c75 struct iosys_map dst_map, vmap; in st7586_xrgb8888_to_gray332() local
82 iosys_map_set_vaddr(&vmap, vaddr); in st7586_xrgb8888_to_gray332()
83 drm_fb_xrgb8888_to_gray8(&dst_map, NULL, &vmap, fb, clip, fmtcnv_state); in st7586_xrgb8888_to_gray332()
/linux/arch/arm/mm/
H A Dfault-armv.c253 p1 = vmap(&page, 1, VM_IOREMAP, prot); in check_writebuffer_bugs()
254 p2 = vmap(&page, 1, VM_IOREMAP, prot); in check_writebuffer_bugs()
/linux/drivers/gpu/drm/xen/
H A Dxen_drm_front_gem.c108 .vmap = xen_drm_front_gem_prime_vmap,
292 vaddr = vmap(xen_obj->pages, xen_obj->num_pages, in xen_drm_front_gem_prime_vmap()
/linux/Documentation/features/vm/huge-vmap/
H A Darch-support.txt2 # Feature name: huge-vmap
/linux/drivers/accel/amdxdna/
H A Damdxdna_ubuf.c113 kva = vmap(ubuf->pages, ubuf->nr_pages, VM_MAP, PAGE_KERNEL); in amdxdna_ubuf_vmap()
131 .vmap = amdxdna_ubuf_vmap,
/linux/drivers/gpu/drm/rockchip/
H A Drockchip_drm_gem.c139 rk_obj->kvaddr = vmap(rk_obj->pages, rk_obj->num_pages, VM_MAP, in rockchip_gem_alloc_iommu()
278 .vmap = rockchip_gem_prime_vmap,
524 vaddr = vmap(rk_obj->pages, rk_obj->num_pages, VM_MAP, in rockchip_gem_prime_vmap()
/linux/Documentation/core-api/
H A Dcachetlb.rst376 vmap/vmalloc API. Since kernel I/O goes via physical pages, the I/O
378 the only aliases. This isn't true for vmap aliases, so anything in
379 the kernel trying to do I/O to vmap areas must manually manage
380 coherency. It must do this by flushing the vmap range before doing
386 the vmap area. This is to make sure that any data the kernel
387 modified in the vmap range is made visible to the physical
394 the cache for a given virtual address range in the vmap area
398 vmap area.
/linux/drivers/gpu/drm/v3d/
H A Dv3d_bo.c71 .vmap = drm_gem_shmem_object_vmap,
198 bo->vaddr = vmap(obj->pages, obj->base.size >> PAGE_SHIFT, VM_MAP, in v3d_get_bo_vaddr()
/linux/arch/hexagon/kernel/
H A Dvdso.c29 vdso = vmap(&vdso_page, 1, 0, PAGE_KERNEL); in vdso_init()

1234567