Home
last modified time | relevance | path

Searched refs:aspace (Results 1 – 25 of 37) sorted by relevance

12

/linux/drivers/gpu/drm/msm/
H A Dmsm_gem_vma.c15 struct msm_gem_address_space *aspace = container_of(kref, in msm_gem_address_space_destroy() local
18 drm_mm_takedown(&aspace->mm); in msm_gem_address_space_destroy()
19 if (aspace->mmu) in msm_gem_address_space_destroy()
20 aspace->mmu->funcs->destroy(aspace->mmu); in msm_gem_address_space_destroy()
21 put_pid(aspace->pid); in msm_gem_address_space_destroy()
22 kfree(aspace); in msm_gem_address_space_destroy()
26 void msm_gem_address_space_put(struct msm_gem_address_space *aspace) in msm_gem_address_space_put() argument
28 if (aspace) in msm_gem_address_space_put()
29 kref_put(&aspace->kref, msm_gem_address_space_destroy); in msm_gem_address_space_put()
33 msm_gem_address_space_get(struct msm_gem_address_space *aspace) in msm_gem_address_space_get() argument
[all …]
H A Dmsm_gem.h51 msm_gem_address_space_get(struct msm_gem_address_space *aspace);
53 void msm_gem_address_space_put(struct msm_gem_address_space *aspace);
64 struct msm_gem_address_space *aspace; member
69 struct msm_gem_vma *msm_gem_vma_new(struct msm_gem_address_space *aspace);
130 struct msm_gem_address_space *aspace);
132 struct msm_gem_address_space *aspace, uint64_t *iova);
134 struct msm_gem_address_space *aspace, uint64_t iova);
136 struct msm_gem_address_space *aspace, uint64_t *iova,
139 struct msm_gem_address_space *aspace, uint64_t *iova);
141 struct msm_gem_address_space *aspace);
[all …]
H A Dmsm_gem.c401 struct msm_gem_address_space *aspace) in add_vma() argument
408 vma = msm_gem_vma_new(aspace); in add_vma()
418 struct msm_gem_address_space *aspace) in lookup_vma() argument
426 if (vma->aspace == aspace) in lookup_vma()
457 if (vma->aspace) { in put_iova_spaces()
480 struct msm_gem_address_space *aspace, in get_vma_locked() argument
487 vma = lookup_vma(obj, aspace); in get_vma_locked()
492 vma = add_vma(obj, aspace); in get_vma_locked()
564 struct msm_gem_address_space *aspace) in msm_gem_get_vma_locked() argument
566 return get_vma_locked(obj, aspace, 0, U64_MAX); in msm_gem_get_vma_locked()
[all …]
H A Dmsm_fb.c79 struct msm_gem_address_space *aspace, in msm_framebuffer_prepare() argument
91 ret = msm_gem_get_and_pin_iova(fb->obj[i], aspace, &msm_fb->iova[i]); in msm_framebuffer_prepare()
102 struct msm_gem_address_space *aspace, in msm_framebuffer_cleanup() argument
112 msm_gem_unpin_iova(fb->obj[i], aspace); in msm_framebuffer_cleanup()
119 struct msm_gem_address_space *aspace, int plane) in msm_framebuffer_iova() argument
H A Dmsm_kms.c169 struct msm_gem_address_space *aspace; in msm_kms_init_aspace() local
193 aspace = msm_gem_address_space_create(mmu, "mdp_kms", in msm_kms_init_aspace()
195 if (IS_ERR(aspace)) { in msm_kms_init_aspace()
196 dev_err(mdp_dev, "aspace create, error %pe\n", aspace); in msm_kms_init_aspace()
200 return aspace; in msm_kms_init_aspace()
H A Dmsm_drv.h266 struct msm_gem_address_space *aspace, bool needs_dirtyfb);
268 struct msm_gem_address_space *aspace, bool needed_dirtyfb);
270 struct msm_gem_address_space *aspace, int plane);
H A Dmsm_drv.c348 ctx->aspace = msm_gpu_create_private_address_space(priv->gpu, current); in context_init()
526 return msm_gem_get_iova(obj, ctx->aspace, iova); in msm_ioctl_gem_info_iova()
540 if (priv->gpu->aspace == ctx->aspace) in msm_ioctl_gem_info_set_iova()
546 return msm_gem_set_iova(obj, ctx->aspace, iova); in msm_ioctl_gem_info_set_iova()
H A Dmsm_gem_submit.c66 submit->aspace = queue->ctx->aspace; in submit_create()
305 vma = msm_gem_get_vma_locked(obj, submit->aspace); in submit_pin_objects()
662 if (unlikely(!ctx->aspace) && !capable(CAP_SYS_RAWIO)) { in msm_ioctl_gem_submit()
/linux/drivers/staging/vme_user/
H A Dvme_fake.c48 u32 aspace; member
56 u32 aspace; member
156 dma_addr_t buf_base, u32 aspace, u32 cycle) in fake_slave_set() argument
168 switch (aspace) { in fake_slave_set()
212 bridge->slaves[i].aspace = aspace; in fake_slave_set()
225 dma_addr_t *buf_base, u32 *aspace, u32 *cycle) in fake_slave_get() argument
240 *aspace = bridge->slaves[i].aspace; in fake_slave_get()
253 u32 aspace, u32 cycle, u32 dwidth) in fake_master_set() argument
296 switch (aspace) { in fake_master_set()
320 bridge->masters[i].aspace = aspace; in fake_master_set()
[all …]
H A Dvme.c137 u32 aspace, cycle, dwidth; in vme_get_size() local
142 &aspace, &cycle, &dwidth); in vme_get_size()
149 &buf_base, &aspace, &cycle); in vme_get_size()
163 int vme_check_window(struct vme_bridge *bridge, u32 aspace, in vme_check_window() argument
171 switch (aspace) { in vme_check_window()
329 dma_addr_t buf_base, u32 aspace, u32 cycle) in vme_slave_set() argument
347 if (!(((image->address_attr & aspace) == aspace) && in vme_slave_set()
353 retval = vme_check_window(bridge, aspace, vme_base, size); in vme_slave_set()
358 aspace, cycle); in vme_slave_set()
379 dma_addr_t *buf_base, u32 *aspace, u32 *cycle) in vme_slave_get() argument
[all …]
H A Dvme_tsi148.c468 dma_addr_t pci_base, u32 aspace, u32 cycle) in tsi148_slave_set() argument
484 switch (aspace) { in tsi148_slave_set()
611 dma_addr_t *pci_base, u32 *aspace, u32 *cycle) in tsi148_slave_get() argument
649 *aspace = 0; in tsi148_slave_get()
657 *aspace |= VME_A16; in tsi148_slave_get()
661 *aspace |= VME_A24; in tsi148_slave_get()
665 *aspace |= VME_A32; in tsi148_slave_get()
669 *aspace |= VME_A64; in tsi148_slave_get()
799 u32 aspace, u32 cycle, u32 dwidth) in tsi148_master_set()
950 switch (aspace) { in tsi148_master_set()
800 tsi148_master_set(struct vme_master_resource * image,int enabled,unsigned long long vme_base,unsigned long long size,u32 aspace,u32 cycle,u32 dwidth) tsi148_master_set() argument
1035 __tsi148_master_get(struct vme_master_resource * image,int * enabled,unsigned long long * vme_base,unsigned long long * size,u32 * aspace,u32 * cycle,u32 * dwidth) __tsi148_master_get() argument
1144 tsi148_master_get(struct vme_master_resource * image,int * enabled,unsigned long long * vme_base,unsigned long long * size,u32 * aspace,u32 * cycle,u32 * dwidth) tsi148_master_get() argument
1163 u32 aspace, cycle, dwidth; tsi148_master_read() local
1248 u32 aspace, cycle, dwidth; tsi148_master_write() local
1401 tsi148_dma_set_vme_src_attributes(struct device * dev,__be32 * attr,u32 aspace,u32 cycle,u32 dwidth) tsi148_dma_set_vme_src_attributes() argument
1499 tsi148_dma_set_vme_dest_attributes(struct device * dev,__be32 * attr,u32 aspace,u32 cycle,u32 dwidth) tsi148_dma_set_vme_dest_attributes() argument
1902 tsi148_lm_set(struct vme_lm_resource * lm,unsigned long long lm_base,u32 aspace,u32 cycle) tsi148_lm_set() argument
1967 tsi148_lm_get(struct vme_lm_resource * lm,unsigned long long * lm_base,u32 * aspace,u32 * cycle) tsi148_lm_get() argument
[all...]
H A Dvme_user.h14 __u32 aspace; /* Address Space */ member
36 __u32 aspace; /* Address Space */ member
H A Dvme_user.c332 &master.size, &master.aspace, in vme_user_ioctl()
362 master.aspace, master.cycle, master.dwidth); in vme_user_ioctl()
378 &slave.aspace, &slave.cycle); in vme_user_ioctl()
402 image[minor].pci_buf, slave.aspace, in vme_user_ioctl()
/linux/drivers/gpu/drm/msm/disp/mdp4/
H A Dmdp4_kms.c123 struct msm_gem_address_space *aspace = kms->aspace; in mdp4_destroy() local
126 msm_gem_unpin_iova(mdp4_kms->blank_cursor_bo, kms->aspace); in mdp4_destroy()
129 if (aspace) { in mdp4_destroy()
130 aspace->mmu->funcs->detach(aspace->mmu); in mdp4_destroy()
131 msm_gem_address_space_put(aspace); in mdp4_destroy()
383 struct msm_gem_address_space *aspace; in mdp4_kms_init() local
452 aspace = NULL; in mdp4_kms_init()
454 aspace = msm_gem_address_space_create(mmu, in mdp4_kms_init()
457 if (IS_ERR(aspace)) { in mdp4_kms_init()
460 ret = PTR_ERR(aspace); in mdp4_kms_init()
[all …]
H A Dmdp4_plane.c90 return msm_framebuffer_prepare(new_state->fb, kms->aspace, false); in mdp4_plane_prepare_fb()
105 msm_framebuffer_cleanup(fb, kms->aspace, false); in mdp4_plane_cleanup_fb()
156 msm_framebuffer_iova(fb, kms->aspace, 0)); in mdp4_plane_set_scanout()
158 msm_framebuffer_iova(fb, kms->aspace, 1)); in mdp4_plane_set_scanout()
160 msm_framebuffer_iova(fb, kms->aspace, 2)); in mdp4_plane_set_scanout()
162 msm_framebuffer_iova(fb, kms->aspace, 3)); in mdp4_plane_set_scanout()
/linux/drivers/gpu/drm/msm/adreno/
H A Dadreno_gpu.c208 struct msm_gem_address_space *aspace; in adreno_iommu_create_address_space() local
227 aspace = msm_gem_address_space_create(mmu, "gpu", in adreno_iommu_create_address_space()
230 if (IS_ERR(aspace) && !IS_ERR(mmu)) in adreno_iommu_create_address_space()
233 return aspace; in adreno_iommu_create_address_space()
265 gpu->aspace->mmu->funcs->resume_translation(gpu->aspace->mmu); in adreno_fault_handler()
359 if (ctx->aspace) in adreno_get_param()
360 *value = gpu->global_faults + ctx->aspace->faults; in adreno_get_param()
368 if (ctx->aspace == gpu->aspace) in adreno_get_param()
370 *value = ctx->aspace->va_start; in adreno_get_param()
373 if (ctx->aspace == gpu->aspace) in adreno_get_param()
[all …]
H A Da5xx_preempt.c257 MSM_BO_WC | MSM_BO_MAP_PRIV, gpu->aspace, &bo, &iova); in preempt_init_ring()
265 MSM_BO_WC, gpu->aspace, &counters_bo, &counters_iova); in preempt_init_ring()
267 msm_gem_kernel_put(bo, gpu->aspace); in preempt_init_ring()
298 msm_gem_kernel_put(a5xx_gpu->preempt_bo[i], gpu->aspace); in a5xx_preempt_fini()
299 msm_gem_kernel_put(a5xx_gpu->preempt_counters_bo[i], gpu->aspace); in a5xx_preempt_fini()
H A Da6xx_gmu.c1264 msm_gem_kernel_put(gmu->hfi.obj, gmu->aspace); in a6xx_gmu_memory_free()
1265 msm_gem_kernel_put(gmu->debug.obj, gmu->aspace); in a6xx_gmu_memory_free()
1266 msm_gem_kernel_put(gmu->icache.obj, gmu->aspace); in a6xx_gmu_memory_free()
1267 msm_gem_kernel_put(gmu->dcache.obj, gmu->aspace); in a6xx_gmu_memory_free()
1268 msm_gem_kernel_put(gmu->dummy.obj, gmu->aspace); in a6xx_gmu_memory_free()
1269 msm_gem_kernel_put(gmu->log.obj, gmu->aspace); in a6xx_gmu_memory_free()
1271 gmu->aspace->mmu->funcs->detach(gmu->aspace->mmu); in a6xx_gmu_memory_free()
1272 msm_gem_address_space_put(gmu->aspace); in a6xx_gmu_memory_free()
1301 ret = msm_gem_get_and_pin_iova_range(bo->obj, gmu->aspace, &bo->iova, in a6xx_gmu_memory_alloc()
1326 gmu->aspace = msm_gem_address_space_create(mmu, "gmu", 0x0, 0x80000000); in a6xx_gmu_memory_probe()
[all …]
H A Da5xx_debugfs.c119 msm_gem_unpin_iova(a5xx_gpu->pm4_bo, gpu->aspace); in reset_set()
125 msm_gem_unpin_iova(a5xx_gpu->pfp_bo, gpu->aspace); in reset_set()
H A Da5xx_gpu.c623 gpu->aspace, &a5xx_gpu->shadow_bo, in a5xx_ucode_load()
1043 msm_gem_unpin_iova(a5xx_gpu->pm4_bo, gpu->aspace); in a5xx_destroy()
1048 msm_gem_unpin_iova(a5xx_gpu->pfp_bo, gpu->aspace); in a5xx_destroy()
1053 msm_gem_unpin_iova(a5xx_gpu->gpmu_bo, gpu->aspace); in a5xx_destroy()
1058 msm_gem_unpin_iova(a5xx_gpu->shadow_bo, gpu->aspace); in a5xx_destroy()
1458 SZ_1M, MSM_BO_WC, gpu->aspace, in a5xx_crashdumper_init()
1558 msm_gem_kernel_put(dumper.bo, gpu->aspace); in a5xx_gpu_state_get_hlsq_regs()
1566 msm_gem_kernel_put(dumper.bo, gpu->aspace); in a5xx_gpu_state_get_hlsq_regs()
1787 if (gpu->aspace) in a5xx_gpu_init()
1788 msm_mmu_set_fault_handler(gpu->aspace->mmu, gpu, a5xx_fault_handler); in a5xx_gpu_init()
/linux/drivers/gpu/drm/msm/disp/mdp5/
H A Dmdp5_kms.c201 struct msm_gem_address_space *aspace = kms->aspace; in mdp5_kms_destroy() local
203 if (aspace) { in mdp5_kms_destroy()
204 aspace->mmu->funcs->detach(aspace->mmu); in mdp5_kms_destroy()
205 msm_gem_address_space_put(aspace); in mdp5_kms_destroy()
503 struct msm_gem_address_space *aspace; in mdp5_kms_init() local
537 aspace = msm_kms_init_aspace(mdp5_kms->dev); in mdp5_kms_init()
538 if (IS_ERR(aspace)) { in mdp5_kms_init()
539 ret = PTR_ERR(aspace); in mdp5_kms_init()
543 kms->aspace = aspace; in mdp5_kms_init()
H A Dmdp5_plane.c147 return msm_framebuffer_prepare(new_state->fb, kms->aspace, needs_dirtyfb); in mdp5_plane_prepare_fb()
162 msm_framebuffer_cleanup(fb, kms->aspace, needed_dirtyfb); in mdp5_plane_cleanup_fb()
481 msm_framebuffer_iova(fb, kms->aspace, 0)); in set_scanout_locked()
483 msm_framebuffer_iova(fb, kms->aspace, 1)); in set_scanout_locked()
485 msm_framebuffer_iova(fb, kms->aspace, 2)); in set_scanout_locked()
487 msm_framebuffer_iova(fb, kms->aspace, 3)); in set_scanout_locked()
/linux/drivers/gpu/drm/msm/disp/dpu1/
H A Ddpu_encoder_phys_wb.c561 struct msm_gem_address_space *aspace; in dpu_encoder_phys_wb_prepare_wb_job() local
571 aspace = phys_enc->dpu_kms->base.aspace; in dpu_encoder_phys_wb_prepare_wb_job()
577 ret = msm_framebuffer_prepare(job->fb, aspace, false); in dpu_encoder_phys_wb_prepare_wb_job()
591 dpu_format_populate_addrs(aspace, job->fb, &wb_cfg->dest); in dpu_encoder_phys_wb_prepare_wb_job()
614 struct msm_gem_address_space *aspace; in dpu_encoder_phys_wb_cleanup_wb_job() local
619 aspace = phys_enc->dpu_kms->base.aspace; in dpu_encoder_phys_wb_cleanup_wb_job()
621 msm_framebuffer_cleanup(job->fb, aspace, false); in dpu_encoder_phys_wb_cleanup_wb_job()
H A Ddpu_kms.c1081 if (!dpu_kms->base.aspace) in _dpu_kms_mmu_destroy()
1084 mmu = dpu_kms->base.aspace->mmu; in _dpu_kms_mmu_destroy()
1087 msm_gem_address_space_put(dpu_kms->base.aspace); in _dpu_kms_mmu_destroy()
1089 dpu_kms->base.aspace = NULL; in _dpu_kms_mmu_destroy()
1094 struct msm_gem_address_space *aspace; in _dpu_kms_mmu_init() local
1096 aspace = msm_kms_init_aspace(dpu_kms->dev); in _dpu_kms_mmu_init()
1097 if (IS_ERR(aspace)) in _dpu_kms_mmu_init()
1098 return PTR_ERR(aspace); in _dpu_kms_mmu_init()
1100 dpu_kms->base.aspace = aspace; in _dpu_kms_mmu_init()
H A Ddpu_plane.h37 struct msm_gem_address_space *aspace; member

12