/linux/drivers/gpu/drm/msm/ |
H A D | msm_gem_vma.c | 15 struct msm_gem_address_space *aspace = container_of(kref, in msm_gem_address_space_destroy() local 18 drm_mm_takedown(&aspace->mm); in msm_gem_address_space_destroy() 19 if (aspace->mmu) in msm_gem_address_space_destroy() 20 aspace->mmu->funcs->destroy(aspace->mmu); in msm_gem_address_space_destroy() 21 put_pid(aspace->pid); in msm_gem_address_space_destroy() 22 kfree(aspace); in msm_gem_address_space_destroy() 26 void msm_gem_address_space_put(struct msm_gem_address_space *aspace) in msm_gem_address_space_put() argument 28 if (aspace) in msm_gem_address_space_put() 29 kref_put(&aspace->kref, msm_gem_address_space_destroy); in msm_gem_address_space_put() 33 msm_gem_address_space_get(struct msm_gem_address_space *aspace) in msm_gem_address_space_get() argument [all …]
|
H A D | msm_gem.h | 51 msm_gem_address_space_get(struct msm_gem_address_space *aspace); 53 void msm_gem_address_space_put(struct msm_gem_address_space *aspace); 64 struct msm_gem_address_space *aspace; member 69 struct msm_gem_vma *msm_gem_vma_new(struct msm_gem_address_space *aspace); 130 struct msm_gem_address_space *aspace); 132 struct msm_gem_address_space *aspace, uint64_t *iova); 134 struct msm_gem_address_space *aspace, uint64_t iova); 136 struct msm_gem_address_space *aspace, uint64_t *iova, 139 struct msm_gem_address_space *aspace, uint64_t *iova); 141 struct msm_gem_address_space *aspace); [all …]
|
H A D | msm_gem.c | 401 struct msm_gem_address_space *aspace) in add_vma() argument 408 vma = msm_gem_vma_new(aspace); in add_vma() 418 struct msm_gem_address_space *aspace) in lookup_vma() argument 426 if (vma->aspace == aspace) in lookup_vma() 457 if (vma->aspace) { in put_iova_spaces() 480 struct msm_gem_address_space *aspace, in get_vma_locked() argument 487 vma = lookup_vma(obj, aspace); in get_vma_locked() 492 vma = add_vma(obj, aspace); in get_vma_locked() 564 struct msm_gem_address_space *aspace) in msm_gem_get_vma_locked() argument 566 return get_vma_locked(obj, aspace, 0, U64_MAX); in msm_gem_get_vma_locked() [all …]
|
H A D | msm_fb.c | 79 struct msm_gem_address_space *aspace, in msm_framebuffer_prepare() argument 91 ret = msm_gem_get_and_pin_iova(fb->obj[i], aspace, &msm_fb->iova[i]); in msm_framebuffer_prepare() 102 struct msm_gem_address_space *aspace, in msm_framebuffer_cleanup() argument 112 msm_gem_unpin_iova(fb->obj[i], aspace); in msm_framebuffer_cleanup() 119 struct msm_gem_address_space *aspace, int plane) in msm_framebuffer_iova() argument
|
H A D | msm_kms.c | 169 struct msm_gem_address_space *aspace; in msm_kms_init_aspace() local 193 aspace = msm_gem_address_space_create(mmu, "mdp_kms", in msm_kms_init_aspace() 195 if (IS_ERR(aspace)) { in msm_kms_init_aspace() 196 dev_err(mdp_dev, "aspace create, error %pe\n", aspace); in msm_kms_init_aspace() 200 return aspace; in msm_kms_init_aspace()
|
H A D | msm_drv.h | 266 struct msm_gem_address_space *aspace, bool needs_dirtyfb); 268 struct msm_gem_address_space *aspace, bool needed_dirtyfb); 270 struct msm_gem_address_space *aspace, int plane);
|
H A D | msm_drv.c | 348 ctx->aspace = msm_gpu_create_private_address_space(priv->gpu, current); in context_init() 526 return msm_gem_get_iova(obj, ctx->aspace, iova); in msm_ioctl_gem_info_iova() 540 if (priv->gpu->aspace == ctx->aspace) in msm_ioctl_gem_info_set_iova() 546 return msm_gem_set_iova(obj, ctx->aspace, iova); in msm_ioctl_gem_info_set_iova()
|
H A D | msm_gem_submit.c | 66 submit->aspace = queue->ctx->aspace; in submit_create() 305 vma = msm_gem_get_vma_locked(obj, submit->aspace); in submit_pin_objects() 662 if (unlikely(!ctx->aspace) && !capable(CAP_SYS_RAWIO)) { in msm_ioctl_gem_submit()
|
/linux/drivers/staging/vme_user/ |
H A D | vme_fake.c | 48 u32 aspace; member 56 u32 aspace; member 156 dma_addr_t buf_base, u32 aspace, u32 cycle) in fake_slave_set() argument 168 switch (aspace) { in fake_slave_set() 212 bridge->slaves[i].aspace = aspace; in fake_slave_set() 225 dma_addr_t *buf_base, u32 *aspace, u32 *cycle) in fake_slave_get() argument 240 *aspace = bridge->slaves[i].aspace; in fake_slave_get() 253 u32 aspace, u32 cycle, u32 dwidth) in fake_master_set() argument 296 switch (aspace) { in fake_master_set() 320 bridge->masters[i].aspace = aspace; in fake_master_set() [all …]
|
H A D | vme.c | 137 u32 aspace, cycle, dwidth; in vme_get_size() local 142 &aspace, &cycle, &dwidth); in vme_get_size() 149 &buf_base, &aspace, &cycle); in vme_get_size() 163 int vme_check_window(struct vme_bridge *bridge, u32 aspace, in vme_check_window() argument 171 switch (aspace) { in vme_check_window() 329 dma_addr_t buf_base, u32 aspace, u32 cycle) in vme_slave_set() argument 347 if (!(((image->address_attr & aspace) == aspace) && in vme_slave_set() 353 retval = vme_check_window(bridge, aspace, vme_base, size); in vme_slave_set() 358 aspace, cycle); in vme_slave_set() 379 dma_addr_t *buf_base, u32 *aspace, u32 *cycle) in vme_slave_get() argument [all …]
|
H A D | vme_tsi148.c | 468 dma_addr_t pci_base, u32 aspace, u32 cycle) in tsi148_slave_set() argument 484 switch (aspace) { in tsi148_slave_set() 611 dma_addr_t *pci_base, u32 *aspace, u32 *cycle) in tsi148_slave_get() argument 649 *aspace = 0; in tsi148_slave_get() 657 *aspace |= VME_A16; in tsi148_slave_get() 661 *aspace |= VME_A24; in tsi148_slave_get() 665 *aspace |= VME_A32; in tsi148_slave_get() 669 *aspace |= VME_A64; in tsi148_slave_get() 799 u32 aspace, u32 cycle, u32 dwidth) in tsi148_master_set() 950 switch (aspace) { in tsi148_master_set() 800 tsi148_master_set(struct vme_master_resource * image,int enabled,unsigned long long vme_base,unsigned long long size,u32 aspace,u32 cycle,u32 dwidth) tsi148_master_set() argument 1035 __tsi148_master_get(struct vme_master_resource * image,int * enabled,unsigned long long * vme_base,unsigned long long * size,u32 * aspace,u32 * cycle,u32 * dwidth) __tsi148_master_get() argument 1144 tsi148_master_get(struct vme_master_resource * image,int * enabled,unsigned long long * vme_base,unsigned long long * size,u32 * aspace,u32 * cycle,u32 * dwidth) tsi148_master_get() argument 1163 u32 aspace, cycle, dwidth; tsi148_master_read() local 1248 u32 aspace, cycle, dwidth; tsi148_master_write() local 1401 tsi148_dma_set_vme_src_attributes(struct device * dev,__be32 * attr,u32 aspace,u32 cycle,u32 dwidth) tsi148_dma_set_vme_src_attributes() argument 1499 tsi148_dma_set_vme_dest_attributes(struct device * dev,__be32 * attr,u32 aspace,u32 cycle,u32 dwidth) tsi148_dma_set_vme_dest_attributes() argument 1902 tsi148_lm_set(struct vme_lm_resource * lm,unsigned long long lm_base,u32 aspace,u32 cycle) tsi148_lm_set() argument 1967 tsi148_lm_get(struct vme_lm_resource * lm,unsigned long long * lm_base,u32 * aspace,u32 * cycle) tsi148_lm_get() argument [all...] |
H A D | vme_user.h | 14 __u32 aspace; /* Address Space */ member 36 __u32 aspace; /* Address Space */ member
|
H A D | vme_user.c | 332 &master.size, &master.aspace, in vme_user_ioctl() 362 master.aspace, master.cycle, master.dwidth); in vme_user_ioctl() 378 &slave.aspace, &slave.cycle); in vme_user_ioctl() 402 image[minor].pci_buf, slave.aspace, in vme_user_ioctl()
|
/linux/drivers/gpu/drm/msm/disp/mdp4/ |
H A D | mdp4_kms.c | 123 struct msm_gem_address_space *aspace = kms->aspace; in mdp4_destroy() local 126 msm_gem_unpin_iova(mdp4_kms->blank_cursor_bo, kms->aspace); in mdp4_destroy() 129 if (aspace) { in mdp4_destroy() 130 aspace->mmu->funcs->detach(aspace->mmu); in mdp4_destroy() 131 msm_gem_address_space_put(aspace); in mdp4_destroy() 383 struct msm_gem_address_space *aspace; in mdp4_kms_init() local 452 aspace = NULL; in mdp4_kms_init() 454 aspace = msm_gem_address_space_create(mmu, in mdp4_kms_init() 457 if (IS_ERR(aspace)) { in mdp4_kms_init() 460 ret = PTR_ERR(aspace); in mdp4_kms_init() [all …]
|
H A D | mdp4_plane.c | 90 return msm_framebuffer_prepare(new_state->fb, kms->aspace, false); in mdp4_plane_prepare_fb() 105 msm_framebuffer_cleanup(fb, kms->aspace, false); in mdp4_plane_cleanup_fb() 156 msm_framebuffer_iova(fb, kms->aspace, 0)); in mdp4_plane_set_scanout() 158 msm_framebuffer_iova(fb, kms->aspace, 1)); in mdp4_plane_set_scanout() 160 msm_framebuffer_iova(fb, kms->aspace, 2)); in mdp4_plane_set_scanout() 162 msm_framebuffer_iova(fb, kms->aspace, 3)); in mdp4_plane_set_scanout()
|
/linux/drivers/gpu/drm/msm/adreno/ |
H A D | adreno_gpu.c | 208 struct msm_gem_address_space *aspace; in adreno_iommu_create_address_space() local 227 aspace = msm_gem_address_space_create(mmu, "gpu", in adreno_iommu_create_address_space() 230 if (IS_ERR(aspace) && !IS_ERR(mmu)) in adreno_iommu_create_address_space() 233 return aspace; in adreno_iommu_create_address_space() 265 gpu->aspace->mmu->funcs->resume_translation(gpu->aspace->mmu); in adreno_fault_handler() 359 if (ctx->aspace) in adreno_get_param() 360 *value = gpu->global_faults + ctx->aspace->faults; in adreno_get_param() 368 if (ctx->aspace == gpu->aspace) in adreno_get_param() 370 *value = ctx->aspace->va_start; in adreno_get_param() 373 if (ctx->aspace == gpu->aspace) in adreno_get_param() [all …]
|
H A D | a5xx_preempt.c | 257 MSM_BO_WC | MSM_BO_MAP_PRIV, gpu->aspace, &bo, &iova); in preempt_init_ring() 265 MSM_BO_WC, gpu->aspace, &counters_bo, &counters_iova); in preempt_init_ring() 267 msm_gem_kernel_put(bo, gpu->aspace); in preempt_init_ring() 298 msm_gem_kernel_put(a5xx_gpu->preempt_bo[i], gpu->aspace); in a5xx_preempt_fini() 299 msm_gem_kernel_put(a5xx_gpu->preempt_counters_bo[i], gpu->aspace); in a5xx_preempt_fini()
|
H A D | a6xx_gmu.c | 1264 msm_gem_kernel_put(gmu->hfi.obj, gmu->aspace); in a6xx_gmu_memory_free() 1265 msm_gem_kernel_put(gmu->debug.obj, gmu->aspace); in a6xx_gmu_memory_free() 1266 msm_gem_kernel_put(gmu->icache.obj, gmu->aspace); in a6xx_gmu_memory_free() 1267 msm_gem_kernel_put(gmu->dcache.obj, gmu->aspace); in a6xx_gmu_memory_free() 1268 msm_gem_kernel_put(gmu->dummy.obj, gmu->aspace); in a6xx_gmu_memory_free() 1269 msm_gem_kernel_put(gmu->log.obj, gmu->aspace); in a6xx_gmu_memory_free() 1271 gmu->aspace->mmu->funcs->detach(gmu->aspace->mmu); in a6xx_gmu_memory_free() 1272 msm_gem_address_space_put(gmu->aspace); in a6xx_gmu_memory_free() 1301 ret = msm_gem_get_and_pin_iova_range(bo->obj, gmu->aspace, &bo->iova, in a6xx_gmu_memory_alloc() 1326 gmu->aspace = msm_gem_address_space_create(mmu, "gmu", 0x0, 0x80000000); in a6xx_gmu_memory_probe() [all …]
|
H A D | a5xx_debugfs.c | 119 msm_gem_unpin_iova(a5xx_gpu->pm4_bo, gpu->aspace); in reset_set() 125 msm_gem_unpin_iova(a5xx_gpu->pfp_bo, gpu->aspace); in reset_set()
|
H A D | a5xx_gpu.c | 623 gpu->aspace, &a5xx_gpu->shadow_bo, in a5xx_ucode_load() 1043 msm_gem_unpin_iova(a5xx_gpu->pm4_bo, gpu->aspace); in a5xx_destroy() 1048 msm_gem_unpin_iova(a5xx_gpu->pfp_bo, gpu->aspace); in a5xx_destroy() 1053 msm_gem_unpin_iova(a5xx_gpu->gpmu_bo, gpu->aspace); in a5xx_destroy() 1058 msm_gem_unpin_iova(a5xx_gpu->shadow_bo, gpu->aspace); in a5xx_destroy() 1458 SZ_1M, MSM_BO_WC, gpu->aspace, in a5xx_crashdumper_init() 1558 msm_gem_kernel_put(dumper.bo, gpu->aspace); in a5xx_gpu_state_get_hlsq_regs() 1566 msm_gem_kernel_put(dumper.bo, gpu->aspace); in a5xx_gpu_state_get_hlsq_regs() 1787 if (gpu->aspace) in a5xx_gpu_init() 1788 msm_mmu_set_fault_handler(gpu->aspace->mmu, gpu, a5xx_fault_handler); in a5xx_gpu_init()
|
/linux/drivers/gpu/drm/msm/disp/mdp5/ |
H A D | mdp5_kms.c | 201 struct msm_gem_address_space *aspace = kms->aspace; in mdp5_kms_destroy() local 203 if (aspace) { in mdp5_kms_destroy() 204 aspace->mmu->funcs->detach(aspace->mmu); in mdp5_kms_destroy() 205 msm_gem_address_space_put(aspace); in mdp5_kms_destroy() 503 struct msm_gem_address_space *aspace; in mdp5_kms_init() local 537 aspace = msm_kms_init_aspace(mdp5_kms->dev); in mdp5_kms_init() 538 if (IS_ERR(aspace)) { in mdp5_kms_init() 539 ret = PTR_ERR(aspace); in mdp5_kms_init() 543 kms->aspace = aspace; in mdp5_kms_init()
|
H A D | mdp5_plane.c | 147 return msm_framebuffer_prepare(new_state->fb, kms->aspace, needs_dirtyfb); in mdp5_plane_prepare_fb() 162 msm_framebuffer_cleanup(fb, kms->aspace, needed_dirtyfb); in mdp5_plane_cleanup_fb() 481 msm_framebuffer_iova(fb, kms->aspace, 0)); in set_scanout_locked() 483 msm_framebuffer_iova(fb, kms->aspace, 1)); in set_scanout_locked() 485 msm_framebuffer_iova(fb, kms->aspace, 2)); in set_scanout_locked() 487 msm_framebuffer_iova(fb, kms->aspace, 3)); in set_scanout_locked()
|
/linux/drivers/gpu/drm/msm/disp/dpu1/ |
H A D | dpu_encoder_phys_wb.c | 561 struct msm_gem_address_space *aspace; in dpu_encoder_phys_wb_prepare_wb_job() local 571 aspace = phys_enc->dpu_kms->base.aspace; in dpu_encoder_phys_wb_prepare_wb_job() 577 ret = msm_framebuffer_prepare(job->fb, aspace, false); in dpu_encoder_phys_wb_prepare_wb_job() 591 dpu_format_populate_addrs(aspace, job->fb, &wb_cfg->dest); in dpu_encoder_phys_wb_prepare_wb_job() 614 struct msm_gem_address_space *aspace; in dpu_encoder_phys_wb_cleanup_wb_job() local 619 aspace = phys_enc->dpu_kms->base.aspace; in dpu_encoder_phys_wb_cleanup_wb_job() 621 msm_framebuffer_cleanup(job->fb, aspace, false); in dpu_encoder_phys_wb_cleanup_wb_job()
|
H A D | dpu_kms.c | 1081 if (!dpu_kms->base.aspace) in _dpu_kms_mmu_destroy() 1084 mmu = dpu_kms->base.aspace->mmu; in _dpu_kms_mmu_destroy() 1087 msm_gem_address_space_put(dpu_kms->base.aspace); in _dpu_kms_mmu_destroy() 1089 dpu_kms->base.aspace = NULL; in _dpu_kms_mmu_destroy() 1094 struct msm_gem_address_space *aspace; in _dpu_kms_mmu_init() local 1096 aspace = msm_kms_init_aspace(dpu_kms->dev); in _dpu_kms_mmu_init() 1097 if (IS_ERR(aspace)) in _dpu_kms_mmu_init() 1098 return PTR_ERR(aspace); in _dpu_kms_mmu_init() 1100 dpu_kms->base.aspace = aspace; in _dpu_kms_mmu_init()
|
H A D | dpu_plane.h | 37 struct msm_gem_address_space *aspace; member
|