Lines Matching refs:ttm

423 	if (WARN(amdgpu_ttm_tt_get_usermm(bo->tbo.ttm),  in amdgpu_amdkfd_bo_validate()
569 struct ttm_tt *src_ttm = mem->bo->tbo.ttm; in kfd_mem_dmamap_userptr()
570 struct ttm_tt *ttm = bo->tbo.ttm; in kfd_mem_dmamap_userptr() local
573 if (WARN_ON(ttm->num_pages != src_ttm->num_pages)) in kfd_mem_dmamap_userptr()
576 ttm->sg = kmalloc_obj(*ttm->sg); in kfd_mem_dmamap_userptr()
577 if (unlikely(!ttm->sg)) in kfd_mem_dmamap_userptr()
581 ret = sg_alloc_table_from_pages(ttm->sg, src_ttm->pages, in kfd_mem_dmamap_userptr()
582 ttm->num_pages, 0, in kfd_mem_dmamap_userptr()
583 (u64)ttm->num_pages << PAGE_SHIFT, in kfd_mem_dmamap_userptr()
588 ret = dma_map_sgtable(adev->dev, ttm->sg, direction, 0); in kfd_mem_dmamap_userptr()
600 dma_unmap_sgtable(adev->dev, ttm->sg, direction, 0); in kfd_mem_dmamap_userptr()
603 sg_free_table(ttm->sg); in kfd_mem_dmamap_userptr()
605 kfree(ttm->sg); in kfd_mem_dmamap_userptr()
606 ttm->sg = NULL; in kfd_mem_dmamap_userptr()
653 struct ttm_tt *ttm = bo->tbo.ttm; in kfd_mem_dmamap_sg_bo() local
661 if (unlikely(ttm->sg)) { in kfd_mem_dmamap_sg_bo()
678 ttm->sg = create_sg_table(dma_addr, mem->bo->tbo.sg->sgl->length); in kfd_mem_dmamap_sg_bo()
679 if (unlikely(!ttm->sg)) { in kfd_mem_dmamap_sg_bo()
692 sg_free_table(ttm->sg); in kfd_mem_dmamap_sg_bo()
693 kfree(ttm->sg); in kfd_mem_dmamap_sg_bo()
694 ttm->sg = NULL; in kfd_mem_dmamap_sg_bo()
730 struct ttm_tt *ttm = bo->tbo.ttm; in kfd_mem_dmaunmap_userptr() local
732 if (unlikely(!ttm->sg)) in kfd_mem_dmaunmap_userptr()
738 dma_unmap_sgtable(adev->dev, ttm->sg, direction, 0); in kfd_mem_dmaunmap_userptr()
739 sg_free_table(ttm->sg); in kfd_mem_dmaunmap_userptr()
740 kfree(ttm->sg); in kfd_mem_dmaunmap_userptr()
741 ttm->sg = NULL; in kfd_mem_dmaunmap_userptr()
776 struct ttm_tt *ttm = bo->tbo.ttm; in kfd_mem_dmaunmap_sg_bo() local
779 if (unlikely(!ttm->sg)) { in kfd_mem_dmaunmap_sg_bo()
789 dma_unmap_resource(adev->dev, ttm->sg->sgl->dma_address, in kfd_mem_dmaunmap_sg_bo()
790 ttm->sg->sgl->length, dir, DMA_ATTR_SKIP_CPU_SYNC); in kfd_mem_dmaunmap_sg_bo()
791 sg_free_table(ttm->sg); in kfd_mem_dmaunmap_sg_bo()
792 kfree(ttm->sg); in kfd_mem_dmaunmap_sg_bo()
793 ttm->sg = NULL; in kfd_mem_dmaunmap_sg_bo()
918 (amdgpu_ttm_tt_get_usermm(mem->bo->tbo.ttm) && reuse_dmamap(adev, bo_adev)) || in kfd_mem_attach()
933 } else if (amdgpu_ttm_tt_get_usermm(mem->bo->tbo.ttm)) { in kfd_mem_attach()
1121 amdgpu_ttm_tt_set_user_pages(bo->tbo.ttm, range); in init_user_pages()
1830 bo->tbo.ttm->sg = sg; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu()
1947 if (amdgpu_ttm_tt_get_usermm(mem->bo->tbo.ttm)) { in amdgpu_amdkfd_gpuvm_free_memory_of_gpu()
2044 if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm)) { in amdgpu_amdkfd_gpuvm_map_memory_to_gpu()
2075 if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm) && in amdgpu_amdkfd_gpuvm_map_memory_to_gpu()
2138 if (entry->bo_va->base.bo->tbo.ttm && in amdgpu_amdkfd_gpuvm_dmaunmap_mem()
2139 !entry->bo_va->base.bo->tbo.ttm->sg) in amdgpu_amdkfd_gpuvm_dmaunmap_mem()
2290 if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm)) { in amdgpu_amdkfd_gpuvm_map_gtt_bo_to_kernel()
2634 amdgpu_ttm_tt_set_user_pages(bo->tbo.ttm, mem->range); in update_invalid_user_pages()
2712 if (bo->tbo.ttm->pages[0]) { in validate_invalid_user_pages()