| /linux/drivers/accel/amdxdna/ |
| H A D | amdxdna_gem.c | 29 amdxdna_gem_heap_alloc(struct amdxdna_gem_obj *abo) in amdxdna_gem_heap_alloc() argument 31 struct amdxdna_client *client = abo->client; in amdxdna_gem_heap_alloc() 33 struct amdxdna_mem *mem = &abo->mem; in amdxdna_gem_heap_alloc() 61 ret = drm_mm_insert_node_generic(&heap->mm, &abo->mm_node, in amdxdna_gem_heap_alloc() 69 mem->dev_addr = abo->mm_node.start; in amdxdna_gem_heap_alloc() 83 amdxdna_gem_destroy_obj(struct amdxdna_gem_obj *abo) in amdxdna_gem_destroy_obj() argument 85 mutex_destroy(&abo->lock); in amdxdna_gem_destroy_obj() 86 kfree(abo); in amdxdna_gem_destroy_obj() 90 amdxdna_gem_heap_free(struct amdxdna_gem_obj *abo) in amdxdna_gem_heap_free() argument 94 mutex_lock(&abo->client->mm_lock); in amdxdna_gem_heap_free() [all …]
|
| H A D | amdxdna_ctx.c | 95 void *amdxdna_cmd_get_payload(struct amdxdna_gem_obj *abo, u32 *size) in amdxdna_cmd_get_payload() argument 97 struct amdxdna_cmd *cmd = abo->mem.kva; in amdxdna_cmd_get_payload() 100 if (amdxdna_cmd_get_op(abo) == ERT_CMD_CHAIN) in amdxdna_cmd_get_payload() 116 u32 amdxdna_cmd_get_cu_idx(struct amdxdna_gem_obj *abo) in amdxdna_cmd_get_cu_idx() argument 118 struct amdxdna_cmd *cmd = abo->mem.kva; in amdxdna_cmd_get_cu_idx() 122 if (amdxdna_cmd_get_op(abo) == ERT_CMD_CHAIN) in amdxdna_cmd_get_cu_idx() 335 struct amdxdna_gem_obj *abo; in amdxdna_hwctx_sync_debug_bo() local 346 abo = to_xdna_obj(gobj); in amdxdna_hwctx_sync_debug_bo() 349 hwctx = xa_load(&client->hwctx_xa, abo->assigned_hwctx); in amdxdna_hwctx_sync_debug_bo() 385 struct amdxdna_gem_obj *abo; in amdxdna_arg_bos_lookup() local [all …]
|
| H A D | aie2_ctx.c | 587 struct amdxdna_gem_obj *abo; in aie2_hwctx_init() local 595 abo = amdxdna_drm_alloc_dev_bo(&xdna->ddev, &args, client->filp); in aie2_hwctx_init() 596 if (IS_ERR(abo)) { in aie2_hwctx_init() 597 ret = PTR_ERR(abo); in aie2_hwctx_init() 602 i, abo->mem.dev_addr, abo->mem.size); in aie2_hwctx_init() 603 priv->cmd_buf[i] = abo; in aie2_hwctx_init() 804 struct amdxdna_gem_obj *abo; in aie2_hwctx_cfg_debug_bo() local 808 abo = amdxdna_gem_get_obj(client, bo_hdl, AMDXDNA_BO_DEV); in aie2_hwctx_cfg_debug_bo() 809 if (!abo) { in aie2_hwctx_cfg_debug_bo() 815 if (abo->assigned_hwctx != AMDXDNA_INVALID_CTX_HANDLE) { in aie2_hwctx_cfg_debug_bo() [all …]
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_gem.c | 229 struct amdgpu_bo *abo = gem_to_amdgpu_bo(obj); in amdgpu_gem_object_open() local 230 struct amdgpu_device *adev = amdgpu_ttm_adev(abo->tbo.bdev); in amdgpu_gem_object_open() 238 mm = amdgpu_ttm_tt_get_usermm(abo->tbo.ttm); in amdgpu_gem_object_open() 242 if (abo->flags & AMDGPU_GEM_CREATE_VM_ALWAYS_VALID && in amdgpu_gem_object_open() 243 !amdgpu_vm_is_bo_always_valid(vm, abo)) in amdgpu_gem_object_open() 248 r = drm_exec_prepare_obj(&exec, &abo->tbo.base, 1); in amdgpu_gem_object_open() 259 amdgpu_vm_bo_update_shared(abo); in amdgpu_gem_object_open() 260 bo_va = amdgpu_vm_bo_find(vm, abo); in amdgpu_gem_object_open() 262 bo_va = amdgpu_vm_bo_add(adev, vm, abo); in amdgpu_gem_object_open() 267 r = amdgpu_eviction_fence_attach(&fpriv->evf_mgr, abo); in amdgpu_gem_object_open() [all …]
|
| H A D | amdgpu_ttm.c | 94 struct amdgpu_bo *abo; in amdgpu_evict_flags() local 115 abo = ttm_to_amdgpu_bo(bo); in amdgpu_evict_flags() 116 if (abo->flags & AMDGPU_GEM_CREATE_DISCARDABLE) { in amdgpu_evict_flags() 133 amdgpu_bo_placement_from_domain(abo, AMDGPU_GEM_DOMAIN_CPU); in amdgpu_evict_flags() 136 !(abo->flags & AMDGPU_GEM_CREATE_CPU_ACCESS_REQUIRED) && in amdgpu_evict_flags() 144 amdgpu_bo_placement_from_domain(abo, AMDGPU_GEM_DOMAIN_VRAM | in amdgpu_evict_flags() 147 abo->placements[0].fpfn = adev->gmc.visible_vram_size >> PAGE_SHIFT; in amdgpu_evict_flags() 148 abo->placements[0].lpfn = 0; in amdgpu_evict_flags() 149 abo->placements[0].flags |= TTM_PL_FLAG_DESIRED; in amdgpu_evict_flags() 152 amdgpu_bo_placement_from_domain(abo, AMDGPU_GEM_DOMAIN_GTT | in amdgpu_evict_flags() [all …]
|
| H A D | amdgpu_display.c | 1865 struct amdgpu_bo *abo = panic_abo; in amdgpu_display_set_pixel() local 1866 struct amdgpu_device *adev = amdgpu_ttm_adev(abo->tbo.bdev); in amdgpu_display_set_pixel() 1870 amdgpu_res_first(abo->tbo.resource, offset, 4, &cursor); in amdgpu_display_set_pixel() 1882 struct amdgpu_bo *abo; in amdgpu_display_get_scanout_buffer() local 1895 abo = gem_to_amdgpu_bo(fb->obj[0]); in amdgpu_display_get_scanout_buffer() 1896 if (!abo) in amdgpu_display_get_scanout_buffer() 1908 if (abo->flags & AMDGPU_GEM_CREATE_NO_CPU_ACCESS) { in amdgpu_display_get_scanout_buffer() 1909 if (abo->tbo.resource->mem_type != TTM_PL_VRAM) { in amdgpu_display_get_scanout_buffer() 1919 panic_abo = abo; in amdgpu_display_get_scanout_buffer() 1922 if (!abo->kmap.virtual && in amdgpu_display_get_scanout_buffer() [all …]
|
| H A D | dce_v8_0.c | 1795 struct amdgpu_bo *abo; in dce_v8_0_crtc_do_set_base() local 1816 abo = gem_to_amdgpu_bo(obj); in dce_v8_0_crtc_do_set_base() 1817 r = amdgpu_bo_reserve(abo, false); in dce_v8_0_crtc_do_set_base() 1821 abo->flags |= AMDGPU_GEM_CREATE_VRAM_CONTIGUOUS; in dce_v8_0_crtc_do_set_base() 1822 r = amdgpu_bo_pin(abo, AMDGPU_GEM_DOMAIN_VRAM); in dce_v8_0_crtc_do_set_base() 1824 amdgpu_bo_unreserve(abo); in dce_v8_0_crtc_do_set_base() 1827 fb_location = amdgpu_bo_gpu_offset(abo); in dce_v8_0_crtc_do_set_base() 1829 amdgpu_bo_get_tiling_flags(abo, &tiling_flags); in dce_v8_0_crtc_do_set_base() 1830 amdgpu_bo_unreserve(abo); in dce_v8_0_crtc_do_set_base() 1994 abo = gem_to_amdgpu_bo(fb->obj[0]); in dce_v8_0_crtc_do_set_base() [all …]
|
| H A D | dce_v6_0.c | 1886 struct amdgpu_bo *abo; in dce_v6_0_crtc_do_set_base() local 1906 abo = gem_to_amdgpu_bo(obj); in dce_v6_0_crtc_do_set_base() 1907 r = amdgpu_bo_reserve(abo, false); in dce_v6_0_crtc_do_set_base() 1911 abo->flags |= AMDGPU_GEM_CREATE_VRAM_CONTIGUOUS; in dce_v6_0_crtc_do_set_base() 1912 r = amdgpu_bo_pin(abo, AMDGPU_GEM_DOMAIN_VRAM); in dce_v6_0_crtc_do_set_base() 1914 amdgpu_bo_unreserve(abo); in dce_v6_0_crtc_do_set_base() 1917 fb_location = amdgpu_bo_gpu_offset(abo); in dce_v6_0_crtc_do_set_base() 1919 amdgpu_bo_get_tiling_flags(abo, &tiling_flags); in dce_v6_0_crtc_do_set_base() 1920 amdgpu_bo_unreserve(abo); in dce_v6_0_crtc_do_set_base() 2082 abo = gem_to_amdgpu_bo(fb->obj[0]); in dce_v6_0_crtc_do_set_base() [all …]
|
| H A D | dce_v10_0.c | 1848 struct amdgpu_bo *abo; in dce_v10_0_crtc_do_set_base() local 1869 abo = gem_to_amdgpu_bo(obj); in dce_v10_0_crtc_do_set_base() 1870 r = amdgpu_bo_reserve(abo, false); in dce_v10_0_crtc_do_set_base() 1874 abo->flags |= AMDGPU_GEM_CREATE_VRAM_CONTIGUOUS; in dce_v10_0_crtc_do_set_base() 1875 r = amdgpu_bo_pin(abo, AMDGPU_GEM_DOMAIN_VRAM); in dce_v10_0_crtc_do_set_base() 1877 amdgpu_bo_unreserve(abo); in dce_v10_0_crtc_do_set_base() 1880 fb_location = amdgpu_bo_gpu_offset(abo); in dce_v10_0_crtc_do_set_base() 1882 amdgpu_bo_get_tiling_flags(abo, &tiling_flags); in dce_v10_0_crtc_do_set_base() 1883 amdgpu_bo_unreserve(abo); in dce_v10_0_crtc_do_set_base() 2067 abo = gem_to_amdgpu_bo(fb->obj[0]); in dce_v10_0_crtc_do_set_base() [all …]
|
| H A D | amdgpu_vm.c | 1302 struct amdgpu_bo *abo = gem_to_amdgpu_bo(gobj); in amdgpu_vm_bo_update() local 1304 if (abo->tbo.resource && in amdgpu_vm_bo_update() 1305 abo->tbo.resource->mem_type == TTM_PL_VRAM) in amdgpu_vm_bo_update()
|
| /linux/scripts/ |
| H A D | extract-ikconfig | 19 if pos=`tr "$cf1\n$cf2" "\n$cf2=" < "$1" | grep -abo "^$cf2"` 33 for pos in `tr "$1\n$2" "\n$2=" < "$img" | grep -abo "^$2"`
|
| /linux/drivers/gpu/drm/amd/display/amdgpu_dm/ |
| H A D | amdgpu_dm.c | 9787 struct amdgpu_bo *abo = gem_to_amdgpu_bo(fb->obj[0]); in get_mem_type() local 9789 return abo->tbo.resource ? abo->tbo.resource->mem_type : 0; in get_mem_type()
|