| /linux/drivers/accel/amdxdna/ |
| H A D | amdxdna_gem.c | 27 amdxdna_gem_heap_alloc(struct amdxdna_gem_obj *abo) in amdxdna_gem_heap_alloc() argument 29 struct amdxdna_client *client = abo->client; in amdxdna_gem_heap_alloc() 31 struct amdxdna_mem *mem = &abo->mem; in amdxdna_gem_heap_alloc() 59 ret = drm_mm_insert_node_generic(&heap->mm, &abo->mm_node, in amdxdna_gem_heap_alloc() 67 mem->dev_addr = abo->mm_node.start; in amdxdna_gem_heap_alloc() 81 amdxdna_gem_destroy_obj(struct amdxdna_gem_obj *abo) in amdxdna_gem_destroy_obj() argument 83 mutex_destroy(&abo->lock); in amdxdna_gem_destroy_obj() 84 kfree(abo); in amdxdna_gem_destroy_obj() 88 amdxdna_gem_heap_free(struct amdxdna_gem_obj *abo) in amdxdna_gem_heap_free() argument 92 mutex_lock(&abo->client->mm_lock); in amdxdna_gem_heap_free() [all …]
|
| H A D | amdxdna_ctx.c | 96 void *amdxdna_cmd_get_payload(struct amdxdna_gem_obj *abo, u32 *size) in amdxdna_cmd_get_payload() argument 98 struct amdxdna_cmd *cmd = abo->mem.kva; in amdxdna_cmd_get_payload() 101 if (amdxdna_cmd_get_op(abo) == ERT_CMD_CHAIN) in amdxdna_cmd_get_payload() 111 abo->mem.size)) { in amdxdna_cmd_get_payload() 120 u32 amdxdna_cmd_get_cu_idx(struct amdxdna_gem_obj *abo) in amdxdna_cmd_get_cu_idx() argument 122 struct amdxdna_cmd *cmd = abo->mem.kva; in amdxdna_cmd_get_cu_idx() 126 if (amdxdna_cmd_get_op(abo) == ERT_CMD_CHAIN) in amdxdna_cmd_get_cu_idx() 139 int amdxdna_cmd_set_error(struct amdxdna_gem_obj *abo, in amdxdna_cmd_set_error() argument 144 struct amdxdna_cmd *cmd = abo->mem.kva; in amdxdna_cmd_set_error() 150 if (amdxdna_cmd_get_op(abo) == ERT_CMD_CHAIN) { in amdxdna_cmd_set_error() [all …]
|
| H A D | amdxdna_ctx.h | 144 amdxdna_cmd_get_op(struct amdxdna_gem_obj *abo) in amdxdna_cmd_get_op() argument 146 struct amdxdna_cmd *cmd = abo->mem.kva; in amdxdna_cmd_get_op() 152 amdxdna_cmd_set_state(struct amdxdna_gem_obj *abo, enum ert_cmd_state s) in amdxdna_cmd_set_state() argument 154 struct amdxdna_cmd *cmd = abo->mem.kva; in amdxdna_cmd_set_state() 161 amdxdna_cmd_get_state(struct amdxdna_gem_obj *abo) in amdxdna_cmd_get_state() argument 163 struct amdxdna_cmd *cmd = abo->mem.kva; in amdxdna_cmd_get_state() 168 void *amdxdna_cmd_get_payload(struct amdxdna_gem_obj *abo, u32 *size); 169 u32 amdxdna_cmd_get_cu_idx(struct amdxdna_gem_obj *abo); 170 int amdxdna_cmd_set_error(struct amdxdna_gem_obj *abo,
|
| H A D | aie2_ctx.c | 574 struct amdxdna_gem_obj *abo; in aie2_hwctx_init() local 582 abo = amdxdna_drm_alloc_dev_bo(&xdna->ddev, &args, client->filp); in aie2_hwctx_init() 583 if (IS_ERR(abo)) { in aie2_hwctx_init() 584 ret = PTR_ERR(abo); in aie2_hwctx_init() 589 i, abo->mem.dev_addr, abo->mem.size); in aie2_hwctx_init() 590 priv->cmd_buf[i] = abo; in aie2_hwctx_init() 791 struct amdxdna_gem_obj *abo; in aie2_hwctx_cfg_debug_bo() local 795 abo = amdxdna_gem_get_obj(client, bo_hdl, AMDXDNA_BO_DEV); in aie2_hwctx_cfg_debug_bo() 796 if (!abo) { in aie2_hwctx_cfg_debug_bo() 802 if (abo->assigned_hwctx != AMDXDNA_INVALID_CTX_HANDLE) { in aie2_hwctx_cfg_debug_bo() [all …]
|
| H A D | aie2_message.c | 501 struct amdxdna_gem_obj *abo; in aie2_config_cu() local 526 abo = to_xdna_obj(gobj); in aie2_config_cu() 528 if (abo->type != AMDXDNA_BO_DEV) { in aie2_config_cu() 535 abo->mem.dev_addr >> shift); in aie2_config_cu() 538 abo->mem.dev_addr, req.cfgs[i]); in aie2_config_cu() 1005 struct amdxdna_gem_obj *abo; in aie2_cmdlist_multi_execbuf() local 1007 abo = amdxdna_gem_get_obj(client, boh, AMDXDNA_BO_CMD); in aie2_cmdlist_multi_execbuf() 1008 if (!abo) { in aie2_cmdlist_multi_execbuf() 1015 abo, &size, &op); in aie2_cmdlist_multi_execbuf() 1016 amdxdna_gem_put_obj(abo); in aie2_cmdlist_multi_execbuf() [all …]
|
| H A D | aie2_pci.h | 353 void aie2_hmm_invalidate(struct amdxdna_gem_obj *abo, unsigned long cur_seq);
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_gem.c | 229 struct amdgpu_bo *abo = gem_to_amdgpu_bo(obj); in amdgpu_gem_object_open() local 230 struct amdgpu_device *adev = amdgpu_ttm_adev(abo->tbo.bdev); in amdgpu_gem_object_open() 238 mm = amdgpu_ttm_tt_get_usermm(abo->tbo.ttm); in amdgpu_gem_object_open() 242 if (abo->flags & AMDGPU_GEM_CREATE_VM_ALWAYS_VALID && in amdgpu_gem_object_open() 243 !amdgpu_vm_is_bo_always_valid(vm, abo)) in amdgpu_gem_object_open() 248 r = drm_exec_prepare_obj(&exec, &abo->tbo.base, 1); in amdgpu_gem_object_open() 259 amdgpu_vm_bo_update_shared(abo); in amdgpu_gem_object_open() 260 bo_va = amdgpu_vm_bo_find(vm, abo); in amdgpu_gem_object_open() 262 bo_va = amdgpu_vm_bo_add(adev, vm, abo); in amdgpu_gem_object_open() 267 r = amdgpu_eviction_fence_attach(&fpriv->evf_mgr, abo); in amdgpu_gem_object_open() [all …]
|
| H A D | amdgpu_ttm.c | 94 struct amdgpu_bo *abo; in amdgpu_evict_flags() local 115 abo = ttm_to_amdgpu_bo(bo); in amdgpu_evict_flags() 116 if (abo->flags & AMDGPU_GEM_CREATE_DISCARDABLE) { in amdgpu_evict_flags() 133 amdgpu_bo_placement_from_domain(abo, AMDGPU_GEM_DOMAIN_CPU); in amdgpu_evict_flags() 136 !(abo->flags & AMDGPU_GEM_CREATE_CPU_ACCESS_REQUIRED) && in amdgpu_evict_flags() 144 amdgpu_bo_placement_from_domain(abo, AMDGPU_GEM_DOMAIN_VRAM | in amdgpu_evict_flags() 147 abo->placements[0].fpfn = adev->gmc.visible_vram_size >> PAGE_SHIFT; in amdgpu_evict_flags() 148 abo->placements[0].lpfn = 0; in amdgpu_evict_flags() 149 abo->placements[0].flags |= TTM_PL_FLAG_DESIRED; in amdgpu_evict_flags() 152 amdgpu_bo_placement_from_domain(abo, AMDGPU_GEM_DOMAIN_GTT | in amdgpu_evict_flags() [all …]
|
| H A D | amdgpu_display.c | 1865 struct amdgpu_bo *abo = panic_abo; in amdgpu_display_set_pixel() local 1866 struct amdgpu_device *adev = amdgpu_ttm_adev(abo->tbo.bdev); in amdgpu_display_set_pixel() 1870 amdgpu_res_first(abo->tbo.resource, offset, 4, &cursor); in amdgpu_display_set_pixel() 1882 struct amdgpu_bo *abo; in amdgpu_display_get_scanout_buffer() local 1895 abo = gem_to_amdgpu_bo(fb->obj[0]); in amdgpu_display_get_scanout_buffer() 1896 if (!abo) in amdgpu_display_get_scanout_buffer() 1908 if (abo->flags & AMDGPU_GEM_CREATE_NO_CPU_ACCESS) { in amdgpu_display_get_scanout_buffer() 1909 if (abo->tbo.resource->mem_type != TTM_PL_VRAM) { in amdgpu_display_get_scanout_buffer() 1919 panic_abo = abo; in amdgpu_display_get_scanout_buffer() 1922 if (!abo->kmap.virtual && in amdgpu_display_get_scanout_buffer() [all …]
|
| H A D | dce_v8_0.c | 1795 struct amdgpu_bo *abo; in dce_v8_0_crtc_do_set_base() local 1816 abo = gem_to_amdgpu_bo(obj); in dce_v8_0_crtc_do_set_base() 1817 r = amdgpu_bo_reserve(abo, false); in dce_v8_0_crtc_do_set_base() 1821 abo->flags |= AMDGPU_GEM_CREATE_VRAM_CONTIGUOUS; in dce_v8_0_crtc_do_set_base() 1822 r = amdgpu_bo_pin(abo, AMDGPU_GEM_DOMAIN_VRAM); in dce_v8_0_crtc_do_set_base() 1824 amdgpu_bo_unreserve(abo); in dce_v8_0_crtc_do_set_base() 1827 fb_location = amdgpu_bo_gpu_offset(abo); in dce_v8_0_crtc_do_set_base() 1829 amdgpu_bo_get_tiling_flags(abo, &tiling_flags); in dce_v8_0_crtc_do_set_base() 1830 amdgpu_bo_unreserve(abo); in dce_v8_0_crtc_do_set_base() 1994 abo = gem_to_amdgpu_bo(fb->obj[0]); in dce_v8_0_crtc_do_set_base() [all …]
|
| H A D | dce_v6_0.c | 1886 struct amdgpu_bo *abo; in dce_v6_0_crtc_do_set_base() local 1906 abo = gem_to_amdgpu_bo(obj); in dce_v6_0_crtc_do_set_base() 1907 r = amdgpu_bo_reserve(abo, false); in dce_v6_0_crtc_do_set_base() 1911 abo->flags |= AMDGPU_GEM_CREATE_VRAM_CONTIGUOUS; in dce_v6_0_crtc_do_set_base() 1912 r = amdgpu_bo_pin(abo, AMDGPU_GEM_DOMAIN_VRAM); in dce_v6_0_crtc_do_set_base() 1914 amdgpu_bo_unreserve(abo); in dce_v6_0_crtc_do_set_base() 1917 fb_location = amdgpu_bo_gpu_offset(abo); in dce_v6_0_crtc_do_set_base() 1919 amdgpu_bo_get_tiling_flags(abo, &tiling_flags); in dce_v6_0_crtc_do_set_base() 1920 amdgpu_bo_unreserve(abo); in dce_v6_0_crtc_do_set_base() 2082 abo = gem_to_amdgpu_bo(fb->obj[0]); in dce_v6_0_crtc_do_set_base() [all …]
|
| H A D | dce_v10_0.c | 1848 struct amdgpu_bo *abo; in dce_v10_0_crtc_do_set_base() local 1869 abo = gem_to_amdgpu_bo(obj); in dce_v10_0_crtc_do_set_base() 1870 r = amdgpu_bo_reserve(abo, false); in dce_v10_0_crtc_do_set_base() 1874 abo->flags |= AMDGPU_GEM_CREATE_VRAM_CONTIGUOUS; in dce_v10_0_crtc_do_set_base() 1875 r = amdgpu_bo_pin(abo, AMDGPU_GEM_DOMAIN_VRAM); in dce_v10_0_crtc_do_set_base() 1877 amdgpu_bo_unreserve(abo); in dce_v10_0_crtc_do_set_base() 1880 fb_location = amdgpu_bo_gpu_offset(abo); in dce_v10_0_crtc_do_set_base() 1882 amdgpu_bo_get_tiling_flags(abo, &tiling_flags); in dce_v10_0_crtc_do_set_base() 1883 amdgpu_bo_unreserve(abo); in dce_v10_0_crtc_do_set_base() 2067 abo = gem_to_amdgpu_bo(fb->obj[0]); in dce_v10_0_crtc_do_set_base() [all …]
|
| H A D | amdgpu_vm.c | 1305 struct amdgpu_bo *abo = gem_to_amdgpu_bo(gobj); in amdgpu_vm_bo_update() local 1307 if (abo->tbo.resource && in amdgpu_vm_bo_update() 1308 abo->tbo.resource->mem_type == TTM_PL_VRAM) in amdgpu_vm_bo_update()
|
| /linux/scripts/ |
| H A D | extract-ikconfig | 19 if pos=`tr "$cf1\n$cf2" "\n$cf2=" < "$1" | grep -abo "^$cf2"` 33 for pos in `tr "$1\n$2" "\n$2=" < "$img" | grep -abo "^$2"`
|
| /linux/drivers/gpu/drm/amd/display/amdgpu_dm/ |
| H A D | amdgpu_dm.c | 9788 struct amdgpu_bo *abo = gem_to_amdgpu_bo(fb->obj[0]); in get_mem_type() local 9790 return abo->tbo.resource ? abo->tbo.resource->mem_type : 0; in get_mem_type()
|