1be462c97SLizhi Hou // SPDX-License-Identifier: GPL-2.0
2be462c97SLizhi Hou /*
3be462c97SLizhi Hou * Copyright (C) 2022-2024, Advanced Micro Devices, Inc.
4be462c97SLizhi Hou */
5be462c97SLizhi Hou
6be462c97SLizhi Hou #include <drm/amdxdna_accel.h>
7be462c97SLizhi Hou #include <drm/drm_device.h>
8be462c97SLizhi Hou #include <drm/drm_drv.h>
9be462c97SLizhi Hou #include <drm/drm_file.h>
10aac24309SLizhi Hou #include <drm/drm_gem.h>
11aac24309SLizhi Hou #include <drm/drm_gem_shmem_helper.h>
12be462c97SLizhi Hou #include <drm/drm_print.h>
13aac24309SLizhi Hou #include <drm/gpu_scheduler.h>
14a37d7847SLizhi Hou #include <linux/xarray.h>
15aac24309SLizhi Hou #include <trace/events/amdxdna.h>
16be462c97SLizhi Hou
17be462c97SLizhi Hou #include "amdxdna_ctx.h"
18aac24309SLizhi Hou #include "amdxdna_gem.h"
19be462c97SLizhi Hou #include "amdxdna_pci_drv.h"
20be462c97SLizhi Hou
21be462c97SLizhi Hou #define MAX_HWCTX_ID 255
22aac24309SLizhi Hou #define MAX_ARG_COUNT 4095
23be462c97SLizhi Hou
24aac24309SLizhi Hou struct amdxdna_fence {
25aac24309SLizhi Hou struct dma_fence base;
26aac24309SLizhi Hou spinlock_t lock; /* for base */
27aac24309SLizhi Hou struct amdxdna_hwctx *hwctx;
28aac24309SLizhi Hou };
29aac24309SLizhi Hou
amdxdna_fence_get_driver_name(struct dma_fence * fence)30aac24309SLizhi Hou static const char *amdxdna_fence_get_driver_name(struct dma_fence *fence)
31aac24309SLizhi Hou {
32aac24309SLizhi Hou return KBUILD_MODNAME;
33aac24309SLizhi Hou }
34aac24309SLizhi Hou
amdxdna_fence_get_timeline_name(struct dma_fence * fence)35aac24309SLizhi Hou static const char *amdxdna_fence_get_timeline_name(struct dma_fence *fence)
36aac24309SLizhi Hou {
37aac24309SLizhi Hou struct amdxdna_fence *xdna_fence;
38aac24309SLizhi Hou
39aac24309SLizhi Hou xdna_fence = container_of(fence, struct amdxdna_fence, base);
40aac24309SLizhi Hou
41aac24309SLizhi Hou return xdna_fence->hwctx->name;
42aac24309SLizhi Hou }
43aac24309SLizhi Hou
44aac24309SLizhi Hou static const struct dma_fence_ops fence_ops = {
45aac24309SLizhi Hou .get_driver_name = amdxdna_fence_get_driver_name,
46aac24309SLizhi Hou .get_timeline_name = amdxdna_fence_get_timeline_name,
47aac24309SLizhi Hou };
48aac24309SLizhi Hou
amdxdna_fence_create(struct amdxdna_hwctx * hwctx)49aac24309SLizhi Hou static struct dma_fence *amdxdna_fence_create(struct amdxdna_hwctx *hwctx)
50aac24309SLizhi Hou {
51aac24309SLizhi Hou struct amdxdna_fence *fence;
52aac24309SLizhi Hou
53aac24309SLizhi Hou fence = kzalloc(sizeof(*fence), GFP_KERNEL);
54aac24309SLizhi Hou if (!fence)
55aac24309SLizhi Hou return NULL;
56aac24309SLizhi Hou
57aac24309SLizhi Hou fence->hwctx = hwctx;
58aac24309SLizhi Hou spin_lock_init(&fence->lock);
59aac24309SLizhi Hou dma_fence_init(&fence->base, &fence_ops, &fence->lock, hwctx->id, 0);
60aac24309SLizhi Hou return &fence->base;
61aac24309SLizhi Hou }
62aac24309SLizhi Hou
amdxdna_hwctx_suspend(struct amdxdna_client * client)63bed4c73eSLizhi Hou void amdxdna_hwctx_suspend(struct amdxdna_client *client)
64bed4c73eSLizhi Hou {
65bed4c73eSLizhi Hou struct amdxdna_dev *xdna = client->xdna;
66bed4c73eSLizhi Hou struct amdxdna_hwctx *hwctx;
67a37d7847SLizhi Hou unsigned long hwctx_id;
68bed4c73eSLizhi Hou
69bed4c73eSLizhi Hou drm_WARN_ON(&xdna->ddev, !mutex_is_locked(&xdna->dev_lock));
70bed4c73eSLizhi Hou mutex_lock(&client->hwctx_lock);
71a37d7847SLizhi Hou amdxdna_for_each_hwctx(client, hwctx_id, hwctx)
72bed4c73eSLizhi Hou xdna->dev_info->ops->hwctx_suspend(hwctx);
73bed4c73eSLizhi Hou mutex_unlock(&client->hwctx_lock);
74bed4c73eSLizhi Hou }
75bed4c73eSLizhi Hou
amdxdna_hwctx_resume(struct amdxdna_client * client)76bed4c73eSLizhi Hou void amdxdna_hwctx_resume(struct amdxdna_client *client)
77bed4c73eSLizhi Hou {
78bed4c73eSLizhi Hou struct amdxdna_dev *xdna = client->xdna;
79bed4c73eSLizhi Hou struct amdxdna_hwctx *hwctx;
80a37d7847SLizhi Hou unsigned long hwctx_id;
81bed4c73eSLizhi Hou
82bed4c73eSLizhi Hou drm_WARN_ON(&xdna->ddev, !mutex_is_locked(&xdna->dev_lock));
83bed4c73eSLizhi Hou mutex_lock(&client->hwctx_lock);
84a37d7847SLizhi Hou amdxdna_for_each_hwctx(client, hwctx_id, hwctx)
85bed4c73eSLizhi Hou xdna->dev_info->ops->hwctx_resume(hwctx);
86bed4c73eSLizhi Hou mutex_unlock(&client->hwctx_lock);
87bed4c73eSLizhi Hou }
88bed4c73eSLizhi Hou
amdxdna_hwctx_destroy_rcu(struct amdxdna_hwctx * hwctx,struct srcu_struct * ss)89aac24309SLizhi Hou static void amdxdna_hwctx_destroy_rcu(struct amdxdna_hwctx *hwctx,
90aac24309SLizhi Hou struct srcu_struct *ss)
91be462c97SLizhi Hou {
92be462c97SLizhi Hou struct amdxdna_dev *xdna = hwctx->client->xdna;
93be462c97SLizhi Hou
94aac24309SLizhi Hou synchronize_srcu(ss);
95aac24309SLizhi Hou
96be462c97SLizhi Hou /* At this point, user is not able to submit new commands */
97be462c97SLizhi Hou mutex_lock(&xdna->dev_lock);
98be462c97SLizhi Hou xdna->dev_info->ops->hwctx_fini(hwctx);
99be462c97SLizhi Hou mutex_unlock(&xdna->dev_lock);
100be462c97SLizhi Hou
101be462c97SLizhi Hou kfree(hwctx->name);
102be462c97SLizhi Hou kfree(hwctx);
103be462c97SLizhi Hou }
104be462c97SLizhi Hou
amdxdna_cmd_get_payload(struct amdxdna_gem_obj * abo,u32 * size)105aac24309SLizhi Hou void *amdxdna_cmd_get_payload(struct amdxdna_gem_obj *abo, u32 *size)
106aac24309SLizhi Hou {
107aac24309SLizhi Hou struct amdxdna_cmd *cmd = abo->mem.kva;
108aac24309SLizhi Hou u32 num_masks, count;
109aac24309SLizhi Hou
110aac24309SLizhi Hou if (amdxdna_cmd_get_op(abo) == ERT_CMD_CHAIN)
111aac24309SLizhi Hou num_masks = 0;
112aac24309SLizhi Hou else
113aac24309SLizhi Hou num_masks = 1 + FIELD_GET(AMDXDNA_CMD_EXTRA_CU_MASK, cmd->header);
114aac24309SLizhi Hou
115aac24309SLizhi Hou if (size) {
116aac24309SLizhi Hou count = FIELD_GET(AMDXDNA_CMD_COUNT, cmd->header);
117aac24309SLizhi Hou if (unlikely(count <= num_masks)) {
118aac24309SLizhi Hou *size = 0;
119aac24309SLizhi Hou return NULL;
120aac24309SLizhi Hou }
121aac24309SLizhi Hou *size = (count - num_masks) * sizeof(u32);
122aac24309SLizhi Hou }
123aac24309SLizhi Hou return &cmd->data[num_masks];
124aac24309SLizhi Hou }
125aac24309SLizhi Hou
amdxdna_cmd_get_cu_idx(struct amdxdna_gem_obj * abo)126aac24309SLizhi Hou int amdxdna_cmd_get_cu_idx(struct amdxdna_gem_obj *abo)
127aac24309SLizhi Hou {
128aac24309SLizhi Hou struct amdxdna_cmd *cmd = abo->mem.kva;
129aac24309SLizhi Hou u32 num_masks, i;
130aac24309SLizhi Hou u32 *cu_mask;
131aac24309SLizhi Hou
132aac24309SLizhi Hou if (amdxdna_cmd_get_op(abo) == ERT_CMD_CHAIN)
133aac24309SLizhi Hou return -1;
134aac24309SLizhi Hou
135aac24309SLizhi Hou num_masks = 1 + FIELD_GET(AMDXDNA_CMD_EXTRA_CU_MASK, cmd->header);
136aac24309SLizhi Hou cu_mask = cmd->data;
137aac24309SLizhi Hou for (i = 0; i < num_masks; i++) {
138aac24309SLizhi Hou if (cu_mask[i])
139aac24309SLizhi Hou return ffs(cu_mask[i]) - 1;
140aac24309SLizhi Hou }
141aac24309SLizhi Hou
142aac24309SLizhi Hou return -1;
143aac24309SLizhi Hou }
144aac24309SLizhi Hou
145be462c97SLizhi Hou /*
146be462c97SLizhi Hou * This should be called in close() and remove(). DO NOT call in other syscalls.
147be462c97SLizhi Hou * This guarantee that when hwctx and resources will be released, if user
148be462c97SLizhi Hou * doesn't call amdxdna_drm_destroy_hwctx_ioctl.
149be462c97SLizhi Hou */
amdxdna_hwctx_remove_all(struct amdxdna_client * client)150be462c97SLizhi Hou void amdxdna_hwctx_remove_all(struct amdxdna_client *client)
151be462c97SLizhi Hou {
152be462c97SLizhi Hou struct amdxdna_hwctx *hwctx;
153a37d7847SLizhi Hou unsigned long hwctx_id;
154be462c97SLizhi Hou
155be462c97SLizhi Hou mutex_lock(&client->hwctx_lock);
156a37d7847SLizhi Hou amdxdna_for_each_hwctx(client, hwctx_id, hwctx) {
157be462c97SLizhi Hou XDNA_DBG(client->xdna, "PID %d close HW context %d",
158be462c97SLizhi Hou client->pid, hwctx->id);
159a37d7847SLizhi Hou xa_erase(&client->hwctx_xa, hwctx->id);
160be462c97SLizhi Hou mutex_unlock(&client->hwctx_lock);
161aac24309SLizhi Hou amdxdna_hwctx_destroy_rcu(hwctx, &client->hwctx_srcu);
162be462c97SLizhi Hou mutex_lock(&client->hwctx_lock);
163be462c97SLizhi Hou }
164be462c97SLizhi Hou mutex_unlock(&client->hwctx_lock);
165be462c97SLizhi Hou }
166be462c97SLizhi Hou
amdxdna_drm_create_hwctx_ioctl(struct drm_device * dev,void * data,struct drm_file * filp)167be462c97SLizhi Hou int amdxdna_drm_create_hwctx_ioctl(struct drm_device *dev, void *data, struct drm_file *filp)
168be462c97SLizhi Hou {
169be462c97SLizhi Hou struct amdxdna_client *client = filp->driver_priv;
170be462c97SLizhi Hou struct amdxdna_drm_create_hwctx *args = data;
171be462c97SLizhi Hou struct amdxdna_dev *xdna = to_xdna_dev(dev);
172be462c97SLizhi Hou struct amdxdna_hwctx *hwctx;
173be462c97SLizhi Hou int ret, idx;
174be462c97SLizhi Hou
175be462c97SLizhi Hou if (args->ext || args->ext_flags)
176be462c97SLizhi Hou return -EINVAL;
177be462c97SLizhi Hou
178be462c97SLizhi Hou if (!drm_dev_enter(dev, &idx))
179be462c97SLizhi Hou return -ENODEV;
180be462c97SLizhi Hou
181be462c97SLizhi Hou hwctx = kzalloc(sizeof(*hwctx), GFP_KERNEL);
182be462c97SLizhi Hou if (!hwctx) {
183be462c97SLizhi Hou ret = -ENOMEM;
184be462c97SLizhi Hou goto exit;
185be462c97SLizhi Hou }
186be462c97SLizhi Hou
187be462c97SLizhi Hou if (copy_from_user(&hwctx->qos, u64_to_user_ptr(args->qos_p), sizeof(hwctx->qos))) {
188be462c97SLizhi Hou XDNA_ERR(xdna, "Access QoS info failed");
189be462c97SLizhi Hou ret = -EFAULT;
190be462c97SLizhi Hou goto free_hwctx;
191be462c97SLizhi Hou }
192be462c97SLizhi Hou
193be462c97SLizhi Hou hwctx->client = client;
194be462c97SLizhi Hou hwctx->fw_ctx_id = -1;
195be462c97SLizhi Hou hwctx->num_tiles = args->num_tiles;
196be462c97SLizhi Hou hwctx->mem_size = args->mem_size;
197be462c97SLizhi Hou hwctx->max_opc = args->max_opc;
198a37d7847SLizhi Hou ret = xa_alloc_cyclic(&client->hwctx_xa, &hwctx->id, hwctx,
199a37d7847SLizhi Hou XA_LIMIT(AMDXDNA_INVALID_CTX_HANDLE + 1, MAX_HWCTX_ID),
200a37d7847SLizhi Hou &client->next_hwctxid, GFP_KERNEL);
201be462c97SLizhi Hou if (ret < 0) {
202be462c97SLizhi Hou XDNA_ERR(xdna, "Allocate hwctx ID failed, ret %d", ret);
203be462c97SLizhi Hou goto free_hwctx;
204be462c97SLizhi Hou }
205be462c97SLizhi Hou
206be462c97SLizhi Hou hwctx->name = kasprintf(GFP_KERNEL, "hwctx.%d.%d", client->pid, hwctx->id);
207be462c97SLizhi Hou if (!hwctx->name) {
208be462c97SLizhi Hou ret = -ENOMEM;
209be462c97SLizhi Hou goto rm_id;
210be462c97SLizhi Hou }
211be462c97SLizhi Hou
212be462c97SLizhi Hou mutex_lock(&xdna->dev_lock);
213be462c97SLizhi Hou ret = xdna->dev_info->ops->hwctx_init(hwctx);
214be462c97SLizhi Hou if (ret) {
215be462c97SLizhi Hou mutex_unlock(&xdna->dev_lock);
216be462c97SLizhi Hou XDNA_ERR(xdna, "Init hwctx failed, ret %d", ret);
217be462c97SLizhi Hou goto free_name;
218be462c97SLizhi Hou }
219be462c97SLizhi Hou args->handle = hwctx->id;
220be462c97SLizhi Hou args->syncobj_handle = hwctx->syncobj_hdl;
221be462c97SLizhi Hou mutex_unlock(&xdna->dev_lock);
222be462c97SLizhi Hou
223be462c97SLizhi Hou XDNA_DBG(xdna, "PID %d create HW context %d, ret %d", client->pid, args->handle, ret);
224be462c97SLizhi Hou drm_dev_exit(idx);
225be462c97SLizhi Hou return 0;
226be462c97SLizhi Hou
227be462c97SLizhi Hou free_name:
228be462c97SLizhi Hou kfree(hwctx->name);
229be462c97SLizhi Hou rm_id:
230a37d7847SLizhi Hou xa_erase(&client->hwctx_xa, hwctx->id);
231be462c97SLizhi Hou free_hwctx:
232be462c97SLizhi Hou kfree(hwctx);
233be462c97SLizhi Hou exit:
234be462c97SLizhi Hou drm_dev_exit(idx);
235be462c97SLizhi Hou return ret;
236be462c97SLizhi Hou }
237be462c97SLizhi Hou
amdxdna_drm_destroy_hwctx_ioctl(struct drm_device * dev,void * data,struct drm_file * filp)238be462c97SLizhi Hou int amdxdna_drm_destroy_hwctx_ioctl(struct drm_device *dev, void *data, struct drm_file *filp)
239be462c97SLizhi Hou {
240be462c97SLizhi Hou struct amdxdna_client *client = filp->driver_priv;
241be462c97SLizhi Hou struct amdxdna_drm_destroy_hwctx *args = data;
242be462c97SLizhi Hou struct amdxdna_dev *xdna = to_xdna_dev(dev);
243be462c97SLizhi Hou struct amdxdna_hwctx *hwctx;
244be462c97SLizhi Hou int ret = 0, idx;
245be462c97SLizhi Hou
246*03c318a0SLizhi Hou if (XDNA_MBZ_DBG(xdna, &args->pad, sizeof(args->pad)))
247*03c318a0SLizhi Hou return -EINVAL;
248*03c318a0SLizhi Hou
249be462c97SLizhi Hou if (!drm_dev_enter(dev, &idx))
250be462c97SLizhi Hou return -ENODEV;
251be462c97SLizhi Hou
252a37d7847SLizhi Hou hwctx = xa_erase(&client->hwctx_xa, args->handle);
253be462c97SLizhi Hou if (!hwctx) {
254be462c97SLizhi Hou ret = -EINVAL;
255be462c97SLizhi Hou XDNA_DBG(xdna, "PID %d HW context %d not exist",
256be462c97SLizhi Hou client->pid, args->handle);
257be462c97SLizhi Hou goto out;
258be462c97SLizhi Hou }
259be462c97SLizhi Hou
260a37d7847SLizhi Hou /*
261a37d7847SLizhi Hou * The pushed jobs are handled by DRM scheduler during destroy.
262a37d7847SLizhi Hou * SRCU to synchronize with exec command ioctls.
263a37d7847SLizhi Hou */
264aac24309SLizhi Hou amdxdna_hwctx_destroy_rcu(hwctx, &client->hwctx_srcu);
265be462c97SLizhi Hou
266be462c97SLizhi Hou XDNA_DBG(xdna, "PID %d destroyed HW context %d", client->pid, args->handle);
267be462c97SLizhi Hou out:
268be462c97SLizhi Hou drm_dev_exit(idx);
269be462c97SLizhi Hou return ret;
270be462c97SLizhi Hou }
271be462c97SLizhi Hou
amdxdna_drm_config_hwctx_ioctl(struct drm_device * dev,void * data,struct drm_file * filp)272be462c97SLizhi Hou int amdxdna_drm_config_hwctx_ioctl(struct drm_device *dev, void *data, struct drm_file *filp)
273be462c97SLizhi Hou {
274be462c97SLizhi Hou struct amdxdna_client *client = filp->driver_priv;
275be462c97SLizhi Hou struct amdxdna_drm_config_hwctx *args = data;
276be462c97SLizhi Hou struct amdxdna_dev *xdna = to_xdna_dev(dev);
277be462c97SLizhi Hou struct amdxdna_hwctx *hwctx;
278aac24309SLizhi Hou int ret, idx;
279be462c97SLizhi Hou u32 buf_size;
280be462c97SLizhi Hou void *buf;
281be462c97SLizhi Hou u64 val;
282be462c97SLizhi Hou
283*03c318a0SLizhi Hou if (XDNA_MBZ_DBG(xdna, &args->pad, sizeof(args->pad)))
284*03c318a0SLizhi Hou return -EINVAL;
285*03c318a0SLizhi Hou
286be462c97SLizhi Hou if (!xdna->dev_info->ops->hwctx_config)
287be462c97SLizhi Hou return -EOPNOTSUPP;
288be462c97SLizhi Hou
289be462c97SLizhi Hou val = args->param_val;
290be462c97SLizhi Hou buf_size = args->param_val_size;
291be462c97SLizhi Hou
292be462c97SLizhi Hou switch (args->param_type) {
293be462c97SLizhi Hou case DRM_AMDXDNA_HWCTX_CONFIG_CU:
294be462c97SLizhi Hou /* For those types that param_val is pointer */
295be462c97SLizhi Hou if (buf_size > PAGE_SIZE) {
296be462c97SLizhi Hou XDNA_ERR(xdna, "Config CU param buffer too large");
297be462c97SLizhi Hou return -E2BIG;
298be462c97SLizhi Hou }
299be462c97SLizhi Hou
300be462c97SLizhi Hou /* Hwctx needs to keep buf */
301be462c97SLizhi Hou buf = kzalloc(PAGE_SIZE, GFP_KERNEL);
302be462c97SLizhi Hou if (!buf)
303be462c97SLizhi Hou return -ENOMEM;
304be462c97SLizhi Hou
305be462c97SLizhi Hou if (copy_from_user(buf, u64_to_user_ptr(val), buf_size)) {
306be462c97SLizhi Hou kfree(buf);
307be462c97SLizhi Hou return -EFAULT;
308be462c97SLizhi Hou }
309be462c97SLizhi Hou
310be462c97SLizhi Hou break;
311be462c97SLizhi Hou case DRM_AMDXDNA_HWCTX_ASSIGN_DBG_BUF:
312be462c97SLizhi Hou case DRM_AMDXDNA_HWCTX_REMOVE_DBG_BUF:
313be462c97SLizhi Hou /* For those types that param_val is a value */
314be462c97SLizhi Hou buf = NULL;
315be462c97SLizhi Hou buf_size = 0;
316be462c97SLizhi Hou break;
317be462c97SLizhi Hou default:
318be462c97SLizhi Hou XDNA_DBG(xdna, "Unknown HW context config type %d", args->param_type);
319be462c97SLizhi Hou return -EINVAL;
320be462c97SLizhi Hou }
321be462c97SLizhi Hou
322be462c97SLizhi Hou mutex_lock(&xdna->dev_lock);
323aac24309SLizhi Hou idx = srcu_read_lock(&client->hwctx_srcu);
324a37d7847SLizhi Hou hwctx = xa_load(&client->hwctx_xa, args->handle);
325be462c97SLizhi Hou if (!hwctx) {
326be462c97SLizhi Hou XDNA_DBG(xdna, "PID %d failed to get hwctx %d", client->pid, args->handle);
327be462c97SLizhi Hou ret = -EINVAL;
328aac24309SLizhi Hou goto unlock_srcu;
329be462c97SLizhi Hou }
330be462c97SLizhi Hou
331be462c97SLizhi Hou ret = xdna->dev_info->ops->hwctx_config(hwctx, args->param_type, val, buf, buf_size);
332be462c97SLizhi Hou
333aac24309SLizhi Hou unlock_srcu:
334aac24309SLizhi Hou srcu_read_unlock(&client->hwctx_srcu, idx);
335be462c97SLizhi Hou mutex_unlock(&xdna->dev_lock);
336be462c97SLizhi Hou kfree(buf);
337be462c97SLizhi Hou return ret;
338be462c97SLizhi Hou }
339aac24309SLizhi Hou
340aac24309SLizhi Hou static void
amdxdna_arg_bos_put(struct amdxdna_sched_job * job)341aac24309SLizhi Hou amdxdna_arg_bos_put(struct amdxdna_sched_job *job)
342aac24309SLizhi Hou {
343aac24309SLizhi Hou int i;
344aac24309SLizhi Hou
345aac24309SLizhi Hou for (i = 0; i < job->bo_cnt; i++) {
346aac24309SLizhi Hou if (!job->bos[i])
347aac24309SLizhi Hou break;
348aac24309SLizhi Hou drm_gem_object_put(job->bos[i]);
349aac24309SLizhi Hou }
350aac24309SLizhi Hou }
351aac24309SLizhi Hou
352aac24309SLizhi Hou static int
amdxdna_arg_bos_lookup(struct amdxdna_client * client,struct amdxdna_sched_job * job,u32 * bo_hdls,u32 bo_cnt)353aac24309SLizhi Hou amdxdna_arg_bos_lookup(struct amdxdna_client *client,
354aac24309SLizhi Hou struct amdxdna_sched_job *job,
355aac24309SLizhi Hou u32 *bo_hdls, u32 bo_cnt)
356aac24309SLizhi Hou {
357aac24309SLizhi Hou struct drm_gem_object *gobj;
358aac24309SLizhi Hou int i, ret;
359aac24309SLizhi Hou
360aac24309SLizhi Hou job->bo_cnt = bo_cnt;
361aac24309SLizhi Hou for (i = 0; i < job->bo_cnt; i++) {
362aac24309SLizhi Hou struct amdxdna_gem_obj *abo;
363aac24309SLizhi Hou
364aac24309SLizhi Hou gobj = drm_gem_object_lookup(client->filp, bo_hdls[i]);
365aac24309SLizhi Hou if (!gobj) {
366aac24309SLizhi Hou ret = -ENOENT;
367aac24309SLizhi Hou goto put_shmem_bo;
368aac24309SLizhi Hou }
369aac24309SLizhi Hou abo = to_xdna_obj(gobj);
370aac24309SLizhi Hou
371aac24309SLizhi Hou mutex_lock(&abo->lock);
372aac24309SLizhi Hou if (abo->pinned) {
373aac24309SLizhi Hou mutex_unlock(&abo->lock);
374aac24309SLizhi Hou job->bos[i] = gobj;
375aac24309SLizhi Hou continue;
376aac24309SLizhi Hou }
377aac24309SLizhi Hou
378aac24309SLizhi Hou ret = amdxdna_gem_pin_nolock(abo);
379aac24309SLizhi Hou if (ret) {
380aac24309SLizhi Hou mutex_unlock(&abo->lock);
381aac24309SLizhi Hou drm_gem_object_put(gobj);
382aac24309SLizhi Hou goto put_shmem_bo;
383aac24309SLizhi Hou }
384aac24309SLizhi Hou abo->pinned = true;
385aac24309SLizhi Hou mutex_unlock(&abo->lock);
386aac24309SLizhi Hou
387aac24309SLizhi Hou job->bos[i] = gobj;
388aac24309SLizhi Hou }
389aac24309SLizhi Hou
390aac24309SLizhi Hou return 0;
391aac24309SLizhi Hou
392aac24309SLizhi Hou put_shmem_bo:
393aac24309SLizhi Hou amdxdna_arg_bos_put(job);
394aac24309SLizhi Hou return ret;
395aac24309SLizhi Hou }
396aac24309SLizhi Hou
amdxdna_sched_job_cleanup(struct amdxdna_sched_job * job)397aac24309SLizhi Hou void amdxdna_sched_job_cleanup(struct amdxdna_sched_job *job)
398aac24309SLizhi Hou {
399aac24309SLizhi Hou trace_amdxdna_debug_point(job->hwctx->name, job->seq, "job release");
400aac24309SLizhi Hou amdxdna_arg_bos_put(job);
401aac24309SLizhi Hou amdxdna_gem_put_obj(job->cmd_bo);
402aac24309SLizhi Hou }
403aac24309SLizhi Hou
amdxdna_cmd_submit(struct amdxdna_client * client,u32 cmd_bo_hdl,u32 * arg_bo_hdls,u32 arg_bo_cnt,u32 hwctx_hdl,u64 * seq)404aac24309SLizhi Hou int amdxdna_cmd_submit(struct amdxdna_client *client,
405aac24309SLizhi Hou u32 cmd_bo_hdl, u32 *arg_bo_hdls, u32 arg_bo_cnt,
406aac24309SLizhi Hou u32 hwctx_hdl, u64 *seq)
407aac24309SLizhi Hou {
408aac24309SLizhi Hou struct amdxdna_dev *xdna = client->xdna;
409aac24309SLizhi Hou struct amdxdna_sched_job *job;
410aac24309SLizhi Hou struct amdxdna_hwctx *hwctx;
411aac24309SLizhi Hou int ret, idx;
412aac24309SLizhi Hou
413aac24309SLizhi Hou XDNA_DBG(xdna, "Command BO hdl %d, Arg BO count %d", cmd_bo_hdl, arg_bo_cnt);
414aac24309SLizhi Hou job = kzalloc(struct_size(job, bos, arg_bo_cnt), GFP_KERNEL);
415aac24309SLizhi Hou if (!job)
416aac24309SLizhi Hou return -ENOMEM;
417aac24309SLizhi Hou
418aac24309SLizhi Hou if (cmd_bo_hdl != AMDXDNA_INVALID_BO_HANDLE) {
419aac24309SLizhi Hou job->cmd_bo = amdxdna_gem_get_obj(client, cmd_bo_hdl, AMDXDNA_BO_CMD);
420aac24309SLizhi Hou if (!job->cmd_bo) {
421aac24309SLizhi Hou XDNA_ERR(xdna, "Failed to get cmd bo from %d", cmd_bo_hdl);
422aac24309SLizhi Hou ret = -EINVAL;
423aac24309SLizhi Hou goto free_job;
424aac24309SLizhi Hou }
425aac24309SLizhi Hou } else {
426aac24309SLizhi Hou job->cmd_bo = NULL;
427aac24309SLizhi Hou }
428aac24309SLizhi Hou
429aac24309SLizhi Hou ret = amdxdna_arg_bos_lookup(client, job, arg_bo_hdls, arg_bo_cnt);
430aac24309SLizhi Hou if (ret) {
431aac24309SLizhi Hou XDNA_ERR(xdna, "Argument BOs lookup failed, ret %d", ret);
432aac24309SLizhi Hou goto cmd_put;
433aac24309SLizhi Hou }
434aac24309SLizhi Hou
435aac24309SLizhi Hou idx = srcu_read_lock(&client->hwctx_srcu);
436a37d7847SLizhi Hou hwctx = xa_load(&client->hwctx_xa, hwctx_hdl);
437aac24309SLizhi Hou if (!hwctx) {
438aac24309SLizhi Hou XDNA_DBG(xdna, "PID %d failed to get hwctx %d",
439aac24309SLizhi Hou client->pid, hwctx_hdl);
440aac24309SLizhi Hou ret = -EINVAL;
441aac24309SLizhi Hou goto unlock_srcu;
442aac24309SLizhi Hou }
443aac24309SLizhi Hou
444aac24309SLizhi Hou if (hwctx->status != HWCTX_STAT_READY) {
445aac24309SLizhi Hou XDNA_ERR(xdna, "HW Context is not ready");
446aac24309SLizhi Hou ret = -EINVAL;
447aac24309SLizhi Hou goto unlock_srcu;
448aac24309SLizhi Hou }
449aac24309SLizhi Hou
450aac24309SLizhi Hou job->hwctx = hwctx;
451aac24309SLizhi Hou job->mm = current->mm;
452aac24309SLizhi Hou
453aac24309SLizhi Hou job->fence = amdxdna_fence_create(hwctx);
454aac24309SLizhi Hou if (!job->fence) {
455aac24309SLizhi Hou XDNA_ERR(xdna, "Failed to create fence");
456aac24309SLizhi Hou ret = -ENOMEM;
457aac24309SLizhi Hou goto unlock_srcu;
458aac24309SLizhi Hou }
459aac24309SLizhi Hou kref_init(&job->refcnt);
460aac24309SLizhi Hou
461aac24309SLizhi Hou ret = xdna->dev_info->ops->cmd_submit(hwctx, job, seq);
462aac24309SLizhi Hou if (ret)
463aac24309SLizhi Hou goto put_fence;
464aac24309SLizhi Hou
465aac24309SLizhi Hou /*
466aac24309SLizhi Hou * The amdxdna_hwctx_destroy_rcu() will release hwctx and associated
467aac24309SLizhi Hou * resource after synchronize_srcu(). The submitted jobs should be
468aac24309SLizhi Hou * handled by the queue, for example DRM scheduler, in device layer.
469aac24309SLizhi Hou * For here we can unlock SRCU.
470aac24309SLizhi Hou */
471aac24309SLizhi Hou srcu_read_unlock(&client->hwctx_srcu, idx);
472aac24309SLizhi Hou trace_amdxdna_debug_point(hwctx->name, *seq, "job pushed");
473aac24309SLizhi Hou
474aac24309SLizhi Hou return 0;
475aac24309SLizhi Hou
476aac24309SLizhi Hou put_fence:
477aac24309SLizhi Hou dma_fence_put(job->fence);
478aac24309SLizhi Hou unlock_srcu:
479aac24309SLizhi Hou srcu_read_unlock(&client->hwctx_srcu, idx);
480aac24309SLizhi Hou amdxdna_arg_bos_put(job);
481aac24309SLizhi Hou cmd_put:
482aac24309SLizhi Hou amdxdna_gem_put_obj(job->cmd_bo);
483aac24309SLizhi Hou free_job:
484aac24309SLizhi Hou kfree(job);
485aac24309SLizhi Hou return ret;
486aac24309SLizhi Hou }
487aac24309SLizhi Hou
488aac24309SLizhi Hou /*
489aac24309SLizhi Hou * The submit command ioctl submits a command to firmware. One firmware command
490aac24309SLizhi Hou * may contain multiple command BOs for processing as a whole.
491aac24309SLizhi Hou * The command sequence number is returned which can be used for wait command ioctl.
492aac24309SLizhi Hou */
amdxdna_drm_submit_execbuf(struct amdxdna_client * client,struct amdxdna_drm_exec_cmd * args)493aac24309SLizhi Hou static int amdxdna_drm_submit_execbuf(struct amdxdna_client *client,
494aac24309SLizhi Hou struct amdxdna_drm_exec_cmd *args)
495aac24309SLizhi Hou {
496aac24309SLizhi Hou struct amdxdna_dev *xdna = client->xdna;
497aac24309SLizhi Hou u32 *arg_bo_hdls;
498aac24309SLizhi Hou u32 cmd_bo_hdl;
499aac24309SLizhi Hou int ret;
500aac24309SLizhi Hou
501aac24309SLizhi Hou if (!args->arg_count || args->arg_count > MAX_ARG_COUNT) {
502aac24309SLizhi Hou XDNA_ERR(xdna, "Invalid arg bo count %d", args->arg_count);
503aac24309SLizhi Hou return -EINVAL;
504aac24309SLizhi Hou }
505aac24309SLizhi Hou
506aac24309SLizhi Hou /* Only support single command for now. */
507aac24309SLizhi Hou if (args->cmd_count != 1) {
508aac24309SLizhi Hou XDNA_ERR(xdna, "Invalid cmd bo count %d", args->cmd_count);
509aac24309SLizhi Hou return -EINVAL;
510aac24309SLizhi Hou }
511aac24309SLizhi Hou
512aac24309SLizhi Hou cmd_bo_hdl = (u32)args->cmd_handles;
513aac24309SLizhi Hou arg_bo_hdls = kcalloc(args->arg_count, sizeof(u32), GFP_KERNEL);
514aac24309SLizhi Hou if (!arg_bo_hdls)
515aac24309SLizhi Hou return -ENOMEM;
516aac24309SLizhi Hou ret = copy_from_user(arg_bo_hdls, u64_to_user_ptr(args->args),
517aac24309SLizhi Hou args->arg_count * sizeof(u32));
518aac24309SLizhi Hou if (ret) {
519aac24309SLizhi Hou ret = -EFAULT;
520aac24309SLizhi Hou goto free_cmd_bo_hdls;
521aac24309SLizhi Hou }
522aac24309SLizhi Hou
523aac24309SLizhi Hou ret = amdxdna_cmd_submit(client, cmd_bo_hdl, arg_bo_hdls,
524aac24309SLizhi Hou args->arg_count, args->hwctx, &args->seq);
525aac24309SLizhi Hou if (ret)
526aac24309SLizhi Hou XDNA_DBG(xdna, "Submit cmds failed, ret %d", ret);
527aac24309SLizhi Hou
528aac24309SLizhi Hou free_cmd_bo_hdls:
529aac24309SLizhi Hou kfree(arg_bo_hdls);
530aac24309SLizhi Hou if (!ret)
531aac24309SLizhi Hou XDNA_DBG(xdna, "Pushed cmd %lld to scheduler", args->seq);
532aac24309SLizhi Hou return ret;
533aac24309SLizhi Hou }
534aac24309SLizhi Hou
amdxdna_drm_submit_cmd_ioctl(struct drm_device * dev,void * data,struct drm_file * filp)535aac24309SLizhi Hou int amdxdna_drm_submit_cmd_ioctl(struct drm_device *dev, void *data, struct drm_file *filp)
536aac24309SLizhi Hou {
537aac24309SLizhi Hou struct amdxdna_client *client = filp->driver_priv;
538aac24309SLizhi Hou struct amdxdna_drm_exec_cmd *args = data;
539aac24309SLizhi Hou
540aac24309SLizhi Hou if (args->ext || args->ext_flags)
541aac24309SLizhi Hou return -EINVAL;
542aac24309SLizhi Hou
543aac24309SLizhi Hou switch (args->type) {
544aac24309SLizhi Hou case AMDXDNA_CMD_SUBMIT_EXEC_BUF:
545aac24309SLizhi Hou return amdxdna_drm_submit_execbuf(client, args);
546aac24309SLizhi Hou }
547aac24309SLizhi Hou
548aac24309SLizhi Hou XDNA_ERR(client->xdna, "Invalid command type %d", args->type);
549aac24309SLizhi Hou return -EINVAL;
550aac24309SLizhi Hou }
551