xref: /linux/drivers/gpu/drm/imagination/pvr_cccb.c (revision 0ea5c948cb64bab5bc7a5516774eb8536f05aa0d)
1*d2d79d29SSarah Walker // SPDX-License-Identifier: GPL-2.0-only OR MIT
2*d2d79d29SSarah Walker /* Copyright (c) 2023 Imagination Technologies Ltd. */
3*d2d79d29SSarah Walker 
4*d2d79d29SSarah Walker #include "pvr_ccb.h"
5*d2d79d29SSarah Walker #include "pvr_cccb.h"
6*d2d79d29SSarah Walker #include "pvr_device.h"
7*d2d79d29SSarah Walker #include "pvr_gem.h"
8*d2d79d29SSarah Walker #include "pvr_hwrt.h"
9*d2d79d29SSarah Walker 
10*d2d79d29SSarah Walker #include <linux/compiler.h>
11*d2d79d29SSarah Walker #include <linux/delay.h>
12*d2d79d29SSarah Walker #include <linux/jiffies.h>
13*d2d79d29SSarah Walker #include <linux/mutex.h>
14*d2d79d29SSarah Walker #include <linux/types.h>
15*d2d79d29SSarah Walker 
16*d2d79d29SSarah Walker static __always_inline u32
get_ccb_space(u32 w_off,u32 r_off,u32 ccb_size)17*d2d79d29SSarah Walker get_ccb_space(u32 w_off, u32 r_off, u32 ccb_size)
18*d2d79d29SSarah Walker {
19*d2d79d29SSarah Walker 	return (((r_off) - (w_off)) + ((ccb_size) - 1)) & ((ccb_size) - 1);
20*d2d79d29SSarah Walker }
21*d2d79d29SSarah Walker 
22*d2d79d29SSarah Walker static void
cccb_ctrl_init(void * cpu_ptr,void * priv)23*d2d79d29SSarah Walker cccb_ctrl_init(void *cpu_ptr, void *priv)
24*d2d79d29SSarah Walker {
25*d2d79d29SSarah Walker 	struct rogue_fwif_cccb_ctl *ctrl = cpu_ptr;
26*d2d79d29SSarah Walker 	struct pvr_cccb *pvr_cccb = priv;
27*d2d79d29SSarah Walker 
28*d2d79d29SSarah Walker 	WRITE_ONCE(ctrl->write_offset, 0);
29*d2d79d29SSarah Walker 	WRITE_ONCE(ctrl->read_offset, 0);
30*d2d79d29SSarah Walker 	WRITE_ONCE(ctrl->dep_offset, 0);
31*d2d79d29SSarah Walker 	WRITE_ONCE(ctrl->wrap_mask, pvr_cccb->wrap_mask);
32*d2d79d29SSarah Walker }
33*d2d79d29SSarah Walker 
34*d2d79d29SSarah Walker /**
35*d2d79d29SSarah Walker  * pvr_cccb_init() - Initialise a Client CCB
36*d2d79d29SSarah Walker  * @pvr_dev: Device pointer.
37*d2d79d29SSarah Walker  * @pvr_cccb: Pointer to Client CCB structure to initialise.
38*d2d79d29SSarah Walker  * @size_log2: Log2 size of Client CCB in bytes.
39*d2d79d29SSarah Walker  * @name: Name of owner of Client CCB. Used for fence context.
40*d2d79d29SSarah Walker  *
41*d2d79d29SSarah Walker  * Return:
42*d2d79d29SSarah Walker  *  * Zero on success, or
43*d2d79d29SSarah Walker  *  * Any error code returned by pvr_fw_object_create_and_map().
44*d2d79d29SSarah Walker  */
45*d2d79d29SSarah Walker int
pvr_cccb_init(struct pvr_device * pvr_dev,struct pvr_cccb * pvr_cccb,u32 size_log2,const char * name)46*d2d79d29SSarah Walker pvr_cccb_init(struct pvr_device *pvr_dev, struct pvr_cccb *pvr_cccb,
47*d2d79d29SSarah Walker 	      u32 size_log2, const char *name)
48*d2d79d29SSarah Walker {
49*d2d79d29SSarah Walker 	size_t size = 1 << size_log2;
50*d2d79d29SSarah Walker 	int err;
51*d2d79d29SSarah Walker 
52*d2d79d29SSarah Walker 	pvr_cccb->size = size;
53*d2d79d29SSarah Walker 	pvr_cccb->write_offset = 0;
54*d2d79d29SSarah Walker 	pvr_cccb->wrap_mask = size - 1;
55*d2d79d29SSarah Walker 
56*d2d79d29SSarah Walker 	/*
57*d2d79d29SSarah Walker 	 * Map CCCB and control structure as uncached, so we don't have to flush
58*d2d79d29SSarah Walker 	 * CPU cache repeatedly when polling for space.
59*d2d79d29SSarah Walker 	 */
60*d2d79d29SSarah Walker 	pvr_cccb->ctrl = pvr_fw_object_create_and_map(pvr_dev, sizeof(*pvr_cccb->ctrl),
61*d2d79d29SSarah Walker 						      PVR_BO_FW_FLAGS_DEVICE_UNCACHED,
62*d2d79d29SSarah Walker 						      cccb_ctrl_init, pvr_cccb,
63*d2d79d29SSarah Walker 						      &pvr_cccb->ctrl_obj);
64*d2d79d29SSarah Walker 	if (IS_ERR(pvr_cccb->ctrl))
65*d2d79d29SSarah Walker 		return PTR_ERR(pvr_cccb->ctrl);
66*d2d79d29SSarah Walker 
67*d2d79d29SSarah Walker 	pvr_cccb->cccb = pvr_fw_object_create_and_map(pvr_dev, size,
68*d2d79d29SSarah Walker 						      PVR_BO_FW_FLAGS_DEVICE_UNCACHED,
69*d2d79d29SSarah Walker 						      NULL, NULL, &pvr_cccb->cccb_obj);
70*d2d79d29SSarah Walker 	if (IS_ERR(pvr_cccb->cccb)) {
71*d2d79d29SSarah Walker 		err = PTR_ERR(pvr_cccb->cccb);
72*d2d79d29SSarah Walker 		goto err_free_ctrl;
73*d2d79d29SSarah Walker 	}
74*d2d79d29SSarah Walker 
75*d2d79d29SSarah Walker 	pvr_fw_object_get_fw_addr(pvr_cccb->ctrl_obj, &pvr_cccb->ctrl_fw_addr);
76*d2d79d29SSarah Walker 	pvr_fw_object_get_fw_addr(pvr_cccb->cccb_obj, &pvr_cccb->cccb_fw_addr);
77*d2d79d29SSarah Walker 
78*d2d79d29SSarah Walker 	return 0;
79*d2d79d29SSarah Walker 
80*d2d79d29SSarah Walker err_free_ctrl:
81*d2d79d29SSarah Walker 	pvr_fw_object_unmap_and_destroy(pvr_cccb->ctrl_obj);
82*d2d79d29SSarah Walker 
83*d2d79d29SSarah Walker 	return err;
84*d2d79d29SSarah Walker }
85*d2d79d29SSarah Walker 
86*d2d79d29SSarah Walker /**
87*d2d79d29SSarah Walker  * pvr_cccb_fini() - Release Client CCB structure
88*d2d79d29SSarah Walker  * @pvr_cccb: Client CCB to release.
89*d2d79d29SSarah Walker  */
90*d2d79d29SSarah Walker void
pvr_cccb_fini(struct pvr_cccb * pvr_cccb)91*d2d79d29SSarah Walker pvr_cccb_fini(struct pvr_cccb *pvr_cccb)
92*d2d79d29SSarah Walker {
93*d2d79d29SSarah Walker 	pvr_fw_object_unmap_and_destroy(pvr_cccb->cccb_obj);
94*d2d79d29SSarah Walker 	pvr_fw_object_unmap_and_destroy(pvr_cccb->ctrl_obj);
95*d2d79d29SSarah Walker }
96*d2d79d29SSarah Walker 
97*d2d79d29SSarah Walker /**
98*d2d79d29SSarah Walker  * pvr_cccb_cmdseq_fits() - Check if a command sequence fits in the CCCB
99*d2d79d29SSarah Walker  * @pvr_cccb: Target Client CCB.
100*d2d79d29SSarah Walker  * @size: Size of the command sequence.
101*d2d79d29SSarah Walker  *
102*d2d79d29SSarah Walker  * Check if a command sequence fits in the CCCB we have at hand.
103*d2d79d29SSarah Walker  *
104*d2d79d29SSarah Walker  * Return:
105*d2d79d29SSarah Walker  *  * true if the command sequence fits in the CCCB, or
106*d2d79d29SSarah Walker  *  * false otherwise.
107*d2d79d29SSarah Walker  */
pvr_cccb_cmdseq_fits(struct pvr_cccb * pvr_cccb,size_t size)108*d2d79d29SSarah Walker bool pvr_cccb_cmdseq_fits(struct pvr_cccb *pvr_cccb, size_t size)
109*d2d79d29SSarah Walker {
110*d2d79d29SSarah Walker 	struct rogue_fwif_cccb_ctl *ctrl = pvr_cccb->ctrl;
111*d2d79d29SSarah Walker 	u32 read_offset, remaining;
112*d2d79d29SSarah Walker 	bool fits = false;
113*d2d79d29SSarah Walker 
114*d2d79d29SSarah Walker 	read_offset = READ_ONCE(ctrl->read_offset);
115*d2d79d29SSarah Walker 	remaining = pvr_cccb->size - pvr_cccb->write_offset;
116*d2d79d29SSarah Walker 
117*d2d79d29SSarah Walker 	/* Always ensure we have enough room for a padding command at the end of the CCCB.
118*d2d79d29SSarah Walker 	 * If our command sequence does not fit, reserve the remaining space for a padding
119*d2d79d29SSarah Walker 	 * command.
120*d2d79d29SSarah Walker 	 */
121*d2d79d29SSarah Walker 	if (size + PADDING_COMMAND_SIZE > remaining)
122*d2d79d29SSarah Walker 		size += remaining;
123*d2d79d29SSarah Walker 
124*d2d79d29SSarah Walker 	if (get_ccb_space(pvr_cccb->write_offset, read_offset, pvr_cccb->size) >= size)
125*d2d79d29SSarah Walker 		fits = true;
126*d2d79d29SSarah Walker 
127*d2d79d29SSarah Walker 	return fits;
128*d2d79d29SSarah Walker }
129*d2d79d29SSarah Walker 
130*d2d79d29SSarah Walker /**
131*d2d79d29SSarah Walker  * pvr_cccb_write_command_with_header() - Write a command + command header to a
132*d2d79d29SSarah Walker  *                                        Client CCB
133*d2d79d29SSarah Walker  * @pvr_cccb: Target Client CCB.
134*d2d79d29SSarah Walker  * @cmd_type: Client CCB command type. Must be one of %ROGUE_FWIF_CCB_CMD_TYPE_*.
135*d2d79d29SSarah Walker  * @cmd_size: Size of command in bytes.
136*d2d79d29SSarah Walker  * @cmd_data: Pointer to command to write.
137*d2d79d29SSarah Walker  * @ext_job_ref: External job reference.
138*d2d79d29SSarah Walker  * @int_job_ref: Internal job reference.
139*d2d79d29SSarah Walker  *
140*d2d79d29SSarah Walker  * Caller must make sure there's enough space in CCCB to queue this command. This
141*d2d79d29SSarah Walker  * can be done by calling pvr_cccb_cmdseq_fits().
142*d2d79d29SSarah Walker  *
143*d2d79d29SSarah Walker  * This function is not protected by any lock. The caller must ensure there's
144*d2d79d29SSarah Walker  * no concurrent caller, which should be guaranteed by the drm_sched model (job
145*d2d79d29SSarah Walker  * submission is serialized in drm_sched_main()).
146*d2d79d29SSarah Walker  */
147*d2d79d29SSarah Walker void
pvr_cccb_write_command_with_header(struct pvr_cccb * pvr_cccb,u32 cmd_type,u32 cmd_size,void * cmd_data,u32 ext_job_ref,u32 int_job_ref)148*d2d79d29SSarah Walker pvr_cccb_write_command_with_header(struct pvr_cccb *pvr_cccb, u32 cmd_type, u32 cmd_size,
149*d2d79d29SSarah Walker 				   void *cmd_data, u32 ext_job_ref, u32 int_job_ref)
150*d2d79d29SSarah Walker {
151*d2d79d29SSarah Walker 	u32 sz_with_hdr = pvr_cccb_get_size_of_cmd_with_hdr(cmd_size);
152*d2d79d29SSarah Walker 	struct rogue_fwif_ccb_cmd_header cmd_header = {
153*d2d79d29SSarah Walker 		.cmd_type = cmd_type,
154*d2d79d29SSarah Walker 		.cmd_size = ALIGN(cmd_size, 8),
155*d2d79d29SSarah Walker 		.ext_job_ref = ext_job_ref,
156*d2d79d29SSarah Walker 		.int_job_ref = int_job_ref,
157*d2d79d29SSarah Walker 	};
158*d2d79d29SSarah Walker 	struct rogue_fwif_cccb_ctl *ctrl = pvr_cccb->ctrl;
159*d2d79d29SSarah Walker 	u32 remaining = pvr_cccb->size - pvr_cccb->write_offset;
160*d2d79d29SSarah Walker 	u32 required_size, cccb_space, read_offset;
161*d2d79d29SSarah Walker 
162*d2d79d29SSarah Walker 	/*
163*d2d79d29SSarah Walker 	 * Always ensure we have enough room for a padding command at the end of
164*d2d79d29SSarah Walker 	 * the CCCB.
165*d2d79d29SSarah Walker 	 */
166*d2d79d29SSarah Walker 	if (remaining < sz_with_hdr + PADDING_COMMAND_SIZE) {
167*d2d79d29SSarah Walker 		/*
168*d2d79d29SSarah Walker 		 * Command would need to wrap, so we need to pad the remainder
169*d2d79d29SSarah Walker 		 * of the CCCB.
170*d2d79d29SSarah Walker 		 */
171*d2d79d29SSarah Walker 		required_size = sz_with_hdr + remaining;
172*d2d79d29SSarah Walker 	} else {
173*d2d79d29SSarah Walker 		required_size = sz_with_hdr;
174*d2d79d29SSarah Walker 	}
175*d2d79d29SSarah Walker 
176*d2d79d29SSarah Walker 	read_offset = READ_ONCE(ctrl->read_offset);
177*d2d79d29SSarah Walker 	cccb_space = get_ccb_space(pvr_cccb->write_offset, read_offset, pvr_cccb->size);
178*d2d79d29SSarah Walker 	if (WARN_ON(cccb_space < required_size))
179*d2d79d29SSarah Walker 		return;
180*d2d79d29SSarah Walker 
181*d2d79d29SSarah Walker 	if (required_size != sz_with_hdr) {
182*d2d79d29SSarah Walker 		/* Add padding command */
183*d2d79d29SSarah Walker 		struct rogue_fwif_ccb_cmd_header pad_cmd = {
184*d2d79d29SSarah Walker 			.cmd_type = ROGUE_FWIF_CCB_CMD_TYPE_PADDING,
185*d2d79d29SSarah Walker 			.cmd_size = remaining - sizeof(pad_cmd),
186*d2d79d29SSarah Walker 		};
187*d2d79d29SSarah Walker 
188*d2d79d29SSarah Walker 		memcpy(&pvr_cccb->cccb[pvr_cccb->write_offset], &pad_cmd, sizeof(pad_cmd));
189*d2d79d29SSarah Walker 		pvr_cccb->write_offset = 0;
190*d2d79d29SSarah Walker 	}
191*d2d79d29SSarah Walker 
192*d2d79d29SSarah Walker 	memcpy(&pvr_cccb->cccb[pvr_cccb->write_offset], &cmd_header, sizeof(cmd_header));
193*d2d79d29SSarah Walker 	memcpy(&pvr_cccb->cccb[pvr_cccb->write_offset + sizeof(cmd_header)], cmd_data, cmd_size);
194*d2d79d29SSarah Walker 	pvr_cccb->write_offset += sz_with_hdr;
195*d2d79d29SSarah Walker }
196*d2d79d29SSarah Walker 
fill_cmd_kick_data(struct pvr_cccb * cccb,u32 ctx_fw_addr,struct pvr_hwrt_data * hwrt,struct rogue_fwif_kccb_cmd_kick_data * k)197*d2d79d29SSarah Walker static void fill_cmd_kick_data(struct pvr_cccb *cccb, u32 ctx_fw_addr,
198*d2d79d29SSarah Walker 			       struct pvr_hwrt_data *hwrt,
199*d2d79d29SSarah Walker 			       struct rogue_fwif_kccb_cmd_kick_data *k)
200*d2d79d29SSarah Walker {
201*d2d79d29SSarah Walker 	k->context_fw_addr = ctx_fw_addr;
202*d2d79d29SSarah Walker 	k->client_woff_update = cccb->write_offset;
203*d2d79d29SSarah Walker 	k->client_wrap_mask_update = cccb->wrap_mask;
204*d2d79d29SSarah Walker 
205*d2d79d29SSarah Walker 	if (hwrt) {
206*d2d79d29SSarah Walker 		u32 cleanup_state_offset = offsetof(struct rogue_fwif_hwrtdata, cleanup_state);
207*d2d79d29SSarah Walker 
208*d2d79d29SSarah Walker 		pvr_fw_object_get_fw_addr_offset(hwrt->fw_obj, cleanup_state_offset,
209*d2d79d29SSarah Walker 						 &k->cleanup_ctl_fw_addr[k->num_cleanup_ctl++]);
210*d2d79d29SSarah Walker 	}
211*d2d79d29SSarah Walker }
212*d2d79d29SSarah Walker 
213*d2d79d29SSarah Walker /**
214*d2d79d29SSarah Walker  * pvr_cccb_send_kccb_kick: Send KCCB kick to trigger command processing
215*d2d79d29SSarah Walker  * @pvr_dev: Device pointer.
216*d2d79d29SSarah Walker  * @pvr_cccb: Pointer to CCCB to process.
217*d2d79d29SSarah Walker  * @cctx_fw_addr: FW virtual address for context owning this Client CCB.
218*d2d79d29SSarah Walker  * @hwrt: HWRT data set associated with this kick. May be %NULL.
219*d2d79d29SSarah Walker  *
220*d2d79d29SSarah Walker  * You must call pvr_kccb_reserve_slot() and wait for the returned fence to
221*d2d79d29SSarah Walker  * signal (if this function didn't return NULL) before calling
222*d2d79d29SSarah Walker  * pvr_cccb_send_kccb_kick().
223*d2d79d29SSarah Walker  */
224*d2d79d29SSarah Walker void
pvr_cccb_send_kccb_kick(struct pvr_device * pvr_dev,struct pvr_cccb * pvr_cccb,u32 cctx_fw_addr,struct pvr_hwrt_data * hwrt)225*d2d79d29SSarah Walker pvr_cccb_send_kccb_kick(struct pvr_device *pvr_dev,
226*d2d79d29SSarah Walker 			struct pvr_cccb *pvr_cccb, u32 cctx_fw_addr,
227*d2d79d29SSarah Walker 			struct pvr_hwrt_data *hwrt)
228*d2d79d29SSarah Walker {
229*d2d79d29SSarah Walker 	struct rogue_fwif_kccb_cmd cmd_kick = {
230*d2d79d29SSarah Walker 		.cmd_type = ROGUE_FWIF_KCCB_CMD_KICK,
231*d2d79d29SSarah Walker 	};
232*d2d79d29SSarah Walker 
233*d2d79d29SSarah Walker 	fill_cmd_kick_data(pvr_cccb, cctx_fw_addr, hwrt, &cmd_kick.cmd_data.cmd_kick_data);
234*d2d79d29SSarah Walker 
235*d2d79d29SSarah Walker 	/* Make sure the writes to the CCCB are flushed before sending the KICK. */
236*d2d79d29SSarah Walker 	wmb();
237*d2d79d29SSarah Walker 
238*d2d79d29SSarah Walker 	pvr_kccb_send_cmd_reserved_powered(pvr_dev, &cmd_kick, NULL);
239*d2d79d29SSarah Walker }
240*d2d79d29SSarah Walker 
241*d2d79d29SSarah Walker void
pvr_cccb_send_kccb_combined_kick(struct pvr_device * pvr_dev,struct pvr_cccb * geom_cccb,struct pvr_cccb * frag_cccb,u32 geom_ctx_fw_addr,u32 frag_ctx_fw_addr,struct pvr_hwrt_data * hwrt,bool frag_is_pr)242*d2d79d29SSarah Walker pvr_cccb_send_kccb_combined_kick(struct pvr_device *pvr_dev,
243*d2d79d29SSarah Walker 				 struct pvr_cccb *geom_cccb,
244*d2d79d29SSarah Walker 				 struct pvr_cccb *frag_cccb,
245*d2d79d29SSarah Walker 				 u32 geom_ctx_fw_addr,
246*d2d79d29SSarah Walker 				 u32 frag_ctx_fw_addr,
247*d2d79d29SSarah Walker 				 struct pvr_hwrt_data *hwrt,
248*d2d79d29SSarah Walker 				 bool frag_is_pr)
249*d2d79d29SSarah Walker {
250*d2d79d29SSarah Walker 	struct rogue_fwif_kccb_cmd cmd_kick = {
251*d2d79d29SSarah Walker 		.cmd_type = ROGUE_FWIF_KCCB_CMD_COMBINED_GEOM_FRAG_KICK,
252*d2d79d29SSarah Walker 	};
253*d2d79d29SSarah Walker 
254*d2d79d29SSarah Walker 	fill_cmd_kick_data(geom_cccb, geom_ctx_fw_addr, hwrt,
255*d2d79d29SSarah Walker 			   &cmd_kick.cmd_data.combined_geom_frag_cmd_kick_data.geom_cmd_kick_data);
256*d2d79d29SSarah Walker 
257*d2d79d29SSarah Walker 	/* If this is a partial-render job, we don't attach resources to cleanup-ctl array,
258*d2d79d29SSarah Walker 	 * because the resources are already retained by the geometry job.
259*d2d79d29SSarah Walker 	 */
260*d2d79d29SSarah Walker 	fill_cmd_kick_data(frag_cccb, frag_ctx_fw_addr, frag_is_pr ? NULL : hwrt,
261*d2d79d29SSarah Walker 			   &cmd_kick.cmd_data.combined_geom_frag_cmd_kick_data.frag_cmd_kick_data);
262*d2d79d29SSarah Walker 
263*d2d79d29SSarah Walker 	/* Make sure the writes to the CCCB are flushed before sending the KICK. */
264*d2d79d29SSarah Walker 	wmb();
265*d2d79d29SSarah Walker 
266*d2d79d29SSarah Walker 	pvr_kccb_send_cmd_reserved_powered(pvr_dev, &cmd_kick, NULL);
267*d2d79d29SSarah Walker }
268