xref: /linux/drivers/gpu/drm/amd/include/kgd_kfd_interface.h (revision 8e07e0e3964ca4e23ce7b68e2096fe660a888942)
1 /*
2  * Copyright 2014 Advanced Micro Devices, Inc.
3  *
4  * Permission is hereby granted, free of charge, to any person obtaining a
5  * copy of this software and associated documentation files (the "Software"),
6  * to deal in the Software without restriction, including without limitation
7  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8  * and/or sell copies of the Software, and to permit persons to whom the
9  * Software is furnished to do so, subject to the following conditions:
10  *
11  * The above copyright notice and this permission notice shall be included in
12  * all copies or substantial portions of the Software.
13  *
14  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
17  * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18  * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19  * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20  * OTHER DEALINGS IN THE SOFTWARE.
21  */
22 
23 /*
24  * This file defines the private interface between the
25  * AMD kernel graphics drivers and the AMD KFD.
26  */
27 
28 #ifndef KGD_KFD_INTERFACE_H_INCLUDED
29 #define KGD_KFD_INTERFACE_H_INCLUDED
30 
31 #include <linux/types.h>
32 #include <linux/bitmap.h>
33 #include <linux/dma-fence.h>
34 #include "amdgpu_irq.h"
35 #include "amdgpu_gfx.h"
36 
37 struct pci_dev;
38 struct amdgpu_device;
39 
40 struct kfd_dev;
41 struct kgd_mem;
42 
43 enum kfd_preempt_type {
44 	KFD_PREEMPT_TYPE_WAVEFRONT_DRAIN = 0,
45 	KFD_PREEMPT_TYPE_WAVEFRONT_RESET,
46 	KFD_PREEMPT_TYPE_WAVEFRONT_SAVE
47 };
48 
49 struct kfd_vm_fault_info {
50 	uint64_t	page_addr;
51 	uint32_t	vmid;
52 	uint32_t	mc_id;
53 	uint32_t	status;
54 	bool		prot_valid;
55 	bool		prot_read;
56 	bool		prot_write;
57 	bool		prot_exec;
58 };
59 
60 /* For getting GPU local memory information from KGD */
61 struct kfd_local_mem_info {
62 	uint64_t local_mem_size_private;
63 	uint64_t local_mem_size_public;
64 	uint32_t vram_width;
65 	uint32_t mem_clk_max;
66 };
67 
68 enum kgd_memory_pool {
69 	KGD_POOL_SYSTEM_CACHEABLE = 1,
70 	KGD_POOL_SYSTEM_WRITECOMBINE = 2,
71 	KGD_POOL_FRAMEBUFFER = 3,
72 };
73 
74 /**
75  * enum kfd_sched_policy
76  *
77  * @KFD_SCHED_POLICY_HWS: H/W scheduling policy known as command processor (cp)
78  * scheduling. In this scheduling mode we're using the firmware code to
79  * schedule the user mode queues and kernel queues such as HIQ and DIQ.
80  * the HIQ queue is used as a special queue that dispatches the configuration
81  * to the cp and the user mode queues list that are currently running.
82  * the DIQ queue is a debugging queue that dispatches debugging commands to the
83  * firmware.
84  * in this scheduling mode user mode queues over subscription feature is
85  * enabled.
86  *
87  * @KFD_SCHED_POLICY_HWS_NO_OVERSUBSCRIPTION: The same as above but the over
88  * subscription feature disabled.
89  *
90  * @KFD_SCHED_POLICY_NO_HWS: no H/W scheduling policy is a mode which directly
91  * set the command processor registers and sets the queues "manually". This
92  * mode is used *ONLY* for debugging proposes.
93  *
94  */
95 enum kfd_sched_policy {
96 	KFD_SCHED_POLICY_HWS = 0,
97 	KFD_SCHED_POLICY_HWS_NO_OVERSUBSCRIPTION,
98 	KFD_SCHED_POLICY_NO_HWS
99 };
100 
101 struct kgd2kfd_shared_resources {
102 	/* Bit n == 1 means VMID n is available for KFD. */
103 	unsigned int compute_vmid_bitmap;
104 
105 	/* number of pipes per mec */
106 	uint32_t num_pipe_per_mec;
107 
108 	/* number of queues per pipe */
109 	uint32_t num_queue_per_pipe;
110 
111 	/* Bit n == 1 means Queue n is available for KFD */
112 	DECLARE_BITMAP(cp_queue_bitmap, AMDGPU_MAX_QUEUES);
113 
114 	/* SDMA doorbell assignments (SOC15 and later chips only). Only
115 	 * specific doorbells are routed to each SDMA engine. Others
116 	 * are routed to IH and VCN. They are not usable by the CP.
117 	 */
118 	uint32_t *sdma_doorbell_idx;
119 
120 	/* From SOC15 onward, the doorbell index range not usable for CP
121 	 * queues.
122 	 */
123 	uint32_t non_cp_doorbells_start;
124 	uint32_t non_cp_doorbells_end;
125 
126 	/* Base address of doorbell aperture. */
127 	phys_addr_t doorbell_physical_address;
128 
129 	/* Size in bytes of doorbell aperture. */
130 	size_t doorbell_aperture_size;
131 
132 	/* Number of bytes at start of aperture reserved for KGD. */
133 	size_t doorbell_start_offset;
134 
135 	/* GPUVM address space size in bytes */
136 	uint64_t gpuvm_size;
137 
138 	/* Minor device number of the render node */
139 	int drm_render_minor;
140 
141 	bool enable_mes;
142 };
143 
144 struct tile_config {
145 	uint32_t *tile_config_ptr;
146 	uint32_t *macro_tile_config_ptr;
147 	uint32_t num_tile_configs;
148 	uint32_t num_macro_tile_configs;
149 
150 	uint32_t gb_addr_config;
151 	uint32_t num_banks;
152 	uint32_t num_ranks;
153 };
154 
155 #define KFD_MAX_NUM_OF_QUEUES_PER_DEVICE_DEFAULT 4096
156 
157 /**
158  * struct kfd2kgd_calls
159  *
160  * @program_sh_mem_settings: A function that should initiate the memory
161  * properties such as main aperture memory type (cache / non cached) and
162  * secondary aperture base address, size and memory type.
163  * This function is used only for no cp scheduling mode.
164  *
165  * @set_pasid_vmid_mapping: Exposes pasid/vmid pair to the H/W for no cp
166  * scheduling mode. Only used for no cp scheduling mode.
167  *
168  * @hqd_load: Loads the mqd structure to a H/W hqd slot. used only for no cp
169  * sceduling mode.
170  *
171  * @hqd_sdma_load: Loads the SDMA mqd structure to a H/W SDMA hqd slot.
172  * used only for no HWS mode.
173  *
174  * @hqd_dump: Dumps CPC HQD registers to an array of address-value pairs.
175  * Array is allocated with kmalloc, needs to be freed with kfree by caller.
176  *
177  * @hqd_sdma_dump: Dumps SDMA HQD registers to an array of address-value pairs.
178  * Array is allocated with kmalloc, needs to be freed with kfree by caller.
179  *
180  * @hqd_is_occupies: Checks if a hqd slot is occupied.
181  *
182  * @hqd_destroy: Destructs and preempts the queue assigned to that hqd slot.
183  *
184  * @hqd_sdma_is_occupied: Checks if an SDMA hqd slot is occupied.
185  *
186  * @hqd_sdma_destroy: Destructs and preempts the SDMA queue assigned to that
187  * SDMA hqd slot.
188  *
189  * @set_scratch_backing_va: Sets VA for scratch backing memory of a VMID.
190  * Only used for no cp scheduling mode
191  *
192  * @set_vm_context_page_table_base: Program page table base for a VMID
193  *
194  * @invalidate_tlbs: Invalidate TLBs for a specific PASID
195  *
196  * @invalidate_tlbs_vmid: Invalidate TLBs for a specific VMID
197  *
198  * @read_vmid_from_vmfault_reg: On Hawaii the VMID is not set in the
199  * IH ring entry. This function allows the KFD ISR to get the VMID
200  * from the fault status register as early as possible.
201  *
202  * @get_cu_occupancy: Function pointer that returns to caller the number
203  * of wave fronts that are in flight for all of the queues of a process
204  * as identified by its pasid. It is important to note that the value
205  * returned by this function is a snapshot of current moment and cannot
206  * guarantee any minimum for the number of waves in-flight. This function
207  * is defined for devices that belong to GFX9 and later GFX families. Care
208  * must be taken in calling this function as it is not defined for devices
209  * that belong to GFX8 and below GFX families.
210  *
211  * This structure contains function pointers to services that the kgd driver
212  * provides to amdkfd driver.
213  *
214  */
215 struct kfd2kgd_calls {
216 	/* Register access functions */
217 	void (*program_sh_mem_settings)(struct amdgpu_device *adev, uint32_t vmid,
218 			uint32_t sh_mem_config,	uint32_t sh_mem_ape1_base,
219 			uint32_t sh_mem_ape1_limit, uint32_t sh_mem_bases,
220 			uint32_t inst);
221 
222 	int (*set_pasid_vmid_mapping)(struct amdgpu_device *adev, u32 pasid,
223 					unsigned int vmid, uint32_t inst);
224 
225 	int (*init_interrupts)(struct amdgpu_device *adev, uint32_t pipe_id,
226 			uint32_t inst);
227 
228 	int (*hqd_load)(struct amdgpu_device *adev, void *mqd, uint32_t pipe_id,
229 			uint32_t queue_id, uint32_t __user *wptr,
230 			uint32_t wptr_shift, uint32_t wptr_mask,
231 			struct mm_struct *mm, uint32_t inst);
232 
233 	int (*hiq_mqd_load)(struct amdgpu_device *adev, void *mqd,
234 			    uint32_t pipe_id, uint32_t queue_id,
235 			    uint32_t doorbell_off, uint32_t inst);
236 
237 	int (*hqd_sdma_load)(struct amdgpu_device *adev, void *mqd,
238 			     uint32_t __user *wptr, struct mm_struct *mm);
239 
240 	int (*hqd_dump)(struct amdgpu_device *adev,
241 			uint32_t pipe_id, uint32_t queue_id,
242 			uint32_t (**dump)[2], uint32_t *n_regs, uint32_t inst);
243 
244 	int (*hqd_sdma_dump)(struct amdgpu_device *adev,
245 			     uint32_t engine_id, uint32_t queue_id,
246 			     uint32_t (**dump)[2], uint32_t *n_regs);
247 
248 	bool (*hqd_is_occupied)(struct amdgpu_device *adev,
249 				uint64_t queue_address, uint32_t pipe_id,
250 				uint32_t queue_id, uint32_t inst);
251 
252 	int (*hqd_destroy)(struct amdgpu_device *adev, void *mqd,
253 				enum kfd_preempt_type reset_type,
254 				unsigned int timeout, uint32_t pipe_id,
255 				uint32_t queue_id, uint32_t inst);
256 
257 	bool (*hqd_sdma_is_occupied)(struct amdgpu_device *adev, void *mqd);
258 
259 	int (*hqd_sdma_destroy)(struct amdgpu_device *adev, void *mqd,
260 				unsigned int timeout);
261 
262 	int (*wave_control_execute)(struct amdgpu_device *adev,
263 					uint32_t gfx_index_val,
264 					uint32_t sq_cmd, uint32_t inst);
265 	bool (*get_atc_vmid_pasid_mapping_info)(struct amdgpu_device *adev,
266 					uint8_t vmid,
267 					uint16_t *p_pasid);
268 
269 	/* No longer needed from GFXv9 onward. The scratch base address is
270 	 * passed to the shader by the CP. It's the user mode driver's
271 	 * responsibility.
272 	 */
273 	void (*set_scratch_backing_va)(struct amdgpu_device *adev,
274 				uint64_t va, uint32_t vmid);
275 
276 	void (*set_vm_context_page_table_base)(struct amdgpu_device *adev,
277 			uint32_t vmid, uint64_t page_table_base);
278 	uint32_t (*read_vmid_from_vmfault_reg)(struct amdgpu_device *adev);
279 
280 	uint32_t (*enable_debug_trap)(struct amdgpu_device *adev,
281 					bool restore_dbg_registers,
282 					uint32_t vmid);
283 	uint32_t (*disable_debug_trap)(struct amdgpu_device *adev,
284 					bool keep_trap_enabled,
285 					uint32_t vmid);
286 	int (*validate_trap_override_request)(struct amdgpu_device *adev,
287 					uint32_t trap_override,
288 					uint32_t *trap_mask_supported);
289 	uint32_t (*set_wave_launch_trap_override)(struct amdgpu_device *adev,
290 					     uint32_t vmid,
291 					     uint32_t trap_override,
292 					     uint32_t trap_mask_bits,
293 					     uint32_t trap_mask_request,
294 					     uint32_t *trap_mask_prev,
295 					     uint32_t kfd_dbg_trap_cntl_prev);
296 	uint32_t (*set_wave_launch_mode)(struct amdgpu_device *adev,
297 					uint8_t wave_launch_mode,
298 					uint32_t vmid);
299 	uint32_t (*set_address_watch)(struct amdgpu_device *adev,
300 					uint64_t watch_address,
301 					uint32_t watch_address_mask,
302 					uint32_t watch_id,
303 					uint32_t watch_mode,
304 					uint32_t debug_vmid,
305 					uint32_t inst);
306 	uint32_t (*clear_address_watch)(struct amdgpu_device *adev,
307 			uint32_t watch_id);
308 	void (*get_iq_wait_times)(struct amdgpu_device *adev,
309 			uint32_t *wait_times,
310 			uint32_t inst);
311 	void (*build_grace_period_packet_info)(struct amdgpu_device *adev,
312 			uint32_t wait_times,
313 			uint32_t grace_period,
314 			uint32_t *reg_offset,
315 			uint32_t *reg_data);
316 	void (*get_cu_occupancy)(struct amdgpu_device *adev, int pasid,
317 			int *wave_cnt, int *max_waves_per_cu, uint32_t inst);
318 	void (*program_trap_handler_settings)(struct amdgpu_device *adev,
319 			uint32_t vmid, uint64_t tba_addr, uint64_t tma_addr,
320 			uint32_t inst);
321 };
322 
323 #endif	/* KGD_KFD_INTERFACE_H_INCLUDED */
324