xref: /linux/drivers/gpu/drm/i915/intel_region_ttm.c (revision da1d9caf95def6f0320819cf941c9fd1069ba9e1)
1 // SPDX-License-Identifier: MIT
2 /*
3  * Copyright © 2021 Intel Corporation
4  */
5 #include <drm/ttm/ttm_bo_driver.h>
6 #include <drm/ttm/ttm_device.h>
7 #include <drm/ttm/ttm_range_manager.h>
8 
9 #include "i915_drv.h"
10 #include "i915_scatterlist.h"
11 #include "i915_ttm_buddy_manager.h"
12 
13 #include "intel_region_ttm.h"
14 
15 #include "gem/i915_gem_region.h"
16 #include "gem/i915_gem_ttm.h" /* For the funcs/ops export only */
17 /**
18  * DOC: TTM support structure
19  *
20  * The code in this file deals with setting up memory managers for TTM
21  * LMEM and MOCK regions and converting the output from
22  * the managers to struct sg_table, Basically providing the mapping from
23  * i915 GEM regions to TTM memory types and resource managers.
24  */
25 
26 /**
27  * intel_region_ttm_device_init - Initialize a TTM device
28  * @dev_priv: Pointer to an i915 device private structure.
29  *
30  * Return: 0 on success, negative error code on failure.
31  */
32 int intel_region_ttm_device_init(struct drm_i915_private *dev_priv)
33 {
34 	struct drm_device *drm = &dev_priv->drm;
35 
36 	return ttm_device_init(&dev_priv->bdev, i915_ttm_driver(),
37 			       drm->dev, drm->anon_inode->i_mapping,
38 			       drm->vma_offset_manager, false, false);
39 }
40 
41 /**
42  * intel_region_ttm_device_fini - Finalize a TTM device
43  * @dev_priv: Pointer to an i915 device private structure.
44  */
45 void intel_region_ttm_device_fini(struct drm_i915_private *dev_priv)
46 {
47 	ttm_device_fini(&dev_priv->bdev);
48 }
49 
50 /*
51  * Map the i915 memory regions to TTM memory types. We use the
52  * driver-private types for now, reserving TTM_PL_VRAM for stolen
53  * memory and TTM_PL_TT for GGTT use if decided to implement this.
54  */
55 int intel_region_to_ttm_type(const struct intel_memory_region *mem)
56 {
57 	int type;
58 
59 	GEM_BUG_ON(mem->type != INTEL_MEMORY_LOCAL &&
60 		   mem->type != INTEL_MEMORY_MOCK &&
61 		   mem->type != INTEL_MEMORY_SYSTEM);
62 
63 	if (mem->type == INTEL_MEMORY_SYSTEM)
64 		return TTM_PL_SYSTEM;
65 
66 	type = mem->instance + TTM_PL_PRIV;
67 	GEM_BUG_ON(type >= TTM_NUM_MEM_TYPES);
68 
69 	return type;
70 }
71 
72 /**
73  * intel_region_ttm_init - Initialize a memory region for TTM.
74  * @mem: The region to initialize.
75  *
76  * This function initializes a suitable TTM resource manager for the
77  * region, and if it's a LMEM region type, attaches it to the TTM
78  * device. MOCK regions are NOT attached to the TTM device, since we don't
79  * have one for the mock selftests.
80  *
81  * Return: 0 on success, negative error code on failure.
82  */
83 int intel_region_ttm_init(struct intel_memory_region *mem)
84 {
85 	struct ttm_device *bdev = &mem->i915->bdev;
86 	int mem_type = intel_region_to_ttm_type(mem);
87 	int ret;
88 
89 	ret = i915_ttm_buddy_man_init(bdev, mem_type, false,
90 				      resource_size(&mem->region),
91 				      mem->io_size,
92 				      mem->min_page_size, PAGE_SIZE);
93 	if (ret)
94 		return ret;
95 
96 	mem->region_private = ttm_manager_type(bdev, mem_type);
97 
98 	return 0;
99 }
100 
101 /**
102  * intel_region_ttm_fini - Finalize a TTM region.
103  * @mem: The memory region
104  *
105  * This functions takes down the TTM resource manager associated with the
106  * memory region, and if it was registered with the TTM device,
107  * removes that registration.
108  */
109 int intel_region_ttm_fini(struct intel_memory_region *mem)
110 {
111 	struct ttm_resource_manager *man = mem->region_private;
112 	int ret = -EBUSY;
113 	int count;
114 
115 	/*
116 	 * Put the region's move fences. This releases requests that
117 	 * may hold on to contexts and vms that may hold on to buffer
118 	 * objects placed in this region.
119 	 */
120 	if (man)
121 		ttm_resource_manager_cleanup(man);
122 
123 	/* Flush objects from region. */
124 	for (count = 0; count < 10; ++count) {
125 		i915_gem_flush_free_objects(mem->i915);
126 
127 		mutex_lock(&mem->objects.lock);
128 		if (list_empty(&mem->objects.list))
129 			ret = 0;
130 		mutex_unlock(&mem->objects.lock);
131 		if (!ret)
132 			break;
133 
134 		msleep(20);
135 		flush_delayed_work(&mem->i915->bdev.wq);
136 	}
137 
138 	/* If we leaked objects, Don't free the region causing use after free */
139 	if (ret || !man)
140 		return ret;
141 
142 	ret = i915_ttm_buddy_man_fini(&mem->i915->bdev,
143 				      intel_region_to_ttm_type(mem));
144 	GEM_WARN_ON(ret);
145 	mem->region_private = NULL;
146 
147 	return ret;
148 }
149 
150 /**
151  * intel_region_ttm_resource_to_rsgt -
152  * Convert an opaque TTM resource manager resource to a refcounted sg_table.
153  * @mem: The memory region.
154  * @res: The resource manager resource obtained from the TTM resource manager.
155  *
156  * The gem backends typically use sg-tables for operations on the underlying
157  * io_memory. So provide a way for the backends to translate the
158  * nodes they are handed from TTM to sg-tables.
159  *
160  * Return: A malloced sg_table on success, an error pointer on failure.
161  */
162 struct i915_refct_sgt *
163 intel_region_ttm_resource_to_rsgt(struct intel_memory_region *mem,
164 				  struct ttm_resource *res)
165 {
166 	if (mem->is_range_manager) {
167 		struct ttm_range_mgr_node *range_node =
168 			to_ttm_range_mgr_node(res);
169 
170 		return i915_rsgt_from_mm_node(&range_node->mm_nodes[0],
171 					      mem->region.start);
172 	} else {
173 		return i915_rsgt_from_buddy_resource(res, mem->region.start);
174 	}
175 }
176 
177 #ifdef CONFIG_DRM_I915_SELFTEST
178 /**
179  * intel_region_ttm_resource_alloc - Allocate memory resources from a region
180  * @mem: The memory region,
181  * @size: The requested size in bytes
182  * @flags: Allocation flags
183  *
184  * This functionality is provided only for callers that need to allocate
185  * memory from standalone TTM range managers, without the TTM eviction
186  * functionality. Don't use if you are not completely sure that's the
187  * case. The returned opaque node can be converted to an sg_table using
188  * intel_region_ttm_resource_to_st(), and can be freed using
189  * intel_region_ttm_resource_free().
190  *
191  * Return: A valid pointer on success, an error pointer on failure.
192  */
193 struct ttm_resource *
194 intel_region_ttm_resource_alloc(struct intel_memory_region *mem,
195 				resource_size_t offset,
196 				resource_size_t size,
197 				unsigned int flags)
198 {
199 	struct ttm_resource_manager *man = mem->region_private;
200 	struct ttm_place place = {};
201 	struct ttm_buffer_object mock_bo = {};
202 	struct ttm_resource *res;
203 	int ret;
204 
205 	if (flags & I915_BO_ALLOC_CONTIGUOUS)
206 		place.flags |= TTM_PL_FLAG_CONTIGUOUS;
207 	if (offset != I915_BO_INVALID_OFFSET) {
208 		place.fpfn = offset >> PAGE_SHIFT;
209 		place.lpfn = place.fpfn + (size >> PAGE_SHIFT);
210 	} else if (mem->io_size && mem->io_size < mem->total) {
211 		if (flags & I915_BO_ALLOC_GPU_ONLY) {
212 			place.flags |= TTM_PL_FLAG_TOPDOWN;
213 		} else {
214 			place.fpfn = 0;
215 			place.lpfn = mem->io_size >> PAGE_SHIFT;
216 		}
217 	}
218 
219 	mock_bo.base.size = size;
220 	mock_bo.bdev = &mem->i915->bdev;
221 
222 	ret = man->func->alloc(man, &mock_bo, &place, &res);
223 	if (ret == -ENOSPC)
224 		ret = -ENXIO;
225 	if (!ret)
226 		res->bo = NULL; /* Rather blow up, then some uaf */
227 	return ret ? ERR_PTR(ret) : res;
228 }
229 
230 #endif
231 
232 /**
233  * intel_region_ttm_resource_free - Free a resource allocated from a resource manager
234  * @mem: The region the resource was allocated from.
235  * @res: The opaque resource representing an allocation.
236  */
237 void intel_region_ttm_resource_free(struct intel_memory_region *mem,
238 				    struct ttm_resource *res)
239 {
240 	struct ttm_resource_manager *man = mem->region_private;
241 	struct ttm_buffer_object mock_bo = {};
242 
243 	mock_bo.base.size = res->num_pages << PAGE_SHIFT;
244 	mock_bo.bdev = &mem->i915->bdev;
245 	res->bo = &mock_bo;
246 
247 	man->func->free(man, res);
248 }
249