1 /* 2 * Copyright 2020 Advanced Micro Devices, Inc. 3 * 4 * Permission is hereby granted, free of charge, to any person obtaining a 5 * copy of this software and associated documentation files (the "Software"), 6 * to deal in the Software without restriction, including without limitation 7 * the rights to use, copy, modify, merge, publish, distribute, sublicense, 8 * and/or sell copies of the Software, and to permit persons to whom the 9 * Software is furnished to do so, subject to the following conditions: 10 * 11 * The above copyright notice and this permission notice shall be included in 12 * all copies or substantial portions of the Software. 13 * 14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR 18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, 19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR 20 * OTHER DEALINGS IN THE SOFTWARE. 21 * 22 * Authors: Christian König 23 */ 24 25 #ifndef _TTM_DEVICE_H_ 26 #define _TTM_DEVICE_H_ 27 28 #include <linux/types.h> 29 #include <linux/workqueue.h> 30 #include <drm/ttm/ttm_allocation.h> 31 #include <drm/ttm/ttm_resource.h> 32 #include <drm/ttm/ttm_pool.h> 33 34 struct ttm_device; 35 struct ttm_placement; 36 struct ttm_buffer_object; 37 struct ttm_operation_ctx; 38 39 /** 40 * struct ttm_global - Buffer object driver global data. 41 */ 42 extern struct ttm_global { 43 44 /** 45 * @dummy_read_page: Pointer to a dummy page used for mapping requests 46 * of unpopulated pages. Constant after init. 47 */ 48 struct page *dummy_read_page; 49 50 /** 51 * @device_list: List of buffer object devices. Protected by 52 * ttm_global_mutex. 53 */ 54 struct list_head device_list; 55 56 /** 57 * @bo_count: Number of buffer objects allocated by devices. 58 */ 59 atomic_t bo_count; 60 } ttm_glob; 61 62 struct ttm_device_funcs { 63 /** 64 * ttm_tt_create 65 * 66 * @bo: The buffer object to create the ttm for. 67 * @page_flags: Page flags as identified by TTM_TT_FLAG_XX flags. 68 * 69 * Create a struct ttm_tt to back data with system memory pages. 70 * No pages are actually allocated. 71 * Returns: 72 * NULL: Out of memory. 73 */ 74 struct ttm_tt *(*ttm_tt_create)(struct ttm_buffer_object *bo, 75 uint32_t page_flags); 76 77 /** 78 * ttm_tt_populate 79 * 80 * @ttm: The struct ttm_tt to contain the backing pages. 81 * 82 * Allocate all backing pages 83 * Returns: 84 * -ENOMEM: Out of memory. 85 */ 86 int (*ttm_tt_populate)(struct ttm_device *bdev, 87 struct ttm_tt *ttm, 88 struct ttm_operation_ctx *ctx); 89 90 /** 91 * ttm_tt_unpopulate 92 * 93 * @ttm: The struct ttm_tt to contain the backing pages. 94 * 95 * Free all backing page 96 */ 97 void (*ttm_tt_unpopulate)(struct ttm_device *bdev, 98 struct ttm_tt *ttm); 99 100 /** 101 * ttm_tt_destroy 102 * 103 * @bdev: Pointer to a ttm device 104 * @ttm: Pointer to a struct ttm_tt. 105 * 106 * Destroy the backend. This will be call back from ttm_tt_destroy so 107 * don't call ttm_tt_destroy from the callback or infinite loop. 108 */ 109 void (*ttm_tt_destroy)(struct ttm_device *bdev, struct ttm_tt *ttm); 110 111 /** 112 * struct ttm_bo_driver member eviction_valuable 113 * 114 * @bo: the buffer object to be evicted 115 * @place: placement we need room for 116 * 117 * Check with the driver if it is valuable to evict a BO to make room 118 * for a certain placement. 119 */ 120 bool (*eviction_valuable)(struct ttm_buffer_object *bo, 121 const struct ttm_place *place); 122 /** 123 * struct ttm_bo_driver member evict_flags: 124 * 125 * @bo: the buffer object to be evicted 126 * 127 * Return the bo flags for a buffer which is not mapped to the hardware. 128 * These will be placed in proposed_flags so that when the move is 129 * finished, they'll end up in bo->mem.flags 130 * This should not cause multihop evictions, and the core will warn 131 * if one is proposed. 132 */ 133 134 void (*evict_flags)(struct ttm_buffer_object *bo, 135 struct ttm_placement *placement); 136 137 /** 138 * struct ttm_bo_driver member move: 139 * 140 * @bo: the buffer to move 141 * @evict: whether this motion is evicting the buffer from 142 * the graphics address space 143 * @ctx: context for this move with parameters 144 * @new_mem: the new memory region receiving the buffer 145 * @hop: placement for driver directed intermediate hop 146 * 147 * Move a buffer between two memory regions. 148 * Returns errno -EMULTIHOP if driver requests a hop 149 */ 150 int (*move)(struct ttm_buffer_object *bo, bool evict, 151 struct ttm_operation_ctx *ctx, 152 struct ttm_resource *new_mem, 153 struct ttm_place *hop); 154 155 /** 156 * Hook to notify driver about a resource delete. 157 */ 158 void (*delete_mem_notify)(struct ttm_buffer_object *bo); 159 160 /** 161 * notify the driver that we're about to swap out this bo 162 */ 163 void (*swap_notify)(struct ttm_buffer_object *bo); 164 165 /** 166 * Driver callback on when mapping io memory (for bo_move_memcpy 167 * for instance). TTM will take care to call io_mem_free whenever 168 * the mapping is not use anymore. io_mem_reserve & io_mem_free 169 * are balanced. 170 */ 171 int (*io_mem_reserve)(struct ttm_device *bdev, 172 struct ttm_resource *mem); 173 void (*io_mem_free)(struct ttm_device *bdev, 174 struct ttm_resource *mem); 175 176 /** 177 * Return the pfn for a given page_offset inside the BO. 178 * 179 * @bo: the BO to look up the pfn for 180 * @page_offset: the offset to look up 181 */ 182 unsigned long (*io_mem_pfn)(struct ttm_buffer_object *bo, 183 unsigned long page_offset); 184 185 /** 186 * Read/write memory buffers for ptrace access 187 * 188 * @bo: the BO to access 189 * @offset: the offset from the start of the BO 190 * @buf: pointer to source/destination buffer 191 * @len: number of bytes to copy 192 * @write: whether to read (0) from or write (non-0) to BO 193 * 194 * If successful, this function should return the number of 195 * bytes copied, -EIO otherwise. If the number of bytes 196 * returned is < len, the function may be called again with 197 * the remainder of the buffer to copy. 198 */ 199 int (*access_memory)(struct ttm_buffer_object *bo, unsigned long offset, 200 void *buf, int len, int write); 201 202 /** 203 * Notify the driver that we're about to release a BO 204 * 205 * @bo: BO that is about to be released 206 * 207 * Gives the driver a chance to do any cleanup, including 208 * adding fences that may force a delayed delete 209 */ 210 void (*release_notify)(struct ttm_buffer_object *bo); 211 }; 212 213 /** 214 * struct ttm_device - Buffer object driver device-specific data. 215 */ 216 struct ttm_device { 217 /** 218 * @device_list: Our entry in the global device list. 219 * Constant after bo device init 220 */ 221 struct list_head device_list; 222 223 /** 224 * @alloc_flags: TTM_ALLOCATION_* flags. 225 */ 226 unsigned int alloc_flags; 227 228 /** 229 * @funcs: Function table for the device. 230 * Constant after bo device init 231 */ 232 const struct ttm_device_funcs *funcs; 233 234 /** 235 * @sysman: Resource manager for the system domain. 236 * Access via ttm_manager_type. 237 */ 238 struct ttm_resource_manager sysman; 239 240 /** 241 * @man_drv: An array of resource_managers, one per resource type. 242 */ 243 struct ttm_resource_manager *man_drv[TTM_NUM_MEM_TYPES]; 244 245 /** 246 * @vma_manager: Address space manager for finding BOs to mmap. 247 */ 248 struct drm_vma_offset_manager *vma_manager; 249 250 /** 251 * @pool: page pool for the device. 252 */ 253 struct ttm_pool pool; 254 255 /** 256 * @lru_lock: Protection for the per manager LRU and ddestroy lists. 257 */ 258 spinlock_t lru_lock; 259 260 /** 261 * @unevictable: Buffer objects which are pinned or swapped and as such 262 * not on an LRU list. 263 */ 264 struct list_head unevictable; 265 266 /** 267 * @dev_mapping: A pointer to the struct address_space for invalidating 268 * CPU mappings on buffer move. Protected by load/unload sync. 269 */ 270 struct address_space *dev_mapping; 271 272 /** 273 * @wq: Work queue structure for the delayed delete workqueue. 274 */ 275 struct workqueue_struct *wq; 276 }; 277 278 int ttm_global_swapout(struct ttm_operation_ctx *ctx, gfp_t gfp_flags); 279 int ttm_device_swapout(struct ttm_device *bdev, struct ttm_operation_ctx *ctx, 280 gfp_t gfp_flags); 281 int ttm_device_prepare_hibernation(struct ttm_device *bdev); 282 283 static inline struct ttm_resource_manager * 284 ttm_manager_type(struct ttm_device *bdev, int mem_type) 285 { 286 BUILD_BUG_ON(__builtin_constant_p(mem_type) 287 && mem_type >= TTM_NUM_MEM_TYPES); 288 return bdev->man_drv[mem_type]; 289 } 290 291 static inline void ttm_set_driver_manager(struct ttm_device *bdev, int type, 292 struct ttm_resource_manager *manager) 293 { 294 BUILD_BUG_ON(__builtin_constant_p(type) && type >= TTM_NUM_MEM_TYPES); 295 bdev->man_drv[type] = manager; 296 } 297 298 int ttm_device_init(struct ttm_device *bdev, const struct ttm_device_funcs *funcs, 299 struct device *dev, struct address_space *mapping, 300 struct drm_vma_offset_manager *vma_manager, 301 unsigned int alloc_flags); 302 void ttm_device_fini(struct ttm_device *bdev); 303 void ttm_device_clear_dma_mappings(struct ttm_device *bdev); 304 305 #endif 306