1 /* 2 * SPDX-License-Identifier: MIT 3 * 4 * Copyright © 2019 Intel Corporation 5 */ 6 7 #include "gem/i915_gem_context.h" 8 #include "gem/i915_gem_pm.h" 9 10 #include "i915_drv.h" 11 #include "i915_globals.h" 12 13 #include "intel_context.h" 14 #include "intel_engine.h" 15 #include "intel_engine_pm.h" 16 #include "intel_ring.h" 17 18 static struct i915_global_context { 19 struct i915_global base; 20 struct kmem_cache *slab_ce; 21 } global; 22 23 static struct intel_context *intel_context_alloc(void) 24 { 25 return kmem_cache_zalloc(global.slab_ce, GFP_KERNEL); 26 } 27 28 void intel_context_free(struct intel_context *ce) 29 { 30 kmem_cache_free(global.slab_ce, ce); 31 } 32 33 struct intel_context * 34 intel_context_create(struct i915_gem_context *ctx, 35 struct intel_engine_cs *engine) 36 { 37 struct intel_context *ce; 38 39 ce = intel_context_alloc(); 40 if (!ce) 41 return ERR_PTR(-ENOMEM); 42 43 intel_context_init(ce, ctx, engine); 44 return ce; 45 } 46 47 int __intel_context_do_pin(struct intel_context *ce) 48 { 49 int err; 50 51 if (mutex_lock_interruptible(&ce->pin_mutex)) 52 return -EINTR; 53 54 if (likely(!atomic_read(&ce->pin_count))) { 55 intel_wakeref_t wakeref; 56 57 if (unlikely(!test_bit(CONTEXT_ALLOC_BIT, &ce->flags))) { 58 err = ce->ops->alloc(ce); 59 if (unlikely(err)) 60 goto err; 61 62 __set_bit(CONTEXT_ALLOC_BIT, &ce->flags); 63 } 64 65 err = 0; 66 with_intel_runtime_pm(ce->engine->uncore->rpm, wakeref) 67 err = ce->ops->pin(ce); 68 if (err) 69 goto err; 70 71 GEM_TRACE("%s context:%llx pin ring:{head:%04x, tail:%04x}\n", 72 ce->engine->name, ce->timeline->fence_context, 73 ce->ring->head, ce->ring->tail); 74 75 i915_gem_context_get(ce->gem_context); /* for ctx->ppgtt */ 76 77 smp_mb__before_atomic(); /* flush pin before it is visible */ 78 } 79 80 atomic_inc(&ce->pin_count); 81 GEM_BUG_ON(!intel_context_is_pinned(ce)); /* no overflow! */ 82 83 mutex_unlock(&ce->pin_mutex); 84 return 0; 85 86 err: 87 mutex_unlock(&ce->pin_mutex); 88 return err; 89 } 90 91 void intel_context_unpin(struct intel_context *ce) 92 { 93 if (likely(atomic_add_unless(&ce->pin_count, -1, 1))) 94 return; 95 96 /* We may be called from inside intel_context_pin() to evict another */ 97 intel_context_get(ce); 98 mutex_lock_nested(&ce->pin_mutex, SINGLE_DEPTH_NESTING); 99 100 if (likely(atomic_dec_and_test(&ce->pin_count))) { 101 GEM_TRACE("%s context:%llx retire\n", 102 ce->engine->name, ce->timeline->fence_context); 103 104 ce->ops->unpin(ce); 105 106 i915_gem_context_put(ce->gem_context); 107 intel_context_active_release(ce); 108 } 109 110 mutex_unlock(&ce->pin_mutex); 111 intel_context_put(ce); 112 } 113 114 static int __context_pin_state(struct i915_vma *vma) 115 { 116 u64 flags; 117 int err; 118 119 flags = i915_ggtt_pin_bias(vma) | PIN_OFFSET_BIAS; 120 flags |= PIN_HIGH | PIN_GLOBAL; 121 122 err = i915_vma_pin(vma, 0, 0, flags); 123 if (err) 124 return err; 125 126 err = i915_active_acquire(&vma->active); 127 if (err) 128 goto err_unpin; 129 130 /* 131 * And mark it as a globally pinned object to let the shrinker know 132 * it cannot reclaim the object until we release it. 133 */ 134 i915_vma_make_unshrinkable(vma); 135 vma->obj->mm.dirty = true; 136 137 return 0; 138 139 err_unpin: 140 i915_vma_unpin(vma); 141 return err; 142 } 143 144 static void __context_unpin_state(struct i915_vma *vma) 145 { 146 i915_vma_make_shrinkable(vma); 147 i915_active_release(&vma->active); 148 __i915_vma_unpin(vma); 149 } 150 151 static int __ring_active(struct intel_ring *ring) 152 { 153 int err; 154 155 err = i915_active_acquire(&ring->vma->active); 156 if (err) 157 return err; 158 159 err = intel_ring_pin(ring); 160 if (err) 161 goto err_active; 162 163 return 0; 164 165 err_active: 166 i915_active_release(&ring->vma->active); 167 return err; 168 } 169 170 static void __ring_retire(struct intel_ring *ring) 171 { 172 intel_ring_unpin(ring); 173 i915_active_release(&ring->vma->active); 174 } 175 176 __i915_active_call 177 static void __intel_context_retire(struct i915_active *active) 178 { 179 struct intel_context *ce = container_of(active, typeof(*ce), active); 180 181 GEM_TRACE("%s context:%llx retire\n", 182 ce->engine->name, ce->timeline->fence_context); 183 184 if (ce->state) 185 __context_unpin_state(ce->state); 186 187 intel_timeline_unpin(ce->timeline); 188 __ring_retire(ce->ring); 189 190 intel_context_put(ce); 191 } 192 193 static int __intel_context_active(struct i915_active *active) 194 { 195 struct intel_context *ce = container_of(active, typeof(*ce), active); 196 int err; 197 198 intel_context_get(ce); 199 200 err = __ring_active(ce->ring); 201 if (err) 202 goto err_put; 203 204 err = intel_timeline_pin(ce->timeline); 205 if (err) 206 goto err_ring; 207 208 if (!ce->state) 209 return 0; 210 211 err = __context_pin_state(ce->state); 212 if (err) 213 goto err_timeline; 214 215 return 0; 216 217 err_timeline: 218 intel_timeline_unpin(ce->timeline); 219 err_ring: 220 __ring_retire(ce->ring); 221 err_put: 222 intel_context_put(ce); 223 return err; 224 } 225 226 int intel_context_active_acquire(struct intel_context *ce) 227 { 228 int err; 229 230 err = i915_active_acquire(&ce->active); 231 if (err) 232 return err; 233 234 /* Preallocate tracking nodes */ 235 if (!i915_gem_context_is_kernel(ce->gem_context)) { 236 err = i915_active_acquire_preallocate_barrier(&ce->active, 237 ce->engine); 238 if (err) { 239 i915_active_release(&ce->active); 240 return err; 241 } 242 } 243 244 return 0; 245 } 246 247 void intel_context_active_release(struct intel_context *ce) 248 { 249 /* Nodes preallocated in intel_context_active() */ 250 i915_active_acquire_barrier(&ce->active); 251 i915_active_release(&ce->active); 252 } 253 254 void 255 intel_context_init(struct intel_context *ce, 256 struct i915_gem_context *ctx, 257 struct intel_engine_cs *engine) 258 { 259 struct i915_address_space *vm; 260 261 GEM_BUG_ON(!engine->cops); 262 263 kref_init(&ce->ref); 264 265 ce->gem_context = ctx; 266 rcu_read_lock(); 267 vm = rcu_dereference(ctx->vm); 268 if (vm) 269 ce->vm = i915_vm_get(vm); 270 else 271 ce->vm = i915_vm_get(&engine->gt->ggtt->vm); 272 rcu_read_unlock(); 273 if (ctx->timeline) 274 ce->timeline = intel_timeline_get(ctx->timeline); 275 276 ce->engine = engine; 277 ce->ops = engine->cops; 278 ce->sseu = engine->sseu; 279 ce->ring = __intel_context_ring_size(SZ_16K); 280 281 INIT_LIST_HEAD(&ce->signal_link); 282 INIT_LIST_HEAD(&ce->signals); 283 284 mutex_init(&ce->pin_mutex); 285 286 i915_active_init(&ce->active, 287 __intel_context_active, __intel_context_retire); 288 } 289 290 void intel_context_fini(struct intel_context *ce) 291 { 292 if (ce->timeline) 293 intel_timeline_put(ce->timeline); 294 i915_vm_put(ce->vm); 295 296 mutex_destroy(&ce->pin_mutex); 297 i915_active_fini(&ce->active); 298 } 299 300 static void i915_global_context_shrink(void) 301 { 302 kmem_cache_shrink(global.slab_ce); 303 } 304 305 static void i915_global_context_exit(void) 306 { 307 kmem_cache_destroy(global.slab_ce); 308 } 309 310 static struct i915_global_context global = { { 311 .shrink = i915_global_context_shrink, 312 .exit = i915_global_context_exit, 313 } }; 314 315 int __init i915_global_context_init(void) 316 { 317 global.slab_ce = KMEM_CACHE(intel_context, SLAB_HWCACHE_ALIGN); 318 if (!global.slab_ce) 319 return -ENOMEM; 320 321 i915_global_register(&global.base); 322 return 0; 323 } 324 325 void intel_context_enter_engine(struct intel_context *ce) 326 { 327 intel_engine_pm_get(ce->engine); 328 intel_timeline_enter(ce->timeline); 329 } 330 331 void intel_context_exit_engine(struct intel_context *ce) 332 { 333 intel_timeline_exit(ce->timeline); 334 intel_engine_pm_put(ce->engine); 335 } 336 337 int intel_context_prepare_remote_request(struct intel_context *ce, 338 struct i915_request *rq) 339 { 340 struct intel_timeline *tl = ce->timeline; 341 int err; 342 343 /* Only suitable for use in remotely modifying this context */ 344 GEM_BUG_ON(rq->hw_context == ce); 345 346 if (rcu_access_pointer(rq->timeline) != tl) { /* timeline sharing! */ 347 /* 348 * Ideally, we just want to insert our foreign fence as 349 * a barrier into the remove context, such that this operation 350 * occurs after all current operations in that context, and 351 * all future operations must occur after this. 352 * 353 * Currently, the timeline->last_request tracking is guarded 354 * by its mutex and so we must obtain that to atomically 355 * insert our barrier. However, since we already hold our 356 * timeline->mutex, we must be careful against potential 357 * inversion if we are the kernel_context as the remote context 358 * will itself poke at the kernel_context when it needs to 359 * unpin. Ergo, if already locked, we drop both locks and 360 * try again (through the magic of userspace repeating EAGAIN). 361 */ 362 if (!mutex_trylock(&tl->mutex)) 363 return -EAGAIN; 364 365 /* Queue this switch after current activity by this context. */ 366 err = i915_active_fence_set(&tl->last_request, rq); 367 mutex_unlock(&tl->mutex); 368 if (err) 369 return err; 370 } 371 372 /* 373 * Guarantee context image and the timeline remains pinned until the 374 * modifying request is retired by setting the ce activity tracker. 375 * 376 * But we only need to take one pin on the account of it. Or in other 377 * words transfer the pinned ce object to tracked active request. 378 */ 379 GEM_BUG_ON(i915_active_is_idle(&ce->active)); 380 return i915_active_add_request(&ce->active, rq); 381 } 382 383 struct i915_request *intel_context_create_request(struct intel_context *ce) 384 { 385 struct i915_request *rq; 386 int err; 387 388 err = intel_context_pin(ce); 389 if (unlikely(err)) 390 return ERR_PTR(err); 391 392 rq = i915_request_create(ce); 393 intel_context_unpin(ce); 394 395 return rq; 396 } 397 398 #if IS_ENABLED(CONFIG_DRM_I915_SELFTEST) 399 #include "selftest_context.c" 400 #endif 401