/linux/drivers/gpu/drm/i915/gt/ |
H A D | intel_ggtt.c | 258 unsigned int pat_index, in mtl_ggtt_pte_encode() 268 if (pat_index & BIT(0)) in mtl_ggtt_pte_encode() 271 if (pat_index & BIT(1)) in mtl_ggtt_pte_encode() 278 unsigned int pat_index, in gen8_ggtt_pte_encode() 438 unsigned int pat_index, in gen8_ggtt_insert_page() 445 gen8_set_pte(pte, ggtt->vm.pte_encode(addr, pat_index, flags)); in gen8_ggtt_insert_page() 452 unsigned int pat_index, u32 flags) in gen8_ggtt_insert_page_bind() 457 pte = ggtt->vm.pte_encode(addr, pat_index, flags); in gen8_ggtt_insert_page_bind() 462 gen8_ggtt_insert_page(vm, addr, offset, pat_index, flags); in gen8_ggtt_insert_page_bind() 467 unsigned int pat_index, in gen8_ggtt_insert_entries() 259 mtl_ggtt_pte_encode(dma_addr_t addr,unsigned int pat_index,u32 flags) mtl_ggtt_pte_encode() argument 279 gen8_ggtt_pte_encode(dma_addr_t addr,unsigned int pat_index,u32 flags) gen8_ggtt_pte_encode() argument 439 gen8_ggtt_insert_page(struct i915_address_space * vm,dma_addr_t addr,u64 offset,unsigned int pat_index,u32 flags) gen8_ggtt_insert_page() argument 453 gen8_ggtt_insert_page_bind(struct i915_address_space * vm,dma_addr_t addr,u64 offset,unsigned int pat_index,u32 flags) gen8_ggtt_insert_page_bind() argument 468 gen8_ggtt_insert_entries(struct i915_address_space * vm,struct i915_vma_resource * vma_res,unsigned int pat_index,u32 flags) gen8_ggtt_insert_entries() argument 507 __gen8_ggtt_insert_entries_bind(struct i915_address_space * vm,struct i915_vma_resource * vma_res,unsigned int pat_index,u32 flags) __gen8_ggtt_insert_entries_bind() argument 538 gen8_ggtt_insert_entries_bind(struct i915_address_space * vm,struct i915_vma_resource * vma_res,unsigned int pat_index,u32 flags) gen8_ggtt_insert_entries_bind() argument 594 gen6_ggtt_insert_page(struct i915_address_space * vm,dma_addr_t addr,u64 offset,unsigned int pat_index,u32 flags) gen6_ggtt_insert_page() argument 614 gen6_ggtt_insert_entries(struct i915_address_space * vm,struct i915_vma_resource * vma_res,unsigned int pat_index,u32 flags) gen6_ggtt_insert_entries() argument 666 unsigned int pat_index; global() member 683 bxt_vtd_ggtt_insert_page__BKL(struct i915_address_space * vm,dma_addr_t addr,u64 offset,unsigned int pat_index,u32 unused) bxt_vtd_ggtt_insert_page__BKL() argument 694 unsigned int pat_index; global() member 711 bxt_vtd_ggtt_insert_entries__BKL(struct i915_address_space * vm,struct i915_vma_resource * vma_res,unsigned int pat_index,u32 flags) bxt_vtd_ggtt_insert_entries__BKL() argument 743 intel_ggtt_bind_vma(struct i915_address_space * vm,struct i915_vm_pt_stash * stash,struct i915_vma_resource * vma_res,unsigned int pat_index,u32 flags) intel_ggtt_bind_vma() argument 920 aliasing_gtt_bind_vma(struct i915_address_space * vm,struct i915_vm_pt_stash * stash,struct i915_vma_resource * vma_res,unsigned int pat_index,u32 flags) aliasing_gtt_bind_vma() argument 1301 snb_pte_encode(dma_addr_t addr,unsigned int pat_index,u32 flags) snb_pte_encode() argument 1322 ivb_pte_encode(dma_addr_t addr,unsigned int pat_index,u32 flags) ivb_pte_encode() argument 1345 byt_pte_encode(dma_addr_t addr,unsigned int pat_index,u32 flags) byt_pte_encode() argument 1360 hsw_pte_encode(dma_addr_t addr,unsigned int pat_index,u32 flags) hsw_pte_encode() argument 1372 iris_pte_encode(dma_addr_t addr,unsigned int pat_index,u32 flags) iris_pte_encode() argument [all...] |
H A D | gen8_ppgtt.c | 33 unsigned int pat_index, in gen8_pte_encode() argument 46 switch (pat_index) { in gen8_pte_encode() 62 unsigned int pat_index, in gen12_pte_encode() argument 73 if (pat_index & BIT(0)) in gen12_pte_encode() 76 if (pat_index & BIT(1)) in gen12_pte_encode() 79 if (pat_index & BIT(2)) in gen12_pte_encode() 82 if (pat_index & BIT(3)) in gen12_pte_encode() 459 unsigned int pat_index, in gen8_ppgtt_insert_pte() argument 463 const gen8_pte_t pte_encode = ppgtt->vm.pte_encode(0, pat_index, flags); in gen8_ppgtt_insert_pte() 506 unsigned int pat_index, in xehp_ppgtt_insert_huge() argument [all …]
|
H A D | intel_ggtt_gmch.c | 21 unsigned int pat_index, in gmch_ggtt_insert_page() argument 24 unsigned int flags = (pat_index == I915_CACHE_NONE) ? in gmch_ggtt_insert_page() 32 unsigned int pat_index, in gmch_ggtt_insert_entries() argument 35 unsigned int flags = (pat_index == I915_CACHE_NONE) ? in gmch_ggtt_insert_entries()
|
H A D | selftest_migrate.c | 140 unsigned int pat_index, in intel_context_copy_ccs() argument 188 len = emit_pte(rq, &it, pat_index, true, offset, CHUNK_SZ); in intel_context_copy_ccs() 226 unsigned int pat_index, in intel_migrate_ccs_copy() argument 246 err = intel_context_copy_ccs(ce, deps, sg, pat_index, in intel_migrate_ccs_copy() 303 obj->pat_index, in clear() 354 obj->pat_index, in clear() 417 src->mm.pages->sgl, src->pat_index, in __migrate_copy() 419 dst->mm.pages->sgl, dst->pat_index, in __migrate_copy() 431 src->mm.pages->sgl, src->pat_index, in __global_copy() 433 dst->mm.pages->sgl, dst->pat_index, in __global_copy() 845 __perf_clear_blt(struct intel_context * ce,struct scatterlist * sg,unsigned int pat_index,bool is_lmem,size_t sz) __perf_clear_blt() argument [all...] |
H A D | intel_gtt.h | 236 unsigned int pat_index, 313 unsigned int pat_index, 328 unsigned int pat_index, 332 unsigned int pat_index, 337 unsigned int pat_index, 341 unsigned int pat_index, 588 unsigned int pat_index, 666 unsigned int pat_index,
|
H A D | intel_migrate.h | 50 unsigned int pat_index, 58 unsigned int pat_index,
|
H A D | gen8_ppgtt.h | 18 unsigned int pat_index,
|
H A D | intel_migrate.c | 364 unsigned int pat_index, in emit_pte() argument 370 const u64 encode = rq->context->vm->pte_encode(0, pat_index, in emit_pte() 987 unsigned int pat_index, in intel_context_migrate_clear() argument 1035 len = emit_pte(rq, &it, pat_index, is_lmem, offset, CHUNK_SZ); in intel_context_migrate_clear() 1121 unsigned int pat_index, in intel_migrate_clear() argument 1142 err = intel_context_migrate_clear(ce, deps, sg, pat_index, in intel_migrate_clear()
|
H A D | gen6_ppgtt.c | 112 unsigned int pat_index, in gen6_ppgtt_insert_entries() argument 120 const u32 pte_encode = vm->pte_encode(0, pat_index, flags); in gen6_ppgtt_insert_entries() 283 unsigned int pat_index, in pd_vma_bind() argument
|
H A D | intel_ppgtt.c | 185 unsigned int pat_index, in ppgtt_bind_vma() argument 203 vm->insert_entries(vm, vma_res, pat_index, pte_flags); in ppgtt_bind_vma()
|
/linux/drivers/gpu/drm/xe/ |
H A D | xe_pt_types.h | 42 u16 pat_index, u32 pt_level); 44 u16 pat_index, u32 pt_level); 46 u16 pat_index, 49 u16 pat_index);
|
H A D | xe_ggtt.c | 67 u16 pat_index) in probe_gsm_size() 81 u16 pat_index) in ggtt_update_access_counter() 86 pte = xelp_ggtt_pte_encode_bo(bo, bo_offset, pat_index); in ggtt_update_access_counter() 88 xe_assert(xe, pat_index <= 3); in ggtt_update_access_counter() 90 if (pat_index & BIT(0)) in ggtt_update_access_counter() 93 if (pat_index & BIT(1)) in xe_ggtt_set_pte() 142 u16 pat_index = tile_to_xe(ggtt->tile)->pat.idx[XE_CACHE_WB]; in primelockdep() 150 pat_index); in primelockdep() 583 u16 pat_index = tile_to_xe(ggtt->tile)->pat.idx[cache_mode]; 593 pte = ggtt->pt_ops->pte_encode_bo(bo, offset, pat_index); 34 xelp_ggtt_pte_encode_bo(struct xe_bo * bo,u64 bo_offset,u16 pat_index) xelp_ggtt_pte_encode_bo() argument 48 xelpg_ggtt_pte_encode_bo(struct xe_bo * bo,u64 bo_offset,u16 pat_index) xelpg_ggtt_pte_encode_bo() argument 109 u16 pat_index = tile_to_xe(ggtt->tile)->pat.idx[XE_CACHE_WB]; xe_ggtt_clear() local 319 u16 pat_index = tile_to_xe(ggtt->tile)->pat.idx[XE_CACHE_WB]; xe_ggtt_printk() local 428 u16 pat_index = tile_to_xe(ggtt->tile)->pat.idx[cache_mode]; xe_ggtt_map_bo() local [all...] |
H A D | xe_vm.c | 912 u16 pat_index, unsigned int flags) in xe_vma_create() 963 vma->pat_index = pat_index; in xe_vma_destroy_late() 1198 static u64 pde_encode_pat_index(u16 pat_index) in xelp_pte_encode_bo() 1202 if (pat_index & BIT(0)) in xelp_pte_encode_bo() 1205 if (pat_index & BIT(1)) in xelp_pte_encode_bo() 1211 static u64 pte_encode_pat_index(u16 pat_index, u32 pt_level) in xelp_pte_encode_vma() 1215 if (pat_index & BIT(0)) in xelp_pte_encode_vma() 1218 if (pat_index & BIT(1)) in xelp_pte_encode_vma() 1221 if (pat_index in xelp_pte_encode_vma() 839 xe_vma_create(struct xe_vm * vm,struct xe_bo * bo,u64 bo_offset_or_userptr,u64 start,u64 end,u16 pat_index,unsigned int flags) xe_vma_create() argument 1125 pde_encode_pat_index(struct xe_device * xe,u16 pat_index) pde_encode_pat_index() argument 1138 pte_encode_pat_index(struct xe_device * xe,u16 pat_index,u32 pt_level) pte_encode_pat_index() argument 1178 xelp_pde_encode_bo(struct xe_bo * bo,u64 bo_offset,const u16 pat_index) xelp_pde_encode_bo() argument 1191 xelp_pte_encode_bo(struct xe_bo * bo,u64 bo_offset,u16 pat_index,u32 pt_level) xelp_pte_encode_bo() argument 1208 xelp_pte_encode_vma(u64 pte,struct xe_vma * vma,u16 pat_index,u32 pt_level) xelp_pte_encode_vma() argument 1227 xelp_pte_encode_addr(struct xe_device * xe,u64 addr,u16 pat_index,u32 pt_level,bool devmem,u64 flags) xelp_pte_encode_addr() argument 2065 vm_bind_ioctl_ops_create(struct xe_vm * vm,struct xe_bo * bo,u64 bo_offset_or_userptr,u64 addr,u64 range,u32 operation,u32 flags,u32 prefetch_region,u16 pat_index) vm_bind_ioctl_ops_create() argument 2138 new_vma(struct xe_vm * vm,struct drm_gpuva_op_map * op,u16 pat_index,unsigned int flags) new_vma() argument 2960 u16 pat_index = (*bind_ops)[i].pat_index; vm_bind_ioctl_check_args() local 3067 xe_vm_bind_ioctl_validate_bo(struct xe_device * xe,struct xe_bo * bo,u64 addr,u64 range,u64 obj_offset,u16 pat_index) xe_vm_bind_ioctl_validate_bo() argument 3187 u16 pat_index = bind_ops[i].pat_index; xe_vm_bind_ioctl() local 3246 u16 pat_index = bind_ops[i].pat_index; xe_vm_bind_ioctl() local [all...] |
H A D | xe_migrate.c | 140 u64 map_ofs, u64 vram_offset, u16 pat_index, u64 pt_2m_ofs) in xe_migrate_prepare_vm() 150 flags = vm->pt_ops->pte_encode_addr(xe, 0, pat_index, level, in xe_migrate_prepare_vm() 164 pat_index); in xe_migrate_prepare_vm() 168 pat_index, in xe_migrate_prepare_vm() 188 u16 pat_index = xe->pat.idx[XE_CACHE_WB]; in xe_migrate_prepare_vm() 218 entry = vm->pt_ops->pde_encode_bo(bo, pt29_ofs, pat_index); in xe_migrate_prepare_vm() 226 pat_index, 0); in xe_migrate_prepare_vm() 243 pat_index, 0); in xe_migrate_prepare_vm() 260 pat_index, 0); in xe_migrate_prepare_vm() 286 XE_PAGE_SIZE, pat_index); in xe_migrate_prepare_vm() 138 u16 pat_index = xe->pat.idx[XE_CACHE_WB]; xe_migrate_prepare_vm() local 522 u16 pat_index; emit_pte() local 1314 u16 pat_index = xe->pat.idx[XE_CACHE_WB]; xe_migrate_update_pgtables() local [all...] |
H A D | xe_vm_types.h | 115 * @pat_index: The pat index to use when encoding the PTEs for this vma. 117 u16 pat_index; 299 /** @pat_index: The pat index to use for this operation. */ 300 u16 pat_index; 105 u16 pat_index; global() member 288 u16 pat_index; global() member
|
H A D | xe_pat.h | 59 u16 xe_pat_index_get_coh_mode(struct xe_device *xe, u16 pat_index);
|
H A D | xe_ggtt_types.h | 78 u64 (*pte_encode_bo)(struct xe_bo *bo, u64 bo_offset, u16 pat_index);
|
H A D | xe_pat.c | 151 u16 xe_pat_index_get_coh_mode(struct xe_device *xe, u16 pat_index) in xe_pat_index_get_coh_mode() argument 153 WARN_ON(pat_index >= xe->pat.n_entries); in xe_pat_index_get_coh_mode() 154 return xe->pat.table[pat_index].coh_mode; in xe_pat_index_get_coh_mode()
|
H A D | xe_pt.c | 60 u16 pat_index = xe->pat.idx[XE_CACHE_WB]; in __xe_pt_empty_pte() 68 0, pat_index); 70 return vm->pt_ops->pte_encode_addr(xe, 0, pat_index, level, IS_DGFX(xe), 0) | in xe_pt_free() 484 u16 pat_index = xe_walk->vma->pat_index; in xe_pt_stage_bind_entry() 501 xe_walk->vma, pat_index, level); in xe_pt_stage_bind_entry() 571 pte = vm->pt_ops->pde_encode_bo(xe_child->bo, 0, pat_index) | flags; in xe_pt_stage_bind_entry() 55 u16 pat_index = xe->pat.idx[XE_CACHE_WB]; __xe_pt_empty_pte() local 478 u16 pat_index = xe_walk->vma->pat_index; xe_pt_stage_bind_entry() local
|
/linux/drivers/gpu/drm/i915/selftests/ |
H A D | mock_gtt.c | 30 unsigned int pat_index, in mock_insert_page() argument 37 unsigned int pat_index, u32 flags) in mock_insert_entries() argument 44 unsigned int pat_index, in mock_bind_ppgtt() argument 97 unsigned int pat_index, in mock_bind_ggtt() argument
|
/linux/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_create.c | 248 unsigned int pat_index; member 417 if (ext.pat_index > max_pat_index) { in ext_set_pat() 419 ext.pat_index); in ext_set_pat() 423 ext_data->pat_index = ext.pat_index; in ext_set_pat() 454 ext_data.pat_index = PAT_INDEX_NOT_SET; in i915_gem_create_ext_ioctl() 491 if (ext_data.pat_index != PAT_INDEX_NOT_SET) { in i915_gem_create_ext_ioctl() 492 i915_gem_object_set_pat_index(obj, ext_data.pat_index); in i915_gem_create_ext_ioctl()
|
H A D | i915_gem_object.c | 73 return obj->pat_index == i915_gem_get_pat_index(obj_to_i915(obj), lvl); in i915_gem_object_has_cache_level() 159 obj->pat_index = i915_gem_get_pat_index(i915, cache_level); in i915_gem_object_set_cache_coherency() 183 unsigned int pat_index) in i915_gem_object_set_pat_index() argument 187 if (obj->pat_index == pat_index) in i915_gem_object_set_pat_index() 190 obj->pat_index = pat_index; in i915_gem_object_set_pat_index() 192 if (pat_index != i915_gem_get_pat_index(i915, I915_CACHE_NONE)) in i915_gem_object_set_pat_index()
|
/linux/drivers/gpu/drm/i915/display/ |
H A D | intel_dpt.c | 44 unsigned int pat_index, in dpt_insert_page() argument 51 vm->pte_encode(addr, pat_index, flags)); in dpt_insert_page() 56 unsigned int pat_index, in dpt_insert_entries() argument 61 const gen8_pte_t pte_encode = vm->pte_encode(0, pat_index, flags); in dpt_insert_entries() 84 unsigned int pat_index, in dpt_bind_vma() argument 99 vm->insert_entries(vm, vma_res, pat_index, pte_flags); in dpt_bind_vma()
|
/linux/drivers/gpu/drm/i915/ |
H A D | i915_debugfs.c | 153 switch (obj->pat_index) { in i915_cache_level_str() 162 switch (obj->pat_index) { in i915_cache_level_str() 170 switch (obj->pat_index) { in i915_cache_level_str()
|
H A D | i915_vma.c | 351 unsigned int pat_index; member 370 vma_res, vw->pat_index, vw->flags); in __vma_bind() 473 unsigned int pat_index, in i915_vma_bind() argument 543 work->pat_index = pat_index; in i915_vma_bind() 573 vma->ops->bind_vma(vma->vm, NULL, vma->resource, pat_index, in i915_vma_bind() 850 color = vma->obj->pat_index; in i915_vma_insert() 1568 vma->obj->pat_index, in i915_vma_pin_ww()
|