| /linux/drivers/gpu/drm/i915/gt/ |
| H A D | intel_ppgtt.c | 188 u32 pte_flags; in ppgtt_bind_vma() local 197 pte_flags = 0; in ppgtt_bind_vma() 199 pte_flags |= PTE_READ_ONLY; in ppgtt_bind_vma() 201 pte_flags |= PTE_LM; in ppgtt_bind_vma() 203 vm->insert_entries(vm, vma_res, pat_index, pte_flags); in ppgtt_bind_vma()
|
| H A D | gen8_ppgtt.c | 828 u32 pte_flags; in gen8_init_scratch() local 852 pte_flags = vm->has_read_only; in gen8_init_scratch() 854 pte_flags |= PTE_LM; in gen8_init_scratch() 860 pte_flags); in gen8_init_scratch()
|
| /linux/drivers/gpu/drm/i915/display/ |
| H A D | intel_dpt.c | 89 u32 pte_flags; in dpt_bind_vma() local 95 pte_flags = 0; in dpt_bind_vma() 97 pte_flags |= PTE_READ_ONLY; in dpt_bind_vma() 99 pte_flags |= PTE_LM; in dpt_bind_vma() 101 vm->insert_entries(vm, vma_res, pat_index, pte_flags); in dpt_bind_vma()
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_gmc.h | 173 uint64_t *pte_flags); 369 #define amdgpu_gmc_get_vm_pte(adev, vm, bo, vm_flags, pte_flags) \ argument 371 (pte_flags))) 372 #define amdgpu_gmc_override_vm_pte_flags(adev, vm, addr, pte_flags) \ argument 374 ((adev), (vm), (addr), (pte_flags))
|
| H A D | amdgpu_amdkfd_gpuvm.c | 988 attachment[i]->pte_flags = get_pte_flags(adev, vm, mem); in kfd_mem_attach() 1327 entry->pte_flags); in map_bo_to_gpuvm()
|
| /linux/arch/arm64/mm/ |
| H A D | proc.S | 255 orr pte, pte, pte_flags 270 pte_flags .req x1 314 mov_q pte_flags, KPTI_NG_PTE_FLAGS 337 bic pte_flags, pte_flags, #PTE_SHARED 410 .unreq pte_flags
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_ggtt_types.h | 63 u64 pte_flags,
|
| H A D | xe_migrate.c | 916 u32 pte_flags; in xe_migrate_copy() local 929 pte_flags = src_is_vram ? PTE_UPDATE_FLAG_IS_VRAM : 0; in xe_migrate_copy() 930 pte_flags |= use_comp_pat ? PTE_UPDATE_FLAG_IS_COMP_PTE : 0; in xe_migrate_copy() 931 batch_size += pte_update_size(m, pte_flags, src, &src_it, &src_L0, in xe_migrate_copy() 937 pte_flags = dst_is_vram ? PTE_UPDATE_FLAG_IS_VRAM : 0; in xe_migrate_copy() 938 batch_size += pte_update_size(m, pte_flags, dst, in xe_migrate_copy() 1315 u32 pte_flags = PTE_UPDATE_FLAG_IS_VRAM; in xe_migrate_vram_copy_chunk() local 1328 pte_flags |= use_comp_pat ? PTE_UPDATE_FLAG_IS_COMP_PTE : 0; in xe_migrate_vram_copy_chunk() 1329 batch_size += pte_update_size(m, pte_flags, vram, &vram_it, &vram_L0, in xe_migrate_vram_copy_chunk() 1537 u32 pte_flags; in xe_migrate_clear() local [all …]
|
| H A D | xe_ggtt.c | 827 struct xe_bo *bo, u64 pte_flags, in xe_ggtt_node_insert_transform() argument 848 transform(ggtt, node, pte_flags, ggtt->pt_ops->ggtt_set_pte, arg); in xe_ggtt_node_insert_transform() 850 xe_ggtt_map_bo(ggtt, node, bo, pte_flags); in xe_ggtt_node_insert_transform()
|
| /linux/arch/x86/mm/ |
| H A D | pti.c | 281 if (pte_flags(*pte) & _PAGE_USER) { in pti_user_pagetable_walk_pte() 401 if (WARN_ON(!(pte_flags(*pte) & _PAGE_PRESENT))) in pti_clone_pgtable()
|
| /linux/drivers/gpu/drm/amd/amdkfd/ |
| H A D | kfd_svm.c | 1218 uint64_t pte_flags; in svm_range_get_pte_flags() local 1334 pte_flags = AMDGPU_PTE_VALID; in svm_range_get_pte_flags() 1335 pte_flags |= (domain == SVM_RANGE_VRAM_DOMAIN) ? 0 : AMDGPU_PTE_SYSTEM; in svm_range_get_pte_flags() 1336 pte_flags |= snoop ? AMDGPU_PTE_SNOOPED : 0; in svm_range_get_pte_flags() 1338 pte_flags |= AMDGPU_PTE_IS_PTE; in svm_range_get_pte_flags() 1340 amdgpu_gmc_get_vm_pte(node->adev, vm, NULL, mapping_flags, &pte_flags); in svm_range_get_pte_flags() 1341 pte_flags |= AMDGPU_PTE_READABLE; in svm_range_get_pte_flags() 1343 pte_flags |= AMDGPU_PTE_WRITEABLE; in svm_range_get_pte_flags() 1347 pte_flags |= AMDGPU_PTE_BUS_ATOMICS; in svm_range_get_pte_flags() 1349 return pte_flags; in svm_range_get_pte_flags() [all …]
|
| /linux/arch/x86/kvm/mmu/ |
| H A D | paging_tmpl.h | 265 pkeys = pte_flags_pkey(pte_flags(pte)); in FNAME()
|