| /linux/arch/arm64/mm/ |
| H A D | contpte.c | 312 pte_t contpte_ptep_get(pte_t *ptep, pte_t orig_pte) in contpte_ptep_get() argument 330 orig_pte = pte_mkdirty(orig_pte); in contpte_ptep_get() 334 orig_pte = pte_mkyoung(orig_pte); in contpte_ptep_get() 342 orig_pte = pte_mkyoung(orig_pte); in contpte_ptep_get() 348 orig_pte = pte_mkdirty(orig_pte); in contpte_ptep_get() 356 return orig_pte; in contpte_ptep_get() 390 pte_t orig_pte; in contpte_ptep_get_lockless() local 396 orig_pte = __ptep_get(orig_ptep); in contpte_ptep_get_lockless() 398 if (!pte_valid_cont(orig_pte)) in contpte_ptep_get_lockless() 399 return orig_pte; in contpte_ptep_get_lockless() [all …]
|
| /linux/mm/ |
| H A D | memory.c | 112 return pte_is_uffd_wp_marker(vmf->orig_pte); in vmf_orig_pte_uffd_wp() 883 pte_t *ptep, pte_t orig_pte) in restore_exclusive_pte() argument 890 if (pte_swp_soft_dirty(orig_pte)) in restore_exclusive_pte() 893 if (pte_swp_uffd_wp(orig_pte)) in restore_exclusive_pte() 916 unsigned long addr, pte_t *ptep, pte_t orig_pte) in try_restore_exclusive_pte() argument 918 const softleaf_t entry = softleaf_from_pte(orig_pte); in try_restore_exclusive_pte() 923 restore_exclusive_pte(vma, folio, page, addr, ptep, orig_pte); in try_restore_exclusive_pte() 943 pte_t orig_pte = ptep_get(src_pte); in copy_nonpresent_pte() local 944 softleaf_t entry = softleaf_from_pte(orig_pte); in copy_nonpresent_pte() 945 pte_t pte = orig_pte; in copy_nonpresent_pte() [all …]
|
| H A D | hugetlb.c | 5777 if (!hugetlb_pte_stable(h, mm, vmf->address, vmf->pte, vmf->orig_pte)) { in hugetlb_no_page() 5806 if (hugetlb_pte_stable(h, mm, vmf->address, vmf->pte, vmf->orig_pte)) in hugetlb_no_page() 5854 if (!hugetlb_pte_stable(h, mm, vmf->address, vmf->pte, vmf->orig_pte)) { in hugetlb_no_page() 5881 if (!pte_same(huge_ptep_get(mm, vmf->address, vmf->pte), vmf->orig_pte)) in hugetlb_no_page() 5893 if (unlikely(pte_is_uffd_wp_marker(vmf->orig_pte))) in hugetlb_no_page() 6018 vmf.orig_pte = huge_ptep_get(mm, vmf.address, vmf.pte); in hugetlb_fault() 6019 if (huge_pte_none(vmf.orig_pte)) in hugetlb_fault() 6026 if (pte_is_marker(vmf.orig_pte)) { in hugetlb_fault() 6028 softleaf_to_marker(softleaf_from_pte(vmf.orig_pte)); in hugetlb_fault() 6046 if (!pte_present(vmf.orig_pte) && !huge_pte_none(vmf.orig_pte)) { in hugetlb_fault() [all …]
|
| H A D | migrate_device.c | 985 pte_t orig_pte; in migrate_vma_insert_page() local 1057 orig_pte = ptep_get(ptep); in migrate_vma_insert_page() 1062 if (pte_present(orig_pte)) { in migrate_vma_insert_page() 1063 unsigned long pfn = pte_pfn(orig_pte); in migrate_vma_insert_page() 1068 } else if (!pte_none(orig_pte)) in migrate_vma_insert_page() 1085 flush_cache_page(vma, addr, pte_pfn(orig_pte)); in migrate_vma_insert_page()
|
| H A D | ksm.c | 1273 pte_t *orig_pte) in write_protect_page() argument 1352 *orig_pte = entry; in write_protect_page() 1373 struct page *kpage, pte_t orig_pte) in replace_page() argument 1410 if (!pte_same(ptep_get(ptep), orig_pte)) { in replace_page() 1479 pte_t orig_pte = __pte(0); in try_to_merge_one_page() local 1509 if (write_protect_page(vma, folio, &orig_pte) == 0) { in try_to_merge_one_page() 1526 err = replace_page(vma, page, kpage, orig_pte); in try_to_merge_one_page()
|
| H A D | khugepaged.c | 1007 vmf.orig_pte = ptep_get_lockless(pte); in __collapse_huge_page_swapin() 1008 if (pte_none(vmf.orig_pte) || in __collapse_huge_page_swapin() 1009 pte_present(vmf.orig_pte)) in __collapse_huge_page_swapin()
|
| /linux/arch/powerpc/kvm/ |
| H A D | trace_pr.h | 33 struct kvmppc_pte *orig_pte), 34 TP_ARGS(rflags, hpteg, va, hpaddr, orig_pte), 49 __entry->eaddr = orig_pte->eaddr; 52 __entry->vpage = orig_pte->vpage;
|
| H A D | book3s_hv_rm_mmu.c | 492 u64 pte, orig_pte, pte_r; in kvmppc_do_h_remove() local 501 pte = orig_pte = be64_to_cpu(hpte[0]); in kvmppc_do_h_remove() 510 __unlock_hpte(hpte, orig_pte); in kvmppc_do_h_remove()
|
| /linux/arch/arm64/include/asm/ |
| H A D | pgtable.h | 1647 extern pte_t contpte_ptep_get(pte_t *ptep, pte_t orig_pte); 1757 pte_t orig_pte = __ptep_get(ptep); in set_pte() local 1759 WARN_ON_ONCE(pte_valid_cont(orig_pte)); in set_pte() 1826 pte_t orig_pte = __ptep_get(ptep); in ptep_test_and_clear_young() local 1828 if (likely(!pte_valid_cont(orig_pte))) in ptep_test_and_clear_young() 1838 pte_t orig_pte = __ptep_get(ptep); in ptep_clear_flush_young() local 1840 if (likely(!pte_valid_cont(orig_pte))) in ptep_clear_flush_young() 1869 pte_t orig_pte = __ptep_get(ptep); in wrprotect_ptes() local 1871 if (unlikely(pte_cont(orig_pte))) { in wrprotect_ptes() 1872 __contpte_try_unfold(mm, addr, ptep, orig_pte); in wrprotect_ptes() [all …]
|
| /linux/arch/x86/kvm/mmu/ |
| H A D | paging_tmpl.h | 203 pt_element_t pte, orig_pte; in FNAME() local 213 pte = orig_pte = walker->ptes[level - 1]; in FNAME() 230 if (pte == orig_pte) in FNAME() 249 ret = __try_cmpxchg_user(ptep_user, &orig_pte, pte, fault); in FNAME()
|
| /linux/fs/proc/ |
| H A D | task_mmu.c | 2085 pte_t *pte, *orig_pte; in pagemap_pmd_range() local 2101 orig_pte = pte = pte_offset_map_lock(walk->mm, pmdp, addr, &ptl); in pagemap_pmd_range() 2114 pte_unmap_unlock(orig_pte, ptl); in pagemap_pmd_range() 3224 pte_t *orig_pte; in gather_pte_stats() local 3240 orig_pte = pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); in gather_pte_stats() 3253 pte_unmap_unlock(orig_pte, ptl); in gather_pte_stats()
|
| /linux/arch/arm64/kvm/hyp/ |
| H A D | pgtable.c | 1252 kvm_pte_t attr_clr, kvm_pte_t *orig_pte, in stage2_update_leaf_attrs() argument 1271 if (orig_pte) in stage2_update_leaf_attrs() 1272 *orig_pte = data.pte; in stage2_update_leaf_attrs()
|
| /linux/include/linux/ |
| H A D | pgtable.h | 1345 pte_t orig_pte) in arch_unmap_one() argument
|
| H A D | mm.h | 714 pte_t orig_pte; /* Value of PTE at the time of fault */ member
|