Searched refs:pmdp_get (Results 1 – 14 of 14) sorted by relevance
63 pmd = pmdp_get(pmdp); in show_pte()220 if (!pmd_present(pmdp_get(pmd_k))) { in vmalloc_fault()224 if (pmd_leaf(pmdp_get(pmd_k))) in vmalloc_fault()
34 if (pmd_none(pmdp_get(pmd))) { in kasan_populate_pte()66 if (pmd_none(pmdp_get(pmdp)) && IS_ALIGNED(vaddr, PMD_SIZE) && in kasan_populate_pmd()
1581 pmd = pmdp_get(pmdp); in remove_pmd_mapping()
149 int huge = pmd_val(pmdp_get(pmd)) & _PAGE_HUGE; in vmemmap_check_pmd()199 if (!pmd_present(pmdp_get(pmd))) { in populate_kernel_pte()
58 if (!pmd_present(pmdp_get(pmd))) in spurious_fault()62 return write ? pmd_write(pmdp_get(pmd)) : 1; in spurious_fault()
51 return (!pmd || pmd_none(pmdp_get(pmd))) ? NULL : (pte_t *) pmd; in huge_pte_offset()
129 #define pmdp_get(pmdp) READ_ONCE(*(pmdp)) macro613 pmd_t old = pmdp_get(pmdp); in pmdp_huge_get_and_clear()
383 pmd = pmdp_get(pmd_offset(&pud, hva)); in get_hva_mapping_size()
1081 pmd_t pmdval = pmdp_get(pmd); in guard_install_pmd_entry()1221 pmd_t pmdval = pmdp_get(pmd); in guard_remove_pmd_entry()
963 pmd = pmdp_get(pmdp); in folio_walk_start()
2121 if (unlikely(!pmd_same(pmdp_get(vmf->pmd), vmf->orig_pmd))) in do_huge_zero_wp_pmd()2274 old_pmd = pmdp_get(vmf->pmd); in do_huge_pmd_numa_page()2319 if (unlikely(!pmd_same(pmdp_get(vmf->pmd), vmf->orig_pmd))) { in do_huge_pmd_numa_page()2325 pmd = pmd_modify(pmdp_get(vmf->pmd), vma->vm_page_prot); in do_huge_pmd_numa_page()
1146 pmd_t entry = pmdp_get(pmd); in page_vma_mkclean_one()2479 pmdval = pmdp_get(pvmw.pmd); in try_to_migrate_one()
561 pmd = pmdp_get(pmdp); in __print_bad_page_map_pgtable()1843 *pmdval = pmdp_get(pmd); in zap_empty_pte_table()6937 pmd = pmdp_get(pmdp); in follow_pfnmap_start()
713 pmd = pmdp_get(pmd_offset(&pud, hva)); in host_pfn_mapping_level()